High-Throughput, Low-Memory Applications on the Pica Architecture

  • Authors:
  • D. Scott Wills;Huy H. Cat;José Cruz-Rivera;W. Stephen Lacy;James M. Baker, Jr.;John C. Eble;Abelardo López-Lagunas;Michael Hopper

  • Affiliations:
  • Georgia Institute of Technology, Atlanta;Georgia Institute of Technology, Atlanta;Univ. of Puerto Rico-Mayaguez, Mayaguez;Univ. of Arkansas, Fayetteville;Georgia Institute of Technology, Atlanta;Georgia Institute of Technology, Atlanta;Georgia Institute of Technology, Atlanta;Georgia Institute of Technology, Atlanta

  • Venue:
  • IEEE Transactions on Parallel and Distributed Systems
  • Year:
  • 1997

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper describes Pica, a fine-grain, message-passing architecture designed to efficiently support high-throughput, low-memory parallel applications, such as image processing, object recognition, and data compression. By specializing the processor and reducing local memory (4,096 36-bit words), multiple nodes can be implemented on a single chip. This allows high-performance systems for high-throughput applications to be realized at lower cost. The architecture minimizes overhead for basic parallel operations. An operand-addressed context cache and round-robin task manager support fast task swapping. Fixed-sized activation contexts simplify storage management. Word-tag synchronization bits provide low-cost synchronization. Several applications have been developed for this architecture, including thermal relaxation, matrix multiplication, JPEG image compression, and Positron Emission Tomography image reconstruction. These applications have been executed using an instrumented instruction-level simulator. The results of these experiments and an evaluation of Pica's architectural features are presented.