skip to main content
10.1145/263764.263787acmconferencesArticle/Chapter ViewAbstractPublication PagesppoppConference Proceedingsconference-collections
Article
Free access

Compilation of parallel multimedia computations—extending retiming theory and Amdahl's law

Published: 21 June 1997 Publication History

Abstract

Multimedia applications (also called multimedia systems) operate on datastreams, which are periodic sequences of data elements, called datasets. A large class of multimedia applications is described by the macro-dataflow graph model, with nodes representing parallelizable tasks, and arcs representing communication. This paper examines how such multimedia applications can be compiled to run efficiently on parallel machines, by optimizing both throughput (T) and latency (L), using two techniques, based on task speedup functions. The first step chooses an appropriate pipeline structure for the system (task clustering). The second step exploits the dataset parallelism intrinsic in the periodic datastream, and runs multiple datasets in parallel (task/cluster multiplicity) for each clustering. The key find-of this research areA The best task clustering depends on system throughput. In general skewed parallelism profiles are desirable i.e. tasks with good speedup and tasks with poor speedup are in separate clusters. Indeed the maximal throughput and minimal latency can be simultaneously attained in the limiting case of a maximally skewed distribution. This result can be viewed as a generalization of Amdahl's law for real-time applications.B Optimal dataset multiplicity for a specific clustering can be determined by extending retiming theory [1] to include parallel resource allocation. In this process, counter-intuitive relaxation regions often appear, wherein by increasing dataset multiplicity, throughput is increased and latency simultaneously reduced (a free lunch).The techniques have been used for compiling real-time image-processing problems on an NCUBE-2 multiprocessor, and show substantial performance gains.

References

[1]
C. Leiserson and J. Saxe. Retiming synchronous circurry. Algorithrnica, 6:5-35, 1991.]]
[2]
T. Yang & A. Gerasoulis. A Fast Static Scheduling Algorithm for DAGs on an Unbounded Number of Processors. In Proc. of Supercomputin9 91, volume 3, pages 633-642, Nov. 1991.]]
[3]
J. Subtdok et al. Communication and memory requirements as the basis for mapping task and data parallel programs. In Proc. o.f Supercomputing '9#, pages 330- 349, November 1994.]]
[4]
J Subhlok and G. Vondran. Optimal mapping of sequences of data parallel tasks. In Proceedings of PPoPP '95, July 1995.]]
[5]
J Subhlok and G. Vondran. Optimal latencythroughput tradeoffs for data parallel pipelines. In Proceedings of $PAA '9#, June 1996.]]
[6]
A. Choudhary et al. Optimal processor assignment for a class of pipelined computations. IEEE Transactions on Parallel and Distributed Systems, 5(4):439-445, April 94.]]
[7]
V. Sarkar. Partitioning and Scheduling Programs for Multiprocessors. Technical Report CSL-TR-87-328, Ph.D Thesis, Computer Systems Lab., Stanford University, April 1987.]]
[8]
K. P. Belkhale and P. Banerjee. Scheduling Algorithms for Parallelizable Tasks. In International Parallel Processing Symposium, June 1993.]]
[9]
G.N.Srinivasa Prasanna and Bruce R. Musicus. Generalized Multiprocessor Scheduling for Directed Acyelic Graphs. In Supercornputing '9#, pages 216-228, November 1994.]]
[10]
C. Leiserson and J. Saxe. A mixed-integer linear programming problem which is efficiently solvable. Journal of Algorithms, 9:114-128, 1986.]]

Cited By

View all

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
PPOPP '97: Proceedings of the sixth ACM SIGPLAN symposium on Principles and practice of parallel programming
June 1997
287 pages
ISBN:0897919068
DOI:10.1145/263764
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 21 June 1997

Permissions

Request permissions for this article.

Check for updates

Qualifiers

  • Article

Conference

PPoPP97
Sponsor:
PPoPP97: Principles & Practices of Parallel Programming
June 18 - 21, 1997
Nevada, Las Vegas, USA

Acceptance Rates

PPOPP '97 Paper Acceptance Rate 26 of 86 submissions, 30%;
Overall Acceptance Rate 230 of 1,014 submissions, 23%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)75
  • Downloads (Last 6 weeks)13
Reflects downloads up to 06 Jan 2025

Other Metrics

Citations

Cited By

View all

View Options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Login options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media