ABSTRACT
Current processor and multiprocessor architectures are almost all based on the Von Neumann paradigm. Based on this paradigm, one can build a general-purpose computer using very few transistors, e.g., 2250 transistors in the first Intel 4004 microprocessor. In other terms, the notion that on-chip space is a scarce resource is at the root of this paradigm which trades on-chip space for program execution time. Today, technology considerably relaxed this space constraint. Still, few research works question this paradigm as the most adequate basis for high-performance computers, even though the paradigm was not initially designed to scale with technology and space.In this article, we propose a different computing model, defining both an architecture and a language, that is intrinsically designed to exploit space; we then investigate the implementation issues of a computer based on this model, and we provide simulation results for small programs and a simplified architecture as a first proof of concept. Through this model, we also want to outline that revisiting some of the principles of today's computing paradigm has the potential of overcoming major limitations of current architectures.
- H. Abelson, D. Allen, D. Coore, C. Hanson, G. Homsy, T. F. Knight, R. Nagpal, E. Rauch, G. J. Sussman, and R. Weiss. Amorphous computing. Communications of the ACM, 43(5):74--82, 2000.]] Google ScholarDigital Library
- L. Adleman. Molecular computation of solutions to combinatorial problems,. Science, 266, November 1994.]] Google ScholarDigital Library
- J-P. BANÂTRE and D. Le MÉTAYER. Gamma and the chemical reaction model : Ten years after. In J.-M. Andreoli, H. Gallaire, and D. Le Métayer, editors, Coordination Programming: Mechanisms, Models and Semantics, pages 1--39, 1996.]] Google ScholarDigital Library
- M. Baron. Microprocessor report. Tidbits, March 2001.]]Google Scholar
- G. Berry and G. Gonthier. The Esterel synchronous programming language: Design, semantics, implementation. Science of Computer Programming, 19(2):87--152, 1992.]] Google ScholarDigital Library
- W. Blume, R. Eigenmann, K. Faigin, J. Grout, J. Hoeflinger, D. Padua, P. Petersen, W. Pottenger, L. Rauchwerger, P. Tu, and S. Weatherford. Parallel programming with Polaris. IEEE Computer, 29(12):78--82, December 1996.]] Google ScholarDigital Library
- F. Bodin, T. Kisuk, P. Knijnenburg, M. O'Boyle, and E. Rohou. Iterative compilation in a non-linear optimisation space. In Workshop on Profile and Feedback-Directed Compilation at PACT, 1998.]]Google Scholar
- J. M.P. Cardoso and H. C. Neto. Fast hardware compilation of behaviors into an FPGA-based dynamic reconfigurable computing system. In Proc. of the XII Symposium on Integrated Circuits and Systems Design, pages 150--153. IEEE Computer Society Press, October 1999.]] Google ScholarDigital Library
- E. Caspi, M. Chu, R. Huang, J. Yeh, J. Wawrzynek, and A. DeHon. Stream computations organized for reconfigurable execution (SCORE). In FPL, pages 605--614, 2000.]] Google ScholarDigital Library
- L. Chen, S. Dropsho, and D. H. Albonesi. Dynamic data dependence tracking and its application to branch prediction. In Proceedings of the 9th International Symposium on High-Performance Computer Architecture, pages 65--77, February 2003.]] Google ScholarDigital Library
- J.-F. Collard, D. Barthou, and P. Feautrier. Fuzzy array dataflow analysis. In ppopp, pages 92--102, Santa Barbara, California, July 1995.]] Google ScholarDigital Library
- M. Conrad. On design principles for a molecular computer. Commun. ACM, 28(5):464--480, 1985.]] Google ScholarDigital Library
- Z. Cvetanovic and D. Bhandarkar. Performance characterization of the alpha 21164 microprocessor using TP and SPEC workloads. In HPCA, pages 270--280, 1996.]] Google ScholarDigital Library
- A. Dandalis and V. K. Prasanna. Run-time performance optimization of an FPGA-based deduction engine for sat solvers. ACM Trans. Des. Autom. Electron. Syst., 7(4):547--562, 2002.]] Google ScholarDigital Library
- A. DeHon. Reconfigurable Architectures for General-Purpose Computing. PhD thesis, Massachusetts Institute of Technology, Department of Electrical Engineering and Computer Science, 1996.]] Google ScholarDigital Library
- A. DeHon. Very large scale spatial computing. In Third International Conference on Unconventional Models of Computation, pages 27--37, October 2002.]] Google ScholarDigital Library
- André DeHon. The density advantage of configurable computing. Computer, 33(4):41--49, 2000.]] Google ScholarDigital Library
- D. Deutsch and R. Jozsa. Rapid solution of problems by quantum computation. Proceedings of the Royal Society of London Ser. A, A439:553--558, 1992.]]Google ScholarCross Ref
- Jr. F. H. Carvalho, R. M. F. Lima, and R. D. Lins. Coordinating functional processes with Haskell#. In Proceedings of the 2002 ACM symposium on Applied computing, pages 393--400. ACM Press, 2002.]] Google ScholarDigital Library
- P. Fradet and D. Le Métayer. Structured Gamma. Science of Computer Programming, 31(2--3):263--289, 1998.]] Google ScholarDigital Library
- T. Gautier, P. Le Guernic, and L. Bernard. Signal: A declarative language for synchronous programming of real time system. Computer Science, pages 257--277, 1987.]]Google Scholar
- S. Ghosh, M. Martonosi, and S. Malik. Cache miss equations: an analytical representation of cache misses. In Proceedings of the 11th international conference on Supercomputing, pages 317--324. ACM Press, 1997.]] Google ScholarDigital Library
- P. N. Glaskowsky. Microprocessor report. Network Processors Mature in 2001, February 2002.]]Google Scholar
- S. C. Goldstein and M. Budiu. Nanofabrics: spatial computing using molecular electronics. In Proceedings of the 28th annual international symposium on Computer architecture, pages 178--191. ACM Press, 2001.]] Google ScholarDigital Library
- F. Gruau. Process of translation and conception of neural networks, based on a logical description of the target problem, us patent en 93 158 92, december 30, 1993.]]Google Scholar
- F. Gruau. Automatic definition of modular neural networks. Adaptive Behaviour, 3(2):151--183, 1995.]] Google ScholarDigital Library
- F. Gruau. Modular genetic neural networks for 6-legged locomotion. Artifical Evolution, pages 201--219, 1995.]] Google ScholarDigital Library
- F. Gruau and P. Malbos. The Blob: A basic topological concept for hardware-free distributed computation. In Unconventional Models of Computation (UMC'02), Kobe, Japan, pages 151--163, 2002.]] Google ScholarDigital Library
- F. Gruau and G. Moszkowski. Time-efficient self-reproduction on a 2-d cellular automaton. In 1st International Workshop on Biologically inspired Approaches To Advanced Information Technology, 2004.]]Google Scholar
- F. Gruau, J-Y. Ratajszczak, and G. Wiber. A neural compiler. Theoretical Computer Science, 141(1, 2):1--52, April 1995.]] Google ScholarDigital Library
- F. Gruau and J. T. Tromp. Cellular gravity. In 741, page~10. Centrum voor Wiskunde en Informatica (CWI), ISSN 1386-3681, 1999.]]Google Scholar
- N. Halbwachs, P. Caspi, P. Raymond, and D. Pilaud. The synchronous data-flow programming language LUSTRE. Proceedings of the IEEE, 79(9):1305--1320, September 1991.]]Google ScholarCross Ref
- C. Hankin, D. Le Métayer, and D. Sands. Refining multiset transformers. Theoretical Computer Science, 192(2):233--258, 1998.]] Google ScholarDigital Library
- T. H. Heil, Z. Smith, and J. E. Smith. Improving branch predictors by correlating on data values. In Proceedings of the 32nd annual ACM/IEEE international symposium on Microarchitecture, pages 28--37. IEEE Computer Society, 1999.]] Google ScholarDigital Library
- G. Hinton, D. Sager, M. Upton, D. Boggs, D. Carmean, A. Kyker, , and P. Roussel. The microarchitecture of the pentium 4 processor. Intel Technology Journal Q1, 2001.]]Google Scholar
- J.W. Klop. Term rewriting systems. In Handbook of Logic in Computer Science, volume 2. Clarendon Press, 1992.]] Google ScholarDigital Library
- H.T. Kung and C.E. Leiserson. Systolic array for VLSI. In Addison Wesley, editor, Introduction to VLSI systems, C. A. Mead and L. A. Conway, 1980.]]Google Scholar
- J. Mazoyer. Computations on one dimensional cellular automata. Annals of Mathematics and Artificial Intelligence, 16:285--309, 1996.]]Google ScholarCross Ref
- J. Moreira. Memory models for the Bluegene/L Supercomputer. http://www.dagstuhl.de/03431/Proceedings/, 2003.]]Google Scholar
- R. Nagarajan, K. Sankaralingam, D. Burger, and S. W. Keckler. A design space evaluation of grid processor architectures. In Proceedings of the 34th annual ACM/IEEE international symposium on Microarchitecture, pages 40--51. IEEE Computer Society, 2001.]] Google ScholarDigital Library
- A. Narayanan and S. Zorbalas. DNA algorithms for computing shortest paths. In John~R. Koza, Wolfgang Banzhaf, Kumar Chellapilla, Kalyanmoy Deb, Marco Dorigo, David~B. Fogel, Max~H. Garzon, David~E. Goldberg, Hitoshi Iba, and Rick Riolo, editors, Genetic Programming 1998: Proceedings of the Third Annual Conference, pages 718--724, University of Wisconsin, Madison, Wisconsin, USA, 22-25 1998. Morgan Kaufmann.]]Google Scholar
- B. Nitzberg and V. Lo. Distributed shared memory: A survey of issues and algorithms. In IEEE Computer, pages 52--60, 1991.]] Google ScholarDigital Library
- G. M. Papadopoulos and D. E. Culler. Monsoon: an explicit token-store architecture. In Proceedings of the 17th annual international symposium on Computer Architecture, pages 82--91. ACM Press, 1990.]] Google ScholarDigital Library
- G. Paun. Computing membranes. Journal of Computer and System Sciences, 1(61):108--143, 2000.]]Google ScholarDigital Library
- J. A. Rose, Y. Gao, M. Garzon, and R. C. Murphy. DNA implementation of finite-state machines. In John R. Koza, Kalyanmoy Deb, Marco Dorigo, David B. Fogel, Max Garzon, Hitoshi Iba, and Rick~L. Riolo, editors, Genetic Programming 1997: Proceedings of the Second Annual Conference, pages 479--490, Stanford University, CA, USA, 13-16 1997. Morgan Kaufmann.]]Google Scholar
- E. Rotenberg, S. Bennett, and J. E. Smith. Trace cache: A low latency approach to high bandwidth instruction fetching. In International Symposium on Microarchitecture, pages 24--35, 1996.]] Google ScholarDigital Library
- K. Sankaralingam, R. Nagarajan, H. Liu, C. Kim, J. Huh, D. Burger, S. W. Keckler, and C. R. Moore. Exploiting ilp, tlp, and dlp with the polymorphous TRIPS architecture. In Proceedings of the 30th annual international symposium on Computer architecture, pages 422--433. ACM Press, 2003.]] Google ScholarDigital Library
- V. Sarkar. Partitioning and scheduling parallel programs for multiprocessors. MIT Press, 1989.]] Google ScholarDigital Library
- G. Sassatelli, L. Torres, P. Benoit, T. Gil., C. Diou, G. Cambon, and J. Galy. Higly scalable dynamically reconfigurable systollic ring-architecture for DSP application. In Design Automation and Test in Europe Conference and Exhibition, march 2002.]] Google ScholarDigital Library
- H. Sharangpani and K. Arora. Itanium processor microarchitecture. IEEE Micro, 20(5):24--43, october 2000.]] Google ScholarDigital Library
- P. W. Shor. Algorithms for quantum computation: Discrete logarithms and factoring. In IEEE Symposium on Foundations of Computer Science, pages 124--134, 1994.]]Google ScholarDigital Library
- M. Sipper. Co-evolving non-uniform cellular automata to perform computations. Physica D, 92:193--208, 1996.]] Google ScholarDigital Library
- R. Stadler, S. Ami, M. Forshaw, and C. Joachim. A memory/adder model based on single c60 molecular transistors. Nanotechnology, 12:350--357, 2001.]]Google ScholarCross Ref
- A. Terechko, E. Le Thenaff, M. Garg, and J. van Eijndhoven. Inter-cluster communication models for clustered VLIW processors. In Proceedings of the 9th International Symposium on High-Performance Computer Architecture, pages 354--364, February 2003.]] Google ScholarDigital Library
- W. Thies, M. Karczmarek, and S. P. Amarasinghe. StreamIt: A language for streaming applications. In Computational Complexity, pages 179--196, 2002.]] Google ScholarDigital Library
- C. D. Thompson. The VLSI complexity of sorting. IEEE Transactions on Computers, C-32(12):1171--1184, 1983.]]Google ScholarDigital Library
- T. Toffoli. Programmable matter methods. Future Generation Computer Systems, 16(2--3):187--201, 1999.]] Google ScholarDigital Library
- T. Toffoli and T. Bach. A common language for "programmable matter" (cellular automata and all that). Bulletin of the Italian Association for Artificial Intelligence, March 2001.]]Google Scholar
- D. M. Tullsen, S. J. Eggers, and H. M. Levy. Simultaneous multithreading: maximizing on-chip parallelism. In 25 years of the international symposia on Computer architecture (selected papers), pages 533--544. ACM Press, 1998.]] Google ScholarDigital Library
- A. M. Turing. The chemical basis of morphogenesis. Phil. Trans. Roy. Soc. of London, Series B: Biological Sciences(237):37--72, 1952.]]Google Scholar
- P. M. B. Vitanyi. Locality, communication, and interconnect length in multi-computers. SIAM Journal of computing, 1988.]] Google ScholarDigital Library
- J. Wahle, L. Neubert, J. Esser, and M. Schreckenberg. A cellular automaton traffic flow model for online simulation of traffic. Parallel Computing, 27(5):719--735, 2001.]] Google ScholarDigital Library
- Z. A. Ye, A. Moshovos, S. Hauck, and P. Banerjee. Chimaera: a high-performance architecture with a tightly-coupled reconfigurable functional unit. In Proceedings of the 27th annual international symposium on Computer architecture, pages 225--235. ACM Press, 2000.]] Google ScholarDigital Library
Index Terms
- BLOB computing
Recommendations
A class of scalable architectures for high-performance, cost-effective parallel computing
SPDP '94: Proceedings of the 1994 6th IEEE Symposium on Parallel and Distributed ProcessingThe family of reconfigurable generalized hypercube (RGH) architectures is proposed for the construction of scalable parallel computers. The objective is to reduce the high VLSI complexity of generalized hypercubes while maintaining to high extent their ...
Applying developmental-inspired principles to the field of developmental robotics
PerMIS '08: Proceedings of the 8th Workshop on Performance Metrics for Intelligent SystemsBio-inspired principles of development and evolution are a special part of the bio-models and principles that can be used to improve intelligent systems (IS). Such principles are central to cognitive developmental robotics (DR) and its effort to ...
Parallel Computer Vision on a Reconfigurable Multiprocessor Network
A novel reconfigurable architecture based on a multiring multiprocessor network is described. The reconfigurability of the architecture is shown to result in a low network diameter and also a low degree of connectivity for each node in the network. The ...
Comments