Technology Trends and Adaptive Computing
System and processor architectures depend on changes in technology. Looking ahead as die density and speed increase, power consumption and on chip interconnection delay become increasingly important in defining architecture tradeoffs. While technology improvements enable increasingly complex processor implementations, there are physical and program behavior limits to the usefulness of this complexity at the processor level. The architecture emphasis then shifts to the system: integrating controllers, signal processors, and other components with the processor to achieve enhanced system performance. In dealing with these elements, adaptability or reconfiguration is essential for optimizing system performance in a changing application environment. A hierarchy of adaptation is proposed based on flexible processor architectures, traditional FPL, and a new coarse grain adaptive arithmetic cell. The adaptive arithmetic cell offers high-performance arithmetic operations while providing computational flexibility. The proposed cell offers efficient and dynamic reconfiguration of the arithmetic units. Hybrid fine and coarse grain techniques may offer the best path to the continued evolution of the processor-based system.
KeywordsArithmetic Operation Technology Trend Instruction Level Parallelism Memory Access Time Adaptive Computing
Unable to display preview. Download preview PDF.
- 1.The National Technology Roadmap for Semiconductors, tech report, Semiconductor Industry Association, San Jose, CA, 1997.Google Scholar
- 3.Hung, P., and Flynn, M.: “Optimum ILP for Superscalar and VLIW Processors,” tech report, Stanford University, EE Dept., CSL-TR-99-737, 1999Google Scholar
- 4.Bakoglu, H.: Circuit, Interconnections, and Packaging for VLSI, Addision-Wesley, Reading, Mass., 1990.Google Scholar
- 6.Schwarz, E.: High Radix Algorithms for Higher-Order Arithmetic Operations, PhD thesis, Stanford University, January 1993.Google Scholar
- 8.Mandelbaum, D.: “A Method for Calculation of the Square Root Using Combinatorial Logic,” Journal of VLSI Signal Processing, December 1993, pp. 233–243.Google Scholar
- 9.Mandelbaum, D.: “A Fast, Efficient, Parallel-Acting Method of Generating Functions Defined by Power Series,” IEEE Trans. on Parallel and Distributed Systems, January 1996, pp. 33–45.Google Scholar
- 10.Liddicoat, A., Flynn, M.: “Reciprocal Approximation Theory with Table Compensation,” In SPIE’s 45th Annual Conference, July 2000.Google Scholar
- 11.Al-Twaijry, H.: Technology Scaling Effects on Multipliers, PhD thesis, Stanford University, November 1998.Google Scholar
- 12.DasSarma, D., and Matula, D.: “Faithful Bipartite ROM Reciprocal Tables,” In Proc. 12th IEEE Symp. Computer Arithmetic, July 1995, pp. 17–28.Google Scholar
- 13.DasSarma, D., and Matula, D.: “Faithful Interpolation in Reciprocal Tables,” In Proc. 13th IEEE Symp. Computer Arithmetic, July 1997, pp. 82–91.Google Scholar
- 14.Takagi, N.: “Generating a Power of an Operand by a Table Look-Up and a Multiplication,” In Proc. 13th IEEE Symp. Computer Arithmetic, July 1997, pp. 126–131.Google Scholar
- 15.Takagi, N.: “Powering by a Table Look-Up and a Multiplication with Operand Modification,” IEEE Trans. on Computers, C-47(11):1216–1222, November 1998.Google Scholar