Layout Decompression Chip for Maskless Lithography

Similar documents
Lossless Compression Algorithms for Direct- Write Lithography Systems

Hardware Implementation of Block GC3 Lossless Compression Algorithm for Direct-Write Lithography Systems

Hardware Implementation of Block GC3 Lossless Compression Algorithm for Direct-Write Lithography Systems

Architecture and Hardware Design of Lossless Compression Algorithms for Direct-Write Maskless Lithography Systems. Hsin-I Liu

Architecture and Hardware Design of Lossless Compression Algorithms for Direct-Write Maskless Lithography Systems

THE USE OF forward error correction (FEC) in optical networks

A video signal processor for motioncompensated field-rate upconversion in consumer television

VLSI Design: 3) Explain the various MOSFET Capacitances & their significance. 4) Draw a CMOS Inverter. Explain its transfer characteristics

A Fast Constant Coefficient Multiplier for the XC6200

Objectives. Combinational logics Sequential logics Finite state machine Arithmetic circuits Datapath

CHAPTER 6 ASYNCHRONOUS QUASI DELAY INSENSITIVE TEMPLATES (QDI) BASED VITERBI DECODER

Optimization of Multi-Channel BCH Error Decoding for Common Cases. Russell Dill Master's Thesis Defense April 20, 2015

A VLSI Architecture for Variable Block Size Video Motion Estimation

Implementation of an MPEG Codec on the Tilera TM 64 Processor

ALONG with the progressive device scaling, semiconductor

Design Project: Designing a Viterbi Decoder (PART I)

A Low Power Delay Buffer Using Gated Driver Tree

Contents Circuits... 1

Microprocessor Design

International Journal of Engineering Trends and Technology (IJETT) - Volume4 Issue8- August 2013

Advanced Data Structures and Algorithms

Reconfigurable Neural Net Chip with 32K Connections

PARALLEL PROCESSOR ARRAY FOR HIGH SPEED PATH PLANNING

PICOSECOND TIMING USING FAST ANALOG SAMPLING

VHDL Design and Implementation of FPGA Based Logic Analyzer: Work in Progress

Low Power VLSI Circuits and Systems Prof. Ajit Pal Department of Computer Science and Engineering Indian Institute of Technology, Kharagpur

nmos transistor Basics of VLSI Design and Test Solution: CMOS pmos transistor CMOS Inverter First-Order DC Analysis CMOS Inverter: Transient Response

Multi-Shaped E-Beam Technology for Mask Writing

A Symmetric Differential Clock Generator for Bit-Serial Hardware

An Alternative Architecture for High Performance Display R. W. Corrigan, B. R. Lang, D.A. LeHoty, P.A. Alioshin Silicon Light Machines, Sunnyvale, CA

Implementation of CRC and Viterbi algorithm on FPGA

Data flow architecture for high-speed optical processors

High Performance Microprocessor Design and Automation: Overview, Challenges and Opportunities IBM Corporation

Asynchronous IC Interconnect Network Design and Implementation Using a Standard ASIC Flow

Combinational vs Sequential

EEC 116 Fall 2011 Lab #5: Pipelined 32b Adder

CacheCompress A Novel Approach for Test Data Compression with cache for IP cores

Sharif University of Technology. SoC: Introduction

FPGA Implementation of Convolutional Encoder And Hard Decision Viterbi Decoder

IMPLEMENTATION OF X-FACTOR CIRCUITRY IN DECOMPRESSOR ARCHITECTURE

COE328 Course Outline. Fall 2007

8/30/2010. Chapter 1: Data Storage. Bits and Bit Patterns. Boolean Operations. Gates. The Boolean operations AND, OR, and XOR (exclusive or)

COPY RIGHT. To Secure Your Paper As Per UGC Guidelines We Are Providing A Electronic Bar Code

Power Reduction Techniques for a Spread Spectrum Based Correlator

ECSE-323 Digital System Design. Datapath/Controller Lecture #1

Challenges in the design of a RGB LED display for indoor applications

Advanced Training Course on FPGA Design and VHDL for Hardware Simulation and Synthesis. 26 October - 20 November, 2009

VGA Controller. Leif Andersen, Daniel Blakemore, Jon Parker University of Utah December 19, VGA Controller Components

FPGA Based Implementation of Convolutional Encoder- Viterbi Decoder Using Multiple Booting Technique

Optimization of memory based multiplication for LUT

Implementation of Memory Based Multiplication Using Micro wind Software

LUT Optimization for Memory Based Computation using Modified OMS Technique

The Design of Efficient Viterbi Decoder and Realization by FPGA

A Low-Power 0.7-V H p Video Decoder

An Overview of the Performance Envelope of Digital Micromirror Device (DMD) Based Projection Display Systems

Reduction of Clock Power in Sequential Circuits Using Multi-Bit Flip-Flops

CMOS Design Analysis of 4 Bit Shifters 1 Baljot Kaur, M.E Scholar, Department of Electronics & Communication Engineering, National

ANALYSIS OF POWER REDUCTION IN 2 TO 4 LINE DECODER DESIGN USING GATE DIFFUSION INPUT TECHNIQUE

Investigation of Look-Up Table Based FPGAs Using Various IDCT Architectures

Recent results of Multi-beam mask writer MBM-1000

Use of Low Power DET Address Pointer Circuit for FIFO Memory Design

Using Embedded Dynamic Random Access Memory to Reduce Energy Consumption of Magnetic Recording Read Channel

Field Programmable Gate Arrays (FPGAs)

Efficient Architecture for Flexible Prescaler Using Multimodulo Prescaler

BITSTREAM COMPRESSION TECHNIQUES FOR VIRTEX 4 FPGAS

The Matched Delay Technique: Wentai Liu, Mark Clements, Ralph Cavin III. North Carolina State University. (919) (ph)

Auto classification and simulation of mask defects using SEM and CAD images

Hardware Implementation of Viterbi Decoder for Wireless Applications

Random Access Scan. Veeraraghavan Ramamurthy Dept. of Electrical and Computer Engineering Auburn University, Auburn, AL

An FPGA Implementation of Shift Register Using Pulsed Latches

A low-power portable H.264/AVC decoder using elastic pipeline

A LOW COMPLEXITY CODE COMPRESSION BASED ON HYBRID RLC-BM CODES

Reading an Image using CMOS Linear Image Sensor. S.R.Shinthu 1, P.Maheswari 2, C.S.Manikandababu 3. 1 Introduction. A.

R Fig. 5 photograph of the image reorganization circuitry. Circuit diagram of output sampling stage.

Part 1: Introduction to Computer Graphics

Inspection of 32nm imprinted patterns with an advanced e-beam inspection system

MUHAMMAD NAEEM LATIF MCS 3 RD SEMESTER KHANEWAL

Low-Power Decimation Filter for 2.5 GHz Operation in Standard-Cell Implementation

Gated Driver Tree Based Power Optimized Multi-Bit Flip-Flops

[Krishna*, 4.(12): December, 2015] ISSN: (I2OR), Publication Impact Factor: 3.785

Pattern Smoothing for Compressed Video Transmission

Digital Blocks Semiconductor IP

NH 67, Karur Trichy Highways, Puliyur C.F, Karur District UNIT-III SEQUENTIAL CIRCUITS

Achieving Faster Time to Tapeout with In-Design, Signoff-Quality Metal Fill

Scan. This is a sample of the first 15 pages of the Scan chapter.

A Low Power Implementation of H.264 Adaptive Deblocking Filter Algorithm

REDUCING DYNAMIC POWER BY PULSED LATCH AND MULTIPLE PULSE GENERATOR IN CLOCKTREE

High Speed Reconfigurable FPGA Architecture for Multi-Technology Applications

Interframe Bus Encoding Technique for Low Power Video Compression

Experiment: FPGA Design with Verilog (Part 4)

AN OPTIMIZED IMPLEMENTATION OF MULTI- BIT FLIP-FLOP USING VERILOG

Contents Slide Set 6. Introduction to Chapter 7 of the textbook. Outline of Slide Set 6. An outline of the first part of Chapter 7

Design of BIST with Low Power Test Pattern Generator

DESIGN AND SIMULATION OF A CIRCUIT TO PREDICT AND COMPENSATE PERFORMANCE VARIABILITY IN SUBMICRON CIRCUIT

Design and FPGA Implementation of 100Gbit/s Scrambler Architectures for OTN Protocol Chethan Kumar M 1, Praveen Kumar Y G 2, Dr. M. Z. Kurian 3.

Chapter 7 Memory and Programmable Logic

Area-efficient high-throughput parallel scramblers using generalized algorithms

CHAPTER 6 DESIGN OF HIGH SPEED COUNTER USING PIPELINING

A CYCLES/MB H.264/AVC MOTION COMPENSATION ARCHITECTURE FOR QUAD-HD APPLICATIONS

CSCB58 - Lab 4. Prelab /3 Part I (in-lab) /1 Part II (in-lab) /1 Part III (in-lab) /2 TOTAL /8

Transcription:

Layout Decompression Chip for Maskless Lithography Borivoje Nikolić, Ben Wild, Vito Dai, Yashesh Shroff, Benjamin Warlick, Avideh Zakhor, William G. Oldham Department of Electrical Engineering and Computer Sciences University of California, Berkeley, CA, USA, 94720-770 ABSTRACT Future maskless lithography systems require data throughputs of the order of tens of terabits per second in order to have comparable performance to today s mask-based lithography systems. This work presents an approach to overcome the throughput problem by compressing the layout data and decompressing it on the chip that interfaces to the writers. To achieve the required throughput, many decompression paths have to operate in parallel. The concept is demonstrated by designing an interface chip for layout decompression, consisting of a Huffman decoder and a Lempel-Ziv systolic decompressor. The 5.5mm x 2.5mm prototype chip, implemented in a 0.µm,.V CMOS process is fully functional at 00MHz dissipating 30mW per decompression row. By scaling the chip size up and implementing it in a 65nm technology, the decompressed data throughput required for writing 60 wafers per hour in 45nm technology is feasible. Keywords: Data compression, maskless, lithography, CMOS, digital design.. INTRODUCTION As minimum feature sizes in CMOS technology scale, the cost of critical dimension masks dramatically increases. Mask costs in 90nm technology are exceeding million dollars. An alternative to mask-based optical lithography is maskless lithography, where the layout data is directly written onto a wafer. Various approaches have been investigated, including e-beam, micro-machined mirror projection, and nano-jet printing []. To achieve the required nm edge placement with 25nm pixels in 45nm technology, a 5-bit per pixel data representation is needed, resulting in a total of over 500Tb of information on a 300mm wafer. To be competitive with conventional optical systems, any future lithography system should be capable of projecting one layer per minute, resulting in approximately 2Tb/s of raw data throughput. Adding the necessary redundancy and communications overhead to the data stream would likely increase the required throughput by another 25%, reaching 5Tb/s ranges. This paper presents a maskless lithography interface and circuitry that provide the required throughput. It is designed to work with a micro-machined mirror array integrated on a chip exposed by extreme ultra-violet (EUV) light in a conceptual system as shown in Figure. In this approach, the conventional mask is replaced by a programmable one, which is reprogrammed between consecutive light flashes. Either storing or continuously feeding the required amount of data to the writer chip is not feasible. The key idea of the work presented here is the compression of rasterized data on a main storage unit and the continuous decompression on-the-fly, as the data is being downloaded on the writer [2]. With compression ratios of about 20, the chip input bandwidth would be 00Gb/s [3]. This is feasible in 90nm or 65nm technologies, which could be used for processing data with 45nm feature sizes. The basic design of a data processing system capable of delivering tera-pixel data rates necessary to achieve nextgeneration maskless lithography is shown in Figure 2. This design consists of storage disks, a processor board with memory, and a decoder-writer chip with data-decoding circuitry fabricated together with a massive array of pixel writers. Layout data for all layers of a single chip is compressed off-line and stored on the disks. Before the writing process begins, only a single compressed layer is transferred from disks to the processor board memory and stored there. As the writers write a stripe across the wafer, compressed data is streamed from the processor board to the decoder-writer chip in real-time as needed. The on-chip decoding circuitry, in real-time, expands the compressed data stream into the data signals necessary to control the writers. The key challenge for such a system is the design of a high-throughput on-chip data decompression architecture and the circuitry that implements it. This paper demonstrates the design of such a highthroughput decompression chip. To simplify the solution, a binary interface to the writers using an SRAM array is implemented. This avoids the handling of 5-bit grayscale values that would require analog control of mirror positions 092 Emerging Lithographic Technologies VIII, edited by R. Scott Mackay, Proceedings of SPIE Vol. 5374 (SPIE, Bellingham, WA, 2004) 0277-76X/04/$5 doi: 0.7/2.5357

using analog memory. However this is not a good choice for a practical solution as it results in 6 times higher required data throughput (5-bit grayscale codes are replaced with 32-bit thermometer codes). 3nm EUV light-source Condenser Optics Mechanical scan Wafer Electronic mask Electronic scroll of pattern Imaging Optics Electronic mask is an array of tiny mirrors (0 5-0 mirrors) Figure. Maskless writing using micromirrors. Gb/s 00Gb/s 5Tb/s Storage Disks 40 GB/layer Processor Processor Board Board 64 4 GBit GB Memory Memory Decompress Writers 20 to all compressed layers 20 to single compressed stripe -Writer Chip Figure 2. Architecture of a data-delivery system for maskless lithography. 00 Gb/s Huffman Variable Rate Buffer LZ Match 5 Tb/s Huffman Variable Rate Buffer LZ Match Figure 3. Multiple LZ77 decoders operating in parallel. Proc. of SPIE Vol. 5374 093

2. DESIGN OF A HIGH-THROUGHPUT DECODER When the data compression and decompression algorithms are being designed, they are optimized for operation on a general-purpose microprocessor, and rarely a hard-wired hardware implementation is targeted. The primary difficulty of using compression to solve the data transmission problem associated with maskless lithography is that in order to be effective, the compressed representation must be decoded in a timely manner to sustain the required throughput, with reasonable computational and memory resources. To achieve this, a compression algorithm has to be designed, and then mapped into an architecture and an implementation that expands the compressed data at a rate of 00 Gb/s to 5 Tb/s of output data. We consider the design of a high-speed hardware decoder based on ZIP, the simplest compression algorithm presented in prior research [4] which achieves good compression on lithography data. The basic algorithm behind ZIP compression is Lempel-Ziv 977 (LZ77) [5], and the design parameters of a LZ77 decoder are optimized to minimize circuit usage while maintaining compression efficiency. 2.. Fast LZ77 Decoding Architecture The design of a fast LZ77 decoder, shown in Figure 3, consists of two decoding blocks, a Huffman decoder, and an LZ match decoder. Because each decoder block has the potential to expand the data, and because the expansion factor varies in time with the compression ratio of the data, buffers are necessary to smooth both input and output data rates. In general, this would require three buffers, one in front of the Huffman decoder, one after the match decoder, and one between the two decoders. To reduce the buffering requirements, we choose an implementation of the Huffman decoding algorithm which observes a constant input rate, and we choose an implementation of the LZ77 decoding algorithm which observes a constant output rate. Consequently the variability in data rate is completely absorbed by a single buffer, capable of both variable input and output data rates. For simplicity, a Huffman decoder with an input data rate of bit per cycle is chosen. Also, a match decoder with a constant output rate of one pixel per cycle is chosen which corresponds to an output rate of 5 bits per cycle for 32 gray level data. To keep the operating frequencies within reasonable limits, a Huffman algorithm that decodes m-bits per cycle and match decoders that output n-pixels per cycle, are used, with a corresponding increase in hardware complexity. The simpler alternative which we adopt is to use k independent LZ77 decoders to multiply both input and output data rate by a factor of k. To accomplish this, the layout data must be divided into blocks and compressed independently, Figure 3, though this may come at some cost to compression efficiency as discussed in [4]. The number of decompression paths needed depends on the maximum operating frequency of the Huffman decoder and LZ match decoder of the final design. 2.2. Huffman Decoding The essence of Huffman encoding is that it assigns shorter codewords to more frequent data, thus reducing the average number of bits required for representation. With layout data tested in [4] it achieves a typical compression ratio of approximately 5. This implementation uses the canonical Huffman table because of its simple representation that lends itself to a less complex decoder implementation. The algorithm description can be found in [6]. The Huffman decoder architecture is shown in Figure 4. The coded data is input sequentially into a shift register. For every bit that is shifted in, the counter is incremented. The output of the counter is used to address mincode, maxcode and index tables. The word in the shift register is compared to the output of the maxcode table. If the shift register word is less than the output of the maxcode table then the decoding is done. To get the decoded word, the shift register word is added to the output of the index table, and then subtracted from the output of the mincode table. This value is then used to address a symbol table which holds the decompressed symbols. To increase the decoding speed, the architecture is pipelined at the cut-sets shown in Figure 4, leaving only the symbol memory lookup time in the critical path. When implemented in 0.µm CMOS technology, the area of this Huffman decoder is about 6mm x 20µm. This large aspect ratio is chosen to allow operation of multiple parallel paths on the same chip. Scaling this design to 65nm technology would make its interfacing to the other parts of the decoder and the writer array feasible. 094 Proc. of SPIE Vol. 5374

Input Reset Shift Reg. Maxcode Table 6 Comparator Reset Counter code + + + + index len Index Table symaddr + Symbol Table output done Mincode Table mincode Pipeline registers Figure 4. Block diagram of the Huffman decoder. 2.3. Lempel-Ziv Matching Algorithm The Lempel-Ziv algorithm replaces repeating sequences of symbols with a pointer to a history buffer. This pointer indicates where in the buffer to retrieve the data (offset), and how many symbols to copy (length). Uncompressed symbols are literals. The architecture chosen for this design consists of a systolic array processor, where the data pass bidirectionally from one processing element to neighboring elements in a regular pattern [7]. Each processing element consists of one forward buffer and one reverse buffer and stores two -bit symbols. Data flows from left to right through the forward registers and then wraps back, flowing from right to left through the dictionary registers. Figure 5 illustrates the functionality of the systolic array processing elements. The decoding is done locally, avoiding any long wires to the history lookup buffer that would be unavoidable in the non-systolic implementation. To get better compression, the literal, offset, and length symbols are independently Huffman coded. To allow this, separate Huffman tables must be multiplexed in, depending on the current symbol being decoded. Furthermore, in the systolic implementation, a runlength decoder is used. This decoder simply sets a flag indicating when an <offset, length> pair is to be decoded by the processing elements. It transmits the offset symbol to the systolic array <length> times. If the input symbol is a literal then this is copied to the systolic array. Two critical parameters in implementing the LZ algorithm are the match length and the size of the buffer. Figure 6 shows the tradeoff between the compression ratio and the buffer size and match length. A buffer size of 024 and a match length of 256 achieve the near optimum compression ratios for layout data [4]. The systolic implementation of the LZ algorithm provides redundancy for the writing process. Since the complete pixel exposure is achieved by multiple EUV light flashes simultaneously with cycling the data through the systolic array, the writing process can be designed to compensate for non-functional mirrors. The layout of this LZ decoder with buffer length of 256 and match length of 256 is about 7.mm x 64µm. This layout aspect ratio is chosen to allow integration of 000 parallel decoding paths in a full-throughput chip in 65nm technology. It should be noted that LZ achieves better compression per unit area in the hardware implementation than the Huffman decoder. However the LZ algorithm alone does not achieve the desired compression ratio, so in this experiment we implemented both the LZ and the Huffman coding [4]. Proc. of SPIE Vol. 5374 095

PE N fdata copy PE k fdata copy PE 0 rdata rdata a) copy_in copy MEM_ID A A=B B rdata_out 0 rdata_in 0 fdata_in<0> fdata_out b) Figure 5. a) Systolic structure, consisting of processing elements, b) Processing element of LZ decompressor. 3. CHIP ARCHITECTURE To meet the throughput requirements for this application, many decompression paths must operate in parallel on the pattern generator chip. The block diagram of each decompression path is shown in Figure 7. The main blocks are a Huffman decoder and a Lempel-Ziv decoder. An asynchronous FIFO is required in between these blocks because the blocks operate at different, time-varying rates. A CRC block is implemented at the end of each frame to check for any data errors. The data is framed into blocks consisting of the 024 bytes of data, a -byte CRC check and an -byte synchronization block to synchronize the writer-interface circuitry to the writers. In the case of a micro-mirror based system, the writer-interface consists of a simple SRAM memory array. 3.. Design Process The design is captured as a high level Simulink model. Datapath blocks are created using the basic building blocks in Simulink such as registers, adders, and multipliers. Memory blocks are created using Matlab M-functions. The design is functionally verified in Simulink and then directly mapped into the standard cell library using the SSHAFT automated design flow []. The datapath blocks in Simulink are mapped into Synopsys Module Compiler blocks. Module Compiler then builds the netlists for the datapath blocks using the standard cell library and generates a VHDL description that can 096 Proc. of SPIE Vol. 5374

be simulated with the same test vectors used in Simulink to verify functionality of each of the datapath blocks. Memory blocks are created using memory generators. From the top-level dataflow graph in Simulink the decompression rows are mapped into standard cells and automatically placed and routed. compression ratio 90 0 70 60 50 40 30 20 0 0 4 6 Best Average Worst 64 256 024 4096 history buffer size (bytes) maximum match length = 256 a) compression ratio 200 0 60 40 20 00 0 60 40 20 0 32 64 Best Average Worst 2 256 52 024 maximum match length buffer size = 204 bytes b) 204 4096 92 Figure 6. Compression performance vs. a) history buffer or b) match length for 6 KB uncompressed data. Proc. of SPIE Vol. 5374 097

Compressed Data Huffman FIFO FIFO Literal/ Offset Length Lempel-Ziv CRC Check Decompressed Data Table select RD/WR SRAM Writer-Interface Lookup Tables Control Synch. Address 0 Figure 7. Decompressor row block diagram. Huffman Lookup Memory FIFO Array Huffman Lempel-Ziv SRAM Writer Interface Single Decompression Path Figure. Die photo of the prototype chip. 4. EXPERIMENTAL RESULTS To demonstrate the feasibility of this approach to data delivery for maskless lithography, a prototype chip in 5-metal.V 0.µm was designed. The chip consists of parallel decompression paths, each with a Huffman block, an LZ block consisting of a 256-byte history lookup and a 256 word FIFO. All the FIFOs and buffers are imported as SRAM blocks. The total chip size is 2.5x5.5mm 2. The history buffer is limited to 256 bytes due to prototype chip area constraints. The chip is fully functional at 00 MHz dissipating 30mW per decompression row. The chip has been designed to operate at 500MHz, although the test setup limits the verification up to 00MHz. Each decompression path loads bits of data per clock cycle; since the chip contains identical paths, a maximum output throughput of 32Gb/s can be achieved. The die photo is shown in Figure. To achieve the required throughput, a complete chip implemented in 65nm technology for printing 45nm features would require two rows of 024 parallel decompression paths, each 09 Proc. of SPIE Vol. 5374

operating at GHz. The LZ symbols would be 6 bits, with appropriately increased buffer size. The decompressed data will be de-multiplexed and fed to two 6,34 x,4 SRAM arrays that will interface to the mirrors. Table compares the performance of the test-chip versus that of the full-scale chip. By scaling up the die size to occupy a full reticle and scaling down the feature sizes to use a 65nm technology, the required throughput is achievable. 5. CONCLUSIONS A real-time layout decompression architecture and its implementation on a silicon chip were presented. The architecture consists of parallel decompression paths, where each path is composed of a Huffman decoder and a Lempel-Ziv systolic decompressor. A fully functional smaller scale prototype chip demonstrates the feasibility of this approach. A scaled up decompression chip would be able to deliver 32Tb/s of data to the array of 256 million mirrors. Besides the micromirror based approach, this architecture can be easily extended to other types of maskless lithography systems such as e- beam direct write, or to any other high data rate application where data compression is possible. ACKNOWLEDGEMENT This research is conducted under the Research Network for Advanced Lithography, supported jointly by the funding of the Semiconductor Research Corporation (0-MC-460) and the Defense Advanced Research Project Agency (MDA972-0--002). REFERENCES [] Y. Shroff, Y. Chen, W.G. Oldham, "Fabrication of parallel-plate nanomirror arrays for EUV maskless lithography", Journal of Vacuum Science and Technology, Nov. 200. [2] V. Dai and A. Zakhor, ``Lossless Layout Compression for Maskless Lithography Systems'' in Emerging Lithographic Technologies IV, Proceedings of the SPIE, San Jose, California, March 2000, Vol. 3997, pp. 467-477. [3] Rambus Redwood Interface, http://www.rambus.com/products/redwood/ [4] V. Dai and A. Zakhor, "Lossless compression techniques for maskless lithography data" in Emerging Lithographic Technologies VI, Proceedings of the SPIE, San Jose, California, March 2002, vol. 46, p. 53-594. [5] J. Ziv, and A. Lempel, A universal algorithm for sequential data compression, IEEE Transasctions on Information Theory, IT-23 (3), pp.337-43, 977. [6] J. Miano, Compressed Image File Formats, pp. 65-75, ACM Press, 999. [7] C. Chen, C. Wei, VLSI design for LZ-based data compression, IEE Proc. Circuits, Devices and Systems, vol. 46, no. 5, pp. 26-277, Oct. 999. [] W. R. Davis, et. al., "A Design Environment for High Throughput, Low Power Dedicated Signal Processing Systems", IEEE Journal of Solid-State Circuits, Mar. 2002. Table : Performance comparison: The test chip, the test chip scaled down to 65nm technology, 65nm design expanded to full reticle. Prototype in Prototype scaled to Full-scale writer 0.µm CMOS 65nm Technology 0nm 65nm 65nm Input bandwidth 00Mb/s N/A 400Gb/s Decompression paths 2 x 024 Path pitch 240µm 0µ 40µm History buffer 256 256 024 Max throughput 32Gb/s 96Gb/s 32Tb/s Mirror array x x 024 x x 024 2 x (6 x 024 x,4) Mirror size 3µm x 3µm µm x µm µm x µm Chip dimensions 2mm x 5mm 0.7mm x.7mm 24mm x 26mm Power 560mW 60mW ~5W Proc. of SPIE Vol. 5374 099