Computational lithography
Computational lithography (also known as computational scaling) is the set of mathematical and algorithmic approaches designed to improve the resolution attainable through photolithography. Computational lithography came to the forefront of photolithography technologies in 2008 when the semiconductor industry faced challenges associated with the transition to a 22 nanometer CMOS microfabrication process and has become instrumental in further shrinking the design nodes and topology of semiconductor transistor manufacturing.
History
Computational lithography means the use of computers to simulate printing of micro-lithography structures. Pioneering work was done by Chris Mack at NSA in developing PROLITH, Rick Dill at IBM and Andy Neureuther at University of California, Berkeley from the early 1980s. These tools were limited to lithography process optimization as the algorithms were limited to a few square micrometres of resist. Commercial full-chip optical proximity correction (OPC), using model forms, was first implemented by TMA (now a subsidiary of Synopsys) and Numerical Technologies (also part of Synopsys) around 1997.[1]
Since then the market and complexity has grown significantly. With the move to sub-wavelength lithography at the 180 nm and 130 nm nodes, RET techniques such as Assist features, phase shift masks started to be used together with OPC. For the transition from 65 nm to 45 nm nodes customers were worrying that not only that design rules were insufficient to guarantee printing without yield limiting hotspots, but also that tape-out time may need thousands of CPUs or weeks of run time. This predicted exponential increase in computational complexity for mask synthesis on moving to the 45 nm process node spawned a significant venture capital investment in design for manufacturing start-up companies.[2]
A number of startup companies promoting their own disruptive solutions to this problem started to appear, techniques from custom hardware acceleration to radical new algorithms such as inverse lithography were touted to resolve the forthcoming bottlenecks. Despite this activity, incumbent OPC suppliers were able to adapt and keep their major customers, with RET and OPC being used together as for previous nodes, but now on more layers and with larger data files, and turn around time concerns were met by new algorithms and improvements in multi-core commodity processors. The term computational lithography was first used by Brion Technology (now a subsidiary of ASML) in 2005[3] to promote their hardware accelerated full chip lithography simulation platform. Since then the term has been used by the industry to describe full chip mask synthesis solutions. As 45 nm goes into full production and EUV lithography introduction is delayed, 32 nm and 22 nm are expected to run on existing 193 nm scanners technology.
Now, not only are throughput and capabilities concerns resurfacing, but also new computational lithography techniques such as Source Mask Optimization (SMO) is seen as a way to squeeze better resolution specific to a given design. Today, all the major mask synthesis vendors have settled on the term "computational lithography" to describe and promote the set of mask synthesis technologies required for 22 nm.
Techniques comprising computational lithography
Computational lithography makes use of a number of numerical simulations to improve the performance (resolution and contrast) of cutting-edge photomasks. The combined techniques include Resolution Enhancement Technology (RET), Optical Proximity Correction (OPC), Source Mask Optimization (SMO), etc.[4] The techniques vary in terms of their technical feasibility and engineering sensible-ness, resulting in the adoption of some and the continual R&D of others.[5]
Resolution enhancement technology
Resolution enhancement technologies, first used in the 90 nanometer generation, using the mathematics of diffraction optics to specify multi-layer phase-shift photomasks that use interference patterns in the photomask that enhance resolution on the printed wafer surface.
Optical proximity correction
Optical proximity correction uses computational methods to counteract the effects of diffraction-related blurring and under-exposure by modifying on-mask geometries with means such as: adjusting linewidths depending on the density of surrounding geometries (a trace surrounded by a large open area will be over-exposed compared with the same trace surrounded by a dense pattern), adding "dog-bone" endcaps to the end of lines to prevent line shortening, correcting for electron beam proximity effects
OPC can be broadly divided into rule-based and model-based.[6] Inverse lithography technology, which treats the OPC as an inverse imaging problem, is also a useful technique because it can provide unintuitive mask patterns.[7]
Complex modeling of the lens system and photoresist
Beyond the models used for RET and OPC, computational lithography attempts to improve chip manufacturability and yields such as by using the signature of the scanner to help improve accuracy of the OPC model:[8] polarization characteristics of the lens pupil, Jones matrix of the stepper lens, optical parameters of the photoresist stack, diffusion through the photoresist, stepper illumination control variables.
Computational effort
The computational effort behind these methods is immense. According to one estimate, the calculations required to adjust OPC geometries to take into account variations to focus and exposure for a state-of-the-art integrated circuit will take approximately 100 CPU-years of computer time.[9] This does not include modeling the 3D polarization of the light source or any of the several other systems that need to be modeled in production computational photolithographic mask making flows. Brion Technologies, a subsidiary of ASML, markets a rack-mounted hardware accelerator dedicated for use in making computational lithographic calculations — a mask-making shop can purchase a large number of their systems to run in parallel. Others have claimed significant acceleration using re-purposed off-the-shelf graphics cards for their high parallel throughput.[10]
193 nm deep UV photolithography
The periodic enhancement in the resolution achieved through photolithography has been a driving force behind Moore's Law. Resolution improvements enable printing of smaller geometries on an integrated circuit. The minimum feature size that a projection system typically used in photolithography can print is given approximately by:
where
- is the minimum feature size (also called the critical dimension).
- is the wavelength of light used.
- is the numerical aperture of the lens as seen from the wafer.
- (commonly called k1 factor) is a coefficient that encapsulates process-related factors.
Historically, resolution enhancements in photolithography have been achieved through the progression of stepper illumination sources to smaller and smaller wavelengths — from "g-line" (436 nm) and "i-line" (365 nm) sources based on mercury lamps, to the current systems based on deep ultraviolet excimer lasers sources at 193 nm. However the progression to yet finer wavelength sources has been stalled by the intractable problems associated with extreme ultraviolet lithography and x-ray lithography, forcing semiconductor manufacturers to extend the current 193 nm optical lithography systems until some form of next-generation lithography proves viable (although 157 nm steppers have also been marketed, they have proven cost-prohibitive at $50M each).[11] Efforts to improve resolution by increasing the numerical aperture have led to the use of immersion lithography. As further improvements in resolution through wavelength reduction or increases in numerical aperture have become either technically challenging or economically unfeasible, much attention has been paid to reducing the k1-factor. The k1 factor can be reduced through process improvements, such as phase-shift photomasks. These techniques have enabled photolithography at the 32 nanometer CMOS process technology node using a wavelength of 193 nm (deep ultraviolet). However, with the ITRS roadmap calling for the 22 nanometer node to be in use by 2011, photolithography researchers have had to develop an additional suite of improvements to make 22 nm technology manufacturable.[12] While the increase in mathematical modeling has been underway for some time, the degree and expense of those calculations has justified the use of a new term to cover the changing landscape: computational lithography.
References
- "Major US Semiconductor Maker Chooses TMA for OPC Software", PRNewswire, 1997-10-16
- McGrath, Dylan (2005-12-16), "DFM pumps up the volume", EETimes
- McGrath, Dylan (2005-02-12), "Litho simulation vendor opens Japanese subsidiary", EETimes
- LaPedus, Mark (2008-09-17), "IBM rolls 'computational scaling' for litho at 22-nm", EETimes
- E. Lam; A. Wong (2009), "Computation lithography: virtual reality and virtual virtuality", Optics Express, 17 (15): 12259–12268, Bibcode:2009OExpr..1712259L, doi:10.1364/OE.17.012259, hdl:10722/62090, PMID 19654627
- A. Wong (2001), Resolution enhancement techniques in optical lithography, SPIE Press
- S. Chan; A. Wong; E. Lam (2008), "Initialization for robust inverse synthesis of phase-shifting masks in optical projection lithography", Optics Express, 16 (19): 14746–14760, Bibcode:2008OExpr..1614746C, doi:10.1364/OE.16.014746, PMID 18795012
- Hand, Aaron (November 2007), "Nikon and Synopsys Deliver on Advanced OPC Promise", Semiconductor International, archived from the original on 2009-08-09, retrieved 2010-01-15
- Wiley, Jim (May 2006), "Future challenges in computational lithography", Solid State Technology
- LaPedus, Mark (2008-02-25), "Gauda claims OPC acceleration breakthrough.", EE Times
- "Reticle enhancement technology will extend life of 193nm litho", Electronics Weekly, 2004-02-25
- Moretti, Gabe (2008-10-13), "Custom litho addresses 22-nm IC manufacturing", EETimes, archived from the original on 2013-01-22