Postavke privatnosti

MIT Stacks Transistors and Memory in 3D to Reduce AI Energy Consumption in Data Centers

MIT researchers have developed a new platform for stacking transistors and memory in the back end of the chip, combining amorphous indium oxide and ferroelectric hafnium-zirconium oxide to accelerate AI calculations and reduce the growing energy consumption of data centers. The solution paves the way for more compact, efficient AI chips.

MIT Stacks Transistors and Memory in 3D to Reduce AI Energy Consumption in Data Centers
Photo by: Domagoj Skledar - illustration/ arhiva (vlastita)

AI-based computing is entering a phase where energy efficiency can no longer be treated as a side issue. Training and running generative artificial intelligence, deep learning, and computer vision models are pushing the electricity consumption of data centers toward levels comparable to the consumption of entire countries. Estimates show that data centers consumed about 415 terawatt-hours of electricity in 2024, and by the end of the decade, that figure could more than double. In this context, every watt saved at the chip level becomes strategically important.


A group of researchers from the Massachusetts Institute of Technology (MIT) has therefore developed a new platform for electronics integration that attempts to solve the problem where it originates – in the chip architecture itself. Instead of keeping logic components and memory separate, as in classic circuits, they stack them vertically into a compact, three-dimensional "stack" built on top of an existing circuit. The new approach allows transistors and memory elements to be placed on top of one another in the back end of the chip, thereby drastically shortening the path data travels and reducing energy losses.


The key to this technology is a combination of a new material – amorphous indium oxide – and a carefully optimized manufacturing process that works at significantly lower temperatures than classic CMOS technology. This opens the possibility of upgrading already finished silicon circuits with additional "floors" of logic and memory, without destroying transistors made in the front end of the line. The result is an electronic platform that promises faster computing, lower energy consumption, and higher integration density, particularly attractive for applications like generative AI, deep learning, and real-time image processing.


The bottleneck of classic architecture: separated logic and memory


Most of today's computing systems still rely on an architecture that places logic (the processor) and memory in physically separate blocks. Logic transistors execute instructions, while memory elements – whether SRAM, DRAM, or flash – serve to store data and model parameters. Every time the processor needs data, it must "travel" through a network of wires, interconnects, and buses. This constant traffic of information between logic and memory is known as the von Neumann bottleneck and is today one of the main sources of inefficiency.


As models become larger and the amount of data grows exponentially, the energy cost of moving bits itself becomes greater than the cost of computing. Analyses show that a significant part of the energy in data centers goes precisely to data transfer within the chip, between chips, and to and from external memory. Additionally, every extra hop between levels of the memory hierarchy introduces latency and limits the total speed of the system. Therefore, an increasing part of research is directed toward the concept of "in-memory" computing and toward three-dimensional integration that shortens the data path.


Standard CMOS chips are strictly divided into the front-end-of-line and back-end-of-line processes. In the front end (front-end-of-line), active components like transistors and capacitors are made, while the back end (back-end-of-line) serves to route wires, interconnects, and metallization that connect these components into functional circuits. Although the back-end process has seen significant improvements over the years, it still serves primarily as a "highway for electrons," rather than an active zone where calculation takes place.


The problem arises when we want to build additional active layers above the already formed transistors in the front end. Classic silicon transistors require high processing temperatures, often hundreds of degrees Celsius, which would irreversibly damage or destroy the devices located underneath. Because of this, vertical stacking of logic and memory within the same chip, with a density suitable for modern AI accelerators, has so far remained largely in the domain of theory and simulations.


New strategy: active components in the back end of the chip


The MIT team decided to "flip" the usual approach and use the back end of the process as a space for embedding additional active layers. Instead of trying to bake silicon again at high temperatures, the researchers developed transistors that can be made at temperatures around 150 degrees Celsius, directly onto the back-end structures. Thus, an additional layer of logic and memory is created that sits above the existing CMOS circuit but does not damage it.


The basic idea is that on an already finished chip, where classic silicon transistors are still responsible for part of the work, new layers of transistors and memory are added in the back end. These layers take over the most energy-intensive tasks – for example, local data processing and storage of neural network parameters – while the basic logic in the front end is used as the "anchor" of the entire system. The result is a hybrid architecture in which different materials and technologies complement each other.


For such an approach to succeed, it is necessary to find materials that retain good electronic characteristics even at very small dimensions, but can simultaneously be deposited at low temperatures. This is exactly where amorphous indium oxide comes onto the scene, a material that has attracted great attention in recent years as a candidate for transistors in the back-end-of-line process and in monolithic 3D integration.


Amorphous indium oxide: a channel two nanometers thick


Amorphous indium oxide belongs to the group of amorphous oxide semiconductors, materials that offer a combination of high charge mobility, good scalability, and the ability to be deposited in thin layers at relatively low temperatures. In the MIT paper, this material takes on the role of the active channel of transistors made in the back end of the chip. The channel is the layer where the key function of the transistor takes place – the controlled switching on and off of the flow of electrons between the source and the drain.


The researchers managed to form a layer of amorphous indium oxide about two nanometers thick, which is roughly a dozen atomic layers. In this regime of ultra-thin films, every defect in the material network has a relatively large impact on the behavior of the device. Particularly important are so-called oxygen vacancies – places in the network where an oxygen atom is missing – because precisely these defects can act as donor sites that provide free electrons needed for conducting current.


For the correct operation of the transistor, it is necessary to find a balance: too few vacancies mean that the channel is too weak a conductor and the device is hard to turn on, while too many defects lead to current leakage, instability, and increased energy consumption in the off state. The MIT team therefore devoted a large part of the work to precise control of the deposition process and subsequent processing of amorphous indium oxide, so that the number and distribution of defects would be exactly as needed for stable operation at nanometer scales.


The result is an extremely small transistor with a channel layer of only a few nanometers, which can reliably switch between on and off states with very little additional energy. According to the researchers, such optimized transistors achieve performance comparable to, and even better than, currently the most advanced solutions in their class, with lower energy consumption per operation.


Memory integrated into the transistor: ferroelectric hafnium-zirconium oxide


Based on the same approach, the researchers also made transistors with built-in memory, also in the back end of the chip. These so-called memory transistors have dimensions of the order of magnitude of 20 nanometers, and use ferroelectric hafnium-zirconium oxide as the key memory layer. It is a material that can retain two stable states of electric polarization, making it suitable for low-energy, non-volatile memories.


Ferroelectric hafnium-zirconium oxide (HfZrO or HZO) has become one of the stars of research in the field of next-generation memories in the last ten years. Unlike older ferroelectric materials, this compound is compatible with the standard CMOS process and can be crystallized at temperatures that are low enough not to damage existing structures in the back-end-of-line process. Because of this, HZO is increasingly mentioned as a candidate for integrating ferroelectric memories directly onto logical chips.


In the MIT demonstration, a ferroelectric layer of hafnium-zirconium oxide was added above the channel of amorphous indium oxide, thereby creating a compact memory transistor. By changing the polarization of the ferroelectric layer, the effective threshold for turning on the transistor changes, which allows the storage of logical states without the need for constant refreshing. Such an approach combines the functions of logic and memory in the same physical element, so the same transistor can be used for both data processing and their storage.


These memory transistors show extremely fast state switching, of the order of magnitude of ten nanoseconds, which is so fast that it approaches the limits of the measuring instruments used in the experiment. Even more importantly, this change of state is achieved at significantly lower voltages than with similar devices, which further reduces overall energy consumption.


Energy gain: less data movement, more computing per joule


Merging logic and memory in a vertically stacked structure dramatically affects the energy profile of the system. When a transistor can locally store data it is currently processing, there is no need for bits to be constantly sent through the chip and to external memory. Every eliminated hop through the memory hierarchy means fewer losses in wires and interconnects, less heating, and less energy spent on cooling.


For AI accelerators, in which the same model parameters and the same blocks of data are read and written billions of times during operation, this difference can be decisive. Estimates by global organizations show that the electricity consumption of data centers could rise to about 950 terawatt-hours annually by 2030, primarily due to the growth of AI workloads. Even a relatively small percentage of savings per chip, at the level of a few tens of percent less consumption per operation, can turn into significant savings at the level of entire data centers and energy systems.


MIT's approach therefore fits into a broader strategy of "efficient AI," which counts not only on renewable energy sources and better infrastructure but also on significantly more efficient computing in the silicon base itself. By vertically stacking transistors and memory, it is possible to achieve more operations per joule of energy, without necessarily increasing the clock speed or adding an ever-increasing number of conventional cores.


Connecting with global trends in the semiconductor industry


The work of the MIT team builds upon a strong research wave that has in recent years been looking for ways to use amorphous oxide semiconductors and ferroelectric materials in the back end of the process for monolithic 3D integration. Numerous academic groups and industrial laboratories have already shown that indium oxide and related materials can be deposited in ultra-thin layers, with high charge carrier mobility and stable operation at low consumption.


In parallel, ferroelectric memories based on hafnium oxide and hafnium-zirconium oxide are being intensively researched. They offer non-volatile data storage, the possibility of operation at low voltages, and compatibility with existing silicon technologies. The latest reviews from literature show that ferroelectric transistors can achieve switching times of the order of a few nanoseconds, operation with voltage levels below five volts, and endurance of billions of cycles, making them serious candidates for future embedded memories and in-memory computing.


MIT's demonstration combines these trends into one system: it uses amorphous indium oxide to build low-temperature transistors in the back end of the chip and hafnium-zirconium oxide for the implementation of compact ferroelectric memory. Additionally, the researchers, in collaboration with partners, have also developed performance models of these transistors, which is a key step toward embedding such elements into larger circuits, such as accelerators for neural networks or specialized processors for computer vision.


From research prototype to industrial application


The new transistors and memory elements were presented at the prestigious IEEE International Electron Devices Meeting (IEDM), which is considered one of the main places where the industry and the academic community exchange results on the future of semiconductor technology. The fact that researchers from MIT, the University of Waterloo, and the industrial giant Samsung Electronics participated in the work shows that the idea of vertically stacking logic and memory in the back end of the chip is already being very seriously considered outside the laboratory as well.


The path from a research prototype to a commercial product is nevertheless long. It is necessary to prove the reliability of the device over billions of cycles, show that new materials can be reproducibly manufactured in large batches, and integrate design tools that will allow engineers to use these transistors and memories in real projects. Performance modeling, which MIT and partners are working on, is one of the first steps: it allows circuits that would use such elements to be simulated today and their benefit compared to classic architectures to be estimated.


The semiconductor industry is simultaneously exploring other approaches to three-dimensional stacking – from advanced memory technologies to logic transistors that are stacked on top of each other in complementary structures. However, solutions that can be integrated into existing CMOS processes, without drastic changes in production plants, have the greatest chances for rapid adoption. In this sense, amorphous oxide semiconductors and hafnium-zirconium oxide have an important advantage because they already fit into the existing infrastructure.


Next steps: scaling, optimization, and new functionalities


The research team from MIT has already announced that the next steps include further improvement of the performance of transistors made in the back end, as well as finer control of the properties of ferroelectric hafnium-zirconium oxide. The goal is to simultaneously increase operating speed, reduce required voltages, and maintain stability during long-term operation. In doing so, understanding the fundamental physics at the level of individual ferroelectric domains in nanometer structures will also play an important role.


It is particularly interesting that these miniature memory transistors serve not only as functional elements but also as an experimental platform for studying the physics of ferroelectrics in extremely scaled dimensions. By observing how domains behave in structures the size of only a few tens of nanometers, researchers can test theoretical models and devise new ways to utilize ferroelectricity in computing, sensors, or neuromorphic circuits.


In a broader sense, the work of the MIT team is part of a global race for new materials and architectures that could replace or upgrade silicon when we approach the physical limits of its scaling. The combination of three-dimensional integration, new semiconductors, and ferroelectric memories offers a path toward chips that simultaneously provide greater processing power, lower energy consumption, and more flexible data organization – exactly what is needed so that generative artificial intelligence and other data-intensive tools can develop without causing an energy crisis.


Although time will pass before such transistors and memories come to life in mass products, the direction is clear: future generations of computing systems will increasingly resemble layered structures in which data is processed where it is created, and the boundary between logic and memory gradually disappears. In that scenario, solutions like MIT's platform for stacking transistors and memory in the back end of the chip can play an important role in maintaining the growth of computing power, while simultaneously keeping energy consumption under control.

Find accommodation nearby

Creation time: 7 hours ago

Science & tech desk

Our Science and Technology Editorial Desk was born from a long-standing passion for exploring, interpreting, and bringing complex topics closer to everyday readers. It is written by employees and volunteers who have followed the development of science and technological innovation for decades, from laboratory discoveries to solutions that change daily life. Although we write in the plural, every article is authored by a real person with extensive editorial and journalistic experience, and deep respect for facts and verifiable information.

Our editorial team bases its work on the belief that science is strongest when it is accessible to everyone. That is why we strive for clarity, precision, and readability, without oversimplifying in a way that would compromise the quality of the content. We often spend hours studying research papers, technical documents, and expert sources in order to present each topic in a way that will interest rather than burden the reader. In every article, we aim to connect scientific insights with real life, showing how ideas from research centres, universities, and technology labs shape the world around us.

Our long experience in journalism allows us to recognize what is truly important for the reader, whether it is progress in artificial intelligence, medical breakthroughs, energy solutions, space missions, or devices that enter our everyday lives before we even imagine their possibilities. Our view of technology is not purely technical; we are also interested in the human stories behind major advances – researchers who spend years completing projects, engineers who turn ideas into functional systems, and visionaries who push the boundaries of what is possible.

A strong sense of responsibility guides our work as well. We want readers to trust the information we provide, so we verify sources, compare data, and avoid rushing to publish when something is not fully clear. Trust is built more slowly than news is written, but we believe that only such journalism has lasting value.

To us, technology is more than devices, and science is more than theory. These are fields that drive progress, shape society, and create new opportunities for everyone who wants to understand how the world works today and where it is heading tomorrow. That is why we approach every topic with seriousness but also with curiosity, because curiosity opens the door to the best stories.

Our mission is to bring readers closer to a world that is changing faster than ever before, with the conviction that quality journalism can be a bridge between experts, innovators, and all those who want to understand what happens behind the headlines. In this we see our true task: to transform the complex into the understandable, the distant into the familiar, and the unknown into the inspiring.

NOTE FOR OUR READERS
Karlobag.eu provides news, analyses and information on global events and topics of interest to readers worldwide. All published information is for informational purposes only.
We emphasize that we are not experts in scientific, medical, financial or legal fields. Therefore, before making any decisions based on the information from our portal, we recommend that you consult with qualified experts.
Karlobag.eu may contain links to external third-party sites, including affiliate links and sponsored content. If you purchase a product or service through these links, we may earn a commission. We have no control over the content or policies of these sites and assume no responsibility for their accuracy, availability or any transactions conducted through them.
If we publish information about events or ticket sales, please note that we do not sell tickets either directly or via intermediaries. Our portal solely informs readers about events and purchasing opportunities through external sales platforms. We connect readers with partners offering ticket sales services, but do not guarantee their availability, prices or purchase conditions. All ticket information is obtained from third parties and may be subject to change without prior notice. We recommend that you thoroughly check the sales conditions with the selected partner before any purchase, as the Karlobag.eu portal does not assume responsibility for transactions or ticket sale conditions.
All information on our portal is subject to change without prior notice. By using this portal, you agree to read the content at your own risk.