Research Highlights

“Smarter” Semiconductor Technology for Training “Smarter” Artificial Intelligence

2024-07-31 237

[POSTECH and Korea University develop next-generation semiconductor technology for high-efficiency, low-power artificial intelligence]

김세영 교수팀 뷰페이지(en)A research team, consisting of Professor Seyoung Kim from the Department of Materials Science and Engineering and the Department of Semiconductor Engineering and alumnus Kyungmi Noh and PhD student Hyunjeong Kwak from the Department of Materials Science and Engineering at POSTECH, and Professor Hyung-Min Lee from Korea University’s School of Electrical Engineering, has recently demonstrated that analog hardware using ECRAM devices can maximize the computational performance of artificial intelligence, showcasing its potential for commercialization. Their research has been published in the esteemed international journal, “Science Advances.”

The rapid advancement of AI technology, including applications like generative AI, has pushed the scalability of existing digital hardware (CPUs, GPUs, ASICs, etc.) to its limits. Consequently, there is active research into analog hardware specialized for AI computation. Analog hardware adjusts the resistance of semiconductors based on external voltage or current and utilizes a cross-point array*1 structure with vertically crossed memory devices to process AI computation in parallel. Although it offers advantages over digital hardware for specific computational tasks and continuous data processing, meeting the diverse requirements for computational learning and inference remains challenging.

To address the limitations of analog hardware memory devices, the research team focused on Electrochemical Random Access Memory (ECRAM)*2, which manage electrical conductivity through ion movement and concentration. Unlike traditional semiconductor memory, these devices feature a three-terminal structure with separate paths for reading and writing data, allowing for operation at relatively low power.

In their study, the team successfully fabricated ECRAM devices using three-terminal-based semiconductors in a 64×64 array. Experiments revealed that the hardware incorporating the team’s devices demonstrated excellent electrical and switching characteristics, along with high yield and uniformity. Additionally, the team applied the Tiki-Taka algorithm*3, a cutting-edge analog-based learning algorithm, to this high-yield hardware, successfully maximizing the accuracy of AI neural network training computations. Notably, the researchers demonstrated the impact of the “weight retention” property of hardware training on learning and confirmed that their technique does not overload artificial neural networks, highlighting the potential for commercializing the technology.

This research is significant because the largest array of ECRAM devices for storing and processing analog signals reported in the literature to date is 10×10. The researchers have now successfully implemented these devices on the largest scale, with varied characteristics for each device.

영본1
Professor Seyoung Kim of POSTECH remarked, “By realizing large-scale arrays based on novel memory device technologies and developing analog-specific AI algorithms, we have identified the potential for AI computational performance and energy efficiency that far surpass current digital methods.”

The research was conducted with support from the Ministry of Trade, Industry and Energy, the Public-Private Partnership for Semiconductor Talent Training Program supported by the Korea Planning & Evaluation Institute of Industrial Technology (KEIT) and Korea Semiconductor Industry Association, and EDA Tool of the IDEC.

DOI: https://doi.org/10.1126/sciadv.adl3350


1. Cross-point array
An arrangement often used in memory device design where memory cells are located at the intersections of vertical and horizontal lines. This design is primarily used in next-generation memory devices such as RRAM and phase change memory (PCM), offering advantages in increasing memory density and reducing power consumption

2. ECRAM
Electrochemical Random Access Memory

3. Tiki-Taka algorithm
Named after a football strategy, this algorithm maximizes efficiency and accuracy in AI learning and inference. It has significant potential in areas requiring high-speed computation and precise data processing