Skip to content

HYMEM-AI: Energy-Efficient Hybrid Memory based Architecture for Next-Generation AI Systems

Information

  • Funder: KAKENHI (Grant-in-Aid for Scientific Research B)
  • Duration: 2026/04-2029/03
  • Summary: Artificial Intelligence (AI) has made great progress, but its high energy consumption remains a serious challenge, especially for small devices such as smartphones, IoT sensors, and battery- or solar-powered systems. This project proposes a new energy-efficient neuromorphic AI accelerator using 3D integrated circuits (3D-ICs) that combine Non-Volatile Memory (NVM) and SRAM to reduce power usage both during operation and idle periods. NVM can retain data without power, significantly lowering idle energy consumption, but it is typically slower and more energy-intensive during read/write operations than SRAM. To address this, we introduce a hybrid memory architecture in which stacked NVM layers are placed above the logic layer to shorten data transfer distance, while a small, fast SRAM layer acts as a cache between logic and NVM to improve speed and reduce read/write energy. In addition, adaptive power management techniques dynamically adjust energy usage based on available power with minimal impact on accuracy and performance. With these technologies, the proposed design aims to reduce energy consumption during operation and during idle time, enabling practical low-power AI chips for real-world applications such as smartphones, IoT devices, and smart sensors.

Image title Image title

Research Technology & Prototyping

Our research group specializes in the design and validation of high-efficiency AI hardware and next-generation memory-compute systems. We leverage industry-leading AMD (formerly Xilinx) FPGA platforms to bridge the gap between architectural theory and physical silicon realization.

Core Methodologies

Full-System Prototyping: We utilize a rigorous RTL-based design flow (Verilog/VHDL) within the Vivado™ and Vitis™ environments. This ensures cycle-accurate hardware validation and reliable performance benchmarking for complex neural accelerators.

3D-IC & Interconnect Emulation: Recognizing the challenges of iterative 3D-IC fabrication, we use advanced Programmable Network-on-Chip (NoC) features to simulate high-bandwidth vertical interconnects and through-silicon via (TSV) latency characteristics.

Advanced Memory Modeling: Our group develops custom behavioral wrappers to model the timing, energy signatures, and data-retention properties of emerging technologies, such as Non-Volatile Memory (NVM) and high-density 3D-stacked SRAM.

By using adaptive computing platforms as high-fidelity emulation vehicles, we can explore innovative architectural trade-offs—such as Memory–Compute Co-Design and Event-Driven AI—with the precision of hardware-level execution. This methodology allows us to accelerate the development cycle for low-power AI chips targeting edge devices, smartphones, and autonomous IoT sensors.

Team Members

  • DANG Nam Khanh (PI)
  • SUZUKI Daisuke
  • BEN ABDALLAH Abderazek
  • WANG Zhishang

Students

  • HANYU Yuga
  • Atharv Sharma
  • Van-Vu Luyen
  • KOMATSUZAKI Aruki

Progress

Early Prototyping

As a early prototyping, we use HumanDATA Kintex 7 EDX-008-70T to validate the read/write of MRAM by Everspin.

Image title

AY2026

(To be updated)

Contact

  • khanh [at] u-aizu.ac.jp

Funder

KAKENHI