SK Hynix AI-Specific Computing Memory Solution AiMX-xPU at Hot Chips 2024

1
SK Hynix Hot Chips 2024 AiMX XPU_Page_12
SK Hynix Hot Chips 2024 AiMX XPU_Page_12

At Hot Chips 2024, SK Hynix was focused on more than standard DRAM for AI accelerators. Instead, the company showed off its latest advancements in in-memory computing, this time for LLM inference with its AiMX-xPU and LPDDR-AiM. The idea is that instead of moving data from memory to compute to perform memory-related transformations, those can be done directly in the memory without having to traverse an interconnect. That makes it more power efficient and potentially faster.

We are doing these live at Hot Chips 2024 this week, so please excuse typos.

SK Hynix AI-Specific Computing Memory Solution AiMX-xPU at Hot Chips 2024

SK Hynix is professing its love LLMs since they are memory-bound.

SK Hynix Hot Chips 2024 AiMX XPU_Page_05
SK Hynix Hot Chips 2024 AiMX XPU_Page_05

The company is showing off its GDDR6 Accelerator-in-Memory card using the Xilinx Virtex FPGAs and the special GDDR6 AiM package.

SK Hynix Hot Chips 2024 AiMX XPU_Page_06
SK Hynix Hot Chips 2024 AiMX XPU_Page_06

Here is a look at the card.

SK Hynix Acceleratior In Memory AiM At OCP Summit 2023 1
SK Hynix Accelerator In Memory AiM At OCP Summit 2023 1

Here is a look at the GDDR6 chip. We saw these again at FMS 2024, but we already had photos of them.

SK Hynix Acceleratior In Memory AiM At OCP Summit 2023 2
SK Hynix Accelerator In Memory AiM At OCP Summit 2023 2

Plus, SK Hynix mentioned the OCP 2023 live demo where we took these photos.

SK Hynix Hot Chips 2024 AiMX XPU_Page_07
SK Hynix Hot Chips 2024 AiMX XPU_Page_07

Here are the fully connected layer and multi-head attention memory and compute bound parts of LLMs.

SK Hynix Hot Chips 2024 AiMX XPU_Page_09
SK Hynix Hot Chips 2024 AiMX XPU_Page_09

Here is how the pressure varies based on batch size.

SK Hynix Hot Chips 2024 AiMX XPU_Page_10
SK Hynix Hot Chips 2024 AiMX XPU_Page_10

SK Hynix mapped multi-head attention into AiM.

SK Hynix Hot Chips 2024 AiMX XPU_Page_11
SK Hynix Hot Chips 2024 AiMX XPU_Page_11

The company also doubled the memory capacity to 32GB using 32 AIM packages up from 16. 32GB is probably not enough for a product, but it is for a prototype. Still, the company was able to show the performance of the technology.

SK Hynix Hot Chips 2024 AiMX XPU_Page_12
SK Hynix Hot Chips 2024 AiMX XPU_Page_12

The next-gen demos are going to show things like Llama-3 and the company is also looking at scaling from 32GB to 256GB per card.

SK Hynix Hot Chips 2024 AiMX XPU_Page_13
SK Hynix Hot Chips 2024 AiMX XPU_Page_13

Beyond just looking at data center AI, the company is looking at on-device AI. We are already seeing companies like Apple, Intel, AMD, and Qualcomm push NPUs for AI.

SK Hynix Hot Chips 2024 AiMX XPU_Page_15
SK Hynix Hot Chips 2024 AiMX XPU_Page_15

On-device AI typically lowers batch size, making those workloads memory-bound. Moving the compute off-SoC means it can be more energy efficient and not take compute die space on the SoC.

SK Hynix Hot Chips 2024 AiMX XPU_Page_16
SK Hynix Hot Chips 2024 AiMX XPU_Page_16

The goal is to optimize AiM in the future in a product for LPDDR5-AiM. The goal is not to change existing LPDDR commands and not to have a negative performance impact. Specs on this table are estimated.

SK Hynix Hot Chips 2024 AiMX XPU_Page_17
SK Hynix Hot Chips 2024 AiMX XPU_Page_17

With LPDDR5, the can be integrated onto the SoC for mobile devices.

SK Hynix Hot Chips 2024 AiMX XPU_Page_18
SK Hynix Hot Chips 2024 AiMX XPU_Page_18

There may be a need to make different trade offs for different applications.

SK Hynix Hot Chips 2024 AiMX XPU_Page_20
SK Hynix Hot Chips 2024 AiMX XPU_Page_20

One of the challenges is to arbitrate between normal usage for LPDDR memory, as well as the compute demands. Also, there is a possibility to change the thermal and power requirements of the chips.

SK Hynix Hot Chips 2024 AiMX XPU_Page_21
SK Hynix Hot Chips 2024 AiMX XPU_Page_21

Another challenge is how to program the AiM.

SK Hynix Hot Chips 2024 AiMX XPU_Page_22
SK Hynix Hot Chips 2024 AiMX XPU_Page_22

It looks like SK Hynix is expanding the line and types of places where AiM/ AiMX can be used.

SK Hynix Hot Chips 2024 AiMX XPU_Page_24
SK Hynix Hot Chips 2024 AiMX XPU_Page_24

SK Hynix said that in GDDR6 AiM occupies around 20% of the die area.

Final Words

This still feels like something that a major SoC/ Chip vendor is going to have to pick up and integrate before it goes mainstream. In many ways, computing in memory may make sense. We shall see if this one goes from prototype to product in the future.

1 COMMENT

  1. Join TopAITrends.io on a journey to discover the power of artificial intelligence. Our platform offers a carefully curated directory of AI tools, updated daily to keep you ahead in this rapidly evolving field. Whether you’re automating tasks, exploring creative tools, or looking to enhance your business, our AI solutions cater to all needs. Dive into personalized recommendations, user-driven reviews, and expert content. Discover the future of AI today with TopAITrends.io.

LEAVE A REPLY

Please enter your comment!
Please enter your name here

This site uses Akismet to reduce spam. Learn how your comment data is processed.