IBM first unveiled the AIU late in 2022, and over the last year, researchers worked to create a system designed specifically for handling AI tasks more efficiently than traditional options. The AIU was built from the ground up for AI, and has been shown to be able to handle inference well.
With the AIU, researchers designed chips with their memory and processing units closer together, leading to lower latency and comparably faster chip speeds. The AIU system-on-a-chip (SOC) has 32 individual AI processing cores and contains 23 billion transistors, and was built using 5 nm node process technology. (For reference, IBM Research is currently investigating the path to chips smaller than 1 nm.) The AIU fits easily on a single-slot, half-length PCIe card, meaning it’s perfect for mounting in server racks.
The AIU is well suited to AI inferencing tasks and can be used in concert with GPUs for training and building a more efficient generative AI infrastructure. This cluster will be the first installation in the CEAIS since it was unveiled last year. Once installed and running, the AIU cluster will be a valuable tool for students, faculty, and researchers to work on the AI technologies that will power the future.