Hardware-Software Co-Design for AI Accelerators and Compilers on RISC-V Platforms


Course Objective
This course aims to equip students with the knowledge and skills required for end-to-end hardware–software co-design on open RISC-V platforms, with a strong focus on AI/ML workloads. Students will build a solid foundation in computer architecture, digital design, and SoC integration with open-source RISC-V hardware while training in the design and implementation of custom AI accelerators, including integration into full RISC-V SoC prototypes using LiteX and FPGA platforms.
The course provides hands-on experience in low-level programming, including bare-metal drivers, Linux kernel modules, and user-space interfaces for hardware accelerators. Participants will develop expertise in compiler toolchains and MLIR-based frameworks, enabling them to extend compilers, add intrinsics, and design accelerator-specific MLIR dialects. Through practical deployment of AI/ML applications on specialized open hardware using runtime frameworks such as IREE, TVM, and TensorFlow Lite, the course creates a skilled talent pipeline of engineers and researchers who can bridge hardware, compilers, and AI runtimes, aligned with industrial projects and collaborations.
Course Outcomes
Upon completion of this course, students will demonstrate familiarity with the complete hardware/software stack – from RISC-V CPU microarchitecture and SoC buses to operating systems, drivers, and compilers. They will be able to design, implement, and integrate custom AI accelerators into RISC-V SoCs and validate them on FPGA platforms, while developing and testing bare-metal HALs, Linux kernel drivers, and user-space APIs to control accelerators.
Students will extend LLVM/Clang and MLIR toolchains with custom instructions, intrinsics, and dialects, enabling compiler-level optimization for specialized hardware. They will deploy and optimize AI/ML workloads (e.g., CNNs, RNNs, transformers) on RISC-V + accelerator platforms using IREE/MLIR and runtime frameworks, while analyzing and evaluating performance, scheduling, and memory trade-offs for AI inference across CPU and custom hardware backends. Finally, students will contribute to industry-relevant open-source projects, CI pipelines, and collaborative research in AI compilers, firmware, and runtime systems.