
EPFL's e-GPU: Open-source RISC-V GPU platform enables TinyAI on edge devices with just 28mW power, delivering 15x performance boost for wearables and IoT sensors.

Drivetech Partners
EPFL researchers have introduced embedded GPU (e-GPU), an innovative open-source and configurable RISC-V GPU platform designed specifically for ultra-low-power TinyAI applications at the edge. This breakthrough technology, released in May 2025, operates with just 28 mW of power while providing parallel processing capabilities that can dramatically accelerate AI workloads in resource-constrained devices.
Key Takeaways
Open-source RISC-V architecture eliminates licensing fees and enables widespread customization for specific applications
The platform achieves up to 15.1x performance acceleration and 3.1x energy reduction in bio-signal processing benchmarks
Tiny-OpenCL framework democratizes GPU parallel programming for ultra-low-power devices without requiring specialized compilers
Implemented in 16nm CMOS technology, the e-GPU integrates advanced power-saving techniques while maintaining high compute capability
This technology enables AI capabilities in previously constrained devices like medical wearables and environmental sensors
The Revolution of Configurable RISC-V GPUs for Edge AI
The embedded GPU (e-GPU) platform represents a fundamental paradigm shift in how we approach computational needs for resource-constrained devices. Implemented in TSMC's 16 nm SVT CMOS technology, the e-GPU operates at 300 MHz and 0.8 V, striking an optimal balance between performance and power efficiency. Despite its tiny footprint, the platform delivers significant parallel processing capabilities while maintaining power consumption within a modest 28 mW budget.
This breakthrough addresses one of the most persistent challenges in edge computing: enabling sophisticated AI processing in devices with strict power and size limitations. Traditional GPUs, while computationally powerful, have been too power-hungry for deployment in IoT sensors, wearables, and other edge devices. The e-GPU changes this equation by bringing parallel computing to ultra-low-power environments without compromising on energy efficiency.

Extensive Hardware Configurability for Optimized TinyAI Applications
One of the e-GPU's most significant advantages is its extensive configurability options that allow developers to tailor the hardware precisely to their application requirements. The high-range configuration features 16 threads for parallel processing, enabling substantial performance gains for suitable workloads. This flexibility ensures that the platform can adapt to various use cases without wasting silicon area or power on unnecessary features.
The e-GPU has been integrated with X-HEEP (eXtendible Heterogeneous Energy-Efficient Platform) to create an accelerated processing unit (APU) specifically designed for TinyAI applications. This integration implements state-of-the-art fine-grained power management strategies including:
Clock-gating to reduce dynamic power consumption
Power-gating to eliminate leakage in inactive components
RAM retention to preserve critical data while minimizing power usage
These power-saving techniques are integrated through the XAIF interface, allowing connected accelerators to leverage the same energy-efficient mechanisms. The result is a platform that enables application-specific customization without sacrificing the energy efficiency that's critical for edge deployment.
Tiny-OpenCL: Democratizing GPU Parallelism for Constrained Environments
The hardware capabilities of the e-GPU are complemented by Tiny-OpenCL, a lightweight programming framework specifically designed for resource-constrained environments. This implementation enables sophisticated parallel computing on ultra-low-power devices while maximizing compatibility with existing GPU software and ensuring cross-platform portability.
Tiny-OpenCL consists of several key components that work together to provide a complete programming environment:
SIMT RISC-V extension API for low-level compute unit functionalities
Startup functions that handle initialization of the e-GPU
Scheduler functions that efficiently distribute workloads across available threads
These components are precompiled into a static library using the standard RISC-V GNU toolchain. One of the most innovative aspects of Tiny-OpenCL is how it transforms OpenCL kernels into standard C functions using a parser script. This approach eliminates the need for a dedicated compiler, allowing seamless integration with existing compilation tools and significantly reducing the barrier to entry for developers working with constrained devices.
Impressive Performance Benchmarks and Energy Efficiency
The e-GPU platform has undergone rigorous benchmarking to validate its performance and efficiency claims. The General Matrix Multiply (GeMM) benchmark shows negligible scheduling overhead for matrix sizes larger than 256×256, demonstrating the platform's efficiency for common computational workloads.
Even more impressive are the results from the TinyBio benchmark, which focuses on bio-signal processing workloads common in healthcare applications. The e-GPU achieves:
Up to 15.1x speed-up compared to the baseline host processor
3.1x reduction in energy consumption
These gains with only 2.5x area overhead
The benchmarks also revealed that at fixed matrix sizes, increasing the number of parallel threads results in decreased data transfer time, showing the scalable performance benefits of the platform's parallel architecture. These metrics prove that even modest hardware can deliver significant computational benefits when optimized for specific workloads.
The RISC-V Advantage for Ultra-Low-Power Computing
The choice of RISC-V as the foundation for the e-GPU platform brings numerous advantages for ultra-low-power computing. RISC-V is a fully open instruction set architecture (ISA), eliminating the licensing fees typically associated with processor development and enabling greater innovation across the industry.
Hundreds of companies are now developing cores and chips based on RISC-V, creating a vibrant ecosystem that benefits all participants. RISC-V International helps develop standards and ensure cross-compatibility, fostering collaboration while maintaining the flexibility that makes RISC-V so appealing.
The base RISC-V instruction set is simple and streamlined, with extensions available for high-performance applications. This modular approach allows RISC-V cores to scale from tiny control CPUs in sensors to multi-hundred-core server processors, making it particularly well-suited for the diverse requirements of edge computing.
The philosophical alignment between RISC-V and open hardware movements promotes innovation in ways that proprietary architectures cannot match. This openness is especially beneficial for specialized applications like TinyAI, where custom optimizations can make the difference between practical and impractical implementations.
Real-World Applications in Healthcare and Beyond
The e-GPU platform has demonstrated real-world applicability in ultra-low-power healthcare applications, where computational demands must be balanced against strict power and size constraints. The X-HEEP platform can be used standalone as a low-cost microcontroller or integrated into existing systems, providing flexibility for various deployment scenarios.
The TinyBio benchmark specifically targets bio-signal processing workloads common in wearable medical devices, such as ECG analysis, motion detection, and vital sign monitoring. The platform's low power consumption makes it ideal for battery-powered medical monitoring, potentially extending device operating time between charges by a factor of three or more.
Applications extend beyond healthcare to numerous other domains that can benefit from edge AI capabilities:
Environmental monitoring sensors with on-device analysis
Agricultural sensors for crop health and irrigation optimization
Smart infrastructure monitoring with local anomaly detection
Consumer wearables with enhanced functionality and battery life
By enabling AI capabilities in devices previously considered too resource-constrained, the e-GPU platform opens up new possibilities for smart, autonomous systems in previously underserved applications.
Accelerating Innovation through Open-Source Hardware
The open-source nature of the e-GPU platform democratizes access to GPU parallelism in resource-constrained environments, fostering innovation across the industry. Developers can evaluate custom parallel processor designs using existing OpenCL applications, reducing the time and cost associated with hardware exploration.
This openness also allows for the exploration of new application-specific compiler optimization techniques, potentially unlocking even greater performance and efficiency improvements. Performance interfaces and tools like LPN offer up to 7821× faster cycle-level performance simulation, enabling rapid iteration and experimentation.
Machine learning compilers can generate optimized code in seconds instead of hours, dramatically accelerating the development cycle for edge AI applications. Perhaps most importantly, the open-source approach creates a feedback loop of innovation as developers share their improvements with the community, benefiting all participants in the ecosystem.
Future Horizons for TinyAI Acceleration
Looking ahead, the e-GPU platform provides a foundation for continued advancement in TinyAI acceleration. Researchers are exploring new ISA extensions to achieve even better power and energy targets, further extending the range of devices that can benefit from on-device AI processing.
The development of application-specific compiler optimizations promises to extract even more performance from the existing hardware, while integration with other accelerators like Coarse-Grained Reconfigurable Arrays (CGRAs) could create hybrid systems with complementary strengths.
The range of TinyAI applications is expected to expand beyond current benchmarks, bringing intelligence to an ever-wider array of edge devices. Further optimization of the Tiny-OpenCL framework could reduce overhead even further, making GPU programming accessible to an even broader range of developers.
Researchers are also exploring formal verification methods to ensure performance before production deployment, reducing the risk associated with hardware customization. The open nature of the platform ensures that it will continue to evolve based on real-world needs, rather than being constrained by proprietary business interests.
Sources
semiengineering.com - Embedded GPU: An Open-Source and Configurable RISC-V GPU Platform for TinyAI Devices
arxiv.org - e-GPU: An Open-Source and Configurable RISC-V GPU Platform for TinyAI Devices
infoscience.epfl.ch - EPFL Publication Repository
semiengineering.com - EPFL's Open-Source Single-Core RISC-V Microcontroller for Edge Computing
github.