Key Benefits
2nd Generation akidaTM expands the benefits of Event-Based, Neuromorphic Processing to a much broader set of complex network models.
It builds on the Technology Foundations, adds 8 bit weights and activations support and key new features that improve energy efficiency, performance and accuracy, while minimizing model storage needed, thereby enabling a much wider set of Intelligent applications that can be run on Edge devices untethered from the cloud.
akidaTM 2nd Generation Platform Brief
Engaging with lead adopters now. General availability in Q3’2023.
Explore new 2nd Generation Features
Which efficiently accelerate Vision, Multi-Sensory and Spatio-Temporal applications such as
- Audio Processing, Filtering, Denoising for Hearing Aids, and Hearables
- Speech Processing for Consumer, Automotive and Industrial Applications
- Vital Signs Prediction & Monitoring in Healthcare
- Time Series Forecasting for Industrial Predictive Maintenance
- Video Object Detection and Tracking in Consumer, Automotive and Industrial
- Vision, LIDAR Analysis for ADAS
- Advanced Sequence Prediction for Robotic Navigation and Autonomy
These capabilities are critically needed in Industrial, Automotive, Digital Health, Smart Home, and Smart City Applications.
On-Chip
Learning
Eliminates Sensitive Data
sent to Cloud and Improves
Security & Privacy.
Multi-Pass
Processing
Supports 1 To 128 Node
Implementations and
Scales Processing.
Configurable
IP Platform
Flexible and Scalable
for Multiple Edge
AI Use Cases.


Temporal Event-Based
Neural Nets
Neural Nets
Definition
A patent-pending innovation that enable an order of magnitude, or more, reduction in model size and resulting computation for Spatio-Temporal Applications while maintaining or improving accuracy.
TENNs are easily trained with back-propagation like a CNN and inference like a RNN. They efficiently process a stream of 2D frames or a stream of 1D values through time.
Key Benefits:
- Reduced Footprint & Power: TENNs radically reduces the number of parameters and resulting computation by orders of magnitude.
- High Accuracy & Speed: Reduced footprint does not affect accuracy of results. Lesser computation needed results in greater speed of execution.
- Simplified Training: Pipeline similar to CNN training, but with benefits of RNN operation.
Visualization
The spatial and temporal aspects of the data seamlessly processed by TENNs. In this case, the input data is a stream of 2D video frames processed through time.

Popular for tasks like
Real-Time Robotics
Quickly processing environmental changes for drone navigation or robotic arms.
Early Seizure Detection
Quickly detect and predict the onset of epileptic seizures in real-time.
Dynamic Vision Sensing
Fast motion tracking and optical flow estimation in changing visual scenes.
Vision
Transformers
Transformers
Definition
Vision Transformers, or ViTs, are a type of deep learning model that uses self-attention mechanisms to process visual data.
ViTs were introduced in a paper by Dosovitskiy in 2020 and have since gained popularity in computer vision research.
In a Vision Transformer, an image is first divided into patches, which are then flattened and fed into a multi-layer transformer network. A multi-head self-attention mechanism in each transformer layer allows the model to focus on the relationships between image patches at differing levels of abstraction to capture local and global features.
The transformer’s output is passed through a final classification layer to obtain the predicted class label.
Key Benefits:
- Complete hardware acceleration: ViT encoder block fully implemented in hardware for execution independent of host CPU.
- Highly Efficient Performance: Execution in hardware without CPU overhead, highly reduces energy consumption while improving speed.
- Low footprint, High scalability: Compact design reduces overhead. Can scale throughput and capacity with up to 12 nodes.
Visualization
Vision Transformers are like teaching a computer to see by breaking down pictures into small pieces, similar to how you’d piece together a jigsaw puzzle.

Instead of analyzing the whole image at once, the computer examines each piece and then combines what it learns from all the pieces to understand the entire picture.
Popular for tasks like
Image Classification
Achieving state-of-the-art results on benchmarks like ImageNet.
Object Detection
Detect and classify multiple objects within an image simultaneously.
Image Generation
Producing new images or enhancing existing ones based on learned patterns.
Skip
Connections
Connections
Definition
The foundational akidaTM technology simultaneously accelerates multiple feed-forward networks completely in hardware.
With the added support for short and long-range skip connections, an akidaTM neural processor can now accelerate complex neural networks such as ResNet completely in hardware without model computations on a host CPU.
Skip connections are impemented by storing data from previous layers in the akidaTM mesh or in a scratchpad memory for combination with data in later layers.
Key Benefits:
- Complex Network Acceleration: Enables complete hardware execution of non feed-forward model architectures such as ResNet and DenseNet.
- Low Latency: Eliminates CPU interaction during network evaluation which minimizes model latency.
Visualization
Skip connections effectively allow a neural network to remember the outputs of earlier layers in the network for use in computation in later layers of the network.
This facilitates combining information from different abstraction levels.
Popular for tasks like
Image Segmentation
Help retain details, crucial for precise tasks like medical imaging.
Residual Learning
Allow training deeper layers without degradation, thereby improving overall model performance.
Object Detection
Combine features to detect objects of varying sizes effectively.
Efficiency
Ideal for always-on, energy-sipping Sensor Applications:
Vibration Detection
Anomaly Detection
Keyword Spotting
Sensor Fusion
Low-Res Presence Detection
Gesture Detection
@Sensor Inference
Either Standalone or with Min-spec MCU.
Configurable to ideal fit:
1 – 2 nodes (4 NPE/node)
Anomaly Detection
Keyword Spotting
Expected implementations:
50 MHz – 200
MHz Up to 100 GOPs
Additional Benefits
- Eliminates need for CPU intervention
- Fully accelerates most feed-forward networks
- Optional skip connection and TENNs support for more complex networks
- Completely customizable to fit very constrained power, thermal, and silicon area budgets
- Enables energy-harvesting and multi-year battery life applications, sub milli-watt sensors
Balanced
Accelerates in hardware most Neural Network Functions:
Advanced Keyword Spotting
Sensor Fusion
Low-Res Presence Detection
Gesture Detection & Recognition
Object Classification
Biometric Recognition
Advanced Speech Recognition
Object Detection & Semantic Segmentation
and Application SoCs
With Min-Spec or Mid-Spec MCU.
Configurable to ideal fit:
3 – 8 nodes (4 NPE/node) 25 KB
100 KB per NPE
Process, physical IP and other optimizations
Expected implementations:
100 – 500 MHz
Up to 1 TOP
Additional Benefits
- CPU is free for most non-NN compute
- CPU runs application with minimal NN-management
- Completely customizable to fit very constrained power, thermal and silicon area budgets
- Enables intelligent, learning-enabled MCUs and SoCs consuming tens to hundreds of milliwatts or less
Performance
Detection, Classification, Segmentation, Tracking, and ViT:
Gesture Detection
Object Classification
Advanced Speech Recognition
Object Detection & Semantic Segmentation
Advanced Sequence Prediction
Video Object Detection & Tracking
Vision Transformer Networks
in a Sensor-Edge Power Envelope
With Mid-Spec MCU or Mid-Spec MPU.
Configurable to ideal fit:
8 – 256 nodes (4 NPE/node) + optional Vision Transformer
100 KB per NPE
Process, physical IP and other optimizations
Expected implementations:
800 MHz – 2 GHz
Up to 131 TOPs
Additional Benefits
- CPU is free for most non-NN compute
- CPU runs application with minimal NN-management
- Builds on Performance product with Vision transformer capability
- akidaTM accelerates most complex spatio-temporal and Vision Transformer networks in hardware