Key Benefits

2nd Generation akidaTM expands the benefits of Event-Based, Neuromorphic Processing to a much broader set of complex network models.

It builds on the Technology Foundations, adds 8 bit weights and activations support and key new features that improve energy efficiency, performance and accuracy, while minimizing model storage needed, thereby enabling a much wider set of Intelligent applications that can be run on Edge devices untethered from the cloud.

akidaTM 2nd Generation Platform Brief

Availability:
Engaging with lead adopters now. General availability in Q3’2023.

Explore new 2nd Generation Features

Which efficiently accelerate Vision, Multi-Sensory and Spatio-Temporal applications such as

  • Audio Processing, Filtering, Denoising for Hearing Aids, and Hearables
  • Speech Processing for Consumer, Automotive and Industrial Applications
  • Vital Signs Prediction & Monitoring in Healthcare
  • Time Series Forecasting for Industrial Predictive Maintenance
  • Video Object Detection and Tracking in Consumer, Automotive and Industrial
  • Vision, LIDAR Analysis for ADAS
  • Advanced Sequence Prediction for Robotic Navigation and Autonomy

These capabilities are critically needed in Industrial, Automotive, Digital Health, Smart Home, and Smart City Applications.

Technology Foundations

On-Chip
Learning

Eliminates Sensitive Data
sent to Cloud and Improves
Security & Privacy.

Multi-Pass
Processing

Supports 1 To 128 Node
Implementations and
Scales Processing.

Configurable
IP Platform

Flexible and Scalable
for Multiple Edge
AI Use Cases.

2nd Generation New Features

Temporal Event-Based
Neural Nets

Definition

A patent-pending innovation that enable an order of magnitude, or more, reduction in model size and resulting computation for Spatio-Temporal Applications while maintaining or improving accuracy.

TENNs are easily trained with back-propagation like a CNN and inference like a RNN. They efficiently process a stream of 2D frames or a stream of 1D values through time.

Key Benefits:

  • Reduced Footprint & Power: TENNs radically reduces the number of parameters and resulting computation by orders of magnitude.
  • High Accuracy & Speed: Reduced footprint does not affect accuracy of results. Lesser computation needed results in greater speed of execution.
  • Simplified Training: Pipeline similar to CNN training, but with benefits of RNN operation.

Visualization

The spatial and temporal aspects of the data seamlessly processed by TENNs. In this case, the input data is a stream of 2D video frames processed through time.

Popular for tasks like

Real-Time Robotics

Quickly processing environmental changes for drone navigation or robotic arms.

Early Seizure Detection

Quickly detect and predict the onset of epileptic seizures in real-time.

Dynamic Vision Sensing

Fast motion tracking and optical flow estimation in changing visual scenes.

Vision
Transformers

Definition

Vision Transformers, or ViTs, are a type of deep learning model that uses self-attention mechanisms to process visual data.

ViTs were introduced in a paper by Dosovitskiy in 2020 and have since gained popularity in computer vision research.

In a Vision Transformer, an image is first divided into patches, which are then flattened and fed into a multi-layer transformer network. A multi-head self-attention mechanism in each transformer layer allows the model to focus on the relationships between image patches at differing levels of abstraction to capture local and global features.

The transformer’s output is passed through a final classification layer to obtain the predicted class label.

Key Benefits:

  • Complete hardware acceleration: ViT encoder block fully implemented in hardware for execution independent of host CPU.
  • Highly Efficient Performance: Execution in hardware without CPU overhead, highly reduces energy consumption while improving speed.
  • Low footprint, High scalability: Compact design reduces overhead. Can scale throughput and capacity with up to 12 nodes.

Visualization

Vision Transformers are like teaching a computer to see by breaking down pictures into small pieces, similar to how you’d piece together a jigsaw puzzle.

Instead of analyzing the whole image at once, the computer examines each piece and then combines what it learns from all the pieces to understand the entire picture.

akida enables highly-efficient, portable implementations for advanced vision applications to scale Edge solutions in a variety of industries.

Popular for tasks like

Image Classification

Achieving state-of-the-art results on benchmarks like ImageNet.

Object Detection

Detect and classify multiple objects within an image simultaneously.

Image Generation

Producing new images or enhancing existing ones based on learned patterns.

Skip
Connections

Definition

The foundational akidaTM technology simultaneously accelerates multiple feed-forward networks completely in hardware.

With the added support for short and long-range skip connections, an akidaTM neural processor can now accelerate complex neural networks such as ResNet completely in hardware without model computations on a host CPU.

Skip connections are impemented by storing data from previous layers in the akidaTM mesh or in a scratchpad memory for combination with data in later layers.

Key Benefits:

  • Complex Network Acceleration: Enables complete hardware execution of non feed-forward model architectures such as ResNet and DenseNet.
  • Low Latency: Eliminates CPU interaction during network evaluation which minimizes model latency.

Visualization

Skip connections effectively allow a neural network to remember the outputs of earlier layers in the network for use in computation in later layers of the network.

This facilitates combining information from different abstraction levels.

Popular for tasks like

Image Segmentation

Help retain details, crucial for precise tasks like medical imaging.

Residual Learning

Allow training deeper layers without degradation, thereby improving overall model performance.

Object Detection

Combine features to detect objects of varying sizes effectively.

One IP Platform, Multiple
Configurable Products

Max
Efficiency

Ideal for always-on, energy-sipping Sensor Applications:

  • Vibration Detection
  • Anomaly Detection
  • Keyword Spotting
  • Sensor Fusion
  • Low-Res Presence Detection
  • Gesture Detection
Extremely Efficient
@Sensor Inference

Either Standalone or with Min-spec MCU.

Configurable to ideal fit:

  • 1 – 2 nodes (4 NPE/node)
  • Anomaly Detection
  • Keyword Spotting

Expected implementations:

  • 50 MHz – 200
  • MHz Up to 100 GOPs

Additional Benefits

  • Eliminates need for CPU intervention
  • Fully accelerates most feed-forward networks
  • Optional skip connection and TENNs support for more complex networks
  • Completely customizable to fit very constrained power, thermal, and silicon area budgets
  • Enables energy-harvesting and multi-year battery life applications, sub milli-watt sensors
Sensor
Balanced

Accelerates in hardware most Neural Network Functions:

  • Advanced Keyword Spotting
  • Sensor Fusion
  • Low-Res Presence Detection
  • Gesture Detection & Recognition
  • Object Classification
  • Biometric Recognition
  • Advanced Speech Recognition
  • Object Detection & Semantic Segmentation
Optimal for Sensor Fusion
and Application SoCs

With Min-Spec or Mid-Spec MCU.

Configurable to ideal fit:

  • 3 – 8 nodes (4 NPE/node) 25 KB
  • 100 KB per NPE
  • Process, physical IP and other optimizations

Expected implementations:

  • 100 – 500 MHz
  • Up to 1 TOP

Additional Benefits

  • CPU is free for most non-NN compute
  • CPU runs application with minimal NN-management
  • Completely customizable to fit very constrained power, thermal and silicon area budgets
  • Enables intelligent, learning-enabled MCUs and SoCs consuming tens to hundreds of milliwatts or less
Max
Performance

Detection, Classification, Segmentation, Tracking, and ViT:

  • Gesture Detection
  • Object Classification
  • Advanced Speech Recognition
  • Object Detection & Semantic Segmentation
  • Advanced Sequence Prediction
  • Video Object Detection & Tracking
  • Vision Transformer Networks
Advanced Network-Edge Performance
in a Sensor-Edge Power Envelope

With Mid-Spec MCU or Mid-Spec MPU.

Configurable to ideal fit:

  • 8 – 256 nodes (4 NPE/node) + optional Vision Transformer
  • 100 KB per NPE
  • Process, physical IP and other optimizations

Expected implementations:

  • 800 MHz – 2 GHz
  • Up to 131 TOPs

Additional Benefits

  • CPU is free for most non-NN compute
  • CPU runs application with minimal NN-management
  • Builds on Performance product with Vision transformer capability
  • akidaTM accelerates most complex spatio-temporal and Vision Transformer networks in hardware