Chip design

for the most efficient Generative AI

Edge or Cloud

all the Hardware you need to run Transformer models

Redefining efficiency

Industry-best token throughput per TB/s.

Overcoming the Generative AI bottleneck:
Memory Bandwidth

Nvidia H100
31
Google TPU v6e
40
RaiderChip NPU
58

Tokens / second / user (Llama 3.1 8B - 1 TB/s)

Performance approaching the physical limit — over 90% efficiency at scale.

Full hardware architecture
& Embedded RISC-V

Reduced Power Consumption and Silicon cost

Maximum Tokens per Watt with Minimum silicon area

More tokens per dollar

from foundry to deployment

Flexibility & Scalability by design

One core, multiple architectures


Discover a Generative AI inference solution that scales linearly.

Multiple memory technologies

DDR LPDDR HBM

Scalable blocks to match memory bandwidth

Extreme scalability

from 250 MHz on FPGA
to 2 GHz on ASIC

Generative AI with no limits

Future-proof target agnostic architecture

12nm, 7nm, and 3nm

One flexible architecture for any target

An open design that adapts to each application

Edge & Cloud

One NPU to power your vision

Plug'n'Play Integration

Minimalism as a principle

100% Hardware accelerated Transformers architecture.

Model support out of the box

Reconfigurable data pipeline through embedded RISC-V

Support future models without hardware redesign

Kairós

Edge AI inference ASIC

Llama, Phi, DeepSeek, Qwen ... all Transformer models on a small form-factor chip.

Generative AI everywhere you need.

GenAI IP

Reprogrammable. Your way.

Configure your own Generative AI accelerator for your Adaptive SoC.

Choose the device that best fits your needs.

Aión

Cloud AI inference ASIC

More tokens per second, less cost per token. Leading cloud inference performance.

+100% Nvidia H100
+50% Google TPU v6e

Custom NPU

Flexibility: The core of our technology

From the fastest Edge inferencer at 9mm2 to HBM server chips: the performance your application needs.

Discover our custom NPU design services

Get Started Today!

Contact us to begin evaluating our accelerators

See firsthand how our AI solutions transform your devices

Experience the future of Generative AI acceleration with RaiderChip