When you purchase through links on our site, we may earn an affiliate commission.Heres how it works.

Current solutions are not optimized for inference and rely on general-purpose CPUs, which were not designed for AI.

NeuReality says this will deliver improved performance and scalability at a lower cost alongside reduced power consumption.

NeuReality NR1 AI Inference Solution

“We didnt have a go at just improve an already flawed system.

This new architecture enables inference through hardware with AI-over-Fabric, an AI hypervisor, and AI-pipeline offload.

The latter is an AI-centric inference server containing NR1-M modules with the NR1 NAPU.

More from TechRadar Pro