Deva-3 ⇒ ❲Original❳
The model hallucinated cars sliding, pedestrians walking cautiously, and brake lights flashing. It had never seen snow, but it had learned friction and low-traction behavior from dry roads. It generalized the concept of slipperiness.
Current AVs rely on "predictive models" that assume other drivers are rational. DEVA-3 simulates irrational behavior. It can predict the "jerk" who cuts across three lanes without a blinker because it has seen that episode 10,000 times in training data. Wayve and Ghost Autonomy are rumored to be testing DEVA-3 variants on public roads in London right now. deva-3
The car that avoids the accident, the robot that doesn't drop the egg, and the drone that navigates the forest—they will all be running something very close to DEVA-3 by 2027. Current AVs rely on "predictive models" that assume
They trained DEVA-3 on nothing but dashcam footage from Phoenix, Arizona. Then, they gave it a single frame from a snowy street in Oslo—something it had never seen. Wayve and Ghost Autonomy are rumored to be
Imagine an NPC that doesn't follow a script. In a sandbox game, a DEVA-3-powered NPC could watch you build a fortress, predict you will attack at dawn, and fortify its own walls accordingly—without a single line of explicit logic code. The "Aha Moment" from the Research Paper I spoke with a researcher on the team (who requested anonymity due to an upcoming IPO). He told me about their internal "Genesis Test."