OpenVLA
Open-source 7B Vision-Language-Action model built on Prismatic VLM and Llama 2. Converts visual observations and language goals into robot actions.
0.7
Response will appear here...
Pricing
API Integration
Use our OpenAI-compatible API to integrate OpenVLA into your application.
npm install railwailimport railwail from "railwail";
const rw = railwail("YOUR_API_KEY");
// Simple — just pass a string
const reply = await rw.run("openvla", "Hello! What can you do?");
console.log(reply);
// With message history
const reply2 = await rw.run("openvla", [
{ role: "system", content: "You are a helpful assistant." },
{ role: "user", content: "Explain quantum computing simply." },
]);
console.log(reply2);
// Full response with usage info
const res = await rw.chat("openvla", [
{ role: "user", content: "Hello!" },
], { temperature: 0.7, max_tokens: 500 });
console.log(res.choices[0].message.content);
console.log(res.usage);Free credits on sign-up
Related Models
View all VLA / RoboticsGR00T N1
NVIDIA's foundation model for humanoid robots. World-model-based VLA enabling whole-body control and human-like manipulation.
Pi0
Physical Intelligence's foundation model for robot control. Combines vision-language understanding with dexterous manipulation across diverse tasks.
RT-2
Google DeepMind's Robotic Transformer 2. Vision-Language-Action model that translates visual observations and language instructions directly into robot actions.
Gemini Robotics
Google DeepMind's Gemini model adapted for robotics. Leverages Gemini's multimodal understanding for zero-shot robot task planning and execution.
Start using OpenVLA today
Get started with free credits. No credit card required. Access OpenVLA and 100+ other models through a single API.