LLARVA
Vision-Language-Action model using LLM backbones for structured robot action prediction. Bridges language models and low-level robot control.
0.7
Response will appear here...
Pricing
API Integration
Use our OpenAI-compatible API to integrate LLARVA into your application.
npm install railwailimport railwail from "railwail";
const rw = railwail("YOUR_API_KEY");
// Simple — just pass a string
const reply = await rw.run("llarva", "Hello! What can you do?");
console.log(reply);
// With message history
const reply2 = await rw.run("llarva", [
{ role: "system", content: "You are a helpful assistant." },
{ role: "user", content: "Explain quantum computing simply." },
]);
console.log(reply2);
// Full response with usage info
const res = await rw.chat("llarva", [
{ role: "user", content: "Hello!" },
], { temperature: 0.7, max_tokens: 500 });
console.log(res.choices[0].message.content);
console.log(res.usage);Free credits on sign-up
Related Models
View all VLA / RoboticsGR00T N1
NVIDIA's foundation model for humanoid robots. World-model-based VLA enabling whole-body control and human-like manipulation.
OpenVLA
Open-source 7B Vision-Language-Action model built on Prismatic VLM and Llama 2. Converts visual observations and language goals into robot actions.
Pi0
Physical Intelligence's foundation model for robot control. Combines vision-language understanding with dexterous manipulation across diverse tasks.
RT-2
Google DeepMind's Robotic Transformer 2. Vision-Language-Action model that translates visual observations and language instructions directly into robot actions.
Start using LLARVA today
Get started with free credits. No credit card required. Access LLARVA and 100+ other models through a single API.