LLaVA 1.6 34B
Open-source multimodal model combining language and vision. Strong visual understanding with conversational capabilities.
0.7
Response will appear here...
Pricing
API Integration
Use our OpenAI-compatible API to integrate LLaVA 1.6 34B into your application.
npm install railwailimport railwail from "railwail";
const rw = railwail("YOUR_API_KEY");
// Simple — just pass a string
const reply = await rw.run("llava-1-6-34b", "Hello! What can you do?");
console.log(reply);
// With message history
const reply2 = await rw.run("llava-1-6-34b", [
{ role: "system", content: "You are a helpful assistant." },
{ role: "user", content: "Explain quantum computing simply." },
]);
console.log(reply2);
// Full response with usage info
const res = await rw.chat("llava-1-6-34b", [
{ role: "user", content: "Hello!" },
], { temperature: 0.7, max_tokens: 500 });
console.log(res.choices[0].message.content);
console.log(res.usage);Free credits on sign-up
Related Models
View all MultimodalGPT-4o (Vision)
GPT-4o's vision capabilities. Analyze images, charts, documents, and screenshots with detailed understanding and reasoning.
Gemini 2.0 Flash (Multimodal)
Google's multimodal model accepting text, images, audio, and video. Native multimodal understanding across input types.
Claude 3.5 Sonnet (Vision)
Claude's vision capabilities. Excellent at analyzing images, documents, and code screenshots with detailed, accurate descriptions.
Pixtral Large
Mistral's vision-language model. 124B parameters with native image understanding, document analysis, and visual reasoning.
Start using LLaVA 1.6 34B today
Get started with free credits. No credit card required. Access LLaVA 1.6 34B and 100+ other models through a single API.