~$Q25VL
Efficient vision-language model at 7B scale. Handles images, documents, and video frames with strong multimodal reasoning. Runs on all 4 universal-a nodes.
Cold start comparison vs similar models. Lower is better.
No subscriptions. Buy credits, pay per inference. Scale to zero when idle.
import cumulus from "cumulus-sdk" // Deploy Qwen 2.5 VL 7B on Ion const client = await cumulus.deploy("qwen2-5-vl-7b") // Run inference const result = await client.run({ prompt: "Your prompt here", // model-specific params... })