Bridge the gap between open-source models and proprietary workflows. A complete local environment for Image Analysis, LLM Prompt Generation, and Batch Video Queuing.
Moving from a reference image to a consistent cinematic prompt—and tracking parameters across hundreds of generations—requires a dedicated space.
No cloud dependencies for generation. i2vStudio ties directly into your local or remote ComfyUI node and localized Ollama models, keeping your assets strictly private.
Let LLMs do the heavy lifting. The built-in orchestrator converts natural language into perfect FLUX image prompts and formats strict 80-word Wan 2.2 cinematic sequences.
Design dozens of variations using dynamically saved taxonomy tokens. Push massive jobs to the ComfyUI queue instantly, or export your sequences to portable ZIP manifests.
Explore the specialized tools designed to streamline your video pipeline.
Gain complete visibility over your local hardware and node states. Real-time Nvidia-SMI GPU polling prevents OOM errors, while directory monitoring tracks your exact asset cache size.
Drop in any reference picture to reverse-engineer its cinematic properties. Using local Vision Models (like LLaVA), the app autonomously extracts themes, visual styles, lighting, and camera language directly into your taxonomy database.
Need a custom starting frame? Use the Image Studio to prompt SDXL or FLUX workflows dynamically. Enable the LLM Assistant to convert short ideas into highly detailed, prompt-engineered structures before generating and saving to your ComfyUI output.
Select your saved elements to synthesize the ultimate Wan 2.2 generative prompt (targeting the optimal 80-120 word count). Adjust frame length, FPS, and target resolutions, then stage the sequence into your pending batch queue.
A built-in file manager connected directly to ComfyUI's /input/ and /output/ directories. Toggle "Flatten" view to see all images deeply nested in subdirectories. Instantly copy outputs back into inputs for looping workflows.
Review your queued definitions. When ready, click execute to push the entire JSON array to ComfyUI seamlessly. Want to share your workload? Export the queue as a ZIP archive containing all modified JSON workflows and their required start frames, ready for import on any other node.
Manage the taxonomy of your prompt elements. Easily bulk-inject CSV lists of Camera Moves, Lighting adjustments, and Visual Styles. Edit or delete elements to keep your workshop environment uncluttered.
A floating AI Assistant lives in the corner of your workspace. Tell it to "make a sci-fi video of a robot," and the agent will intelligently formulate a multi-step JSON plan—rendering the image, moving it to your inputs, and staging the final video queue entirely autonomously.
i2vStudio acts as a centralized brain, meaning you must configure your worker engines properly to accept incoming API instructions.
Standard API port is expected at 8188.
input/ and output/ directory paths match what you have set in app.py configuration (Default: /media/tish/inference/ComfyUI/).Default API port is expected at 11434. Required for VLM extraction and Agent logic.
Click any command block to copy