Files
ComfyUI_frontend/apps/hub/knowledge/models/kandinsky.md
dante01yoon bbd0a6b201 feat: migrate workflow template site as apps/hub
Migrate workflow_templates/site into the frontend monorepo as apps/hub
so the hub can use @comfyorg/design-system and shared packages.

Changes to existing files:
- pnpm-workspace.yaml: add @astrojs/sitemap, @astrojs/vercel, lucide-vue-next
- eslint.config.ts: add hub ignores and i18n/import rule overrides
- .oxlintrc.json: add hub scripts to ignore patterns
- knip.config.ts: add hub workspace config

apps/hub adaptations from source:
- Replace local cn() with @comfyorg/tailwind-utils (19 files)
- Integrate @comfyorg/design-system/css/base.css in global.css
- Make TEMPLATES_DIR configurable via HUB_TEMPLATES_DIR env var
- Add HUB_SKIP_SYNC flag for builds without template data
- Remove Vite 8-incompatible rollupOptions.output.manualChunks
- Fix stylelint violations (modern color notation, number precision)
- Gitignore generated content (thumbnails, synced templates, AI cache)
2026-04-06 20:53:13 +09:00

1.6 KiB

Kandinsky

Kandinsky is a family of open-source diffusion models for video and image generation, developed by Kandinsky Lab (Sber AI, Russia). The models support both English and Russian text prompts.

Model Variants

Kandinsky 5.0 Video Pro (19B)

  • HD video at 1280x768, 24fps (5 or 10 seconds)
  • Controllable camera motion via LoRA
  • Top-1 open-source T2V model on LMArena

Kandinsky 5.0 Video Lite (2B)

  • Lightweight model, #1 among open-source in its class
  • CFG-distilled (2x faster) and diffusion-distilled (6x faster) variants
  • Best Russian concept understanding in open source

Kandinsky 5.0 Image Lite (6B)

  • HD image output (1280x768, 1024x1024)
  • Strong text rendering; image editing variant available

Key Features

  • Bilingual support (English and Russian prompts)
  • Flow Matching architecture with MIT license
  • Camera control via trained LoRAs
  • ComfyUI and Diffusers integration
  • MagCache acceleration for faster inference

Hardware Requirements

  • Video Lite: 12GB VRAM minimum with optimizations
  • Video Pro: 24GB+ VRAM recommended
  • NF4 quantization and FlashAttention 2/3 or SDPA supported

Common Use Cases

  • Open-source video generation research
  • Russian and English bilingual content creation
  • Camera-controlled video synthesis
  • Image generation with text rendering
  • Fine-tuning with custom LoRAs

Key Parameters

  • prompt: Text description in English or Russian
  • num_frames: Number of output frames (5s or 10s)
  • resolution: Output resolution (up to 1280x768)
  • steps: Inference steps (varies by distillation level)