Files
ComfyUI_frontend/apps/hub/knowledge/models/chronoedit.md
dante01yoon bbd0a6b201 feat: migrate workflow template site as apps/hub
Migrate workflow_templates/site into the frontend monorepo as apps/hub
so the hub can use @comfyorg/design-system and shared packages.

Changes to existing files:
- pnpm-workspace.yaml: add @astrojs/sitemap, @astrojs/vercel, lucide-vue-next
- eslint.config.ts: add hub ignores and i18n/import rule overrides
- .oxlintrc.json: add hub scripts to ignore patterns
- knip.config.ts: add hub workspace config

apps/hub adaptations from source:
- Replace local cn() with @comfyorg/tailwind-utils (19 files)
- Integrate @comfyorg/design-system/css/base.css in global.css
- Make TEMPLATES_DIR configurable via HUB_TEMPLATES_DIR env var
- Add HUB_SKIP_SYNC flag for builds without template data
- Remove Vite 8-incompatible rollupOptions.output.manualChunks
- Fix stylelint violations (modern color notation, number precision)
- Gitignore generated content (thumbnails, synced templates, AI cache)
2026-04-06 20:53:13 +09:00

2.2 KiB

ChronoEdit

ChronoEdit is an image editing framework by NVIDIA that reframes editing as a video generation task, using temporal reasoning to ensure physically plausible and consistent edits.

Model Variants

ChronoEdit-14B

  • Full 14 billion parameter model for maximum quality
  • Built on pretrained video diffusion model architecture
  • Requires ~34GB VRAM (38GB with temporal reasoning enabled)

ChronoEdit-2B

  • Compact 2 billion parameter variant for efficiency
  • Maintains core temporal reasoning capabilities
  • Lower VRAM requirements for broader hardware compatibility

ChronoEdit-14B 8-Step Distilled LoRA

  • Distilled variant requiring only 8 inference steps
  • Faster generation with minimal quality loss
  • Uses flow-shift 2.0 and guidance-scale 1.0

Key Features

  • Treats image editing as a video generation task for temporal consistency
  • Temporal reasoning tokens simulate intermediate editing trajectories
  • Ensures physically plausible edits (object interactions, lighting, shadows)
  • Two-stage pipeline: temporal reasoning stage followed by editing frame generation
  • Prompt enhancer integration for improved editing instructions
  • LoRA fine-tuning support via DiffSynth-Studio
  • Upscaler LoRA available for super-resolution editing
  • PaintBrush LoRA for sketch-to-object editing
  • Apache-2.0 license

Hardware Requirements

  • 14B model: 34GB VRAM minimum (38GB with temporal reasoning)
  • 2B model: 12GB+ VRAM estimated
  • Supports model offloading to reduce peak VRAM
  • Linux only (not supported on Windows/macOS)

Common Use Cases

  • Physically consistent image editing (add/remove/modify objects)
  • World simulation for autonomous driving and robotics
  • Visualizing editing trajectories and reasoning
  • Image super-resolution via upscaler LoRA
  • Sketch-to-object conversion via PaintBrush LoRA

Key Parameters

  • prompt: Text description of the desired edit
  • num_inference_steps: Denoising steps (default ~50, or 8 with distilled LoRA)
  • guidance_scale: Prompt adherence strength (default ~7.5, or 1.0 with distilled LoRA)
  • flow_shift: Flow matching shift parameter (2.0 for distilled LoRA)
  • enable_temporal_reasoning: Toggle temporal reasoning stage for better consistency