--- license: other license_name: lucy-edit-dev-model-non-commercial-license-v1.0 license_link: >- https://d2drjpuinn46lb.cloudfront.net/LUCY_EDIT-Non_Commercial_License_17_Sep_2025.pdf pipeline_tag: video-to-video library_name: diffusers --- # Lucy Edit Dev (5B)

Lucy Edit Dev Logo

๐Ÿงช GitHub  |  ๐Ÿ“– Playground  |  ๐Ÿ“‘ Technical Paper  |  ๐Ÿ’ฌ Discord

---

Put the woman in gothic black jeans and leather jacket and crop top under it.

1.2) Put her in a clown outfit.

1.3) Put the woman in a red bikini with an open thick coat above it.
**Lucy Edit Dev** is an open-weight **video editing** model that performs **instruction-guided edits** on videos using free-text prompts โ€” it supports a variety of edits, such as **clothing & accessory changes**, **character changes**, **object insertions**, and **scene replacements** while preserving the motion and composition perfectly. - ๐Ÿš€ **First open-source instruction-guided video editing model** - ๐Ÿงฉ **Built on Wan2.2 5B architecture** โ€” inherits high-compression VAE + DiT stack, making adapting existing scripts and workflows easy. - ๐Ÿƒโ€โ™‚๏ธ **Motion Preservation** - preserves the motion and composition of videos perfectly, allowing precise edits. - ๐ŸŽฏ **Edit reliability** โ€” edits are more robust when compared to common inference time methods. - ๐Ÿงข **Wardrobe & accessories** โ€” change outfits, add glasses/earrings/hats/etc. - ๐ŸงŒ **Character Changes** โ€” replace characters with monsters, animals and known characters. (e.g., "Replace the person with a polar bear") - ๐Ÿ—บ๏ธ **Scenery swap** โ€” move the scene (e.g., "transform the scene into a 2D cartoon,") - ๐Ÿ“ **Pure text instructions** โ€” no finetuning, no masks required for common edits โ„น๏ธ Model size: **~5B params**. Build on top of **Wan2.2 5B**. --- ## ๐ŸŽฌ Demos
### Sample 1

1.1) Turn the man into an alien

1.2) Turn the man into a bear

1.3) Make it snowy
### Sample 2

2.1) Turn the woman into Harley Quinn

2.2) Turn the woman into Lego

2.3) Turn the shirt into a sports jersey
Note: The prompts above are not enriched, the model will react better to enriched prompts - as described in the prompt guideline section below. --- ## ๐Ÿ”ฅ Latest News - **[2025-09-18]**: Initial **Lucy Edit Dev** weights & reference code released. - **[2025-09-16]**: Diffusers integration PR opened and merged. PR #12340. --- ## ๐Ÿ› ๏ธ Quickstart ### Installation ```bash pip install git+https://github.com/huggingface/diffusers ``` ### Inference Please refer to the "Prompting Guidelines & Supported Edits" section for the best experience. ```python from typing import List import torch from PIL import Image from diffusers import AutoencoderKLWan, LucyEditPipeline from diffusers.utils import export_to_video, load_video # Arguments url = "https://d2drjpuinn46lb.cloudfront.net/painter_original_edit.mp4" prompt = "Change the apron and blouse to a classic clown costume: satin polka-dot jumpsuit in bright primary colors, ruffled white collar, oversized pom-pom buttons, white gloves, oversized red shoes, red foam nose; soft window light from left, eye-level medium shot, natural folds and fabric highlights." negative_prompt = "" num_frames = 81 height = 480 width = 832 # Load video def convert_video(video: List[Image.Image]) -> List[Image.Image]: video = load_video(url)[:num_frames] video = [video[i].resize((width, height)) for i in range(num_frames)] return video video = load_video(url, convert_method=convert_video) # Load model model_id = "decart-ai/Lucy-Edit-Dev" vae = AutoencoderKLWan.from_pretrained(model_id, subfolder="vae", torch_dtype=torch.float32) pipe = LucyEditPipeline.from_pretrained(model_id, vae=vae, torch_dtype=torch.bfloat16) pipe.to("cuda") # Generate video output = pipe( prompt=prompt, video=video, negative_prompt=negative_prompt, height=480, width=832, num_frames=81, guidance_scale=5.0 ).frames[0] # Export video export_to_video(output, "output.mp4", fps=24) ``` --- ## Prompting Guidelines & Supported Edits Lucy Edit is built for **precise, realistic, and identity-preserving video edits.** Prompts with ~20โ€“30 descriptive words work best. Using the right **trigger words** helps the model understand your intent. ### Trigger Words - **Change** โ†’ Clothing or color modifications - **Add** โ†’ Adding animals or objects - **Replace** โ†’ Object substitution or subject swap - **Transform to** โ†’ Global scene or style transformations ### Supported Edit Types #### 1. Clothing Changes โœ… **Best performance.** Lucy Edit excels at swapping outfits while preserving motion, pose, and identity. *Example*: *โ€œChange the shirt to a kimono with wide sleeves and patterned fabric.โ€* #### 2. Human/Character Replacement โœ… **Strong results.** Works well for transforming people into new characters or creatures. Detailed prompts are key. *Example*: *โ€œReplace the person with a tiger, striped orange fur, muscular build, and glowing green eyes.โ€* *Example*: *โ€œReplace the person with an 2D anime character, big eyes, blue gown and battle scars.โ€* #### 3. Replace Objects โœ… **Reliable for structure-preserving swaps.** Ideal when replacing one object with another of similar scale. *Example*: *โ€œReplace the apple with a glowing crystal ball emitting blue light.โ€* #### 4. Color Changes โš ๏ธ **Mixed reliability.** Sometimes subtle, sometimes exaggerated. Works best with precise descriptions. *Example*: *โ€œChange the jacket color to deep red leather with a glossy finish.โ€* #### 5. Add Objects โš ๏ธ **Often attaches to the subject.** Works best for wearable or handheld props. *Example*: *โ€œAdd a golden crown on the personโ€™s head, decorated with ornate jewels.โ€* #### 6. Global Transformations โš ๏ธ **Effective for backgrounds or scene-wide changes, might alter the subject** Alter environment or style, might, Often changes the identity of the subject. *Example*: *โ€œTransform the sunny beach into a snowy tundra with falling snowflakes.โ€* ### Additional Notes - **Strengths:** Lucy Edit excels at **identity conservation, edit precision, realism, and prompt adherence.** - **Detail matters:** Longer prompts (20โ€“30 words) describing style, appearance, and context improve results. - **Frame count:** 81-frame generations produce better temporal consistency than shorter clips. --- ## ๐Ÿ“ฆ Integrations * โ˜๏ธ **Hosted API:** You can access the model on our API and get 5000 free credits here. * ๐Ÿงจ **Diffusers:** *Coming soon* * ๐Ÿงฉ **ComfyUI:** *Coming soon* ## ๐Ÿงญ Roadmap * โœ… Public Batch API. * โœ… Diffusers pipeline (`LucyEditPipeline`) * โœ… Remote ComfyUI custom nodes. * โœ… Technical Report * [ ] Local Inference ComfyUI Nodes. * [ ] LoRA and fine-tuning scripts. --- ## ๐Ÿ”’ License This model falls under the LUCY EDIT DEV MODEL Non-Commercial License v1.0 --- ## ๐Ÿ“ฃ Citation ```bibtex @article{decart2025lucyedit, title = {Lucy Edit: Open-Weight Text-Guided Video Editing}, author = {DecartAI Team}, year = {2025} url = { https://d2drjpuinn46lb.cloudfront.net/Lucy_Edit__High_Fidelity_Text_Guided_Video_Editing.pdf} } ``` --- ## ๐Ÿ™ Acknowledgements Lucy Edit Dev builds on the excellent foundations of **Wan2.2** (5B), and thanks the broader open-source community including **diffusers** and **Hugging Face**. --- ## ๐Ÿ“ฌ Contact * GitHub Issues: DecartAI/lucy-edit. * Discord: Join our discord server, here.