Skip to content

mattcurf/video_diffusion_play

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

video_diffusion_play

Generate videos from text using the latest Wan2.2 model from https://wan.video and Alibaba's Tongyi Lab

IMAGE ALT Sample Video

"Over-the-shoulder shot, two shot, telephoto lens, warm colors, high contrast lighting, soft lighting, daylight, daylight, close-up shot, center composition.In an eye-level shot, a foreign girl sits by a window. She is wearing a red and black striped sweater over a blue turtleneck, and her short, smooth golden hair is tucked behind her ears. Her gaze is gentle as she looks at someone off-camera with a slight smile, her expression natural and friendly. Her hands are gently folded on the table, where an open book and a white ceramic cup with steam gently rising are placed. The background is out-of-focus, with soft daylight streaming in through the window. Faint warm yellow lights and scattered red decorations are visible, creating a cozy atmosphere."

Hardware Configuration

  • Intel/AMD system with at least 24 cores
  • 256 GB DDR5
  • NVIDIA 5090 w/ 32GB VRAM
  • 150GB free storage space for models

Prerequisites

System setup with:

  • kernel drivers for NVIDIA GPU
  • docker with NVIDIA-Runtime support
  • 'make' installed with your package manager (e.g. apt install make)

Steps for setup and execution

  1. Build docker container
$ make image
  1. Run sample text to video script
$ make run

On NVIDIA 5090 card, it takes about 45 minutes to generate about 81 frames of 1280x720p video

References

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published