Skip to content

CISLab-HKUST/ID-Driven-SDS

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

10 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Implementation of Score Distillation Sampling Loss with IP-Adapter

This repo is based on IP-Adapter and DreamFusion. Thanks for their work.


Update

  • [2024/03/05] 💦 Release face detection (Retinaface) and recognition (ArcFace) model with tensor input, but not useful for fine-grained details. Maybe an image feature extraction model (Eg. DINO V2) should be added.
  • [2024/03/03] 💦 Release SDS Loss with IP-Adapter-FaceID-PlusV2.

INSTALL

Please refer to IP-Adapter and follow their guidance. If you have any problems with this repo, welcome to add an issue.

Notice: This implementation is based on the special version of IP-Adapter - IP-Adapter-FaceID-PlusV2. You can find some details through their huggingface, link.

Models

Some third-party models should be installed, here is the list.

Face Recognition Model:

链接:https://pan.baidu.com/s/1b2EUPJ3t9n4_3EX91efEmQ 
提取码:av25 
--来自百度网盘超级会员V4的分享

Base Model: Realistic_Vision_V4.0_noVAE

Face Detector and Feature Extractor: insightface - buffalo_l

Image Encoder: CLIP-ViT-H-14-laion2B-s32B-b79K

VAE: sd-vae-ft-mse Better vae model finetuned on LAION Dataset.

IP-Adapter-FaceID-PlusV2 ckpt: Please go to the IP-Adapter Huggingface link and download ip-adapter-faceid-plusv2_sd15.bin .

OpenPose ControlNet: OpenPose ControlNet Annotator and ControlNet are included in this link.

Configs

After downloading all the related weights, please check the config file in configs/guidance_config.py carefully, and change the related weight path to your own path.

About Text to 3D/Avatar Generation

This repo is aimed to be applied on text to 3D avatar generation, we are working hard on it. If you are familiar with Score Distillation Sampling (SDS) Loss-based text to 3d generation, this repo is easy for you to use. Just take the rendered image with gradient as the input, SDS loss will give the proper guidance for 3D optimization.

About ControlNet

I don't check the code with OpenPose ControlNet, if you are interested, you can test on the 2d_playground.py. Just change

guidance = StableDiffusionIPAdapterGuidance(
    device=device,
    config=config,
)

to

guidance = StableDiffusionIPAdapterGuidance(
    device=device,
    config=config,
    use_controlnet=True,
)

and add OpenPose Skeleton Image Path to skeleton_image_path.

Results

Here are the test results, you can generate your own results through 2d_playground.py. Face ID: face

Results: image1 image2 video1 video2

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages