Kaynağa Gözat

update readme

Yuwei Guo 2 yıl önce
ebeveyn
işleme
3742e3e524

+ 13 - 6
README.md

@@ -31,12 +31,10 @@ Our approach takes around 60 GB GPU memory to inference. NVIDIA A100 is recomman
 
 ```
 git clone https://github.com/guoyww/animatediff.git
-cd animatediff
+cd AnimateDiff
 
-conda create -n animatediff python=3.8
+conda env create -f environment.yaml
 conda activate animatediff
-
-pip install -r requirements.txt
 ```
 
 ### Download Base T2I & Motion Module Checkpoints
@@ -65,7 +63,7 @@ bash download_bashscripts/8-GhibliBackground.sh
 ```
 
 ### Inference
-After downloading the above peronalized T2I checkpoints, run the following commands to generate animations.
+After downloading the above peronalized T2I checkpoints, run the following commands to generate animations. The results will automatically be saved to `samples/` folder.
 ```
 python -m scripts.animate --config configs/prompts/1-ToonYou.yaml
 python -m scripts.animate --config configs/prompts/2-Lyriel.yaml
@@ -100,7 +98,16 @@ python -m scripts.animate --prompt configs/prompts/lora.yaml
 ``` -->
 
 ## Gallery
-Here we demonstrate several best results we got in previous experiments.
+Here we demonstrate several best results we found in our experiments or generated by other artists.
+<table class="center">
+    <tr>
+    <td><img src="__assets__/animations/model_07/01.gif"></td>
+    <td><img src="__assets__/animations/model_07/02.gif"></td>
+    <td><img src="__assets__/animations/model_07/03.gif"></td>
+    <td><img src="__assets__/animations/model_07/04.gif"></td>
+    </tr>
+</table>
+<p style="margin-left: 2em; margin-top: -1em">Model:<a href="https://civitai.com/models/107295?modelVersionId=115371">holding_sign</a> (samples are contributed by CivitAI artists)</p>
 
 <table class="center">
     <tr>

BIN
__assets__/animations/model_07/01.gif


BIN
__assets__/animations/model_07/02.gif


BIN
__assets__/animations/model_07/03.gif


BIN
__assets__/animations/model_07/04.gif


+ 20 - 0
environment.yaml

@@ -0,0 +1,20 @@
+name: animatediff
+channels:
+  - pytorch
+  - xformers
+dependencies:
+  - python=3.10
+  - pytorch==1.12.1
+  - torchvision==0.13.1
+  - torchaudio==0.12.1
+  - cudatoolkit=11.3
+  - xformers
+  - pip
+  - pip:
+    - diffusers[torch]==0.11.1
+    - transformers==4.25.1
+    - imageio==2.27.0
+    - gdown
+    - einops
+    - omegaconf
+    - safetensors

+ 0 - 10
requirements.txt

@@ -1,10 +0,0 @@
---extra-index-url https://download.pytorch.org/whl/cu113
-torch==1.12.1+cu113
-torchvision==0.13.1+cu113
-diffusers[torch]==0.11.1
-transformers==4.25.1
-imageio==2.27.0
-gdown
-einops
-omegaconf
-safetensors