diff --git a/README.md b/README.md index e84a1a6..ae77ee7 100644 --- a/README.md +++ b/README.md @@ -82,6 +82,12 @@ of the **CogVideoX** open-source model. + [cli_demo](inference/cli_demo.py): A more detailed explanation of the inference code, mentioning the significance of common parameters. + [cli_vae_demo](inference/cli_vae_demo.py): Executing the VAE inference code alone currently requires 71GB of memory, but it will be optimized in the future. + [convert_demo](inference/convert_demo.py): How to convert user input into a format suitable for CogVideoX. Because CogVideoX is trained on long caption, we need to convert the input text to be consistent with the training distribution using a LLM. By default, the script uses GLM4, but it can also be replaced with any other LLM such as GPT, Gemini, etc. ++ [gradio_demo](gradio_demo.py): A simple gradio web UI demonstrating how to use the CogVideoX-2B model to generate videos. + +
+ +
+ + [web_demo](inference/web_demo.py): A simple streamlit web application demonstrating how to use the CogVideoX-2B model to generate videos.
diff --git a/README_zh.md b/README_zh.md index 13b509d..f487cab 100644 --- a/README_zh.md +++ b/README_zh.md @@ -77,6 +77,12 @@ CogVideoX是 [清影](https://chatglm.cn/video?fr=osm_cogvideox) 同源的开源 + [cli_demo](inference/cli_demo.py): 更详细的推理代码讲解,常见参数的意义,在这里都会提及。 + [cli_vae_demo](inference/cli_vae_demo.py): 单独执行VAE的推理代码,目前需要71GB显存,将来会优化。 + [convert_demo](inference/convert_demo.py): 如何将用户的输入转换成适合 CogVideoX的长输入。因为CogVideoX是在长文本上训练的,所以我们需要把输入文本的分布通过LLM转换为和训练一致的长文本。脚本中默认使用GLM4,也可以替换为GPT、Gemini等任意大语言模型。 ++ [gradio_demo](gradio_demo.py): 一个简单的gradio网页应用,展示如何使用 CogVideoX-2B 模型生成视频。 + +
+ +
+ + [web_demo](inference/web_demo.py): 一个简单的streamlit网页应用,展示如何使用 CogVideoX-2B 模型生成视频。
diff --git a/gradio_demo.py b/gradio_demo.py index 65eeb48..ea0b020 100644 --- a/gradio_demo.py +++ b/gradio_demo.py @@ -9,7 +9,6 @@ import torch from diffusers import CogVideoXPipeline from datetime import datetime, timedelta from openai import OpenAI -import spaces import imageio import moviepy.editor as mp from typing import List, Union @@ -88,7 +87,6 @@ def convert_prompt(prompt: str, retry_times: int = 3) -> str: return prompt -@spaces.GPU(duration=240) def infer( prompt: str, num_inference_steps: int, diff --git a/resources/gradio_demo.png b/resources/gradio_demo.png new file mode 100644 index 0000000..a78b78c Binary files /dev/null and b/resources/gradio_demo.png differ