diff --git a/README.md b/README.md
new file mode 100644
index 0000000..7c7b3f7
--- /dev/null
+++ b/README.md
@@ -0,0 +1,340 @@
+# Image-to-Video
+
+本项目**MS-Image2Video**旨在解决根据输入图像生成高清视频任务。**MS-Image2Video**由达摩院研发的高清视频生成基础模型,其核心部分包含两个阶段,分别解决语义一致性和清晰度的问题,参数量共计约37亿,模型经过在大规模视频和图像数据混合预训练,并在少量精品数据上微调得到,该数据分布广泛、类别多样化,模型对不同的数据均有良好的泛化性。项目于现有的视频生成模型,**MS-Image2Video**在清晰度、质感、语义、时序连续性等方面均具有明显的优势。
+
+此外,**MS-Image2Video**的许多设计理念继承于我们以公开工作**VideoComposer**,您可以参考我们的[VideoComposer](https://videocomposer.github.io)和本项目的Github代码库了解详细细节
+
+The **MS-Image2Video** project aims to address the task of generating high-definition videos based on input images. Developed by Alibaba Cloud, the **MS-Image2Video** is a fundamental model for generating high-definition videos. Its core components consist of two stages that address the issues of semantic consistency and clarity, totaling approximately 3.7 billion parameters. The model is pre-trained on a large-scale mix of video and image data and fine-tuned on a small number of high-quality data sets with a wide range of distributions and diverse categories. The model demonstrates good generalization capabilities for different data types. Compared to existing video generation models, **MS-Image2Video** has significant advantages in terms of clarity, texture, semantics, and temporal continuity.
+
+Additionally, many of the design concepts for **MS-Image2Video** are inherited from our publicly available work, **VideoComposer**. For detailed information, please refer to our [VideoComposer](https://videocomposer.github.io) and the Github code repository for this project.
+
+
+
+
+
+ Fig.1 MS-Image2Video
+
+
+## 模型介绍 (Introduction)
+
+**MS-Image2Video**建立在Stable Diffusion之上,如图Fig.2所示,通过专门设计的时空UNet在隐空间中进行时空建模并通过解码器将重建出最终视频。为能够生成720P视频,我们将**MS-Image2Video**分为两个阶段,第一阶段保证语义一致性但低分辨率,第二阶段通过DDIM逆运算并在新的VLDM上进行去噪以提高视频分辨率已经时间和空间上的一致性。通过在模型、训练和数据上的联合优化,本项目主要具有以下几个特点:
+
+- 高清&宽屏,可以直接生成720P(1280*720)分辨率的视频,且相比于现有的开源项目,不仅分辨率得到有效提高,其生产的宽屏视频可以适合更多的场景
+- 无水印,模型通过我们内部大规模无水印视频/图像训练,并在高质量数据微调得到,生成的无水印视频可适用更多视频平台,减少许多限制
+- 连续性,通过特定训练和推理策略,在视频的细节生成的稳定性上(时间和空间维度)有明显提高
+- 质感好,通过收集特定的风格的视频数据训练,使得生成的模型在质感得到明显提升,可以生成科技感、电影色、卡通风格和素描等类型视频
+
+以下为生成的部分案例:
+
+**MS-Image2Video** is built on Stable Diffusion, as shown in Fig.2, and uses a specially designed spatiotemporal UNet to perform spatiotemporal modeling in the latent space, and then reconstructs the final video through the decoder. In order to generate 720P videos, **MS-Image2Video** is divided into two stages. The first stage guarantees semantic consistency but with low resolution, while the second stage uses the DDIM inverse operation and applies denoising on a new VLDM to improve the resolution and spatiotemporal consistency of the video. Through joint optimization of the model, training, and data, this project has the following characteristics:
+
+- High-definition & widescreen, can directly generate 720P (1280*720) resolution videos, and compared to existing open source projects, not only is the resolution effectively improved, but the widescreen videos it produces can also be suitable for more scenarios.
+- No watermark, the model is trained on a large-scale watermark-free video/image dataset internally and fine-tuned on high-quality data, generating watermark-free videos that can be applied to more video platforms and reducing many restrictions.
+- Continuity, through specific training and inference strategies, there is a significant improvement in the stability of detail generation in videos (in the time and space dimensions).
+- Good texture, by collecting specific style video data for training, the generated model has a significant improvement in texture and can generate technology, film color, cartoon style, sketch and other types of videos.
+
+
+Below are some examples generated by the model:
+
+
+
+
+
+ Fig.2 VLDM
+
+
+
+**为方便展示,本页面展示为低分辨率GIF格式,但是GIF会下降视频质量,具体效果可以参下面的视频链接**
+
+**For display purposes, this page shows low-resolution GIF format. However, GIF format may reduce video quality. For specific effects, please refer to the video link below.**
+
+
+
+
+
+
+### 依赖项 (Dependency)
+
+本**MS-Image2Video**项目适配ModelScope代码库,以下是本项目需要安装的部分依赖项:
+
+The **MS-Image2Video** project is compatible with the ModelScope codebase, and the following are some of the dependencies that need to be installed for this project.
+
+```bash
+pip install modelscope==1.4.2
+pip install torch==1.11.0+cu113
+pip install torchvision==0.12.0+cu113
+pip install open_clip_torch>=2.0.2
+pip install easydict
+pip install numpy
+pip install opencv-python-headless
+pip install opencv-python
+pip install einops>=0.4
+pip install rotary-embedding-torch
+pip install fairscale
+pip install scipy
+pip install imageio
+pip install oss2
+pip install pytorch-lightning
+```
+
+
+### 快速使用 (Inference)
+
+关于更多的尝试,请关注我们将公开的技术报告和开源代码。
+
+For more experiments, please stay tuned for our upcoming technical report and open-source code release.
+
+
+### 代码范例 (Code example)
+```python
+
+```
+
+
+### 模型局限 (Limitation)
+
+
+本**MS-Image2Video**项目的模型在处理以下情况会存在局限性:
+- 小目标生成能力有限,在生成较小目标的时候,会存在一定的错误
+- 快速运动目标生成能力有限,当生成快速运动目标时,会存在一定的假象
+- 生成速度较慢,生成高清视频会明显导致生成速度减慢
+
+此外,我们研究也发现,生成的视频空间上的质量和时序上的变化速度在一定程度上存在互斥现象,在本项目我们选择了其折中的模型,兼顾两则的平衡。
+
+
+The model of the **MS-Image2Video** project has limitations in the following scenarios:
+- Limited ability to generate small objects: There may be some errors when generating smaller objects.
+- Limited ability to generate fast-moving objects: There may be some artifacts when generating fast-moving objects.
+- Slow generation speed: Generating high-definition videos significantly slows down the generation speed.
+
+Additionally, our research has found that there is a trade-off between the spatial quality and temporal variability of the generated videos. In this project, we have chosen a model that strikes a balance between the two.
+
+
+*如果您正在尝试使用我们的模型,我们建议您首先在使用第一阶段得到满意的符合语义的视频之后,再尝试第二阶段的调整(因为该过程比较耗时),这样可以提高您的使用效率,更容易得到更好的结果。*
+
+*If you are trying to use our model, we recommend that you first focus on obtaining satisfactory semantic-consistent videos using the first stage before attempting adjustments in the second stage (as this process can be time-consuming). This approach will improve your efficiency and increase the likelihood of achieving better results.*
+
+
+## 训练数据介绍 (Training Data)
+
+我们训练数据主要来源来源广泛,具备以下几个属性:
+
+- 混合训练,模型有按照视频图像比7:1训练模型,以此保证视频生成质量
+- 类别分布广,数据数十亿的总体量涵盖人、动物、机车、科幻、场景等等绝大多数的实际数据
+- 来源分布广,数据来源于开源数据、视频网站以及其他内部数据,具有多分辨率、长宽比等
+- 精品数据构建,为了提升模型生成的质量,我们构建了约20w的精品数据对预训练模型进行微调
+
+Our training data mainly comes from various sources and has the following attributes:
+
+- Mixed training: The model is trained with a 7:1 ratio of video to image to ensure the quality of video generation.
+- Wide class distribution: The data set covers most real-world categories, including people, animals, locomotives, science fiction, scenes, etc. with a total volume of billions of data points.
+- Wide source distribution: The data comes from open-source data, video websites, and other internal sources, with varying resolutions and aspect ratios.
+- High-quality data construction: To improve the quality of the model-generated videos, we constructed approximately 200,000 high-quality data pairs for fine-tuning the pre-training model.
+
+
+相关的技术文档正在撰写中,欢迎及时关注。
+
+The relevant technical report is currently being written, and we welcome you to stay tuned for updates.
+
+
+## 相关论文以及引用信息 (Reference)
+```
+@article{videocomposer2023,
+ title={VideoComposer: Compositional Video Synthesis with Motion Controllability},
+ author={Wang, Xiang* and Yuan, Hangjie* and Zhang, Shiwei* and Chen, Dayou* and Wang, Jiuniu and Zhang, Yingya and Shen, Yujun and Zhao, Deli and Zhou, Jingren},
+ journal={arXiv preprint arXiv:2306.02018},
+ year={2023}
+}
+
+@inproceedings{videofusion2023,
+ title={VideoFusion: Decomposed Diffusion Models for High-Quality Video Generation},
+ author={Luo, Zhengxiong and Chen, Dayou and Zhang, Yingya and Huang, Yan and Wang, Liang and Shen, Yujun and Zhao, Deli and Zhou, Jingren and Tan, Tieniu},
+ booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
+ year={2023}
+}
+```
+
+## 使用协议 (License Agreement)
+我们的代码和模型权重仅可用于个人/学术研究,暂不支持商用。
+
+Our code and model weights are only available for personal/academic research use and are currently not supported for commercial use.
+
+## 联系我们 (Contact Us)
+如果你想联系我们的算法/产品同学, 或者想加入我们的算法团队(实习/正式), 欢迎发邮件至: 。