The editor of Downcodes learned that the Zhipu technical team has recently open sourced its latest video generation model CogVideoX v1.5, which is another major update of the CogVideoX series since August. This update brings significant performance improvements, provides developers with more powerful video generation tools, and marks a new stage in AI video generation technology. CogVideoX v1.5 has greatly improved video length, resolution and frame rate, and enhanced its ability to understand complex semantics, further expanding its application scenarios.
It is understood that this update has greatly improved the video generation capabilities, including supporting 5-second and 10-second video lengths, 768P resolution, and 16-frame generation capabilities. At the same time, the I2V (image to video) model also supports any size ratio, further enhancing the ability to understand complex semantics.
CogVideoX v1.5 contains two main models: CogVideoX v1.5-5B and CogVideoX v1.5-5B-I2V, which are designed to provide developers with more powerful video generation tools.
What is even more noteworthy is that CogVideoX v1.5 will be simultaneously launched on the Qingying platform and combined with the newly launched CogSound sound effect model to become the "New Qingying" . New Qingying will provide a number of special services, including significant improvements in video quality, aesthetic performance and motion rationality, and support the generation of 10-second, 4K, 60-frame ultra-high-definition videos.
The official introduction is as follows:
The open source of CogVideoX v1.5 brings new vitality to the field of AI video generation and provides developers with more possibilities. I believe that in the future, we will see more innovative applications based on CogVideoX v1.5. The editor of Downcodes will continue to pay attention to the latest developments in this field and bring you more exciting reports.