HappyHorse maybe will be open weights soon (it beat seedance 2.0 on Artificial Analysis!)

Posted by External_Mood4719@reddit | LocalLLaMA | View on Reddit | 12 comments

source unified large model for text-to-video/image-to-video + audio) has recently been making waves on the international stage. After verification from multiple sources, the team behind it has been revealed: they are from the obao and Tmall Group (TTG) Future Life Labled by ang Di(The lab was created by the ATH-AI Innovation Business Department and has since become an independent entity).

ofile of Zhang Di: He holds both a Bachelor's and Master's degree from Shanghai Jiao Tong University. He is the head of the TTG Future Life Lab (Rank: P11) and reports to Zheng Bo, Chief Scientist of TTG and CTO of Alimama. He previously served as the lead (No. 1 position) for Kuaishou’s ing.d prior to that, he was the head of Big Data and Machine Learning Engineering Architecture at Alimama.

P.S. 1. It is rumored that HappyHorse 1.0 will be officially released on the 10th of this month. (It has been undergoing intensive testing recently; in fact, information was leaked back in March, but Alibaba PR immediately deleted the relevant sources). Word is that the team will also release several different types of models, so stay tuned. 2. Alimama is the algorithm platform within the Taobao and Tmall ecosystem and has produced many renowned algorithm experts (this is also the birthplace of the Wan model). After honing his skills at Kuaishou’s Kling, Zhang Di’s return is described as "a fish back in water." He is reportedly extremely excited lately. The team at Xixi District C works late every night and is even happily putting in overtime on Saturdays.

[Basic Information]

  1. Model Type: Open-source unified model for Text-to-Video / Image-to-Video + Audio.

  2. Inference Paradigm: Single Transformer Transfusion, CFG-less (Classifier-Free Guidance-less).

  3. Inference Steps: 8 steps.

[Video Parameters]

* **Resolution:** 1280×720 (720p)

* **Frame Rate:** 24fps

* **Duration:** 5 seconds

**[Audio Capabilities]**

* **Native Synchronous Generation:** Sound effects / Ambient sound / Voiceover

* **Supported Languages:** Chinese, English, Japanese, Korean, German, French

**[Open Source Status]**

* **Fully Open Source:** Base model + Distilled model + Super-resolution + Inference code

Source: https://mp.weixin.qq.com/s/n66lk5q_Mm10UYTnpEOf3w?poc_token=HKwe1mmjFX-RhveuVjk_MbRgFTcirVE2tKrRP_gS