#1 on Artificial Analysis
Top of both text-to-video (Elo 1,357) and image-to-video (Elo 1,415) leaderboards on the Artificial Analysis blind-test arena as of April 2026.
HappyHorse-1.0 from Alibaba's ATH unit launched anonymously on Artificial Analysis in April 2026 and topped both the text-to-video and image-to-video leaderboards. 15-billion parameters, native 1080p, 7-language lip sync. Try it on TwoShot with free credits - no Alibaba Cloud Bailian account needed.
HappyHorse-1.0 is Alibaba's frontier AI video model, built by the ATH (Alibaba Token Hub) AI Innovation Unit inside Taotian Group's Future Life Lab. It is a 15-billion-parameter unified multimodal architecture that produces native 1080p video with synchronized lip-synced audio in seven languages: Mandarin, Cantonese, English, Japanese, Korean, German, and French.
The model first appeared on Artificial Analysis - the independent blind-test benchmarking platform - on April 7, 2026, with no team affiliation listed. Within days it climbed to #1 on both the text-to-video and image-to-video Elo leaderboards. On April 10, the developers revealed via a new X account that HappyHorse was an Alibaba ATH project. The public beta launched on April 27 through the official site and Alibaba Cloud Bailian's Model Studio API, with full commercial release scheduled for May 2026.
HappyHorse is a separate effort from Alibaba's Tongyi Lab (which ships the Wan video series). Tongyi focuses on productivity features like Wan 2.7's "Thinking Mode"; ATH focuses on raw benchmark quality. The two-track strategy is paying off - between them, Alibaba now leads multiple AI video benchmarks heading into mid-2026.
Why HappyHorse-1.0 sits at the top of the Artificial Analysis arena.
Top of both text-to-video (Elo 1,357) and image-to-video (Elo 1,415) leaderboards on the Artificial Analysis blind-test arena as of April 2026.
Crisp 1080p video out of the box - no upscaling, no post-processing. Built on a 15-billion-parameter unified multimodal architecture.
Synchronized audio with phoneme-accurate lip movement across Mandarin, Cantonese, English, Japanese, Korean, German, and French.
Highest I2V Elo on the leaderboard. Drop in a still and HappyHorse animates it with consistent character identity and natural motion.
Pure prompt-driven creation. Coherent physics, natural camera work, and stable subjects across the full clip.
The official API ships through Alibaba Cloud Bailian. TwoShot gives you global access on free credits - no enterprise onboarding, no Chinese cloud account.
Skip the Alibaba Cloud Bailian onboarding. TwoShot puts HappyHorse-style video creation behind a simple browser interface with free credits on signup.
Sign up at twoshot.app. No credit card needed - free credits land in your account immediately.
Type the scene in plain language. Include camera direction, dialogue, mood. Or upload a reference image to drive image-to-video.
The assistant picks the best video model for your prompt and ships native 1080p output. No queue, no waitlist.
How HappyHorse-1.0 stacks up against the AI video models people compare it to. Sora is included for historical context - OpenAI has shut it down.
HappyHorse leads on both leaderboards because it handles every starting point - text, image, and audio.
Pure prompt-driven creation. Coherent physics, natural movement, stable subjects. Currently #1 on Artificial Analysis text-to-video Elo.
Drop in a still and HappyHorse animates it with consistent character identity. The strongest category at Elo 1,415.
Phoneme-accurate dialogue in 7 languages. Speech audio is generated alongside the video, not bolted on after.
On April 7, 2026, a model called "HappyHorse-1.0" appeared on Artificial Analysis's video generation arena with no team affiliation. The platform runs blind-test Elo rankings: users see two AI-generated videos and pick the better one without knowing the source.
HappyHorse climbed fast. Within three days it sat at #1 for both text-to-video (Elo 1,357) and image-to-video (Elo 1,415), beating every named model on the platform - Seedance 2, Veo 3, Kling 3, and the existing top-ranked entries. The community speculated for nearly a week before the developers came forward.
On April 10, a fresh X account confirmed: HappyHorse-1.0 was an Alibaba project, built by the ATH (Alibaba Token Hub) AI Innovation Unit inside Taotian Group's Future Life Lab. The team said the anonymous launch was deliberate - they wanted unbiased benchmark feedback before attaching the Alibaba name.
The public beta opened on April 27, 2026, via the official happyhorsesai.com site and Alibaba Cloud Bailian's Model Studio API. Full commercial release is scheduled for May.
The official Alibaba Cloud Bailian API is enterprise-tiered with a 10% early-access discount. TwoShot gives you HappyHorse-style video creation on free credits with no Bailian onboarding.
Create your first HappyHorse-style video in your browser. Free credits, no waitlist, no Bailian account.
HappyHorse-1.0 is the AI video generation model from Alibaba's ATH (Future Life Lab) unit, part of Taotian Group. It is a 15-billion-parameter unified multimodal model that outputs 1080p video with synchronized lip-synced audio in seven languages. HappyHorse went viral in April 2026 after launching anonymously on the Artificial Analysis benchmark platform and climbing to #1 on both the text-to-video and image-to-video leaderboards. Alibaba confirmed authorship on April 10, 2026, and the public beta launched on April 27 via the official site and Alibaba Cloud Bailian's Model Studio API.
Artificial Analysis runs blind-test Elo rankings where users vote on side-by-side video pairs without knowing which model produced them. HappyHorse-1.0 reached an Elo of 1,357 for text-to-video and 1,415 for image-to-video - the highest scores on either leaderboard at launch. The unified multimodal architecture, 15B parameter scale, and native 1080p output combine to consistently win the blind comparisons against Veo, Kling, and Seedance.
The official HappyHorse API on Alibaba Cloud Bailian is paid (Alibaba ran a 10% early-access discount for enterprise customers at launch). TwoShot gives you free credits on signup so you can try HappyHorse-style video creation without a credit card or a Bailian account. When your free credits run out you can subscribe or buy a credit pack.
Alibaba's ATH (Alibaba Token Hub) AI Innovation Unit, run inside Taotian Group's Future Life Lab. The team launched HappyHorse-1.0 anonymously on April 7, 2026 to gather unbiased benchmark feedback. They revealed Alibaba authorship via a new X account on April 10 once the model topped the leaderboards. ATH is a separate effort from Alibaba's Tongyi Lab (which builds the Wan video series), reflecting Alibaba's two-track AI video strategy.
Sora is no longer available - OpenAI shut it down. While Sora was running, it set the benchmark for AI video physics and led on maximum clip duration (25 seconds). HappyHorse leads on independent blind-test benchmarks (Elo 1,357 / 1,415 on Artificial Analysis) and outputs native 1080p with synchronized 7-language audio out of the box. For practical creation in 2026, HappyHorse is the relevant comparison; Sora is a historical reference.
On the Artificial Analysis arena, HappyHorse-1.0 currently sits above Seedance 2, Veo 3, and Kling 3 in both the text-to-video and image-to-video Elo rankings (no-audio category). Seedance 2 still leads on lip-sync language coverage (8+ vs 7) and multimodal input flexibility (12 reference files). Kling 3 leads on resolution (4K vs 1080p). The honest answer is that HappyHorse wins blind perceptual votes today; the other models lead on specific capability axes. TwoShot lets you try all of them.
Seven languages with phoneme-level lip sync: Mandarin Chinese, Cantonese, English, Japanese, Korean, German, and French. This is fewer than Seedance 2's 8+ but covers the major content markets in Asia, North America, and Europe. The model produces lip-synchronized speech, not just musical or ambient audio.
Both ship out of Alibaba but from different teams. Wan 2.7 is from Tongyi Lab and was released April 6, 2026, focused on a new 'Thinking Mode' for video editing and long-text rendering. HappyHorse-1.0 is from the ATH unit at Taotian Group and was released April 27, 2026, focused on raw generation quality (it tops the blind-test leaderboards). They reflect Alibaba's two-track approach: Tongyi for productivity-focused models, ATH for frontier benchmark performance.
Yes. The official HappyHorse API is available through Alibaba Cloud Bailian (Model Studio) since the April 27, 2026 beta launch. Enterprise tiers include early-access discounts. For developers who want to call HappyHorse-style models without setting up a Bailian account or navigating Chinese cloud onboarding, TwoShot provides assistant-driven access through HTTP endpoints.
Any of the use cases that drove its #1 leaderboard position: cinematic text-to-video shorts, image animation with character consistency, multilingual dialogue scenes (7 languages with lip sync), product showcases, music-video footage, social-format clips. Native 1080p output makes the results suitable for direct upload to YouTube, TikTok, Instagram, and X without upscaling.
Pure prompt-driven creation - the #1 T2V model on Artificial Analysis
Animate any image with consistent character identity - Elo 1,415
Phoneme-synced dialogue in 7 languages, generated alongside the video
How to use HappyHorse for free - no Bailian account required
Developer access without Alibaba Cloud onboarding
Skip the Bailian setup - use HappyHorse-style models on TwoShot
Everything you need to create, transform, and perfect your audio, images, and video
Create original music, beats, and sounds from text descriptions using AI. Any genre, any style.
Create stunning visuals, album covers, thumbnails, and art from text descriptions. Edit and upscale existing images.
Create videos from text or images. Animate photos, create music videos, and produce motion content for social media.
Text-to-speech, voice enhancement, and vocal transformation.
Isolate vocals, drums, bass, and instruments from any track in seconds.
Remove background noise, upscale images, enhance video quality, and polish your media.
Generate custom SFX and foley for games, videos, and podcasts.
Remix tracks in new styles or extend songs seamlessly with AI.
Cowrite lyrics and scripts - draft, refine, and iterate together until every line is right.
Arrange, compose, and produce directly in your browser. Audio, video, images — all in one workspace.
200,000+ royalty-free sounds and samples ready for commercial use.
AI tools for music, video, images, and voice
Turn ideas into tracks faster. Create beats, sounds, and full productions with AI assistance.
Complete video production with AI. Generate videos, images, music, and voiceovers.
Studio-quality audio from any recording. Clean up interviews, enhance voices, and add music.
Production-ready AI for audio, video, and visuals. Full rights clearance, API access, team collaboration.
Transform your creative ideas into tangible sounds with our AI powered tools. Simply describe what you want - "fast drum & bass jungle-style drum loop" or "layered flutes inspired by nature" - and see the magic unfold.
Create stunning visuals from text descriptions. Design album covers, thumbnails, portraits, and art — all through conversation.
Create videos from text or images. Animate photos, produce music videos, and make motion content for social media.
Upload any photo and watch it move. AI-powered motion control turns still images into dynamic dance videos and animations.
Change backgrounds, remove objects, upscale resolution, and edit images through simple conversation. No Photoshop needed.
A creative partner for lyrics and scripts. Get a draft, then go back and forth - refine lines, try new angles, iterate together until it's exactly what you envisioned.
Text-to-speech, voice enhancement, and vocal transformation. Create professional voiceovers in any style or voice.
Isolate vocals, drums, bass, and instruments from any track in seconds. Perfect for remixing, sampling, or creating karaoke versions.
Remove background noise, upscale images, enhance video quality, and polish your media.
Generate custom sound effects and foley for games, videos, and podcasts. From explosions to footsteps, create exactly what you need.
Leverage the power of our AI to reimagine existing samples. Extract particular elements from a sample, or create a completely new sample based on a reference.
Arrange, compose, and produce directly in your browser with our online DAW. Drag and drop samples, add effects, and export your creations.




Explore our library of 200,000+ royalty-free samples. From old-school chops to hyper-pop melodies - chat naturally with stem to find exactly what you need.
From Grammy-winning producers to major labels, see who's creating with TwoShot



