This guide details integrating Azure OpenAI's Sora 1 and Sora 2 text-to-video models into web applications. Sora 1 offers extensive resolution and duration options, while Sora 2 introduces native audio and additional generation modes. Understanding the distinct API structures for Sora 1 and Sora 2 is crucial for correct request formatting. Key API parameters include the model name, prompt, resolution, and duration, with Sora 2 using OpenAI's standard format and Sora 1 using Azure-specific parameters. Sora 2 supports unique modes like image-to-video and video-to-video remix, which require specific handling. Cost analysis reveals Sora 1 is generally more economical for silent videos, whereas Sora 2 is preferred for its advanced features. Sora 1 provides a wider range of resolutions and durations, while Sora 2's current preview is limited to 720p with fixed durations. Both models enforce a limit of two concurrent video generation jobs and make generated videos available for 24 hours. Best practices include implementing adaptive backoff for job status polling and handling differing response structures for downloading videos. Robust error handling and image preprocessing for image-to-video are also essential for successful implementation.
techcommunity.microsoft.com
techcommunity.microsoft.com
Create attached notes ...
