TechFlow News, March 30: According to JINSHI Data, Alibaba Qwen has announced the launch of its new multimodal large language model, Qwen3.5-Omni. The Qwen3.5-Omni series includes three Instruct versions—Plus, Flash, and Light—supporting a context length of up to 256K tokens. The model accepts audio inputs exceeding 10 hours in duration and video-audio inputs up to 400 seconds long at 720p resolution (1 frame per second). Trained natively on massive amounts of text, visual, and over 100 million hours of audio-video data, Qwen3.5-Omni demonstrates exceptional multimodal perception and generation capabilities. Compared to Qwen3-Omni, Qwen3.5-Omni significantly enhances multilingual support, enabling speech recognition in 113 languages and dialects and speech synthesis in 36 languages and dialects.
Navigating Web3 tides with focused insights
Contribute An Article
Media Requests
Risk Disclosure: This website's content is not investment advice and offers no trading guidance or related services. Per regulations from the PBOC and other authorities, users must be aware of virtual currency risks. Contact us / support@techflowpost.com ICP License: 琼ICP备2022009338号




