Following the DeepSeek R1, the Alibaba Cloud Tongyi Qianwen team has just announced the launch of its latest open-source model
The newly released Qwen2.5-1M series includes two open-source models:
Following the DeepSeek R1, the Alibaba Cloud Tongyi Qianwen team has just announced the launch of its latest open-source model
The newly released Qwen2.5-1M series includes two open-source models:
Welcome to the [AI Daily] column! This is your daily guide to exploring the world of artificial intelligence. Every day, we present you with hot topics in the AI field, focusing on developers, helping you understand technical trends, and learning about innovative AI product applications.
Welcome to the AIbase [AI Daily Report] section! Spend three minutes a day to learn about the latest AI events, helping you understand AI industry trends and innovative AI product applications. For more AI news, visit: https://www.aibase.com/zh1. Baidu officially releases the WENXIN Large Model 4.5 series and fully opens it to the public, featuring ten new models with various parameter configurations. These models are trained and inferred using the PaddlePaddle framework, achieving a FLOPs utilization rate of 47%, and perform well in multi-modal text tasks.
Recently, Google announced that the API of its flagship AI model, Gemini2.5Pro, has been reintroduced to the free tier of Google AI Studio. This news has triggered widespread attention and enthusiastic discussions within the developer community. According to AIbase, this move marks another important advancement in Google's efforts to popularize AI technology, offering developers lower barriers to innovation. As the most advanced AI model from Google so far, Gemini2.5Pro is known for its exceptional multimodal capabilities and strong reasoning power.
On June 29, 2025, the Alibaba International AI Team officially released the new multi-modal large model **Ovis-U1**, marking another major breakthrough in the field of multi-modal artificial intelligence. As the latest masterpiece of the Ovis series, Ovis-U1 integrates multi-modal understanding, image generation, and image editing functions, demonstrating powerful cross-modal processing capabilities, providing new possibilities for developers, researchers, and industry applications. This is a detailed report on Ovis-U1 by AIbase. Ovis-U1
Recently, the Qwen VLo multimodal large model was officially released, achieving significant advancements in image content understanding and generation, offering users a brand-new visual creation experience. According to the introduction, Qwen VLo has been comprehensively upgraded based on the advantages of the original Qwen-VL series models. The model not only can accurately understand the "world", but also can perform high-quality re-creation based on understanding, truly achieving the transition from perception to generation. Users can now access Qwen Chat (chat.qwen.ai)
OpenAI has officially announced the date and location of its next developer conference (DevDay), which will be held on October 6, 2025, in San Francisco. This conference is expected to attract more than 1,500 developers and is anticipated to be the largest developer event to date. The agenda for this DevDay will be rich and diverse, featuring multiple important sessions. The conference will include live-streamed keynote speeches, during which OpenAI will share its latest developments and future vision in the field of artificial intelligence. In addition, participants will also be able to
Black Forest Labs officially announced that its new image editing model FLUX.1Kontext [dev] is now open source, drawing widespread attention from the AI community. As the latest member of the FLUX.1 series, this model is praised as an open-source alternative comparable to GPT-4o, thanks to its powerful image editing capabilities and efficient performance. FLUX.1Kontext [dev] is based on a 1.2 billion parameter flow matching transformer architecture, specifically designed for image editing tasks, and supports consumer-grade hardware.
New Oriental officially launched its first consumer-facing original AI education product - New Oriental AI 1-on-1 today. This is not only a major breakthrough in teaching methods, but also marks a critical step in New Oriental's strategic layout of "Education + AI." The core competitiveness of New Oriental AI 1-on-1 lies in providing learners with a high-frequency interactive 1-on-1 learning experience. The AI teacher can realistically reproduce the learning environment, achieving real interaction and real Q&A. At the same time, the AI teacher is patient, responsible, proficient in teaching, and can provide timely feedback, as well as praise and encourage students.
Recently, a research team from the University of Hong Kong, The Chinese University of Hong Kong, and SenseTime has released a groundbreaking framework - GoT-R1. This new multimodal large model significantly enhances the semantic and spatial reasoning capabilities of AI in visual generation tasks by introducing reinforcement learning (RL), successfully generating high-fidelity and semantically consistent images from complex text prompts. This advancement marks another leap in image generation technology. Currently, although existing multimodal large models have made significant progress in generating images based on text prompts
Recently, Google officially launched a new command-line tool - Gemini CLI. This tool is based on Google's self-developed Gemini2.5Pro AI model, aiming to provide developers with convenient AI Q&A and content generation services. With Gemini CLI, developers can directly call the power of AI in their own terminal interface, thereby improving programming efficiency and work convenience. One of the highlights of Gemini CLI is its support for up to 1 million tokens