Dark | Light
# ![@ModelScope2022 Avatar](https://lunarcrush.com/gi/w:26/cr:twitter::1784494412913049600.png) @ModelScope2022 ModelScope

ModelScope posts on X about native, agentic, ai, inference the most. They currently have [-----] followers and [---] posts still getting attention that total [------] engagements in the last [--] hours.

### Engagements: [------] [#](/creator/twitter::1784494412913049600/interactions)
![Engagements Line Chart](https://lunarcrush.com/gi/w:600/cr:twitter::1784494412913049600/c:line/m:interactions.svg)

- [--] Week [------] -12%
- [--] Month [-------] +72%
- [--] Year [---------] +210,270%

### Mentions: [--] [#](/creator/twitter::1784494412913049600/posts_active)
![Mentions Line Chart](https://lunarcrush.com/gi/w:600/cr:twitter::1784494412913049600/c:line/m:posts_active.svg)

- [--] Year [--] +9,400%

### Followers: [-----] [#](/creator/twitter::1784494412913049600/followers)
![Followers Line Chart](https://lunarcrush.com/gi/w:600/cr:twitter::1784494412913049600/c:line/m:followers.svg)

- [--] Week [-----] +4.50%
- [--] Month [-----] +52%

### CreatorRank: [-------] [#](/creator/twitter::1784494412913049600/influencer_rank)
![CreatorRank Line Chart](https://lunarcrush.com/gi/w:600/cr:twitter::1784494412913049600/c:line/m:influencer_rank.svg)

### Social Influence

**Social category influence**
[technology brands](/list/technology-brands)  9% [finance](/list/finance)  2% [travel destinations](/list/travel-destinations)  2% [stocks](/list/stocks)  2% [products](/list/products)  1% [gaming](/list/gaming)  1%

**Social topic influence**
[native](/topic/native) #117, [agentic](/topic/agentic) #156, [ai](/topic/ai) 10%, [inference](/topic/inference) #190, [realtime](/topic/realtime) 7%, [model](/topic/model) 7%, [math](/topic/math) 6%, [flow](/topic/flow) 5%, [tencent](/topic/tencent) 4%, [science](/topic/science) 4%

**Top accounts mentioned or mentioned by**
[@furaidosu](/creator/undefined) [@alitongyilab](/creator/undefined) [@alibabacloud](/creator/undefined) [@alibabaqwen](/creator/undefined) [@ali_tongyilab](/creator/undefined) [@barrakali](/creator/undefined) [@dx8152](/creator/undefined) [@1328](/creator/undefined) [@ultralytics](/creator/undefined) [@sean60133791259](/creator/undefined) [@bdsqlsz](/creator/undefined) [@nnnnmailcom](/creator/undefined) [@blankbraindead](/creator/undefined) [@wolfyblair](/creator/undefined) [@agibotofficial](/creator/undefined) [@stepfunai](/creator/undefined) [@stevewarnerfl](/creator/undefined) [@kimimoonshots](/creator/undefined) [@yashasgunderia](/creator/undefined) [@zaiorg](/creator/undefined)
### Top Social Posts
Top posts by engagements in the last [--] hours

"๐Ÿฟ Meet Sirchmunk: The agentic search engine that operates directly on your raw data. Turn files into evolving intelligence in real-time Build information retrieval pipelines that require ZERO vector databases ZERO pre-indexing and ZERO chunking: ๐ŸŽฏ Embedding-Free Retrieval Drop files in and search instantly. No ETL pipelines no vector silos and no setup hours. ๐ŸŽฒ Monte Carlo Evidence Sampling It doesn't "read" full docsit explores focuses and synthesizes. 500-page manual Only the most relevant snippets hit the LLM. ๐Ÿง  Self-Evolving Knowledge Every search builds reusable knowledge clusters."  
[X Link](https://x.com/ModelScope2022/status/2022370189157831078)  2026-02-13T18:00Z [----] followers, [----] engagements


"Love seeing what the MiniMax team shipped with M2.5 ๐Ÿš€ Real-world productivity strong agentic workflows and impressive efficiency numbers this is serious engineering. SOTA across coding search and tool use while getting faster is a big deal. ๐Ÿ”ฅ Excited to share that the model will be coming to ModelScope soon for the open-source community. Stay tuned ๐ŸŒโœจ Introducing M2.5 an open-source frontier model designed for real-world productivity. - SOTA performance at coding (SWE-Bench Verified 80.2%) search (BrowseComp 76.3%) agentic tool-calling (BFCL 76.8%) & office work. - Optimized for efficient"  
[X Link](https://x.com/ModelScope2022/status/2021994964125987267)  2026-02-12T17:09Z [----] followers, [----] engagements


"Nanbeige just dropped Nanbeige4.1-3B a tiny powerhouse that outclasses models 10x its size.๐Ÿš€ โœ… 256k Context Window โœ… "Deep Search" Agent capabilities built-in โœ… Two-stage Code RL (Correctness โžก Efficiency) โœ… Beats Qwen3-32B on key benchmarks The era of high-performance low-cost local deployment is here. ๐Ÿค–: https://modelscope.cn/models/nanbeige/Nanbeige4.1-3B https://modelscope.cn/models/nanbeige/Nanbeige4.1-3B"  
[X Link](https://x.com/ModelScope2022/status/2022629650178941435)  2026-02-14T11:11Z [----] followers, 11.9K engagements


"Introducing LingBot-World: An open-source world simulator pushing the boundaries of video generation. ๐Ÿš€ ๐ŸŒ High-Fidelity: Realistic scientific & stylized. ๐Ÿง  Long-Term Memory: Minute-level consistency. โšก Real-Time: 1s latency at [--] FPS. ๐Ÿ“œ Apache [---] Licensed. Model: Github: https://github.com/Robbyant/lingbot-world https://modelscope.cn/models/Robbyant/lingbot-world-base-cam https://github.com/Robbyant/lingbot-world https://modelscope.cn/models/Robbyant/lingbot-world-base-cam"  
[X Link](https://x.com/ModelScope2022/status/2016564553816088624)  2026-01-28T17:30Z [----] followers, 29.3K engagements


"Tencent YouTu Research open-sourcedYoutu-VL-4B-Instruct a compact VLM redefining visual perception viaVLUAS. ๐Ÿ‘โšก ๐ŸŒŸHighlights: โœ…All-in-One Vision: SOTA performance in detection segmentation depth estimation and pose estimationwithouttask-specific heads. โœ…OCR & Reasoning: Exceptional at complex document parsing and multimodal math reasoning. โœ…GUI Agent Ready: Optimized for world-understanding and interface navigation. โœ…Efficiency: 4B parameters make it perfect for edge deployment and fast inference. ๐Ÿ”งPerformance: Outperforms many larger models inOmniDocBenchandVision-Centrictasks while"  
[X Link](https://x.com/ModelScope2022/status/2017191058262700328)  2026-01-30T11:00Z [----] followers, [----] engagements


"Stepfun open-sourcedStep-3.5-Flash a powerhouse model specifically architected for high-speed reasoning and complex Agentic workflows. ๐Ÿš€ Model: Key Technical Specs: โœ…Sparse MoE Architecture: 196B total params but only11B activeper token. SOTA efficiency. โœ…MTP-3 (Multi-Token Prediction): It predicts3 tokens at once hitting a blistering350 TPSfor code-heavy tasks. โšก โœ…Hybrid Attention (SWA + Full): A 3:1 mix that masters256K contextwindows while keeping compute costs low. โœ…Parallel Thinking: Massively boosted performance for multi-step reasoning and deep search. Why Devs should care: - Built"  
[X Link](https://x.com/ModelScope2022/status/2018215995740164168)  2026-02-02T06:52Z [----] followers, 34K engagements


"Beyond-Reality-Z-Image [---] is here.๐Ÿš€ Highlights: โœ…Texture SOTA: Optimized high-frequency details for what might be the best out-of-the-box skin and fabric textures in the Z-model family. โœ…Cinematic Lighting: Re-engineered color balance to achieve that authentic "film look" without heavy prompting. โœ…Speed x Versatility: Generates at Turbo speeds (10-15 steps) but supports broader artistic styles and 2D/Anime aesthetics than the official release. โœ…Efficiency: Optimized for low CFG (1.0) making it highly stable and responsive to LoRA integration. Available now on ModelScope: Recommended Stack:"  
[X Link](https://x.com/ModelScope2022/status/2018309394719805444)  2026-02-02T13:03Z [----] followers, 16.4K engagements


"How do Agents "Grow from Experience" A dialogue on Key Issues and Technical Paths of Self-Evolution AI AMA Deep Roundtable Vol.1 [----] was about Reasoning. [----] is the "Mirror Moment"Agents are shifting from static tools to digital lifeforms that learn iterate and evolve from every failure. ๐Ÿš€ https://twitter.com/i/web/status/2018712297829732601 https://twitter.com/i/web/status/2018712297829732601"  
[X Link](https://x.com/ModelScope2022/status/2018712297829732601)  2026-02-03T15:44Z [----] followers, [----] engagements


"Meetthe1stAuthorsbehindthemostinfluentialpapers:earlyexperienceDREAMGYMAgentEvolverSPICER-ZeroEvo-MemoryandSEAgent. TheyarenowatoralumnioftheseeliteAIlabs: ๐Ÿ”นAlibabaTongyiLabMetaFAIRMetaSuperintelligenceLabsGoogleDeepMindTencentAISeattleLab ๐Ÿ”นSJTUOSUUChicagoNUSWashU 6leadresearchers1stage.Unrivaledtechnicaldepth.๐Ÿง  https://twitter.com/i/web/status/2018712300744847796 https://twitter.com/i/web/status/2018712300744847796"  
[X Link](https://x.com/ModelScope2022/status/2018712300744847796)  2026-02-03T15:44Z [----] followers, [---] engagements


"MiniCPM-o [---] brings true "Omni" capabilities (Vision + Audio + Text) into a single 9B end-to-end architecture (Qwen3-8B + SigLip2 + Whisper + CosyVoice2).๐Ÿš€ โœ… Vision SOTA: OpenCompass score of 77.6surpassing GPT-4o and Gemini [---] Pro in VLM tasks with just 9B params. โœ… Real-time Full Duplex: It "sees" "hears" and "speaks" simultaneously. Supports active interaction (it can interrupt or initiate conversation based on live video). โœ… World-Class OCR: Beats DeepSeek-OCR [--] and GPT-4o in document parsing (OmniDocBench leader). โœ… Voice Cloning: Clone any voice with just a short reference"  
[X Link](https://x.com/ModelScope2022/status/2018925461649396202)  2026-02-04T05:51Z [----] followers, [----] engagements


"๐Ÿš€ ACE-Step v1.5 is out: an open-source music foundation model that runs locally on consumer GPUs (4GB VRAM) and generates full songs in 2s (A100) or 10s (RTX 3090). โœ… Beats most commercial models in quality โœ… Train a personalized LoRA from just a few tracks โœ… Built on a novel LM+DiT hybrid architecture with internal RLno external reward bias โœ… Supports 50+ languages + editing (covers repainting vocalBGM) ๐Ÿค– https://modelscope.cn/models/ACE-Step/Ace-Step1.5 https://modelscope.cn/models/ACE-Step/Ace-Step1.5"  
[X Link](https://x.com/ModelScope2022/status/2018977724048777344)  2026-02-04T09:19Z [----] followers, [----] engagements


"๐Ÿš€ Meet Intern-S1-Pro: A massive 1T parameter MoE model for Multimodal Science Reasoning โœ… [---] Experts (22B active) โœ… SOTA in AI4Science (Chemistry Materials Bio) โœ… FoPE + Time-series modeling (up to [--] points) โœ… Native "Thinking Mode" support Open-source science just leveled up. ๐Ÿงช๐Ÿ’ป Model: Github: https://github.com/InternLM/Intern-S1 https://www.modelscope.cn/models/Shanghai_AI_Laboratory/Intern-S1-Pro https://github.com/InternLM/Intern-S1 https://www.modelscope.cn/models/Shanghai_AI_Laboratory/Intern-S1-Pro"  
[X Link](https://x.com/ModelScope2022/status/2019068430041919653)  2026-02-04T15:20Z [----] followers, [----] engagements


"Introducing Singularity Cinema (MS-Agent framework)๐Ÿš€an LLM-powered tool that transforms documents or prompts into storyboards generating stunning 5-minute+ videos via its Manim/Remotion dual-engine. By fusing nano-banana-pro๐ŸŒ Qwen-Image๐ŸŽจ Sora and edge-tts it delivers high-quality information-dense contentperfect for science economics and history creators looking to scale their social media presence. ๐Ÿ”—: ๐ŸŽฆPrompt: Convert xx/deploy.md into a short video in a blue-themed style making sure to use the important images from the document. The short video must be in English. ๐ŸŽฌOutput Video๐Ÿ‘‡"  
[X Link](https://x.com/ModelScope2022/status/2019443081461002520)  2026-02-05T16:08Z [----] followers, [----] engagements


"From a local designer in Chengdu to a global AI sensation with 117k+ downloadsDaniels journey shows whats possible when open-source meets imagination. Now its your turn. Join the Qwen-Image LoRA Competition and start building your own legacy. The next top creator could be you. https://modelscope.ai/active/qwenimagelora LoRA Training Contest: Win iPhone/PS5 and other prizes Gaussian splash LoRA creation process shared. Competition information: https://t.co/MEEZHQcr83 Training address: https://t.co/uxwtMlJzoP @Ali_TongyiLab @ModelScope2022 #HappyQwensday #QwenImageLora https://t.co/5igMVZUO5i"  
[X Link](https://x.com/ModelScope2022/status/2019612106086314106)  2026-02-06T03:20Z [----] followers, [----] engagements


"ChenkinNoob-XL v0.2 is here A high-fidelity SDXL-to-Rectified Flow conversion delivering vivid color range and superior lighting adherence. โœ… Performance: Eliminates "base model greyness"; Superior contrast/dark scene rendering; High stability across wide CFG ranges (3-6); Optimal convergence at 20-28 steps. โœ… Ecosystem: Native ComfyUI workflow included; ReForge-ready with built-in Flow support; LoRA-compatible via Pochi.toml (Refresh scripts). โœ… Architecture: Rectified Flow (RF) linear ODE trajectory; Trained on 47M samples using 8xH20 nodes; Leverages Cosine Optimal Transport and Protected"  
[X Link](https://x.com/ModelScope2022/status/2019637177471173090)  2026-02-06T05:00Z [----] followers, [----] engagements


"StepFun open-sourced Step-3.5-Flashfeaturing powerful reasoning and Agent intelligence. Speeds up to [---] TPS Competes with closed-source models in Agent and Math tasks Excels at complex long-chain reasoning. Dont miss out https://x.com/i/broadcasts/1OyKAjZgZPaGb https://x.com/i/broadcasts/1OyKAjZgZPaGb"  
[X Link](https://x.com/ModelScope2022/status/2019744317792018474)  2026-02-06T12:05Z [----] followers, [----] engagements


"Meet X-Coder ๐Ÿš€IIGroup drops a powerhouse suite for code LLMs โœ… X-Coder-SFT-376k: Massive synthetic dataset via GPT-o3-mini/DeepSeek-R1. โœ… X-Coder-RL-40k: Targeted RL alignment for logic & verification. โœ… X-Coder-8B Models: SOTA reasoning built on Qwen3. Experience the power of fully synthetic code evolution. ๐Ÿ’ป๐Ÿ”ฅ ๐Ÿ“Š Models & Dataset: #LLM #OpenCode #Qwen3 #DeepSeek #Coding https://modelscope.cn/models/IIGroup/X-Coder-SFT-Qwen3-8B https://modelscope.cn/datasets/IIGroup/X-Coder-RL-40k https://modelscope.cn/datasets/IIGroup/X-Coder-SFT-376k"  
[X Link](https://x.com/ModelScope2022/status/2020745735059443936)  2026-02-09T06:25Z [----] followers, [----] engagements


"YOLO26 is here๐Ÿš€ The most advanced and deployment-ready vision model from @ultralytics redefining real-time edge computing performance. โœ… Performance: 43% CPU inference speed boost vs YOLO11; Real-time performance on low-power hardware; [--] scalable sizes (n/s/m/l/x) for optimized speed-accuracy trade-offs. โœ… Ecosystem: Unified framework for Detection Segmentation Pose OBB and Tracking; Simplified export for seamless hardware integration. โœ… Logic/Architecture: Removed DFL for enhanced hardware compatibility; End-to-end NMS-free inference; ProgLoss + STAL for small target precision; MuSGD"  
[X Link](https://x.com/ModelScope2022/status/2020777189575467356)  2026-02-09T08:30Z [----] followers, [----] engagements


"2/4 Key Technical Highlights: - Architecture: Native single-stream design integrating generation editing and segmentation. - Audio: First model to unify Speech Audio and Music generation via 12.5Hz continuous representation. - Video: Time-Interleaved VideoRoPE for enhanced temporal modeling. https://twitter.com/i/web/status/2021580867009671592 https://twitter.com/i/web/status/2021580867009671592"  
[X Link](https://x.com/ModelScope2022/status/2021580867009671592)  2026-02-11T13:43Z [----] followers, [---] engagements


"๐Ÿš€ GLM-5 is now open-sourceand ranked #4 globally #1 among open models on Artificial Analysis Beyond "vibe coding" it delivers Agentic Engineering: autonomously architects debugs and ships production systems. Coding performance rivals Claude Opus [---] (SWE-bench: [----] ๐Ÿ”“ SOTA) while dominating long-horizon agent tasks (BrowseComp MCP-Atlas ๐Ÿ”). 744B params MIT License ๐Ÿ‘‰ ModelScope: https://modelscope.cn/models/ZhipuAI/GLM-5 https://modelscope.cn/models/ZhipuAI/GLM-5"  
[X Link](https://x.com/ModelScope2022/status/2021792323227820181)  2026-02-12T03:43Z [----] followers, [----] engagements


"๐Ÿฟ Meet Sirchmunk: The agentic search engine that operates directly on your raw data. Turn files into evolving intelligence in real-time Build information retrieval pipelines that require ZERO vector databases ZERO pre-indexing and ZERO chunking: ๐ŸŽฏ Embedding-Free Retrieval Drop files in and search instantly. No ETL pipelines no vector silos and no setup hours. ๐ŸŽฒ Monte Carlo Evidence Sampling It doesn't "read" full docsit explores focuses and synthesizes. 500-page manual Only the most relevant snippets hit the LLM. ๐Ÿง  Self-Evolving Knowledge Every search builds reusable knowledge clusters."  
[X Link](https://x.com/ModelScope2022/status/2022304582207238586)  2026-02-13T13:39Z [----] followers, [---] engagements


"LongCat-Flash-LiteisNow Live Meituan's 68.5B MoE model pioneering "Embedding Expansion" for leading Agent and Coding performance. โœ…Performance: 2.9B4.5B activated params; 500-700 tokens/s throughput; 256K context via YARN; SOTA SWE-Bench (54.4%) and -Bench scores. โœ…Ecosystem: Optimized SGLang-FluentLLM engine with custom CUDA kernels; Open API with 50M free daily tokens. โœ…Architecture: N-gram embedding layer for localized semantic precision; Hash-collision mitigation via sub-table decomposition; Triple-step Speculative Decoding synergy. ๐Ÿค– ๐Ÿ”ง ๐Ÿ“„ Tech"  
[X Link](https://x.com/ModelScope2022/status/2020383958819832285)  2026-02-08T06:27Z [----] followers, [----] engagements


"๐Ÿš€1/4 Introducing Ming-flash-omni-2.0: An open-source Omni-MLLM breakthrough. Based on the Ling-2.0 MoE architecture (100B total / 6B active) it unifies vision speech and image editing into a single model matching specialist performance across all modalities. ๐Ÿงต๐Ÿ‘‡ ๐Ÿค– ModelScope: ๐Ÿ“ฆ GitHub: https://github.com/inclusionAI/Ming https://www.modelscope.cn/models/inclusionAI/Ming-flash-omni-2.0 https://github.com/inclusionAI/Ming https://www.modelscope.cn/models/inclusionAI/Ming-flash-omni-2.0"  
[X Link](https://x.com/ModelScope2022/status/2021580863277084888)  2026-02-11T13:43Z [----] followers, [----] engagements


"๐Ÿš€Introducing Z-Image the full non-distilled version of Z-Image for high-quality customizable image generation. โœ… Model: โœ… Demo: โœจ Capabilities: Photorealistic + diverse art styles (anime illustration digital art) Full CFG & precise negative prompts for reliable prompt following Multi-subject generation with distinct identitiesno "clone faces" Community-friendly: ready for LoRA ControlNet and other fine-tuning workflows ๐Ÿ”ฅ Highlights: Full architecture (vs. distilled Turbo): better diversity & training stability ideal for customization High output variance across seeds = less homogenization"  
[X Link](https://x.com/ModelScope2022/status/2016178781522890948)  2026-01-27T15:57Z [----] followers, 59.5K engagements


"Z-Image-Fun-Lora-Distill is here A high-efficiency Distill LoRA enabling CFG-free high-fidelity generation for Z-Image in just 4-8 steps. โœ… 4-step and 8-step variants; CFG-free inference (CFG=1.0); [----] update delivers superior skin texture and color consistency compared to previous iterations. โœ… Native ComfyUI weights available; full integration with the VideoX-Fun framework; compatible with Z-Image ControlNet Union [---] and third-party LoRAs. โœ… Dual-distillation of both sampling steps and CFG; trained from scratch without Z-Image-Turbo weights; optimized for rapid generation and derivative"  
[X Link](https://x.com/ModelScope2022/status/2021412342723207600)  2026-02-11T02:33Z [----] followers, [----] engagements


"๐Ÿš€ Ring-2.5-1T is open-source The first hybrid linear 1T thinking model. ๐Ÿ“Š Gold Tier reasoning IMO25 (35/42) & CMO25 (105/126) scores. ๐Ÿ›  Agent-native compatible with Claude Code OpenClaw and SGLang. โšกHybrid Linear breakthrough 10x lower memory & 3x throughput for long context. ๐Ÿง  1T MoE structure with 63B active parameters (1:7 MLA + Lightning Linear). ๐Ÿ’ป Long-Horizon mastery autonomously built a 32-bit OS from scratch in [--] hours. Model: GitHub: https://github.com/inclusionAI/Ring-V2.5 https://modelscope.cn/models/inclusionAI/Ring-2.5-1T https://github.com/inclusionAI/Ring-V2.5"  
[X Link](https://x.com/ModelScope2022/status/2022223579421839523)  2026-02-13T08:17Z [----] followers, 13K engagements


"Introducing FireRed-Image-Edit-1.0 from FireRedTeam ๐Ÿš€ Its officially the new SOTA for general image editing. โœ… Better than Closed-Source: Outperforms Nano-Banana & Seedream4.0 on GEdit benchmarks. โœ… Native Evolution: Built from T2I foundations not just a "patch" on existing models. โœ… Style Mastery: Scored a record-breaking 4.97/5.0 in style transfer. โœ… High-Fidelity Text: Keeps original font styles perfectly. โœ… Virtual Try-on: Native support for multi-image joint editing. โœ… Bilingual: Native support for both English & Chinese prompts. Apache [---] license. Local deployment ready. ๐Ÿค– Model: ๐ŸŽ "  
[X Link](https://x.com/ModelScope2022/status/2022687312442704320)  2026-02-14T15:00Z [----] followers, 69.4K engagements


"How 0.9B Models Are Outperforming the Giants: A Comparison of Four New Chinese VLM-OCR Models 1/3 ๐Ÿ† Early [----] has brought a massive surge of open-source models from DeepSeek Z-org PaddlePaddle and Tencent YouTu. A tiny 0.9B parameter model GLM-OCR topped the global OmniDocBench v1.5 leaderboard outscoring giants like Gemini-3 Pro and Qwen3-VL. We are moving away from rigid "detect and recognize" pipelines toward Vision-Language Models (VLMs) that actually "read" and understand document structures like humans do. https://twitter.com/i/web/status/2018615132134551565"  
[X Link](https://x.com/ModelScope2022/status/2018615132134551565)  2026-02-03T09:18Z [----] followers, 30.1K engagements


"Z-Image-Fun-Controlnet-Union & Lora-Distill are now open-source ๐Ÿš€Expanding the Z-Image ecosystem with ultra-fast inference and high-precision spatial control. โœ… Performance: Achieve 8-step inference (CFG: 1.0) via Distill LoRA and professional-grade multi-condition control (Canny Depth Pose Inpaint) across 15+ layer blocks. โœ… Ecosystem: Full cross-compatibility between Distill LoRA and Union ControlNet; includes "Lite" versions for low-VRAM edge devices and seamless natural blending. โœ… Unified Control: Support for 7+ spatial conditions in a single model with adjustable control_context_scale"  
[X Link](https://x.com/ModelScope2022/status/2019339552650092830)  2026-02-05T09:17Z [----] followers, 11K engagements


"Qwen3-ASR & Qwen3-ForcedAligner are now open-source ๐ŸŽ™๐Ÿš€ High-performance streaming-ready speech AI for [--] languages and dialects. ๐Ÿ”—: Demo: ๐Ÿค–The Models โœ… ASR-1.7B: Open-source SOTA. Outperforms Whisper-v3; rivals GPT-4o & Gemini. โœ… ASR-0.6B: Insane efficiency92ms latency; transcribes 2000s of audio in 1s. โœ… ForcedAligner-0.6B: Superior timestamp accuracy across [--] languages. ๐ŸŒŸKey Features โœ…Universal: Supports [--] languages + [--] Chinese dialects. โœ…Robust: Expertly handles singing heavy BGM and low SNR. โœ…Unified: Single model for both streaming & offline (up to 20min audio). โšกPerformance &"  
[X Link](https://x.com/ModelScope2022/status/2016857817597301174)  2026-01-29T12:55Z [----] followers, [----] engagements


"๐Ÿš€ FantasyWorld is now open source Recently accepted to ICLR [----] and ranking #1 on the WorldScore leaderboard this framework offers a unified approach to video and 3D generation. โœจKey Technical Highlights: โœ…Direct 3D Inference: Input an image and camera trajectory to get frame-consistent video dense depth maps and point clouds instantly. โœ…Dual-Branch IRG Architecture: An "Imagination" branch handles visual appearance while a "Geometry" branch ensures spatial accuracy. โœ…Wan [---] Integration: The released 14B models leverage the latest Wan-DiT architectures for superior temporal consistency and"  
[X Link](https://x.com/ModelScope2022/status/2022264617871757514)  2026-02-13T11:00Z [----] followers, [----] engagements


"Say hello to Ling-2.5-1T: 1T params 63B active MIT licensed โšก Hybrid Linear Attention: 1:7 MLA + Lightning Linear. Beats Kimi K2 on long-context throughput. ๐Ÿง  4x token efficiency: Composite rewards match frontier thinking models using 4x fewer tokens. ๐Ÿ“š 1M context: YaRN to 1M. Beats Kimi K2.5 & DeepSeek V3.2 on RULER/MRCR. Perfect NIAH scores. ๐Ÿ›  SOTA tool use: Agentic RL trained. Leads BFCL-V4. Native support for Claude Code OpenCode OpenClaw. ๐ŸŽฏ Follows instructions: Bidirectional RL + agent verification. High-density zero fluff. Honest: Gap vs GPT-5.2/Gemini [--] Pro on long-horizon tasks."  
[X Link](https://x.com/ModelScope2022/status/2023240865649447320)  2026-02-16T03:39Z [----] followers, [----] engagements


"New Z-Image-Turbo control model released ๐ŸŽ‰ More control blocks. Inpaint mode. Better details. - 6-block ControlNet - Trained from scratch on 1M images @1328 res - Canny / HED / Depth / Pose / MLSD - Best control_context_scale: 0.65-0.80 Model: Demo: Code: https://github.com/aigc-apps/VideoX-Fun https://modelscope.cn/studios/AI-ModelScope/ZIT-Controlnet https://modelscope.cn/models/PAI/Z-Image-Turbo-Fun-Controlnet-Union https://modelscope.cn/models/PAI/Z-Image-Turbo-Fun-Controlnet-Union https://github.com/aigc-apps/VideoX-Fun https://modelscope.cn/studios/AI-ModelScope/ZIT-Controlnet"  
[X Link](https://x.com/ModelScope2022/status/2000766764922401091)  2025-12-16T03:15Z [----] followers, 33.1K engagements


"๐Ÿš€Model & Demo are now live on our international site Model: - - - Demo: https://modelscope.ai/studios/FunAudioLLM/Fun-CosyVoice3-0.5B https://modelscope.ai/models/FunAudioLLM/Fun-CosyVoice3-0.5B-2512 https://modelscope.ai/models/FunAudioLLM/Fun-ASR-Nano-2512 https://modelscope.ai/models/FunAudioLLM/Fun-ASR-MLT-Nano-2512 https://modelscope.ai/studios/FunAudioLLM/Fun-CosyVoice3-0.5B https://modelscope.ai/models/FunAudioLLM/Fun-CosyVoice3-0.5B-2512 https://modelscope.ai/models/FunAudioLLM/Fun-ASR-Nano-2512 https://modelscope.ai/models/FunAudioLLM/Fun-ASR-MLT-Nano-2512"  
[X Link](https://x.com/ModelScope2022/status/2000864318758511044)  2025-12-16T09:43Z [----] followers, [---] engagements


"๐Ÿค–Meet Nanbeige4-3B from Boss Zhipina 3B-parameter LLM that outperforms Qwen3-32B on math (AIME) science (GPQA) and tool calling (BFCL-V4) while matching Qwen3-30B-A3B on human preference alignment (Arena-Hard-V2). How โœ… 23T tokens of ultra-curated data โœ… Fine-grained WSD scheduler โœ… 30M+ high-quality SFT instructions โœ… Multi-stage RL + innovative distillation (DPD) โœ… Chain-of-thought reconstruction & deliberative generation It even ranks top [--] on WritingBench & EQ-Bench3beating models 100x larger like GLM-4.5 and Deepseek-R1. All models + tech report now open-source: ๐Ÿ”— Weights: ๐Ÿ“„ Paper:"  
[X Link](https://x.com/ModelScope2022/status/2000890664486363362)  2025-12-16T11:28Z [----] followers, 25.8K engagements


"๐ŸŽ‰ MiMo-V2-Flash FREE API is now live on ModelScope The first major release since Fuli Luo joined Xiaomiand its built for real-world agentic AI. โšก MiMo-V2-Flash: an open high-performance MoE model with 309B total / 15B active parameters 256K context window 150+ tokens/s generation thanks to native Multi-Token Prediction ๐Ÿ”ฅ Key wins for developers: โœ… Hybrid attention (5:1 SWA + Global) [--] less KV cache full long-context recall โœ… 73.4% on SWE-Bench Verified new SOTA for open-source models โœ… Matches DeepSeek-V3.2 on reasoning but much faster in practice โœจ API-readyperfect for building smart"  
[X Link](https://x.com/ModelScope2022/status/2001479829095620906)  2025-12-18T02:29Z [----] followers, 16K engagements


"๐Ÿš€Qwen-Image-Layered is now live on ModelScopean incredible model that can intelligently decompose any image into multiple RGBA layers ๐Ÿค– Model: ๐ŸŒŸ Try Demo ๐Ÿ“‘ Paper: https://modelscope.cn/papers/2512.15603 https://modelscope.cn/studios/Qwen/Qwen-Image-Layered https://modelscope.ai/models/Qwen/Qwen-Image-Layered https://modelscope.cn/papers/2512.15603 https://modelscope.cn/studios/Qwen/Qwen-Image-Layered https://modelscope.ai/models/Qwen/Qwen-Image-Layered ๐ŸŽจ Qwen-Image-Layered is LIVE native image decomposition fully open-sourced โœจ Why it stands out โœ… Photoshop-grade layering Physically"  
[X Link](https://x.com/ModelScope2022/status/2002038699878695146)  2025-12-19T15:29Z [----] followers, [----] engagements


"Its still Sunday and Im already hyped for next week. ๐Ÿ˜ŠThe mysterious Eastern power is dropping a trilogy of magic: ๐Ÿ”ฎ Next week: A secret image-generation model ๐Ÿค– Next week: A secret code-agent model ๐ŸŽ™ The week after: A secret voice model Buckle upthings are about to get interesting"  
[X Link](https://x.com/ModelScope2022/status/2002679068203028809)  2025-12-21T09:54Z [----] followers, 10.1K engagements


"๐Ÿš€ Meet GLM-4.7 your new coding partner supercharged Try it now on ModelScope free API access for a limited time โœ… +5.8% on SWE-bench (73.8%) โœ… +12.9% on Multilingual coding (66.7%) โœ… +16.5% on Terminal Bench [---] (41%) โœ… Smarter agent reasoning & tool use โœ… Better UI cleaner webpages sharper slides โœ… +12.4% on complex reasoning (HLE: 42.8%)From coding to creativity GLM-4.7 delivers across the board. ๐Ÿค– Model: https://modelscope.cn/models/ZhipuAI/GLM-4.7 https://modelscope.cn/models/ZhipuAI/GLM-4.7"  
[X Link](https://x.com/ModelScope2022/status/2003290346135187635)  2025-12-23T02:23Z [----] followers, [----] engagements


"๐Ÿš€ New on ModelScope: QwenLong-L1.5 is now fully open-source A 30B model (3B active params) that matches GPT-5 & Gemini-2.5-Pro in long-context reasoning. ๐Ÿ”ฅ Key wins: โœ… +31.7 pts on OpenAIs MRCR (128K context SOTA across all models) โœ… Matches Gemini-2.5-Pro on [--] major long-QA benchmarks โœ… +9.69 on CorpusQA +6.16 on LongBench-V2 How Three breakthroughs: [--] Synthetic data at scale: 14.1K long-reasoning samples from 9.2B tokens no human labeling. Avg. length: 34K tokens (max: 119K). [--] Stable RL training: Task-balanced sampling + Adaptive Entropy-Controlled Policy Optimization (AEPO) for reliable"  
[X Link](https://x.com/ModelScope2022/status/2003370363590226313)  2025-12-23T07:41Z [----] followers, 44.2K engagements


"@Sean60133791259 @bdsqlsz Could I ask which country you are in Well test the speed. You can DM me"  
[X Link](https://x.com/ModelScope2022/status/2003648027375075799)  2025-12-24T02:04Z [----] followers, [--] engagements


"Merry Christmas everyone ๐ŸŽ„โœจ ๐Ÿš€Introducing BEYOND REALITY Z IMAGE 1.0: a community-tuned checkpoint merging LoRA enhancements into Z-Image optimized for film-style portrait aesthetics and high-frequency texture detail think skin pores fabric weaves and wall surfaces rendered with striking realism. โœ… Keeps Zs photorealism intact โœ… Adds cinematic color depth & tonal nuance โœ… Significantly upgrades portrait + environmental textures โœ… Runs on 8GB VRAM (FP8) tested on laptops Note: Training focused on high-quality synthetic portrait photography so complex poses or non-human textures may vary."  
[X Link](https://x.com/ModelScope2022/status/2003737630245343345)  2025-12-24T08:00Z [----] followers, 13.5K engagements


"We shouldve shouted this louder: Youre officially missing out if you aren't using Civision ๐Ÿคซ ๐ŸŽจ FREE Image & Video Gen & FREE LoRA Training โœจ A UI so clean you dont need a PhD to master it. Weve got the heavy hitters ready: Z-Image Qwen-Image Qwen-Image-Edit Wan2.2 and more. All the power zero cost. Right here on ModelScope: ๐Ÿš€โœจ https://modelscope.ai/civision/imageGeneration https://modelscope.ai/civision/imageGeneration"  
[X Link](https://x.com/ModelScope2022/status/2004110641687417306)  2025-12-25T08:43Z [----] followers, [----] engagements


"๐Ÿš€New on ModelScope: ChenkinNoob-XL v0.2 is live Independently developed by the ChenkinNoob team as a fine-tuned extension of noobai-XL-1.1 v0.2 incorporates Danbooru data up to Nov [--] 2025enhancing performance while preserving the original aesthetic. โœจ Key improvements over v0.1: โœ… Stronger character consistency โœ… Refined details (hands accessories layered outfits) โœ… Higher prompt adherence ๐ŸŽจ Faithful to the noob visual DNA: clean composition solid anatomy and harmonious color palettes. โš™ Recommended config: CFG: [--] Steps: [----] Sampler: Euler a Resolution: [----] (e.g. [-------] 10241024)"  
[X Link](https://x.com/ModelScope2022/status/2004378356645814448)  2025-12-26T02:26Z [----] followers, 30.7K engagements


"๐Ÿš€ New on ModelScope: MiniMax M2.1 is open-source โœ… SOTA in 8+ languages (Rust Go Java C++ TS Kotlin Obj-C JS) โœ… Full-stack Web & mobile dev: Android/iOS 3D visuals vibe coding that actually ships โœ… Smarter faster 30% fewer tokens with lightning mode (M2.1-lightning) for high-TPS workflows โœ… Top-tier on SWE-bench VIBE and custom coding/review benchmarks โœ… Works flawlessly in Cursor Cline Droid BlackBox and more Its not just better code its AI-native development end to end. ๐Ÿ”— Model: https://modelscope.cn/models/MiniMax/MiniMax-M2.1 https://modelscope.cn/models/MiniMax/MiniMax-M2.1"  
[X Link](https://x.com/ModelScope2022/status/2004462984698253701)  2025-12-26T08:03Z [----] followers, 17.3K engagements


"@nnnnmailcom Hi there SDXL models should be supported. Could you let me know exactly what issue you're running into"  
[X Link](https://x.com/ModelScope2022/status/2005487313158279461)  2025-12-29T03:53Z [----] followers, [--] engagements


"๐Ÿš€400+ Free APIs are now available on We currently support: โœ… DeepSeek series โœ… Qwen3 series & Qwen-Image โœ… Z-Image-Turbo and many more Check it out: Go to - Models - Filter by API-Inference. (Currently available on PC only.) ๐Ÿ‘‰ Give them a spin and let us know what you think โœ‰DM us for support. https://modelscope.ai/modelsfilter=inference_type&page=1&tabKey=task http://modelscope.ai http://ModelScope.ai http://modelscope.ai http://ModelScope.ai https://modelscope.ai/modelsfilter=inference_type&page=1&tabKey=task http://modelscope.ai http://ModelScope.ai http://modelscope.ai"  
[X Link](https://x.com/ModelScope2022/status/2005549668886040939)  2025-12-29T08:01Z [----] followers, 17.8K engagements


"@blankbraindead oh I see. May I know which region you are in"  
[X Link](https://x.com/ModelScope2022/status/2005557772797215085)  2025-12-29T08:33Z [----] followers, [---] engagements


"๐Ÿš€ Tencent HunYuan open-sourced HY-MT1.5 A lightweight powerhouse for translation: โœจ 1.8B model: 0.18s latency runs offline on mobile & outperforms most commercial APIs. Even rivals Gemini-3.0-Pro at 90% quality โšก 7B model: Upgraded with WMT25-champion accuracy & fewer artifacts. ๐ŸŒ [--] languages + [--] Chinese dialects ๐Ÿ›  Context-aware custom glossaries format-preserving output ๐Ÿ’ก Perfect for edge + cloud deployments Models: https://modelscope.cn/collections/Tencent-Hunyuan/HY-MT15 https://modelscope.cn/collections/Tencent-Hunyuan/HY-MT15"  
[X Link](https://x.com/ModelScope2022/status/2005944042853667257)  2025-12-30T10:08Z [----] followers, 17K engagements


"One of your New Year open-source gifts ๐ŸŽจ is on its way"  
[X Link](https://x.com/ModelScope2022/status/2005968451538759734)  2025-12-30T11:45Z [----] followers, 10.7K engagements


"๐Ÿš€New on ModelScope: Qwen-Image-2512 is here ๐ŸŽ‰ This version delivers a massive leap in realism and control: โœจ Photorealistic humans dramatically reduced AI look with accurate aging natural expressions and fine details (yes even individual hair strands) ๐ŸŒฟ Ultra-detailed nature fur foliage water and mist rendered with unprecedented texture fidelity ๐Ÿ”ค Reliable text & layout generates complex infographics timelines and multilingual slides with correct structure and legible text ๐Ÿ† Evaluated in 10k+ blind tests: now the strongest open-source image modelrivaling closed alternatives. ๐Ÿ’ก Supports"  
[X Link](https://x.com/ModelScope2022/status/2006302672073507194)  2025-12-31T09:53Z [----] followers, [----] engagements


"2026 is almost here Wishing you a very Happy New Year ๐ŸŽ‰ What open-source models or product features are at the top of your wishlist for the coming year ๐Ÿš€ Wed love to see if we can help make those wishes come true"  
[X Link](https://x.com/ModelScope2022/status/2006341719198163109)  2025-12-31T12:28Z [----] followers, [----] engagements


"๐ŸŒŸQwen-Image-2512 is now live on Model: Try: https://modelscope.ai/civision/imageGeneration/tab=default https://modelscope.ai/models/Qwen/Qwen-Image-2512/summary http://ModelScope.ai https://modelscope.ai/civision/imageGeneration/tab=default https://modelscope.ai/models/Qwen/Qwen-Image-2512/summary http://ModelScope.ai"  
[X Link](https://x.com/ModelScope2022/status/2006419353856782394)  2025-12-31T17:37Z [----] followers, [----] engagements


"๐Ÿš€ New on ModelScope: IQuest Research just dropped IQuest-Coder-V1 a new family of open-source code LLMs (7B to 40B) with 128K context GQA and two specialized variants: - Instruct: for coding assistance - Thinking: RL-tuned for agentic reasoning & error recovery - Plus: a novel Loop architecture (40B only) for efficient deployment via parameter sharing across [--] iterations. ๐ŸŒŸ All models support repo-level understanding via Code-Flow training learning from commit sequences and code evolution not just static snippets. ๐Ÿ” Benchmarks: - [----] on SWE-Bench Verified - [----] on LiveCodeBench v6"  
[X Link](https://x.com/ModelScope2022/status/2007712347373351418)  2026-01-04T07:14Z [----] followers, 10.9K engagements


"Big news ๐Ÿš€ now officially supports FREE LoRA training for Qwen-Image-2512 Z-Image-Turbo and Qwen-Image-Edit-2511. Time to build ๐Ÿ›  ๐ŸŽจModelScope Civision: https://modelscope.ai/civision/modelTraining http://ModelScope.ai https://modelscope.ai/civision/modelTraining http://ModelScope.ai https://modelscope.ai/civision/modelTraining http://ModelScope.ai https://modelscope.ai/civision/modelTraining http://ModelScope.ai"  
[X Link](https://x.com/ModelScope2022/status/2007787503802294735)  2026-01-04T12:13Z [----] followers, 21.5K engagements


"Qwen-Image-2512-Fast is officially here โšก ๐Ÿš€ Try the Demo on ModelScope now: Experience a massive 20x speedup generating high-quality images in just [--] to [--] steps Using CFG distillation weve achieved lightning-fast inference without compromising quality. https://modelscope.cn/studios/kelseye/Qwen-Image-2512-Turbo-LoRA-Demo/summary https://modelscope.cn/studios/kelseye/Qwen-Image-2512-Turbo-LoRA-Demo/summary https://modelscope.cn/studios/kelseye/Qwen-Image-2512-Turbo-LoRA-Demo/summary https://modelscope.cn/studios/kelseye/Qwen-Image-2512-Turbo-LoRA-Demo/summary"  
[X Link](https://x.com/ModelScope2022/status/2007792938248708207)  2026-01-04T12:35Z [----] followers, 30.5K engagements


"@WolfyBlair Yes you'll get daily free quota (magicubes) and you can earn more by participating in community activities"  
[X Link](https://x.com/ModelScope2022/status/2008002207099293874)  2026-01-05T02:26Z [----] followers, [---] engagements


"๐Ÿš€ New on ModelScope: The Wuli Team just released a Turbo LoRA that makes Qwen-Image-2512 inference 20x fasterโšก without sacrificing quality. Key updates: โœ… 4-step turbo inference โœ… Enhanced texture & detail โœ… Native ComfyUI support Model: Demo: https://modelscope.cn/studios/kelseye/Qwen-Image-2512-Turbo-LoRA-Demo https://modelscope.ai/models/Wuli-Art/Qwen-Image-2512-Turbo-LoRA https://modelscope.ai/models/Wuli-Art/Qwen-Image-2512-Turbo-LoRA Qwen-Image-2512-Fast is officially here โšก ๐Ÿš€ Try the Demo on ModelScope now: https://t.co/LRShwwbiNh Experience a massive 20x speedup generating"  
[X Link](https://x.com/ModelScope2022/status/2008020913309298809)  2026-01-05T03:41Z [----] followers, 11.3K engagements


"๐Ÿค– Introducing InternVLA-A1 now fully open-sourced Many VLA models follow instructions well in static scenes but struggle in dynamic environments (conveyor belts rotating platforms multi-robot setups). Why They see the presentbut cant imagine the future. InternVLA-A1 solution: unify perception imagination and action in one model: โœ… Scene understanding: Image + text task parsing โœ… Task imagination: Predict future frames reason about dynamics โœ… Guided control: Execute actions steered by visual foresight Powered by InternData-A1 - Large-scale high-quality simulated dataset InternVLA-A1 stays"  
[X Link](https://x.com/ModelScope2022/status/2008137224575992238)  2026-01-05T11:23Z [----] followers, 38.2K engagements


"๐Ÿš€ Small model BIG agency Meet Youtu-LLM a native agentic 1.96B LLM that thinks plans and acts by itself no external frameworks needed. โœ… Outperforms 4B8B models on math coding & agent benchmarks โœ… Trained on 200B+ agentic reasoning trajectories โœ… 128K context open-weight & ready to run anywhere Think lightweight cant be smart Think again. ๐Ÿค– Model: ๐Ÿ”— GitHub: ๐Ÿ“„ Paper: https://www.modelscope.cn/papers/2512.24618 https://github.com/TencentCloudADP/youtu-tip/blob/master/youtu-llm https://modelscope.cn/collections/Tencent-YouTu-Research/Youtu-LLM https://www.modelscope.cn/papers/2512.24618"  
[X Link](https://x.com/ModelScope2022/status/2008447725600469060)  2026-01-06T07:57Z [----] followers, 11.8K engagements


"๐Ÿš€ TeleChat3-105B-A4.7B-Thinking is now open source A 105B sparse MoE model with fine-grained routing: - [---] experts only [--] activated per token (4.7B active params) - Trained end-to-end on domestic compute - Strong across code math agents writing check HumanEval-X (92.7%) & SWE-Bench (51%) It can even build a working shooter game ๐Ÿ•น or a video site frontendfully autonomously. Also released: TeleChat3-36B-Thinking (dense 36B) optimized for reasoning & roleplay. โœ… Supports vLLM SGLang LLaMA-Factory โœ… Full Thinking mode with think./think reasoning traces โœ… Tuned for long-context and multi-turn"  
[X Link](https://x.com/ModelScope2022/status/2008499004158411187)  2026-01-06T11:20Z [----] followers, 16.8K engagements


"The #EmbodiedAI data gap just got a lot smaller. ๐Ÿš€ @AGIBOTofficial released the Genie Sim [---] datasetthe largest open-source sim dataset in the field ๐Ÿค– 10000+ hours of high-quality sim demos ๐ŸŽฏ 200+ tasks & 100k scenarios ๐Ÿ“Š Multi-sensor: RGB-D Stereo & Kinematics Standardized in LeRobot format. Ready to train โœ… Dataset: โœ… GitHub: https://github.com/AgibotTech/genie_sim https://modelscope.cn/datasets/agibot_world/GenieSim3.0-Dataset https://github.com/AgibotTech/genie_sim https://modelscope.cn/datasets/agibot_world/GenieSim3.0-Dataset"  
[X Link](https://x.com/ModelScope2022/status/2008795887666147377)  2026-01-07T07:00Z [----] followers, [----] engagements


"AI for all ๐ŸŒ Hosted by @alibaba_cloud the Alibaba Cloud AIGC Championship @ Milano Cortina [----] is officially live Create your Winter Olympics videos with the tools and resources powered by ModelScope as an official partner. โ›ธโ„ ๐Ÿ† Win Big ๐ŸŽŸ Milan [----] Tickets: A trip to the Winter Olympics ๐Ÿ› Legacy: Works archived in the Olympic Museum. ๐Ÿ’ฐ Cash Prizes: Reward your creative storytelling. ๐ŸŽฌ Pro Tech: High-speed AI video via Wan [---]. Global Entrance Creating #ModelScope #AlibabaCloud #MilanoCortina2026 #Wan26 #Olympics https://modelscope.ai/civision/videoGeneration"  
[X Link](https://x.com/ModelScope2022/status/2008824590517321874)  2026-01-07T08:54Z [----] followers, [----] engagements


"๐Ÿ‘ComfyUI now supports Z-Image-Turbo LoRAs trained with ModelScope Civision Huge thanks to our amazing community members for the incredibly fast integration ๐Ÿ‘‰ Everyone is welcome to train LoRAs for free on Civision and thanks to your valuable feedback training speed has been significantly improved ๐Ÿš€ https://modelscope.ai/civision/modelTraining https://github.com/Comfy-Org/ComfyUI/pull/11805 https://github.com/Comfy-Org/ComfyUI/pull/11805 https://modelscope.ai/civision/modelTraining https://github.com/Comfy-Org/ComfyUI/pull/11805 https://github.com/Comfy-Org/ComfyUI/pull/11805"  
[X Link](https://x.com/ModelScope2022/status/2010551381233709197)  2026-01-12T03:16Z [----] followers, 11.9K engagements


"GLM-Image is here Try it out and let us know what you think: https://x.com/ModelScope2022/status/2011262011997651194 ๐Ÿš€GLM-Image is now open-source a breakthrough in "Cognitive Generation." By combining Autoregressive models with Diffusion Decoders it finally solves the struggle of generating posters PPTs and knowledge-heavy infographics. ๐ŸŒŸ Highlights: โœ… #1 on CVTG-2K & LongText-Bench https://t.co/cF3z0VFvc9 https://x.com/ModelScope2022/status/2011262011997651194 ๐Ÿš€GLM-Image is now open-source a breakthrough in "Cognitive Generation." By combining Autoregressive models with Diffusion"  
[X Link](https://x.com/ModelScope2022/status/2011267263043944540)  2026-01-14T02:40Z [----] followers, [----] engagements


"Step-Audio-R1.1 by @StepFun_ai just set a new SOTA on the Artificial Analysis Speech Reasoning leaderboard ๐Ÿ† It outperforms Grok Gemini and GPT-Realtime with a 96.4% accuracy rate. โœ… Native Audio Reasoning (End-to-End) โœ… Audio-native CoT (Chain of Thought) โœ… Real-time streaming inference โœ… FULLY OPEN SOURCE ๐ŸŒŸ Demo: ๐Ÿค– Model: https://modelscope.cn/models/stepfun-ai/Step-Audio-R1.1 https://modelscope.cn/studios/stepfun-ai/Step-Audio-R1 https://modelscope.cn/models/stepfun-ai/Step-Audio-R1.1 https://modelscope.cn/studios/stepfun-ai/Step-Audio-R1"  
[X Link](https://x.com/ModelScope2022/status/2011687986338136089)  2026-01-15T06:32Z [----] followers, 15.6K engagements


"@SteveWarnerFL Hi would you mind sharing more about how you use it in your work Are you trying to modify a specific layer or add/remove one Wed love to better understand your situation"  
[X Link](https://x.com/ModelScope2022/status/2012050686171889776)  2026-01-16T06:34Z [----] followers, [--] engagements


"Step3-VL-10B Demo: https://modelscope.cn/studios/stepfun-ai/step3-vl-10b https://modelscope.cn/studios/stepfun-ai/step3-vl-10b"  
[X Link](https://x.com/ModelScope2022/status/2013213460721475715)  2026-01-19T11:34Z [----] followers, [----] engagements


"๐Ÿš€ Introducing GLM-4.7-Flash the strongest 30B-class open model that balances performance & efficiency โœ… Top scores on AIME GPQA SWE-bench -Bench & more โœ… Optimized for agentic coding tool use and reasoning โœ… Runs locally with vLLM / SGLang / Transformers Perfect for developers who want Claude-level coding power at a fraction of the cost ๐Ÿ’ก Try it now: ๐Ÿค– ๐Ÿ”— ๐Ÿ“„ ๐Ÿ™ https://github.com/zai-org/GLM-4.5 https://z.ai/blog/glm-4.7 https://chat.z.ai https://modelscope.cn/models/ZhipuAI/GLM-4.7-Flash https://github.com/zai-org/GLM-4.5 https://z.ai/blog/glm-4.7 https://chat.z.ai"  
[X Link](https://x.com/ModelScope2022/status/2013273444679590292)  2026-01-19T15:32Z [----] followers, 11.3K engagements


"How reliable is your model's tool usage. REALLY ๐Ÿง Inspired by @Kimi_Moonshot's K2 Vendor Verifier we are leveling up Agentic Evaluation with EvalScope. The standardizing Function Calling evaluation allows us to bridge the gap between "claims of support" and "RELIABLE execution" in the field. ๐Ÿ›  ๐Ÿ“– Docs: ๐Ÿ’ป Code: ๐Ÿ”น Standardized: Quantifies MoonshotAI K2 verification logic. ๐Ÿ”น Comprehensive: Measures decision-making + JSON schema compliance. ๐Ÿ”น Simple: Plug-and-play with your data or our benchmarks. Don't let your Agent go live without a proper check-up #LLM #AIAgents #FunctionCalling"  
[X Link](https://x.com/ModelScope2022/status/2014224614252253541)  2026-01-22T06:32Z [----] followers, [----] engagements


"@YashasGunderia Hi you'll need to request to join the organization first and then you'll have access to the GPU resources. Here is the link: https://www.modelscope.ai/organization/xGPU-Explorers https://www.modelscope.ai/organization/xGPU-Explorers"  
[X Link](https://x.com/ModelScope2022/status/2014937070008205667)  2026-01-24T05:43Z [----] followers, [---] engagements


"Thinking Machines Lab proved On-Policy Distillation slashes LLM training costs by 10x and we show you how to reproduce their research. Invest [--] minutes in this guideas we unpack the theory tech details experiment results and code to instantly transform your fine-tuning budget๐Ÿ“š Related Resources๐Ÿ‘‡): โœ… Slash training compute by 10X. โœ… Achieve robust RL performance with zero forgetting. โœ… Get the ready-to-use ms-SWIFT + vLLM code for deployment. Related Resources - TML Blog: - (Open-source implementation for reproducing On-Policy Distillation) - On-Policy Distillation Documentation: - Example"  
[X Link](https://x.com/ModelScope2022/status/1983159127934488946)  2025-10-28T13:09Z [----] followers, [----] engagements


"VibeThinker-1.5B is here ๐Ÿš€ and it flips the bigger = smarter myth on its head. โœ… Just 1.5B params โœ… Trained via novel Spectrum-to-Signal Principle (SSP) โœ… Beats models 400x larger (e.g. 671B DeepSeek-R1) on hard math benchmarks (AIME24/25 HMMT25) โœ… Matches 456B MiniMax-M1 and ties Mistrals Magistral-Medium on LiveCodeBench v6 โœ… Post-training cost: $8K (vs. $290K+ for others) ๐Ÿ’ก SSP first encourages divergent exploration of solution paths then uses RL to converge on optimal strategies smarter training not bigger models. ๐ŸŒ Open-sourced to empower researchers & smaller teams whove been priced"  
[X Link](https://x.com/ModelScope2022/status/1988955029064278216)  2025-11-13T13:00Z [----] followers, [----] engagements


"๐Ÿš€ GLM-4.6V is live @Zai_org 's new open-source multimodal models just dropped on ModelScope and we provide a FREE API for immediate accessโ— GLM-4.6V (106B) for cloud & high-performance workloads GLM-4.6V-Flash (9B) lightweight fast great for local inference ๐Ÿ”ฅ Key dev perks: โœ… Native multimodal tool calling pass images/docs directly as function args no OCR detour โœ… 128K context handles 150-page docs or hour-long videos in one go โœ… Visual Action pipeline powers real multimodal agents (e.g. find this outfit online returns structured shopping list) โœ… 50% cheaper than GLM-4.5V $1/million input"  
[X Link](https://x.com/ModelScope2022/status/1997995482459865290)  2025-12-08T11:43Z [----] followers, [----] engagements


"Weve prepared ModelScope and Tongyi swag for everyone attending the workshopready for a day full of great takeaways ๐Ÿš€ Join Us in Seoul: AI Innovation Meets Creativity @AMD @Alibaba_Qwen @ModelScope2022 @alibaba_cloud AMDs AI Developer Meetup in Seoul (Dec 10) is filling FAST. As a key partner were bringing you the future of generative AI. ๐Ÿ“… Dec [--] ๐Ÿ“ Seoul ๐ŸŽ Free https://t.co/b5XcmEmbPf ๐Ÿš€ Join Us in Seoul: AI Innovation Meets Creativity @AMD @Alibaba_Qwen @ModelScope2022 @alibaba_cloud AMDs AI Developer Meetup in Seoul (Dec 10) is filling FAST. As a key partner were bringing you the"  
[X Link](https://x.com/ModelScope2022/status/1998565458912620942)  2025-12-10T01:28Z [----] followers, [----] engagements


"Huge update for Z-Image-Turbo-Fun-Controlnet-Union [---] ๐Ÿš€ Key technical upgrades: โœ… New 1.9GB Lite model for low-VRAM & natural blending โœ… Fixed mask leakage in inpainting โœ… Multi-resolution dataset refactor (up to 1536px) โœ… 8-step distillation for crisp blur-free Turbo gen No more bright spot artifacts. Just pure control. ๐Ÿค– Model: https://modelscope.ai/models/PAI/Z-Image-Turbo-Fun-Controlnet-Union-2.1 https://modelscope.ai/models/PAI/Z-Image-Turbo-Fun-Controlnet-Union-2.1"  
[X Link](https://x.com/ModelScope2022/status/2011356116119334957)  2026-01-14T08:34Z [----] followers, 20.5K engagements


"Qwen-Image-Layered-Control is now supported by ComfyUI Try it out: https://modelscope.ai/models/DiffSynth-Studio/Qwen-Image-Layered-Control/files DiffSynth-Studio has open-sourced Qwen-Image-Layered-Control ๐Ÿš€ It enables precise text-controlled layer extractionperfect for poster decomposition and graphic design workflows. Control image layers like never before: โœ… Single-layer output via text prompts โœ… Native https://t.co/4QwOKqTFLs https://modelscope.ai/models/DiffSynth-Studio/Qwen-Image-Layered-Control/files DiffSynth-Studio has open-sourced Qwen-Image-Layered-Control ๐Ÿš€ It enables precise"  
[X Link](https://x.com/ModelScope2022/status/2011733445324783764)  2026-01-15T09:33Z [----] followers, 22.9K engagements


"Real-Qwen-Image-V2 is here ๐Ÿ“ธ Developed by wikeeyang this fine-tuned version of Qwen-Image-2512 is built for those who demand peak realism and production-grade quality. ๐Ÿค– Model: Key Highlights: โœจ Sharper details & enhanced realism โœจ Optimized for Asian facial aesthetics โœจ Superior text & image generation quality โœจ Highly compatible with LoRAs & custom workflows Tech Specs: โœ… CFG: [---] - [---] โœ… Steps: [--] - [--] โœ… Sampler: Euler / Simple (flexible) โœ… Model Shift: [---] - [---] https://www.modelscope.cn/models/wikeeyang/Real-Qwen-Image-V2 https://www.modelscope.cn/models/wikeeyang/Real-Qwen-Image-V2"  
[X Link](https://x.com/ModelScope2022/status/2012752195939168522)  2026-01-18T05:01Z [----] followers, 28.4K engagements


"๐Ÿš€ Meet STEP3-VL-10Bit delivers SOTA-level visual perception complex reasoning and human-aligned intelligenceredefining efficiency in open multimodal AI. โœ… Beats or matches models [----] larger (like GLM-4.6V Qwen3-VL even Gemini [---] Pro) โœ… Achieves SOTA on MMMU MathVision OCRBench ScreenSpot and more โœ… Trained on 1.2T tokens + 1400+ RL rounds (RLHF + RLVR) โœ… Supports PaCoRe: parallel collaborative reasoning (128K context) Despite its compact size it leads the 10B class in: STEM reasoning (94.43% on AIME [----] w/ PaCoRe) Visual perception (92.05 on MMBench) GUI understanding & OCR Spatial"  
[X Link](https://x.com/ModelScope2022/status/2013213019690639431)  2026-01-19T11:32Z [----] followers, 217.6K engagements


"Latency or intelligence Step-Audio-R1.1 ends the trade-off. Fei Tian Multimodal Large Model Researcher at StepFun introduces Mind-Paced Speakingenabling real-time voice AI that thinks while speaking. Fast responses. Deep reasoning. No compromise. https://x.com/i/broadcasts/1mrGmBabZavJy https://x.com/i/broadcasts/1mrGmBabZavJy"  
[X Link](https://x.com/ModelScope2022/status/2013219373948039628)  2026-01-19T11:57Z [----] followers, [----] engagements


"Introducing AgentCPM-Report: An 8B on-device agent that rivals closed-source giants in deep research and professional report writing. The Performance: ๐Ÿ† Ranked #1 in Insight Depth across DeepResearch Bench Deep Consult and DeepResearch Gymoutperforming top-tier closed models. Highlights: โœ… 100% Private: Zero data leakage. Your sensitive data stays on your hardware. โœ… UltraRAG Stack: Efficient local indexing for private PDFs and docs. โœ… Production Ready: One-click Docker deployment. How does an 8B model hit SOTA ๐Ÿง  It breaks the "one-shot" generation bottleneck with a "Writing as Reasoning""  
[X Link](https://x.com/ModelScope2022/status/2013586921046344058)  2026-01-20T12:18Z [----] followers, [----] engagements


"๐Ÿš€ Chroma [---] is here and its open From @flashlabsdotai : the worlds first open-source end-to-end real-time speech-to-speech dialogue model with personalized voice cloning. โœจ 150ms end-to-end latency ๐Ÿงฌ High-fidelity voice cloning from just seconds of audio ๐Ÿ“ˆ Speaker similarity (SIM) = [-----] +10.96% over human baseline ๐Ÿง  Strong reasoning in only 4B params ๐Ÿ”“ Fully open weights + code A true open alternative to closed commercial systems and already optimized with @lmsysorgs SGLang for even faster inference ๐Ÿ”— Paper: ๐Ÿค– Model: ๐Ÿ’ป Code: Were excited to see what the community builds with"  
[X Link](https://x.com/ModelScope2022/status/2014006971855466640)  2026-01-21T16:07Z [----] followers, 14.3K engagements


"๐ŸŒพ Meet Sinong (): The first heavy-duty LLM suite for Agriculture ๐Ÿšœ ๐Ÿงฌ Expertise: Crop breeding animal medicine smart farming & ag-economics. ๐Ÿค– Models: 8B & 32B parameters ๐Ÿ“š Stack: 4B+ tokens (240k+ papers 8k+ books patents) ๐Ÿง  Tech: Native CoT reasoning + Multi-Agent RAG for zero-hallucination scientific retrieval. ๐Ÿš€ Open Source: https://modelscope.cn/models/NAULLM/Sinong1.0-32B https://modelscope.cn/models/NAULLM/Sinong1.0-8B https://modelscope.cn/models/NAULLM/Sinong1.0-32B https://modelscope.cn/models/NAULLM/Sinong1.0-8B"  
[X Link](https://x.com/ModelScope2022/status/2014545364347715871)  2026-01-23T03:46Z [----] followers, [----] engagements


"@noctus91 Hi there This isn't actually an error. To use API inference you just need to bind your Alibaba Cloud account (no extra cost involved)"  
[X Link](https://x.com/ModelScope2022/status/2014754313999155689)  2026-01-23T17:37Z [----] followers, [---] engagements


"โšก๐ŸŽจTomorrow. This week is going to be intriguing"  
[X Link](https://x.com/ModelScope2022/status/2015613317088522594)  2026-01-26T02:30Z [----] followers, 51.8K engagements


"DeepSeek-OCR-2 is now on ModelScope โœจ Meet DeepSeek-OCR-2 a vision-language model that reimagines document intelligence with human-like visual understanding. Key Highlights ๐Ÿง  Visual Causal Flow: Understands layout semantics and structurenot just pixels ๐Ÿ“ Dynamic Resolution: Processes documents at (0-6)768 + [----] with adaptive token generation โœ Flexible Prompts: One model for multiple tasks: - Convert to markdown preserve tables & formatting - Parse the figure extract chart insights - Locate "text" pixel-level grounding - OCR this image 100+ language support โšก Production Ready:"  
[X Link](https://x.com/ModelScope2022/status/2016035065298616539)  2026-01-27T06:26Z [----] followers, [----] engagements


"๐Ÿš€ Meet Kimi K2.5 ๐ŸŒ™ This is Kimis most intelligent and versatile model to date achieving SOTA performance across coding vision and agentic workflows. Model: Paper: Highlights: โœ… Native Multimodal Architecture: Seamlessly integrates vision & text. From screenshots to screen recordings K2.5 "sees" and reasons across images and videos to break the limits of text-only prompts. โœ… Agent Clusters: A world-first capability. K2.5 can autonomously spawn up to [---] "clones" to work in parallel on complex tasks slashing execution time by 4.5x. โœ… The Ultimate Office/Coding Assistant: * Office Pro: Masters"  
[X Link](https://x.com/ModelScope2022/status/2016043017610957071)  2026-01-27T06:58Z [----] followers, [----] engagements


"The DiffSynth-Studio team also dropped Z-Image-i2L (Image-to-LoRA). ๐ŸŽจโœจ This model takes a single image as input and instantly outputs a custom LoRA tailored to that specific style or feature. Model: Demo: โœ… Zero Manual Effort: No captioning or training process required. โœ… One-Shot Magic: Get a style LoRA from just one reference image. โœ… Style Transfer: Seamlessly extract and apply unique features to new prompts. โœ… Lower Barrier: Personalized model creation is now accessible to everyone. https://modelscope.cn/studios/DiffSynth-Studio/Z-Image-i2L"  
[X Link](https://x.com/ModelScope2022/status/2016179961779126461)  2026-01-27T16:02Z [----] followers, [----] engagements


"๐Ÿš€ Meet LingBot-VLA: A pragmatic Vision-Language-Action model designed to bridge the gap between perception and execution in robotics. ๐Ÿค– โœ…LingBot-VLA-4B: Lightweight & versatile. โœ…LingBot-VLA-4B-Depth: Enhanced for high-precision spatial tasks. Highlights: ๐Ÿง  Powerful Core: Built on the Qwen2.5-VL-3B foundation mastering multi-tasking and dual-arm coordination across 9+ robot configs. ๐Ÿ“ˆ Elite Performance: Outperforms competitors like [---] and GR00T in success rates (SR) on both GM-100 (Real-world) and RoboTwin [---] (Sim). ๐Ÿ”‹ Hyper-Efficient: 1.52.8x faster training than existing VLA codebases"  
[X Link](https://x.com/ModelScope2022/status/2016209598471196691)  2026-01-27T18:00Z [----] followers, [----] engagements


"HunyuanImage-3.0 is here: A high-performance multimodal model for unified text-to-image generation and creative editing. ๐Ÿš€๐Ÿงต๐Ÿ‘‡ ๐Ÿค–Model: 1/ The Model Suite ๐Ÿ“ฆA versatile lineup for any workflow: โœ…Base 3.0: The core T2I powerhouse. โœ…3.0-Instruct: Advanced I2I and creative editing. โœ…Instruct-Distil: Optimized for 8-step sampling and rapid deployment. 2/ Core Capabilities ๐ŸŽจBeyond renderingits a creative partner. Features Unified Generation (T2I I2I and 3-image fusion) with built-in CoT reasoning and Prompt Enhancement to turn sparse ideas into professional descriptions. 3/ Technical Highlights"  
[X Link](https://x.com/ModelScope2022/status/2016841562274066648)  2026-01-29T11:51Z [----] followers, [----] engagements


"Meet LingBot-VA: The future of robot learning is visual. ๐Ÿค–๐ŸŽฅ Its an autoregressive diffusion framework that predicts future video frames and decodes actions simultaneously. Instead of just reacting it reasons: "I expect the world to look like this next so I should move like that." Key Highlights: ๐Ÿง  Infinite Memory: Unlike memoryless models that get stuck in loops LingBot-VA handles "recurrent states" (like opening the same box twice) with ease by remembering the full history. ๐Ÿณ Generalist Skills: From folding clothes to making breakfast and high-precision screw pickingits a true"  
[X Link](https://x.com/ModelScope2022/status/2017067214919389569)  2026-01-30T02:47Z [----] followers, [----] engagements


"PaddleOCR-VL-1.5 is open-sourced A 0.9B parameter multimodal model that masters the "curved & distorted" document challenge. ๐Ÿš€ ๐Ÿ”— ModelScope: ๐Ÿ”— Demo: Key Technical Breakthroughs: โœ…Polygon-based Localization: No more rigid rectangles. It natively supports irregular boxes to fit warped tilted or screen-captured text perfectly. ๐Ÿ“ โœ…Seal & Stamp Recognition: New dedicated capability to handle complex official document verification. ๐Ÿ’ฎ โœ…Cross-Page Logic: Automatically merges tables and identifies headings across physical page breakscrucial for full-document semantics. โœ…Performance Giant:"  
[X Link](https://x.com/ModelScope2022/status/2017113730866434213)  2026-01-30T05:52Z [----] followers, [----] engagements


"Ready to make this February legendary ๐Ÿš€โœจThe Qwen-Image LoRA Training Competition is officially LIVE Weve teamed up with @Ali_TongyiLab and @Alibaba_Qwen to bring you some absolutely grand prizes:๐Ÿ“ฑ iPhone [--] Pro Max (512GB) ๐ŸŽฎ Sony PS5 Pro ๐Ÿ› $800 Shopping Cards (or equivalent prizes) Stop scrolling and start training ๐Ÿ› ๐ŸŽจJoin here: #HappyQwensday #QwenImageLoRA https://modelscope.ai/active/qwenimagelora https://modelscope.ai/active/qwenimagelora"  
[X Link](https://x.com/ModelScope2022/status/2017939891846512907)  2026-02-01T12:35Z [----] followers, [----] engagements


"2/3 ๐Ÿ› Four specialist models were released in just one week each mastering a specific real-world challenge: โœ… DeepSeek-OCR2: Uses "Visual Causal Flow" to handle reading order in complex multi-column layouts. โœ… GLM-OCR (Z-org): An efficiency powerhouse using Multi-Token Prediction (MTP) to reduce costs by 90% while excelling at formulas. โœ… PaddleOCR-VL-1.5 (PaddlePaddle): Built for "Real5" messy conditionsperfect for curved pages and camera glare. โœ… Youtu-Parsing (Tencent): Delivers 22x faster inference and converts flowcharts directly into Mermaid code."  
[X Link](https://x.com/ModelScope2022/status/2018615134479057183)  2026-02-03T09:18Z [----] followers, [----] engagements


"Here they come They are now at or alumni of these elite AI labs: Alibaba Tongyi Meta (FAIR/Superintelligence) Google DeepMind Tencent AI. Academia: SJTU OSU UChicago NUS WashU. https://x.com/i/broadcasts/1ZkJzZRdQNoJv https://x.com/i/broadcasts/1ZkJzZRdQNoJv"  
[X Link](https://x.com/ModelScope2022/status/2019288799780770192)  2026-02-05T05:55Z [----] followers, [----] engagements


"@pranaysuyash http://ModelScope.ai http://ModelScope.ai"  
[X Link](https://x.com/ModelScope2022/status/2019390286355656859)  2026-02-05T12:38Z [----] followers, [---] engagements


"1/4๐Ÿงต MS-Agent is Rebuilding A massive functional update to ModelScopes agent frameworkintroducing autonomous skill scaling and industrial-grade toolkits. โœ… Agent Skills: Skill Discovery via local/Hub repo_id matching; โœ… Multi-skill Collaboration via DAG-based planning; โœ… Progressive Analysis to optimize context window for small models. โœ… Robust Executor: Secure Docker/Sandbox execution; autonomous self-reflection and bug-fix cycles. Github: https://github.com/modelscope/ms-agent/tree/main/ms_agent/skill https://github.com/modelscope/ms-agent/tree/main/ms_agent/skill"  
[X Link](https://x.com/ModelScope2022/status/2019600619443744930)  2026-02-06T02:34Z [----] followers, [---] engagements


"2/4 ๐Ÿ’ป Code Genesis is Now Live A SOTA end-to-end synthesis engine specializing in zero-to-one development over standard "vibe-coding." โœ… Capabilities: Industrial-grade Code Genesis for full-stack synthesis; Language Server integration for stability and clean code aesthetics. โœ… Technical Edge: Superior benchmarks in end-to-end logic flow; minimalist architecture for rapid project scaffolding. Repo: Tech Report: https://github.com/modelscope/ms-agent/tree/main/projects/code_genesis/TECH_REPORT.md https://github.com/modelscope/ms-agent/tree/main/projects/code_genesis"  
[X Link](https://x.com/ModelScope2022/status/2019600621608022159)  2026-02-06T02:34Z [----] followers, [---] engagements


"๐ŸŽฌ Singularity Cinema is Breaking Benchmarks The GLCC-winning workshop for high-density long-form automated video production. โœ… Generation: LLM-driven storyboards; Manim/Remotion dual-engine animation; 5min+ cinematic output for Science Econ and History. โœ… Model Fusion: Seamlessly integrates Qwen-Image Sora edge-tts and nano-banana-pro; transforms static docs into dynamic media. Demo: https://github.com/modelscope/ms-agent/tree/main/projects/singularity_cinema https://github.com/modelscope/ms-agent/tree/main/projects/singularity_cinema"  
[X Link](https://x.com/ModelScope2022/status/2019600623738765722)  2026-02-06T02:34Z [----] followers, [---] engagements


"4/4 ๐ŸŒ ms-agent WebUI is Live A self-bootstrapped interface generated entirely by the Code Genesis engine. โœ… Usage: pip install ms-agent -U - ms-agent ui; instant GUI access for agent orchestration. โœ… Roadmap: DeepResearch capability launching soonSOTA autonomous retrieval and synthesis incoming. Full Framework: https://github.com/modelscope/ms-agent https://github.com/modelscope/ms-agent"  
[X Link](https://x.com/ModelScope2022/status/2019600625835864392)  2026-02-06T02:34Z [----] followers, [---] engagements

Limited data mode. Full metrics available with subscription: lunarcrush.com/pricing

@ModelScope2022 Avatar @ModelScope2022 ModelScope

ModelScope posts on X about native, agentic, ai, inference the most. They currently have [-----] followers and [---] posts still getting attention that total [------] engagements in the last [--] hours.

Engagements: [------] #

Engagements Line Chart

  • [--] Week [------] -12%
  • [--] Month [-------] +72%
  • [--] Year [---------] +210,270%

Mentions: [--] #

Mentions Line Chart

  • [--] Year [--] +9,400%

Followers: [-----] #

Followers Line Chart

  • [--] Week [-----] +4.50%
  • [--] Month [-----] +52%

CreatorRank: [-------] #

CreatorRank Line Chart

Social Influence

Social category influence technology brands 9% finance 2% travel destinations 2% stocks 2% products 1% gaming 1%

Social topic influence native #117, agentic #156, ai 10%, inference #190, realtime 7%, model 7%, math 6%, flow 5%, tencent 4%, science 4%

Top accounts mentioned or mentioned by @furaidosu @alitongyilab @alibabacloud @alibabaqwen @ali_tongyilab @barrakali @dx8152 @1328 @ultralytics @sean60133791259 @bdsqlsz @nnnnmailcom @blankbraindead @wolfyblair @agibotofficial @stepfunai @stevewarnerfl @kimimoonshots @yashasgunderia @zaiorg

Top Social Posts

Top posts by engagements in the last [--] hours

"๐Ÿฟ Meet Sirchmunk: The agentic search engine that operates directly on your raw data. Turn files into evolving intelligence in real-time Build information retrieval pipelines that require ZERO vector databases ZERO pre-indexing and ZERO chunking: ๐ŸŽฏ Embedding-Free Retrieval Drop files in and search instantly. No ETL pipelines no vector silos and no setup hours. ๐ŸŽฒ Monte Carlo Evidence Sampling It doesn't "read" full docsit explores focuses and synthesizes. 500-page manual Only the most relevant snippets hit the LLM. ๐Ÿง  Self-Evolving Knowledge Every search builds reusable knowledge clusters."
X Link 2026-02-13T18:00Z [----] followers, [----] engagements

"Love seeing what the MiniMax team shipped with M2.5 ๐Ÿš€ Real-world productivity strong agentic workflows and impressive efficiency numbers this is serious engineering. SOTA across coding search and tool use while getting faster is a big deal. ๐Ÿ”ฅ Excited to share that the model will be coming to ModelScope soon for the open-source community. Stay tuned ๐ŸŒโœจ Introducing M2.5 an open-source frontier model designed for real-world productivity. - SOTA performance at coding (SWE-Bench Verified 80.2%) search (BrowseComp 76.3%) agentic tool-calling (BFCL 76.8%) & office work. - Optimized for efficient"
X Link 2026-02-12T17:09Z [----] followers, [----] engagements

"Nanbeige just dropped Nanbeige4.1-3B a tiny powerhouse that outclasses models 10x its size.๐Ÿš€ โœ… 256k Context Window โœ… "Deep Search" Agent capabilities built-in โœ… Two-stage Code RL (Correctness โžก Efficiency) โœ… Beats Qwen3-32B on key benchmarks The era of high-performance low-cost local deployment is here. ๐Ÿค–: https://modelscope.cn/models/nanbeige/Nanbeige4.1-3B https://modelscope.cn/models/nanbeige/Nanbeige4.1-3B"
X Link 2026-02-14T11:11Z [----] followers, 11.9K engagements

"Introducing LingBot-World: An open-source world simulator pushing the boundaries of video generation. ๐Ÿš€ ๐ŸŒ High-Fidelity: Realistic scientific & stylized. ๐Ÿง  Long-Term Memory: Minute-level consistency. โšก Real-Time: 1s latency at [--] FPS. ๐Ÿ“œ Apache [---] Licensed. Model: Github: https://github.com/Robbyant/lingbot-world https://modelscope.cn/models/Robbyant/lingbot-world-base-cam https://github.com/Robbyant/lingbot-world https://modelscope.cn/models/Robbyant/lingbot-world-base-cam"
X Link 2026-01-28T17:30Z [----] followers, 29.3K engagements

"Tencent YouTu Research open-sourcedYoutu-VL-4B-Instruct a compact VLM redefining visual perception viaVLUAS. ๐Ÿ‘โšก ๐ŸŒŸHighlights: โœ…All-in-One Vision: SOTA performance in detection segmentation depth estimation and pose estimationwithouttask-specific heads. โœ…OCR & Reasoning: Exceptional at complex document parsing and multimodal math reasoning. โœ…GUI Agent Ready: Optimized for world-understanding and interface navigation. โœ…Efficiency: 4B parameters make it perfect for edge deployment and fast inference. ๐Ÿ”งPerformance: Outperforms many larger models inOmniDocBenchandVision-Centrictasks while"
X Link 2026-01-30T11:00Z [----] followers, [----] engagements

"Stepfun open-sourcedStep-3.5-Flash a powerhouse model specifically architected for high-speed reasoning and complex Agentic workflows. ๐Ÿš€ Model: Key Technical Specs: โœ…Sparse MoE Architecture: 196B total params but only11B activeper token. SOTA efficiency. โœ…MTP-3 (Multi-Token Prediction): It predicts3 tokens at once hitting a blistering350 TPSfor code-heavy tasks. โšก โœ…Hybrid Attention (SWA + Full): A 3:1 mix that masters256K contextwindows while keeping compute costs low. โœ…Parallel Thinking: Massively boosted performance for multi-step reasoning and deep search. Why Devs should care: - Built"
X Link 2026-02-02T06:52Z [----] followers, 34K engagements

"Beyond-Reality-Z-Image [---] is here.๐Ÿš€ Highlights: โœ…Texture SOTA: Optimized high-frequency details for what might be the best out-of-the-box skin and fabric textures in the Z-model family. โœ…Cinematic Lighting: Re-engineered color balance to achieve that authentic "film look" without heavy prompting. โœ…Speed x Versatility: Generates at Turbo speeds (10-15 steps) but supports broader artistic styles and 2D/Anime aesthetics than the official release. โœ…Efficiency: Optimized for low CFG (1.0) making it highly stable and responsive to LoRA integration. Available now on ModelScope: Recommended Stack:"
X Link 2026-02-02T13:03Z [----] followers, 16.4K engagements

"How do Agents "Grow from Experience" A dialogue on Key Issues and Technical Paths of Self-Evolution AI AMA Deep Roundtable Vol.1 [----] was about Reasoning. [----] is the "Mirror Moment"Agents are shifting from static tools to digital lifeforms that learn iterate and evolve from every failure. ๐Ÿš€ https://twitter.com/i/web/status/2018712297829732601 https://twitter.com/i/web/status/2018712297829732601"
X Link 2026-02-03T15:44Z [----] followers, [----] engagements

"Meetthe1stAuthorsbehindthemostinfluentialpapers:earlyexperienceDREAMGYMAgentEvolverSPICER-ZeroEvo-MemoryandSEAgent. TheyarenowatoralumnioftheseeliteAIlabs: ๐Ÿ”นAlibabaTongyiLabMetaFAIRMetaSuperintelligenceLabsGoogleDeepMindTencentAISeattleLab ๐Ÿ”นSJTUOSUUChicagoNUSWashU 6leadresearchers1stage.Unrivaledtechnicaldepth.๐Ÿง  https://twitter.com/i/web/status/2018712300744847796 https://twitter.com/i/web/status/2018712300744847796"
X Link 2026-02-03T15:44Z [----] followers, [---] engagements

"MiniCPM-o [---] brings true "Omni" capabilities (Vision + Audio + Text) into a single 9B end-to-end architecture (Qwen3-8B + SigLip2 + Whisper + CosyVoice2).๐Ÿš€ โœ… Vision SOTA: OpenCompass score of 77.6surpassing GPT-4o and Gemini [---] Pro in VLM tasks with just 9B params. โœ… Real-time Full Duplex: It "sees" "hears" and "speaks" simultaneously. Supports active interaction (it can interrupt or initiate conversation based on live video). โœ… World-Class OCR: Beats DeepSeek-OCR [--] and GPT-4o in document parsing (OmniDocBench leader). โœ… Voice Cloning: Clone any voice with just a short reference"
X Link 2026-02-04T05:51Z [----] followers, [----] engagements

"๐Ÿš€ ACE-Step v1.5 is out: an open-source music foundation model that runs locally on consumer GPUs (4GB VRAM) and generates full songs in 2s (A100) or 10s (RTX 3090). โœ… Beats most commercial models in quality โœ… Train a personalized LoRA from just a few tracks โœ… Built on a novel LM+DiT hybrid architecture with internal RLno external reward bias โœ… Supports 50+ languages + editing (covers repainting vocalBGM) ๐Ÿค– https://modelscope.cn/models/ACE-Step/Ace-Step1.5 https://modelscope.cn/models/ACE-Step/Ace-Step1.5"
X Link 2026-02-04T09:19Z [----] followers, [----] engagements

"๐Ÿš€ Meet Intern-S1-Pro: A massive 1T parameter MoE model for Multimodal Science Reasoning โœ… [---] Experts (22B active) โœ… SOTA in AI4Science (Chemistry Materials Bio) โœ… FoPE + Time-series modeling (up to [--] points) โœ… Native "Thinking Mode" support Open-source science just leveled up. ๐Ÿงช๐Ÿ’ป Model: Github: https://github.com/InternLM/Intern-S1 https://www.modelscope.cn/models/Shanghai_AI_Laboratory/Intern-S1-Pro https://github.com/InternLM/Intern-S1 https://www.modelscope.cn/models/Shanghai_AI_Laboratory/Intern-S1-Pro"
X Link 2026-02-04T15:20Z [----] followers, [----] engagements

"Introducing Singularity Cinema (MS-Agent framework)๐Ÿš€an LLM-powered tool that transforms documents or prompts into storyboards generating stunning 5-minute+ videos via its Manim/Remotion dual-engine. By fusing nano-banana-pro๐ŸŒ Qwen-Image๐ŸŽจ Sora and edge-tts it delivers high-quality information-dense contentperfect for science economics and history creators looking to scale their social media presence. ๐Ÿ”—: ๐ŸŽฆPrompt: Convert xx/deploy.md into a short video in a blue-themed style making sure to use the important images from the document. The short video must be in English. ๐ŸŽฌOutput Video๐Ÿ‘‡"
X Link 2026-02-05T16:08Z [----] followers, [----] engagements

"From a local designer in Chengdu to a global AI sensation with 117k+ downloadsDaniels journey shows whats possible when open-source meets imagination. Now its your turn. Join the Qwen-Image LoRA Competition and start building your own legacy. The next top creator could be you. https://modelscope.ai/active/qwenimagelora LoRA Training Contest: Win iPhone/PS5 and other prizes Gaussian splash LoRA creation process shared. Competition information: https://t.co/MEEZHQcr83 Training address: https://t.co/uxwtMlJzoP @Ali_TongyiLab @ModelScope2022 #HappyQwensday #QwenImageLora https://t.co/5igMVZUO5i"
X Link 2026-02-06T03:20Z [----] followers, [----] engagements

"ChenkinNoob-XL v0.2 is here A high-fidelity SDXL-to-Rectified Flow conversion delivering vivid color range and superior lighting adherence. โœ… Performance: Eliminates "base model greyness"; Superior contrast/dark scene rendering; High stability across wide CFG ranges (3-6); Optimal convergence at 20-28 steps. โœ… Ecosystem: Native ComfyUI workflow included; ReForge-ready with built-in Flow support; LoRA-compatible via Pochi.toml (Refresh scripts). โœ… Architecture: Rectified Flow (RF) linear ODE trajectory; Trained on 47M samples using 8xH20 nodes; Leverages Cosine Optimal Transport and Protected"
X Link 2026-02-06T05:00Z [----] followers, [----] engagements

"StepFun open-sourced Step-3.5-Flashfeaturing powerful reasoning and Agent intelligence. Speeds up to [---] TPS Competes with closed-source models in Agent and Math tasks Excels at complex long-chain reasoning. Dont miss out https://x.com/i/broadcasts/1OyKAjZgZPaGb https://x.com/i/broadcasts/1OyKAjZgZPaGb"
X Link 2026-02-06T12:05Z [----] followers, [----] engagements

"Meet X-Coder ๐Ÿš€IIGroup drops a powerhouse suite for code LLMs โœ… X-Coder-SFT-376k: Massive synthetic dataset via GPT-o3-mini/DeepSeek-R1. โœ… X-Coder-RL-40k: Targeted RL alignment for logic & verification. โœ… X-Coder-8B Models: SOTA reasoning built on Qwen3. Experience the power of fully synthetic code evolution. ๐Ÿ’ป๐Ÿ”ฅ ๐Ÿ“Š Models & Dataset: #LLM #OpenCode #Qwen3 #DeepSeek #Coding https://modelscope.cn/models/IIGroup/X-Coder-SFT-Qwen3-8B https://modelscope.cn/datasets/IIGroup/X-Coder-RL-40k https://modelscope.cn/datasets/IIGroup/X-Coder-SFT-376k"
X Link 2026-02-09T06:25Z [----] followers, [----] engagements

"YOLO26 is here๐Ÿš€ The most advanced and deployment-ready vision model from @ultralytics redefining real-time edge computing performance. โœ… Performance: 43% CPU inference speed boost vs YOLO11; Real-time performance on low-power hardware; [--] scalable sizes (n/s/m/l/x) for optimized speed-accuracy trade-offs. โœ… Ecosystem: Unified framework for Detection Segmentation Pose OBB and Tracking; Simplified export for seamless hardware integration. โœ… Logic/Architecture: Removed DFL for enhanced hardware compatibility; End-to-end NMS-free inference; ProgLoss + STAL for small target precision; MuSGD"
X Link 2026-02-09T08:30Z [----] followers, [----] engagements

"2/4 Key Technical Highlights: - Architecture: Native single-stream design integrating generation editing and segmentation. - Audio: First model to unify Speech Audio and Music generation via 12.5Hz continuous representation. - Video: Time-Interleaved VideoRoPE for enhanced temporal modeling. https://twitter.com/i/web/status/2021580867009671592 https://twitter.com/i/web/status/2021580867009671592"
X Link 2026-02-11T13:43Z [----] followers, [---] engagements

"๐Ÿš€ GLM-5 is now open-sourceand ranked #4 globally #1 among open models on Artificial Analysis Beyond "vibe coding" it delivers Agentic Engineering: autonomously architects debugs and ships production systems. Coding performance rivals Claude Opus [---] (SWE-bench: [----] ๐Ÿ”“ SOTA) while dominating long-horizon agent tasks (BrowseComp MCP-Atlas ๐Ÿ”). 744B params MIT License ๐Ÿ‘‰ ModelScope: https://modelscope.cn/models/ZhipuAI/GLM-5 https://modelscope.cn/models/ZhipuAI/GLM-5"
X Link 2026-02-12T03:43Z [----] followers, [----] engagements

"๐Ÿฟ Meet Sirchmunk: The agentic search engine that operates directly on your raw data. Turn files into evolving intelligence in real-time Build information retrieval pipelines that require ZERO vector databases ZERO pre-indexing and ZERO chunking: ๐ŸŽฏ Embedding-Free Retrieval Drop files in and search instantly. No ETL pipelines no vector silos and no setup hours. ๐ŸŽฒ Monte Carlo Evidence Sampling It doesn't "read" full docsit explores focuses and synthesizes. 500-page manual Only the most relevant snippets hit the LLM. ๐Ÿง  Self-Evolving Knowledge Every search builds reusable knowledge clusters."
X Link 2026-02-13T13:39Z [----] followers, [---] engagements

"LongCat-Flash-LiteisNow Live Meituan's 68.5B MoE model pioneering "Embedding Expansion" for leading Agent and Coding performance. โœ…Performance: 2.9B4.5B activated params; 500-700 tokens/s throughput; 256K context via YARN; SOTA SWE-Bench (54.4%) and -Bench scores. โœ…Ecosystem: Optimized SGLang-FluentLLM engine with custom CUDA kernels; Open API with 50M free daily tokens. โœ…Architecture: N-gram embedding layer for localized semantic precision; Hash-collision mitigation via sub-table decomposition; Triple-step Speculative Decoding synergy. ๐Ÿค– ๐Ÿ”ง ๐Ÿ“„ Tech"
X Link 2026-02-08T06:27Z [----] followers, [----] engagements

"๐Ÿš€1/4 Introducing Ming-flash-omni-2.0: An open-source Omni-MLLM breakthrough. Based on the Ling-2.0 MoE architecture (100B total / 6B active) it unifies vision speech and image editing into a single model matching specialist performance across all modalities. ๐Ÿงต๐Ÿ‘‡ ๐Ÿค– ModelScope: ๐Ÿ“ฆ GitHub: https://github.com/inclusionAI/Ming https://www.modelscope.cn/models/inclusionAI/Ming-flash-omni-2.0 https://github.com/inclusionAI/Ming https://www.modelscope.cn/models/inclusionAI/Ming-flash-omni-2.0"
X Link 2026-02-11T13:43Z [----] followers, [----] engagements

"๐Ÿš€Introducing Z-Image the full non-distilled version of Z-Image for high-quality customizable image generation. โœ… Model: โœ… Demo: โœจ Capabilities: Photorealistic + diverse art styles (anime illustration digital art) Full CFG & precise negative prompts for reliable prompt following Multi-subject generation with distinct identitiesno "clone faces" Community-friendly: ready for LoRA ControlNet and other fine-tuning workflows ๐Ÿ”ฅ Highlights: Full architecture (vs. distilled Turbo): better diversity & training stability ideal for customization High output variance across seeds = less homogenization"
X Link 2026-01-27T15:57Z [----] followers, 59.5K engagements

"Z-Image-Fun-Lora-Distill is here A high-efficiency Distill LoRA enabling CFG-free high-fidelity generation for Z-Image in just 4-8 steps. โœ… 4-step and 8-step variants; CFG-free inference (CFG=1.0); [----] update delivers superior skin texture and color consistency compared to previous iterations. โœ… Native ComfyUI weights available; full integration with the VideoX-Fun framework; compatible with Z-Image ControlNet Union [---] and third-party LoRAs. โœ… Dual-distillation of both sampling steps and CFG; trained from scratch without Z-Image-Turbo weights; optimized for rapid generation and derivative"
X Link 2026-02-11T02:33Z [----] followers, [----] engagements

"๐Ÿš€ Ring-2.5-1T is open-source The first hybrid linear 1T thinking model. ๐Ÿ“Š Gold Tier reasoning IMO25 (35/42) & CMO25 (105/126) scores. ๐Ÿ›  Agent-native compatible with Claude Code OpenClaw and SGLang. โšกHybrid Linear breakthrough 10x lower memory & 3x throughput for long context. ๐Ÿง  1T MoE structure with 63B active parameters (1:7 MLA + Lightning Linear). ๐Ÿ’ป Long-Horizon mastery autonomously built a 32-bit OS from scratch in [--] hours. Model: GitHub: https://github.com/inclusionAI/Ring-V2.5 https://modelscope.cn/models/inclusionAI/Ring-2.5-1T https://github.com/inclusionAI/Ring-V2.5"
X Link 2026-02-13T08:17Z [----] followers, 13K engagements

"Introducing FireRed-Image-Edit-1.0 from FireRedTeam ๐Ÿš€ Its officially the new SOTA for general image editing. โœ… Better than Closed-Source: Outperforms Nano-Banana & Seedream4.0 on GEdit benchmarks. โœ… Native Evolution: Built from T2I foundations not just a "patch" on existing models. โœ… Style Mastery: Scored a record-breaking 4.97/5.0 in style transfer. โœ… High-Fidelity Text: Keeps original font styles perfectly. โœ… Virtual Try-on: Native support for multi-image joint editing. โœ… Bilingual: Native support for both English & Chinese prompts. Apache [---] license. Local deployment ready. ๐Ÿค– Model: ๐ŸŽ "
X Link 2026-02-14T15:00Z [----] followers, 69.4K engagements

"How 0.9B Models Are Outperforming the Giants: A Comparison of Four New Chinese VLM-OCR Models 1/3 ๐Ÿ† Early [----] has brought a massive surge of open-source models from DeepSeek Z-org PaddlePaddle and Tencent YouTu. A tiny 0.9B parameter model GLM-OCR topped the global OmniDocBench v1.5 leaderboard outscoring giants like Gemini-3 Pro and Qwen3-VL. We are moving away from rigid "detect and recognize" pipelines toward Vision-Language Models (VLMs) that actually "read" and understand document structures like humans do. https://twitter.com/i/web/status/2018615132134551565"
X Link 2026-02-03T09:18Z [----] followers, 30.1K engagements

"Z-Image-Fun-Controlnet-Union & Lora-Distill are now open-source ๐Ÿš€Expanding the Z-Image ecosystem with ultra-fast inference and high-precision spatial control. โœ… Performance: Achieve 8-step inference (CFG: 1.0) via Distill LoRA and professional-grade multi-condition control (Canny Depth Pose Inpaint) across 15+ layer blocks. โœ… Ecosystem: Full cross-compatibility between Distill LoRA and Union ControlNet; includes "Lite" versions for low-VRAM edge devices and seamless natural blending. โœ… Unified Control: Support for 7+ spatial conditions in a single model with adjustable control_context_scale"
X Link 2026-02-05T09:17Z [----] followers, 11K engagements

"Qwen3-ASR & Qwen3-ForcedAligner are now open-source ๐ŸŽ™๐Ÿš€ High-performance streaming-ready speech AI for [--] languages and dialects. ๐Ÿ”—: Demo: ๐Ÿค–The Models โœ… ASR-1.7B: Open-source SOTA. Outperforms Whisper-v3; rivals GPT-4o & Gemini. โœ… ASR-0.6B: Insane efficiency92ms latency; transcribes 2000s of audio in 1s. โœ… ForcedAligner-0.6B: Superior timestamp accuracy across [--] languages. ๐ŸŒŸKey Features โœ…Universal: Supports [--] languages + [--] Chinese dialects. โœ…Robust: Expertly handles singing heavy BGM and low SNR. โœ…Unified: Single model for both streaming & offline (up to 20min audio). โšกPerformance &"
X Link 2026-01-29T12:55Z [----] followers, [----] engagements

"๐Ÿš€ FantasyWorld is now open source Recently accepted to ICLR [----] and ranking #1 on the WorldScore leaderboard this framework offers a unified approach to video and 3D generation. โœจKey Technical Highlights: โœ…Direct 3D Inference: Input an image and camera trajectory to get frame-consistent video dense depth maps and point clouds instantly. โœ…Dual-Branch IRG Architecture: An "Imagination" branch handles visual appearance while a "Geometry" branch ensures spatial accuracy. โœ…Wan [---] Integration: The released 14B models leverage the latest Wan-DiT architectures for superior temporal consistency and"
X Link 2026-02-13T11:00Z [----] followers, [----] engagements

"Say hello to Ling-2.5-1T: 1T params 63B active MIT licensed โšก Hybrid Linear Attention: 1:7 MLA + Lightning Linear. Beats Kimi K2 on long-context throughput. ๐Ÿง  4x token efficiency: Composite rewards match frontier thinking models using 4x fewer tokens. ๐Ÿ“š 1M context: YaRN to 1M. Beats Kimi K2.5 & DeepSeek V3.2 on RULER/MRCR. Perfect NIAH scores. ๐Ÿ›  SOTA tool use: Agentic RL trained. Leads BFCL-V4. Native support for Claude Code OpenCode OpenClaw. ๐ŸŽฏ Follows instructions: Bidirectional RL + agent verification. High-density zero fluff. Honest: Gap vs GPT-5.2/Gemini [--] Pro on long-horizon tasks."
X Link 2026-02-16T03:39Z [----] followers, [----] engagements

"New Z-Image-Turbo control model released ๐ŸŽ‰ More control blocks. Inpaint mode. Better details. - 6-block ControlNet - Trained from scratch on 1M images @1328 res - Canny / HED / Depth / Pose / MLSD - Best control_context_scale: 0.65-0.80 Model: Demo: Code: https://github.com/aigc-apps/VideoX-Fun https://modelscope.cn/studios/AI-ModelScope/ZIT-Controlnet https://modelscope.cn/models/PAI/Z-Image-Turbo-Fun-Controlnet-Union https://modelscope.cn/models/PAI/Z-Image-Turbo-Fun-Controlnet-Union https://github.com/aigc-apps/VideoX-Fun https://modelscope.cn/studios/AI-ModelScope/ZIT-Controlnet"
X Link 2025-12-16T03:15Z [----] followers, 33.1K engagements

"๐Ÿš€Model & Demo are now live on our international site Model: - - - Demo: https://modelscope.ai/studios/FunAudioLLM/Fun-CosyVoice3-0.5B https://modelscope.ai/models/FunAudioLLM/Fun-CosyVoice3-0.5B-2512 https://modelscope.ai/models/FunAudioLLM/Fun-ASR-Nano-2512 https://modelscope.ai/models/FunAudioLLM/Fun-ASR-MLT-Nano-2512 https://modelscope.ai/studios/FunAudioLLM/Fun-CosyVoice3-0.5B https://modelscope.ai/models/FunAudioLLM/Fun-CosyVoice3-0.5B-2512 https://modelscope.ai/models/FunAudioLLM/Fun-ASR-Nano-2512 https://modelscope.ai/models/FunAudioLLM/Fun-ASR-MLT-Nano-2512"
X Link 2025-12-16T09:43Z [----] followers, [---] engagements

"๐Ÿค–Meet Nanbeige4-3B from Boss Zhipina 3B-parameter LLM that outperforms Qwen3-32B on math (AIME) science (GPQA) and tool calling (BFCL-V4) while matching Qwen3-30B-A3B on human preference alignment (Arena-Hard-V2). How โœ… 23T tokens of ultra-curated data โœ… Fine-grained WSD scheduler โœ… 30M+ high-quality SFT instructions โœ… Multi-stage RL + innovative distillation (DPD) โœ… Chain-of-thought reconstruction & deliberative generation It even ranks top [--] on WritingBench & EQ-Bench3beating models 100x larger like GLM-4.5 and Deepseek-R1. All models + tech report now open-source: ๐Ÿ”— Weights: ๐Ÿ“„ Paper:"
X Link 2025-12-16T11:28Z [----] followers, 25.8K engagements

"๐ŸŽ‰ MiMo-V2-Flash FREE API is now live on ModelScope The first major release since Fuli Luo joined Xiaomiand its built for real-world agentic AI. โšก MiMo-V2-Flash: an open high-performance MoE model with 309B total / 15B active parameters 256K context window 150+ tokens/s generation thanks to native Multi-Token Prediction ๐Ÿ”ฅ Key wins for developers: โœ… Hybrid attention (5:1 SWA + Global) [--] less KV cache full long-context recall โœ… 73.4% on SWE-Bench Verified new SOTA for open-source models โœ… Matches DeepSeek-V3.2 on reasoning but much faster in practice โœจ API-readyperfect for building smart"
X Link 2025-12-18T02:29Z [----] followers, 16K engagements

"๐Ÿš€Qwen-Image-Layered is now live on ModelScopean incredible model that can intelligently decompose any image into multiple RGBA layers ๐Ÿค– Model: ๐ŸŒŸ Try Demo ๐Ÿ“‘ Paper: https://modelscope.cn/papers/2512.15603 https://modelscope.cn/studios/Qwen/Qwen-Image-Layered https://modelscope.ai/models/Qwen/Qwen-Image-Layered https://modelscope.cn/papers/2512.15603 https://modelscope.cn/studios/Qwen/Qwen-Image-Layered https://modelscope.ai/models/Qwen/Qwen-Image-Layered ๐ŸŽจ Qwen-Image-Layered is LIVE native image decomposition fully open-sourced โœจ Why it stands out โœ… Photoshop-grade layering Physically"
X Link 2025-12-19T15:29Z [----] followers, [----] engagements

"Its still Sunday and Im already hyped for next week. ๐Ÿ˜ŠThe mysterious Eastern power is dropping a trilogy of magic: ๐Ÿ”ฎ Next week: A secret image-generation model ๐Ÿค– Next week: A secret code-agent model ๐ŸŽ™ The week after: A secret voice model Buckle upthings are about to get interesting"
X Link 2025-12-21T09:54Z [----] followers, 10.1K engagements

"๐Ÿš€ Meet GLM-4.7 your new coding partner supercharged Try it now on ModelScope free API access for a limited time โœ… +5.8% on SWE-bench (73.8%) โœ… +12.9% on Multilingual coding (66.7%) โœ… +16.5% on Terminal Bench [---] (41%) โœ… Smarter agent reasoning & tool use โœ… Better UI cleaner webpages sharper slides โœ… +12.4% on complex reasoning (HLE: 42.8%)From coding to creativity GLM-4.7 delivers across the board. ๐Ÿค– Model: https://modelscope.cn/models/ZhipuAI/GLM-4.7 https://modelscope.cn/models/ZhipuAI/GLM-4.7"
X Link 2025-12-23T02:23Z [----] followers, [----] engagements

"๐Ÿš€ New on ModelScope: QwenLong-L1.5 is now fully open-source A 30B model (3B active params) that matches GPT-5 & Gemini-2.5-Pro in long-context reasoning. ๐Ÿ”ฅ Key wins: โœ… +31.7 pts on OpenAIs MRCR (128K context SOTA across all models) โœ… Matches Gemini-2.5-Pro on [--] major long-QA benchmarks โœ… +9.69 on CorpusQA +6.16 on LongBench-V2 How Three breakthroughs: [--] Synthetic data at scale: 14.1K long-reasoning samples from 9.2B tokens no human labeling. Avg. length: 34K tokens (max: 119K). [--] Stable RL training: Task-balanced sampling + Adaptive Entropy-Controlled Policy Optimization (AEPO) for reliable"
X Link 2025-12-23T07:41Z [----] followers, 44.2K engagements

"@Sean60133791259 @bdsqlsz Could I ask which country you are in Well test the speed. You can DM me"
X Link 2025-12-24T02:04Z [----] followers, [--] engagements

"Merry Christmas everyone ๐ŸŽ„โœจ ๐Ÿš€Introducing BEYOND REALITY Z IMAGE 1.0: a community-tuned checkpoint merging LoRA enhancements into Z-Image optimized for film-style portrait aesthetics and high-frequency texture detail think skin pores fabric weaves and wall surfaces rendered with striking realism. โœ… Keeps Zs photorealism intact โœ… Adds cinematic color depth & tonal nuance โœ… Significantly upgrades portrait + environmental textures โœ… Runs on 8GB VRAM (FP8) tested on laptops Note: Training focused on high-quality synthetic portrait photography so complex poses or non-human textures may vary."
X Link 2025-12-24T08:00Z [----] followers, 13.5K engagements

"We shouldve shouted this louder: Youre officially missing out if you aren't using Civision ๐Ÿคซ ๐ŸŽจ FREE Image & Video Gen & FREE LoRA Training โœจ A UI so clean you dont need a PhD to master it. Weve got the heavy hitters ready: Z-Image Qwen-Image Qwen-Image-Edit Wan2.2 and more. All the power zero cost. Right here on ModelScope: ๐Ÿš€โœจ https://modelscope.ai/civision/imageGeneration https://modelscope.ai/civision/imageGeneration"
X Link 2025-12-25T08:43Z [----] followers, [----] engagements

"๐Ÿš€New on ModelScope: ChenkinNoob-XL v0.2 is live Independently developed by the ChenkinNoob team as a fine-tuned extension of noobai-XL-1.1 v0.2 incorporates Danbooru data up to Nov [--] 2025enhancing performance while preserving the original aesthetic. โœจ Key improvements over v0.1: โœ… Stronger character consistency โœ… Refined details (hands accessories layered outfits) โœ… Higher prompt adherence ๐ŸŽจ Faithful to the noob visual DNA: clean composition solid anatomy and harmonious color palettes. โš™ Recommended config: CFG: [--] Steps: [----] Sampler: Euler a Resolution: [----] (e.g. [-------] 10241024)"
X Link 2025-12-26T02:26Z [----] followers, 30.7K engagements

"๐Ÿš€ New on ModelScope: MiniMax M2.1 is open-source โœ… SOTA in 8+ languages (Rust Go Java C++ TS Kotlin Obj-C JS) โœ… Full-stack Web & mobile dev: Android/iOS 3D visuals vibe coding that actually ships โœ… Smarter faster 30% fewer tokens with lightning mode (M2.1-lightning) for high-TPS workflows โœ… Top-tier on SWE-bench VIBE and custom coding/review benchmarks โœ… Works flawlessly in Cursor Cline Droid BlackBox and more Its not just better code its AI-native development end to end. ๐Ÿ”— Model: https://modelscope.cn/models/MiniMax/MiniMax-M2.1 https://modelscope.cn/models/MiniMax/MiniMax-M2.1"
X Link 2025-12-26T08:03Z [----] followers, 17.3K engagements

"@nnnnmailcom Hi there SDXL models should be supported. Could you let me know exactly what issue you're running into"
X Link 2025-12-29T03:53Z [----] followers, [--] engagements

"๐Ÿš€400+ Free APIs are now available on We currently support: โœ… DeepSeek series โœ… Qwen3 series & Qwen-Image โœ… Z-Image-Turbo and many more Check it out: Go to - Models - Filter by API-Inference. (Currently available on PC only.) ๐Ÿ‘‰ Give them a spin and let us know what you think โœ‰DM us for support. https://modelscope.ai/modelsfilter=inference_type&page=1&tabKey=task http://modelscope.ai http://ModelScope.ai http://modelscope.ai http://ModelScope.ai https://modelscope.ai/modelsfilter=inference_type&page=1&tabKey=task http://modelscope.ai http://ModelScope.ai http://modelscope.ai"
X Link 2025-12-29T08:01Z [----] followers, 17.8K engagements

"@blankbraindead oh I see. May I know which region you are in"
X Link 2025-12-29T08:33Z [----] followers, [---] engagements

"๐Ÿš€ Tencent HunYuan open-sourced HY-MT1.5 A lightweight powerhouse for translation: โœจ 1.8B model: 0.18s latency runs offline on mobile & outperforms most commercial APIs. Even rivals Gemini-3.0-Pro at 90% quality โšก 7B model: Upgraded with WMT25-champion accuracy & fewer artifacts. ๐ŸŒ [--] languages + [--] Chinese dialects ๐Ÿ›  Context-aware custom glossaries format-preserving output ๐Ÿ’ก Perfect for edge + cloud deployments Models: https://modelscope.cn/collections/Tencent-Hunyuan/HY-MT15 https://modelscope.cn/collections/Tencent-Hunyuan/HY-MT15"
X Link 2025-12-30T10:08Z [----] followers, 17K engagements

"One of your New Year open-source gifts ๐ŸŽจ is on its way"
X Link 2025-12-30T11:45Z [----] followers, 10.7K engagements

"๐Ÿš€New on ModelScope: Qwen-Image-2512 is here ๐ŸŽ‰ This version delivers a massive leap in realism and control: โœจ Photorealistic humans dramatically reduced AI look with accurate aging natural expressions and fine details (yes even individual hair strands) ๐ŸŒฟ Ultra-detailed nature fur foliage water and mist rendered with unprecedented texture fidelity ๐Ÿ”ค Reliable text & layout generates complex infographics timelines and multilingual slides with correct structure and legible text ๐Ÿ† Evaluated in 10k+ blind tests: now the strongest open-source image modelrivaling closed alternatives. ๐Ÿ’ก Supports"
X Link 2025-12-31T09:53Z [----] followers, [----] engagements

"2026 is almost here Wishing you a very Happy New Year ๐ŸŽ‰ What open-source models or product features are at the top of your wishlist for the coming year ๐Ÿš€ Wed love to see if we can help make those wishes come true"
X Link 2025-12-31T12:28Z [----] followers, [----] engagements

"๐ŸŒŸQwen-Image-2512 is now live on Model: Try: https://modelscope.ai/civision/imageGeneration/tab=default https://modelscope.ai/models/Qwen/Qwen-Image-2512/summary http://ModelScope.ai https://modelscope.ai/civision/imageGeneration/tab=default https://modelscope.ai/models/Qwen/Qwen-Image-2512/summary http://ModelScope.ai"
X Link 2025-12-31T17:37Z [----] followers, [----] engagements

"๐Ÿš€ New on ModelScope: IQuest Research just dropped IQuest-Coder-V1 a new family of open-source code LLMs (7B to 40B) with 128K context GQA and two specialized variants: - Instruct: for coding assistance - Thinking: RL-tuned for agentic reasoning & error recovery - Plus: a novel Loop architecture (40B only) for efficient deployment via parameter sharing across [--] iterations. ๐ŸŒŸ All models support repo-level understanding via Code-Flow training learning from commit sequences and code evolution not just static snippets. ๐Ÿ” Benchmarks: - [----] on SWE-Bench Verified - [----] on LiveCodeBench v6"
X Link 2026-01-04T07:14Z [----] followers, 10.9K engagements

"Big news ๐Ÿš€ now officially supports FREE LoRA training for Qwen-Image-2512 Z-Image-Turbo and Qwen-Image-Edit-2511. Time to build ๐Ÿ›  ๐ŸŽจModelScope Civision: https://modelscope.ai/civision/modelTraining http://ModelScope.ai https://modelscope.ai/civision/modelTraining http://ModelScope.ai https://modelscope.ai/civision/modelTraining http://ModelScope.ai https://modelscope.ai/civision/modelTraining http://ModelScope.ai"
X Link 2026-01-04T12:13Z [----] followers, 21.5K engagements

"Qwen-Image-2512-Fast is officially here โšก ๐Ÿš€ Try the Demo on ModelScope now: Experience a massive 20x speedup generating high-quality images in just [--] to [--] steps Using CFG distillation weve achieved lightning-fast inference without compromising quality. https://modelscope.cn/studios/kelseye/Qwen-Image-2512-Turbo-LoRA-Demo/summary https://modelscope.cn/studios/kelseye/Qwen-Image-2512-Turbo-LoRA-Demo/summary https://modelscope.cn/studios/kelseye/Qwen-Image-2512-Turbo-LoRA-Demo/summary https://modelscope.cn/studios/kelseye/Qwen-Image-2512-Turbo-LoRA-Demo/summary"
X Link 2026-01-04T12:35Z [----] followers, 30.5K engagements

"@WolfyBlair Yes you'll get daily free quota (magicubes) and you can earn more by participating in community activities"
X Link 2026-01-05T02:26Z [----] followers, [---] engagements

"๐Ÿš€ New on ModelScope: The Wuli Team just released a Turbo LoRA that makes Qwen-Image-2512 inference 20x fasterโšก without sacrificing quality. Key updates: โœ… 4-step turbo inference โœ… Enhanced texture & detail โœ… Native ComfyUI support Model: Demo: https://modelscope.cn/studios/kelseye/Qwen-Image-2512-Turbo-LoRA-Demo https://modelscope.ai/models/Wuli-Art/Qwen-Image-2512-Turbo-LoRA https://modelscope.ai/models/Wuli-Art/Qwen-Image-2512-Turbo-LoRA Qwen-Image-2512-Fast is officially here โšก ๐Ÿš€ Try the Demo on ModelScope now: https://t.co/LRShwwbiNh Experience a massive 20x speedup generating"
X Link 2026-01-05T03:41Z [----] followers, 11.3K engagements

"๐Ÿค– Introducing InternVLA-A1 now fully open-sourced Many VLA models follow instructions well in static scenes but struggle in dynamic environments (conveyor belts rotating platforms multi-robot setups). Why They see the presentbut cant imagine the future. InternVLA-A1 solution: unify perception imagination and action in one model: โœ… Scene understanding: Image + text task parsing โœ… Task imagination: Predict future frames reason about dynamics โœ… Guided control: Execute actions steered by visual foresight Powered by InternData-A1 - Large-scale high-quality simulated dataset InternVLA-A1 stays"
X Link 2026-01-05T11:23Z [----] followers, 38.2K engagements

"๐Ÿš€ Small model BIG agency Meet Youtu-LLM a native agentic 1.96B LLM that thinks plans and acts by itself no external frameworks needed. โœ… Outperforms 4B8B models on math coding & agent benchmarks โœ… Trained on 200B+ agentic reasoning trajectories โœ… 128K context open-weight & ready to run anywhere Think lightweight cant be smart Think again. ๐Ÿค– Model: ๐Ÿ”— GitHub: ๐Ÿ“„ Paper: https://www.modelscope.cn/papers/2512.24618 https://github.com/TencentCloudADP/youtu-tip/blob/master/youtu-llm https://modelscope.cn/collections/Tencent-YouTu-Research/Youtu-LLM https://www.modelscope.cn/papers/2512.24618"
X Link 2026-01-06T07:57Z [----] followers, 11.8K engagements

"๐Ÿš€ TeleChat3-105B-A4.7B-Thinking is now open source A 105B sparse MoE model with fine-grained routing: - [---] experts only [--] activated per token (4.7B active params) - Trained end-to-end on domestic compute - Strong across code math agents writing check HumanEval-X (92.7%) & SWE-Bench (51%) It can even build a working shooter game ๐Ÿ•น or a video site frontendfully autonomously. Also released: TeleChat3-36B-Thinking (dense 36B) optimized for reasoning & roleplay. โœ… Supports vLLM SGLang LLaMA-Factory โœ… Full Thinking mode with think./think reasoning traces โœ… Tuned for long-context and multi-turn"
X Link 2026-01-06T11:20Z [----] followers, 16.8K engagements

"The #EmbodiedAI data gap just got a lot smaller. ๐Ÿš€ @AGIBOTofficial released the Genie Sim [---] datasetthe largest open-source sim dataset in the field ๐Ÿค– 10000+ hours of high-quality sim demos ๐ŸŽฏ 200+ tasks & 100k scenarios ๐Ÿ“Š Multi-sensor: RGB-D Stereo & Kinematics Standardized in LeRobot format. Ready to train โœ… Dataset: โœ… GitHub: https://github.com/AgibotTech/genie_sim https://modelscope.cn/datasets/agibot_world/GenieSim3.0-Dataset https://github.com/AgibotTech/genie_sim https://modelscope.cn/datasets/agibot_world/GenieSim3.0-Dataset"
X Link 2026-01-07T07:00Z [----] followers, [----] engagements

"AI for all ๐ŸŒ Hosted by @alibaba_cloud the Alibaba Cloud AIGC Championship @ Milano Cortina [----] is officially live Create your Winter Olympics videos with the tools and resources powered by ModelScope as an official partner. โ›ธโ„ ๐Ÿ† Win Big ๐ŸŽŸ Milan [----] Tickets: A trip to the Winter Olympics ๐Ÿ› Legacy: Works archived in the Olympic Museum. ๐Ÿ’ฐ Cash Prizes: Reward your creative storytelling. ๐ŸŽฌ Pro Tech: High-speed AI video via Wan [---]. Global Entrance Creating #ModelScope #AlibabaCloud #MilanoCortina2026 #Wan26 #Olympics https://modelscope.ai/civision/videoGeneration"
X Link 2026-01-07T08:54Z [----] followers, [----] engagements

"๐Ÿ‘ComfyUI now supports Z-Image-Turbo LoRAs trained with ModelScope Civision Huge thanks to our amazing community members for the incredibly fast integration ๐Ÿ‘‰ Everyone is welcome to train LoRAs for free on Civision and thanks to your valuable feedback training speed has been significantly improved ๐Ÿš€ https://modelscope.ai/civision/modelTraining https://github.com/Comfy-Org/ComfyUI/pull/11805 https://github.com/Comfy-Org/ComfyUI/pull/11805 https://modelscope.ai/civision/modelTraining https://github.com/Comfy-Org/ComfyUI/pull/11805 https://github.com/Comfy-Org/ComfyUI/pull/11805"
X Link 2026-01-12T03:16Z [----] followers, 11.9K engagements

"GLM-Image is here Try it out and let us know what you think: https://x.com/ModelScope2022/status/2011262011997651194 ๐Ÿš€GLM-Image is now open-source a breakthrough in "Cognitive Generation." By combining Autoregressive models with Diffusion Decoders it finally solves the struggle of generating posters PPTs and knowledge-heavy infographics. ๐ŸŒŸ Highlights: โœ… #1 on CVTG-2K & LongText-Bench https://t.co/cF3z0VFvc9 https://x.com/ModelScope2022/status/2011262011997651194 ๐Ÿš€GLM-Image is now open-source a breakthrough in "Cognitive Generation." By combining Autoregressive models with Diffusion"
X Link 2026-01-14T02:40Z [----] followers, [----] engagements

"Step-Audio-R1.1 by @StepFun_ai just set a new SOTA on the Artificial Analysis Speech Reasoning leaderboard ๐Ÿ† It outperforms Grok Gemini and GPT-Realtime with a 96.4% accuracy rate. โœ… Native Audio Reasoning (End-to-End) โœ… Audio-native CoT (Chain of Thought) โœ… Real-time streaming inference โœ… FULLY OPEN SOURCE ๐ŸŒŸ Demo: ๐Ÿค– Model: https://modelscope.cn/models/stepfun-ai/Step-Audio-R1.1 https://modelscope.cn/studios/stepfun-ai/Step-Audio-R1 https://modelscope.cn/models/stepfun-ai/Step-Audio-R1.1 https://modelscope.cn/studios/stepfun-ai/Step-Audio-R1"
X Link 2026-01-15T06:32Z [----] followers, 15.6K engagements

"@SteveWarnerFL Hi would you mind sharing more about how you use it in your work Are you trying to modify a specific layer or add/remove one Wed love to better understand your situation"
X Link 2026-01-16T06:34Z [----] followers, [--] engagements

"Step3-VL-10B Demo: https://modelscope.cn/studios/stepfun-ai/step3-vl-10b https://modelscope.cn/studios/stepfun-ai/step3-vl-10b"
X Link 2026-01-19T11:34Z [----] followers, [----] engagements

"๐Ÿš€ Introducing GLM-4.7-Flash the strongest 30B-class open model that balances performance & efficiency โœ… Top scores on AIME GPQA SWE-bench -Bench & more โœ… Optimized for agentic coding tool use and reasoning โœ… Runs locally with vLLM / SGLang / Transformers Perfect for developers who want Claude-level coding power at a fraction of the cost ๐Ÿ’ก Try it now: ๐Ÿค– ๐Ÿ”— ๐Ÿ“„ ๐Ÿ™ https://github.com/zai-org/GLM-4.5 https://z.ai/blog/glm-4.7 https://chat.z.ai https://modelscope.cn/models/ZhipuAI/GLM-4.7-Flash https://github.com/zai-org/GLM-4.5 https://z.ai/blog/glm-4.7 https://chat.z.ai"
X Link 2026-01-19T15:32Z [----] followers, 11.3K engagements

"How reliable is your model's tool usage. REALLY ๐Ÿง Inspired by @Kimi_Moonshot's K2 Vendor Verifier we are leveling up Agentic Evaluation with EvalScope. The standardizing Function Calling evaluation allows us to bridge the gap between "claims of support" and "RELIABLE execution" in the field. ๐Ÿ›  ๐Ÿ“– Docs: ๐Ÿ’ป Code: ๐Ÿ”น Standardized: Quantifies MoonshotAI K2 verification logic. ๐Ÿ”น Comprehensive: Measures decision-making + JSON schema compliance. ๐Ÿ”น Simple: Plug-and-play with your data or our benchmarks. Don't let your Agent go live without a proper check-up #LLM #AIAgents #FunctionCalling"
X Link 2026-01-22T06:32Z [----] followers, [----] engagements

"@YashasGunderia Hi you'll need to request to join the organization first and then you'll have access to the GPU resources. Here is the link: https://www.modelscope.ai/organization/xGPU-Explorers https://www.modelscope.ai/organization/xGPU-Explorers"
X Link 2026-01-24T05:43Z [----] followers, [---] engagements

"Thinking Machines Lab proved On-Policy Distillation slashes LLM training costs by 10x and we show you how to reproduce their research. Invest [--] minutes in this guideas we unpack the theory tech details experiment results and code to instantly transform your fine-tuning budget๐Ÿ“š Related Resources๐Ÿ‘‡): โœ… Slash training compute by 10X. โœ… Achieve robust RL performance with zero forgetting. โœ… Get the ready-to-use ms-SWIFT + vLLM code for deployment. Related Resources - TML Blog: - (Open-source implementation for reproducing On-Policy Distillation) - On-Policy Distillation Documentation: - Example"
X Link 2025-10-28T13:09Z [----] followers, [----] engagements

"VibeThinker-1.5B is here ๐Ÿš€ and it flips the bigger = smarter myth on its head. โœ… Just 1.5B params โœ… Trained via novel Spectrum-to-Signal Principle (SSP) โœ… Beats models 400x larger (e.g. 671B DeepSeek-R1) on hard math benchmarks (AIME24/25 HMMT25) โœ… Matches 456B MiniMax-M1 and ties Mistrals Magistral-Medium on LiveCodeBench v6 โœ… Post-training cost: $8K (vs. $290K+ for others) ๐Ÿ’ก SSP first encourages divergent exploration of solution paths then uses RL to converge on optimal strategies smarter training not bigger models. ๐ŸŒ Open-sourced to empower researchers & smaller teams whove been priced"
X Link 2025-11-13T13:00Z [----] followers, [----] engagements

"๐Ÿš€ GLM-4.6V is live @Zai_org 's new open-source multimodal models just dropped on ModelScope and we provide a FREE API for immediate accessโ— GLM-4.6V (106B) for cloud & high-performance workloads GLM-4.6V-Flash (9B) lightweight fast great for local inference ๐Ÿ”ฅ Key dev perks: โœ… Native multimodal tool calling pass images/docs directly as function args no OCR detour โœ… 128K context handles 150-page docs or hour-long videos in one go โœ… Visual Action pipeline powers real multimodal agents (e.g. find this outfit online returns structured shopping list) โœ… 50% cheaper than GLM-4.5V $1/million input"
X Link 2025-12-08T11:43Z [----] followers, [----] engagements

"Weve prepared ModelScope and Tongyi swag for everyone attending the workshopready for a day full of great takeaways ๐Ÿš€ Join Us in Seoul: AI Innovation Meets Creativity @AMD @Alibaba_Qwen @ModelScope2022 @alibaba_cloud AMDs AI Developer Meetup in Seoul (Dec 10) is filling FAST. As a key partner were bringing you the future of generative AI. ๐Ÿ“… Dec [--] ๐Ÿ“ Seoul ๐ŸŽ Free https://t.co/b5XcmEmbPf ๐Ÿš€ Join Us in Seoul: AI Innovation Meets Creativity @AMD @Alibaba_Qwen @ModelScope2022 @alibaba_cloud AMDs AI Developer Meetup in Seoul (Dec 10) is filling FAST. As a key partner were bringing you the"
X Link 2025-12-10T01:28Z [----] followers, [----] engagements

"Huge update for Z-Image-Turbo-Fun-Controlnet-Union [---] ๐Ÿš€ Key technical upgrades: โœ… New 1.9GB Lite model for low-VRAM & natural blending โœ… Fixed mask leakage in inpainting โœ… Multi-resolution dataset refactor (up to 1536px) โœ… 8-step distillation for crisp blur-free Turbo gen No more bright spot artifacts. Just pure control. ๐Ÿค– Model: https://modelscope.ai/models/PAI/Z-Image-Turbo-Fun-Controlnet-Union-2.1 https://modelscope.ai/models/PAI/Z-Image-Turbo-Fun-Controlnet-Union-2.1"
X Link 2026-01-14T08:34Z [----] followers, 20.5K engagements

"Qwen-Image-Layered-Control is now supported by ComfyUI Try it out: https://modelscope.ai/models/DiffSynth-Studio/Qwen-Image-Layered-Control/files DiffSynth-Studio has open-sourced Qwen-Image-Layered-Control ๐Ÿš€ It enables precise text-controlled layer extractionperfect for poster decomposition and graphic design workflows. Control image layers like never before: โœ… Single-layer output via text prompts โœ… Native https://t.co/4QwOKqTFLs https://modelscope.ai/models/DiffSynth-Studio/Qwen-Image-Layered-Control/files DiffSynth-Studio has open-sourced Qwen-Image-Layered-Control ๐Ÿš€ It enables precise"
X Link 2026-01-15T09:33Z [----] followers, 22.9K engagements

"Real-Qwen-Image-V2 is here ๐Ÿ“ธ Developed by wikeeyang this fine-tuned version of Qwen-Image-2512 is built for those who demand peak realism and production-grade quality. ๐Ÿค– Model: Key Highlights: โœจ Sharper details & enhanced realism โœจ Optimized for Asian facial aesthetics โœจ Superior text & image generation quality โœจ Highly compatible with LoRAs & custom workflows Tech Specs: โœ… CFG: [---] - [---] โœ… Steps: [--] - [--] โœ… Sampler: Euler / Simple (flexible) โœ… Model Shift: [---] - [---] https://www.modelscope.cn/models/wikeeyang/Real-Qwen-Image-V2 https://www.modelscope.cn/models/wikeeyang/Real-Qwen-Image-V2"
X Link 2026-01-18T05:01Z [----] followers, 28.4K engagements

"๐Ÿš€ Meet STEP3-VL-10Bit delivers SOTA-level visual perception complex reasoning and human-aligned intelligenceredefining efficiency in open multimodal AI. โœ… Beats or matches models [----] larger (like GLM-4.6V Qwen3-VL even Gemini [---] Pro) โœ… Achieves SOTA on MMMU MathVision OCRBench ScreenSpot and more โœ… Trained on 1.2T tokens + 1400+ RL rounds (RLHF + RLVR) โœ… Supports PaCoRe: parallel collaborative reasoning (128K context) Despite its compact size it leads the 10B class in: STEM reasoning (94.43% on AIME [----] w/ PaCoRe) Visual perception (92.05 on MMBench) GUI understanding & OCR Spatial"
X Link 2026-01-19T11:32Z [----] followers, 217.6K engagements

"Latency or intelligence Step-Audio-R1.1 ends the trade-off. Fei Tian Multimodal Large Model Researcher at StepFun introduces Mind-Paced Speakingenabling real-time voice AI that thinks while speaking. Fast responses. Deep reasoning. No compromise. https://x.com/i/broadcasts/1mrGmBabZavJy https://x.com/i/broadcasts/1mrGmBabZavJy"
X Link 2026-01-19T11:57Z [----] followers, [----] engagements

"Introducing AgentCPM-Report: An 8B on-device agent that rivals closed-source giants in deep research and professional report writing. The Performance: ๐Ÿ† Ranked #1 in Insight Depth across DeepResearch Bench Deep Consult and DeepResearch Gymoutperforming top-tier closed models. Highlights: โœ… 100% Private: Zero data leakage. Your sensitive data stays on your hardware. โœ… UltraRAG Stack: Efficient local indexing for private PDFs and docs. โœ… Production Ready: One-click Docker deployment. How does an 8B model hit SOTA ๐Ÿง  It breaks the "one-shot" generation bottleneck with a "Writing as Reasoning""
X Link 2026-01-20T12:18Z [----] followers, [----] engagements

"๐Ÿš€ Chroma [---] is here and its open From @flashlabsdotai : the worlds first open-source end-to-end real-time speech-to-speech dialogue model with personalized voice cloning. โœจ 150ms end-to-end latency ๐Ÿงฌ High-fidelity voice cloning from just seconds of audio ๐Ÿ“ˆ Speaker similarity (SIM) = [-----] +10.96% over human baseline ๐Ÿง  Strong reasoning in only 4B params ๐Ÿ”“ Fully open weights + code A true open alternative to closed commercial systems and already optimized with @lmsysorgs SGLang for even faster inference ๐Ÿ”— Paper: ๐Ÿค– Model: ๐Ÿ’ป Code: Were excited to see what the community builds with"
X Link 2026-01-21T16:07Z [----] followers, 14.3K engagements

"๐ŸŒพ Meet Sinong (): The first heavy-duty LLM suite for Agriculture ๐Ÿšœ ๐Ÿงฌ Expertise: Crop breeding animal medicine smart farming & ag-economics. ๐Ÿค– Models: 8B & 32B parameters ๐Ÿ“š Stack: 4B+ tokens (240k+ papers 8k+ books patents) ๐Ÿง  Tech: Native CoT reasoning + Multi-Agent RAG for zero-hallucination scientific retrieval. ๐Ÿš€ Open Source: https://modelscope.cn/models/NAULLM/Sinong1.0-32B https://modelscope.cn/models/NAULLM/Sinong1.0-8B https://modelscope.cn/models/NAULLM/Sinong1.0-32B https://modelscope.cn/models/NAULLM/Sinong1.0-8B"
X Link 2026-01-23T03:46Z [----] followers, [----] engagements

"@noctus91 Hi there This isn't actually an error. To use API inference you just need to bind your Alibaba Cloud account (no extra cost involved)"
X Link 2026-01-23T17:37Z [----] followers, [---] engagements

"โšก๐ŸŽจTomorrow. This week is going to be intriguing"
X Link 2026-01-26T02:30Z [----] followers, 51.8K engagements

"DeepSeek-OCR-2 is now on ModelScope โœจ Meet DeepSeek-OCR-2 a vision-language model that reimagines document intelligence with human-like visual understanding. Key Highlights ๐Ÿง  Visual Causal Flow: Understands layout semantics and structurenot just pixels ๐Ÿ“ Dynamic Resolution: Processes documents at (0-6)768 + [----] with adaptive token generation โœ Flexible Prompts: One model for multiple tasks: - Convert to markdown preserve tables & formatting - Parse the figure extract chart insights - Locate "text" pixel-level grounding - OCR this image 100+ language support โšก Production Ready:"
X Link 2026-01-27T06:26Z [----] followers, [----] engagements

"๐Ÿš€ Meet Kimi K2.5 ๐ŸŒ™ This is Kimis most intelligent and versatile model to date achieving SOTA performance across coding vision and agentic workflows. Model: Paper: Highlights: โœ… Native Multimodal Architecture: Seamlessly integrates vision & text. From screenshots to screen recordings K2.5 "sees" and reasons across images and videos to break the limits of text-only prompts. โœ… Agent Clusters: A world-first capability. K2.5 can autonomously spawn up to [---] "clones" to work in parallel on complex tasks slashing execution time by 4.5x. โœ… The Ultimate Office/Coding Assistant: * Office Pro: Masters"
X Link 2026-01-27T06:58Z [----] followers, [----] engagements

"The DiffSynth-Studio team also dropped Z-Image-i2L (Image-to-LoRA). ๐ŸŽจโœจ This model takes a single image as input and instantly outputs a custom LoRA tailored to that specific style or feature. Model: Demo: โœ… Zero Manual Effort: No captioning or training process required. โœ… One-Shot Magic: Get a style LoRA from just one reference image. โœ… Style Transfer: Seamlessly extract and apply unique features to new prompts. โœ… Lower Barrier: Personalized model creation is now accessible to everyone. https://modelscope.cn/studios/DiffSynth-Studio/Z-Image-i2L"
X Link 2026-01-27T16:02Z [----] followers, [----] engagements

"๐Ÿš€ Meet LingBot-VLA: A pragmatic Vision-Language-Action model designed to bridge the gap between perception and execution in robotics. ๐Ÿค– โœ…LingBot-VLA-4B: Lightweight & versatile. โœ…LingBot-VLA-4B-Depth: Enhanced for high-precision spatial tasks. Highlights: ๐Ÿง  Powerful Core: Built on the Qwen2.5-VL-3B foundation mastering multi-tasking and dual-arm coordination across 9+ robot configs. ๐Ÿ“ˆ Elite Performance: Outperforms competitors like [---] and GR00T in success rates (SR) on both GM-100 (Real-world) and RoboTwin [---] (Sim). ๐Ÿ”‹ Hyper-Efficient: 1.52.8x faster training than existing VLA codebases"
X Link 2026-01-27T18:00Z [----] followers, [----] engagements

"HunyuanImage-3.0 is here: A high-performance multimodal model for unified text-to-image generation and creative editing. ๐Ÿš€๐Ÿงต๐Ÿ‘‡ ๐Ÿค–Model: 1/ The Model Suite ๐Ÿ“ฆA versatile lineup for any workflow: โœ…Base 3.0: The core T2I powerhouse. โœ…3.0-Instruct: Advanced I2I and creative editing. โœ…Instruct-Distil: Optimized for 8-step sampling and rapid deployment. 2/ Core Capabilities ๐ŸŽจBeyond renderingits a creative partner. Features Unified Generation (T2I I2I and 3-image fusion) with built-in CoT reasoning and Prompt Enhancement to turn sparse ideas into professional descriptions. 3/ Technical Highlights"
X Link 2026-01-29T11:51Z [----] followers, [----] engagements

"Meet LingBot-VA: The future of robot learning is visual. ๐Ÿค–๐ŸŽฅ Its an autoregressive diffusion framework that predicts future video frames and decodes actions simultaneously. Instead of just reacting it reasons: "I expect the world to look like this next so I should move like that." Key Highlights: ๐Ÿง  Infinite Memory: Unlike memoryless models that get stuck in loops LingBot-VA handles "recurrent states" (like opening the same box twice) with ease by remembering the full history. ๐Ÿณ Generalist Skills: From folding clothes to making breakfast and high-precision screw pickingits a true"
X Link 2026-01-30T02:47Z [----] followers, [----] engagements

"PaddleOCR-VL-1.5 is open-sourced A 0.9B parameter multimodal model that masters the "curved & distorted" document challenge. ๐Ÿš€ ๐Ÿ”— ModelScope: ๐Ÿ”— Demo: Key Technical Breakthroughs: โœ…Polygon-based Localization: No more rigid rectangles. It natively supports irregular boxes to fit warped tilted or screen-captured text perfectly. ๐Ÿ“ โœ…Seal & Stamp Recognition: New dedicated capability to handle complex official document verification. ๐Ÿ’ฎ โœ…Cross-Page Logic: Automatically merges tables and identifies headings across physical page breakscrucial for full-document semantics. โœ…Performance Giant:"
X Link 2026-01-30T05:52Z [----] followers, [----] engagements

"Ready to make this February legendary ๐Ÿš€โœจThe Qwen-Image LoRA Training Competition is officially LIVE Weve teamed up with @Ali_TongyiLab and @Alibaba_Qwen to bring you some absolutely grand prizes:๐Ÿ“ฑ iPhone [--] Pro Max (512GB) ๐ŸŽฎ Sony PS5 Pro ๐Ÿ› $800 Shopping Cards (or equivalent prizes) Stop scrolling and start training ๐Ÿ› ๐ŸŽจJoin here: #HappyQwensday #QwenImageLoRA https://modelscope.ai/active/qwenimagelora https://modelscope.ai/active/qwenimagelora"
X Link 2026-02-01T12:35Z [----] followers, [----] engagements

"2/3 ๐Ÿ› Four specialist models were released in just one week each mastering a specific real-world challenge: โœ… DeepSeek-OCR2: Uses "Visual Causal Flow" to handle reading order in complex multi-column layouts. โœ… GLM-OCR (Z-org): An efficiency powerhouse using Multi-Token Prediction (MTP) to reduce costs by 90% while excelling at formulas. โœ… PaddleOCR-VL-1.5 (PaddlePaddle): Built for "Real5" messy conditionsperfect for curved pages and camera glare. โœ… Youtu-Parsing (Tencent): Delivers 22x faster inference and converts flowcharts directly into Mermaid code."
X Link 2026-02-03T09:18Z [----] followers, [----] engagements

"Here they come They are now at or alumni of these elite AI labs: Alibaba Tongyi Meta (FAIR/Superintelligence) Google DeepMind Tencent AI. Academia: SJTU OSU UChicago NUS WashU. https://x.com/i/broadcasts/1ZkJzZRdQNoJv https://x.com/i/broadcasts/1ZkJzZRdQNoJv"
X Link 2026-02-05T05:55Z [----] followers, [----] engagements

"@pranaysuyash http://ModelScope.ai http://ModelScope.ai"
X Link 2026-02-05T12:38Z [----] followers, [---] engagements

"1/4๐Ÿงต MS-Agent is Rebuilding A massive functional update to ModelScopes agent frameworkintroducing autonomous skill scaling and industrial-grade toolkits. โœ… Agent Skills: Skill Discovery via local/Hub repo_id matching; โœ… Multi-skill Collaboration via DAG-based planning; โœ… Progressive Analysis to optimize context window for small models. โœ… Robust Executor: Secure Docker/Sandbox execution; autonomous self-reflection and bug-fix cycles. Github: https://github.com/modelscope/ms-agent/tree/main/ms_agent/skill https://github.com/modelscope/ms-agent/tree/main/ms_agent/skill"
X Link 2026-02-06T02:34Z [----] followers, [---] engagements

"2/4 ๐Ÿ’ป Code Genesis is Now Live A SOTA end-to-end synthesis engine specializing in zero-to-one development over standard "vibe-coding." โœ… Capabilities: Industrial-grade Code Genesis for full-stack synthesis; Language Server integration for stability and clean code aesthetics. โœ… Technical Edge: Superior benchmarks in end-to-end logic flow; minimalist architecture for rapid project scaffolding. Repo: Tech Report: https://github.com/modelscope/ms-agent/tree/main/projects/code_genesis/TECH_REPORT.md https://github.com/modelscope/ms-agent/tree/main/projects/code_genesis"
X Link 2026-02-06T02:34Z [----] followers, [---] engagements

"๐ŸŽฌ Singularity Cinema is Breaking Benchmarks The GLCC-winning workshop for high-density long-form automated video production. โœ… Generation: LLM-driven storyboards; Manim/Remotion dual-engine animation; 5min+ cinematic output for Science Econ and History. โœ… Model Fusion: Seamlessly integrates Qwen-Image Sora edge-tts and nano-banana-pro; transforms static docs into dynamic media. Demo: https://github.com/modelscope/ms-agent/tree/main/projects/singularity_cinema https://github.com/modelscope/ms-agent/tree/main/projects/singularity_cinema"
X Link 2026-02-06T02:34Z [----] followers, [---] engagements

"4/4 ๐ŸŒ ms-agent WebUI is Live A self-bootstrapped interface generated entirely by the Code Genesis engine. โœ… Usage: pip install ms-agent -U - ms-agent ui; instant GUI access for agent orchestration. โœ… Roadmap: DeepResearch capability launching soonSOTA autonomous retrieval and synthesis incoming. Full Framework: https://github.com/modelscope/ms-agent https://github.com/modelscope/ms-agent"
X Link 2026-02-06T02:34Z [----] followers, [---] engagements

Limited data mode. Full metrics available with subscription: lunarcrush.com/pricing

@ModelScope2022
/creator/twitter::ModelScope2022