#  @ModelScope2022 ModelScope ModelScope posts on X about native, agentic, ai, inference the most. They currently have [-----] followers and [---] posts still getting attention that total [------] engagements in the last [--] hours. ### Engagements: [------] [#](/creator/twitter::1784494412913049600/interactions)  - [--] Week [------] -12% - [--] Month [-------] +72% - [--] Year [---------] +210,270% ### Mentions: [--] [#](/creator/twitter::1784494412913049600/posts_active)  - [--] Year [--] +9,400% ### Followers: [-----] [#](/creator/twitter::1784494412913049600/followers)  - [--] Week [-----] +4.50% - [--] Month [-----] +52% ### CreatorRank: [-------] [#](/creator/twitter::1784494412913049600/influencer_rank)  ### Social Influence **Social category influence** [technology brands](/list/technology-brands) 9% [finance](/list/finance) 2% [travel destinations](/list/travel-destinations) 2% [stocks](/list/stocks) 2% [products](/list/products) 1% [gaming](/list/gaming) 1% **Social topic influence** [native](/topic/native) #117, [agentic](/topic/agentic) #156, [ai](/topic/ai) 10%, [inference](/topic/inference) #190, [realtime](/topic/realtime) 7%, [model](/topic/model) 7%, [math](/topic/math) 6%, [flow](/topic/flow) 5%, [tencent](/topic/tencent) 4%, [science](/topic/science) 4% **Top accounts mentioned or mentioned by** [@furaidosu](/creator/undefined) [@alitongyilab](/creator/undefined) [@alibabacloud](/creator/undefined) [@alibabaqwen](/creator/undefined) [@ali_tongyilab](/creator/undefined) [@barrakali](/creator/undefined) [@dx8152](/creator/undefined) [@1328](/creator/undefined) [@ultralytics](/creator/undefined) [@sean60133791259](/creator/undefined) [@bdsqlsz](/creator/undefined) [@nnnnmailcom](/creator/undefined) [@blankbraindead](/creator/undefined) [@wolfyblair](/creator/undefined) [@agibotofficial](/creator/undefined) [@stepfunai](/creator/undefined) [@stevewarnerfl](/creator/undefined) [@kimimoonshots](/creator/undefined) [@yashasgunderia](/creator/undefined) [@zaiorg](/creator/undefined) ### Top Social Posts Top posts by engagements in the last [--] hours "๐ฟ Meet Sirchmunk: The agentic search engine that operates directly on your raw data. Turn files into evolving intelligence in real-time Build information retrieval pipelines that require ZERO vector databases ZERO pre-indexing and ZERO chunking: ๐ฏ Embedding-Free Retrieval Drop files in and search instantly. No ETL pipelines no vector silos and no setup hours. ๐ฒ Monte Carlo Evidence Sampling It doesn't "read" full docsit explores focuses and synthesizes. 500-page manual Only the most relevant snippets hit the LLM. ๐ง Self-Evolving Knowledge Every search builds reusable knowledge clusters." [X Link](https://x.com/ModelScope2022/status/2022370189157831078) 2026-02-13T18:00Z [----] followers, [----] engagements "Love seeing what the MiniMax team shipped with M2.5 ๐ Real-world productivity strong agentic workflows and impressive efficiency numbers this is serious engineering. SOTA across coding search and tool use while getting faster is a big deal. ๐ฅ Excited to share that the model will be coming to ModelScope soon for the open-source community. Stay tuned ๐โจ Introducing M2.5 an open-source frontier model designed for real-world productivity. - SOTA performance at coding (SWE-Bench Verified 80.2%) search (BrowseComp 76.3%) agentic tool-calling (BFCL 76.8%) & office work. - Optimized for efficient" [X Link](https://x.com/ModelScope2022/status/2021994964125987267) 2026-02-12T17:09Z [----] followers, [----] engagements "Nanbeige just dropped Nanbeige4.1-3B a tiny powerhouse that outclasses models 10x its size.๐ โ 256k Context Window โ "Deep Search" Agent capabilities built-in โ Two-stage Code RL (Correctness โก Efficiency) โ Beats Qwen3-32B on key benchmarks The era of high-performance low-cost local deployment is here. ๐ค: https://modelscope.cn/models/nanbeige/Nanbeige4.1-3B https://modelscope.cn/models/nanbeige/Nanbeige4.1-3B" [X Link](https://x.com/ModelScope2022/status/2022629650178941435) 2026-02-14T11:11Z [----] followers, 11.9K engagements "Introducing LingBot-World: An open-source world simulator pushing the boundaries of video generation. ๐ ๐ High-Fidelity: Realistic scientific & stylized. ๐ง Long-Term Memory: Minute-level consistency. โก Real-Time: 1s latency at [--] FPS. ๐ Apache [---] Licensed. Model: Github: https://github.com/Robbyant/lingbot-world https://modelscope.cn/models/Robbyant/lingbot-world-base-cam https://github.com/Robbyant/lingbot-world https://modelscope.cn/models/Robbyant/lingbot-world-base-cam" [X Link](https://x.com/ModelScope2022/status/2016564553816088624) 2026-01-28T17:30Z [----] followers, 29.3K engagements "Tencent YouTu Research open-sourcedYoutu-VL-4B-Instruct a compact VLM redefining visual perception viaVLUAS. ๐โก ๐Highlights: โ All-in-One Vision: SOTA performance in detection segmentation depth estimation and pose estimationwithouttask-specific heads. โ OCR & Reasoning: Exceptional at complex document parsing and multimodal math reasoning. โ GUI Agent Ready: Optimized for world-understanding and interface navigation. โ Efficiency: 4B parameters make it perfect for edge deployment and fast inference. ๐งPerformance: Outperforms many larger models inOmniDocBenchandVision-Centrictasks while" [X Link](https://x.com/ModelScope2022/status/2017191058262700328) 2026-01-30T11:00Z [----] followers, [----] engagements "Stepfun open-sourcedStep-3.5-Flash a powerhouse model specifically architected for high-speed reasoning and complex Agentic workflows. ๐ Model: Key Technical Specs: โ Sparse MoE Architecture: 196B total params but only11B activeper token. SOTA efficiency. โ MTP-3 (Multi-Token Prediction): It predicts3 tokens at once hitting a blistering350 TPSfor code-heavy tasks. โก โ Hybrid Attention (SWA + Full): A 3:1 mix that masters256K contextwindows while keeping compute costs low. โ Parallel Thinking: Massively boosted performance for multi-step reasoning and deep search. Why Devs should care: - Built" [X Link](https://x.com/ModelScope2022/status/2018215995740164168) 2026-02-02T06:52Z [----] followers, 34K engagements "Beyond-Reality-Z-Image [---] is here.๐ Highlights: โ Texture SOTA: Optimized high-frequency details for what might be the best out-of-the-box skin and fabric textures in the Z-model family. โ Cinematic Lighting: Re-engineered color balance to achieve that authentic "film look" without heavy prompting. โ Speed x Versatility: Generates at Turbo speeds (10-15 steps) but supports broader artistic styles and 2D/Anime aesthetics than the official release. โ Efficiency: Optimized for low CFG (1.0) making it highly stable and responsive to LoRA integration. Available now on ModelScope: Recommended Stack:" [X Link](https://x.com/ModelScope2022/status/2018309394719805444) 2026-02-02T13:03Z [----] followers, 16.4K engagements "How do Agents "Grow from Experience" A dialogue on Key Issues and Technical Paths of Self-Evolution AI AMA Deep Roundtable Vol.1 [----] was about Reasoning. [----] is the "Mirror Moment"Agents are shifting from static tools to digital lifeforms that learn iterate and evolve from every failure. ๐ https://twitter.com/i/web/status/2018712297829732601 https://twitter.com/i/web/status/2018712297829732601" [X Link](https://x.com/ModelScope2022/status/2018712297829732601) 2026-02-03T15:44Z [----] followers, [----] engagements "Meetthe1stAuthorsbehindthemostinfluentialpapers:earlyexperienceDREAMGYMAgentEvolverSPICER-ZeroEvo-MemoryandSEAgent. TheyarenowatoralumnioftheseeliteAIlabs: ๐นAlibabaTongyiLabMetaFAIRMetaSuperintelligenceLabsGoogleDeepMindTencentAISeattleLab ๐นSJTUOSUUChicagoNUSWashU 6leadresearchers1stage.Unrivaledtechnicaldepth.๐ง https://twitter.com/i/web/status/2018712300744847796 https://twitter.com/i/web/status/2018712300744847796" [X Link](https://x.com/ModelScope2022/status/2018712300744847796) 2026-02-03T15:44Z [----] followers, [---] engagements "MiniCPM-o [---] brings true "Omni" capabilities (Vision + Audio + Text) into a single 9B end-to-end architecture (Qwen3-8B + SigLip2 + Whisper + CosyVoice2).๐ โ Vision SOTA: OpenCompass score of 77.6surpassing GPT-4o and Gemini [---] Pro in VLM tasks with just 9B params. โ Real-time Full Duplex: It "sees" "hears" and "speaks" simultaneously. Supports active interaction (it can interrupt or initiate conversation based on live video). โ World-Class OCR: Beats DeepSeek-OCR [--] and GPT-4o in document parsing (OmniDocBench leader). โ Voice Cloning: Clone any voice with just a short reference" [X Link](https://x.com/ModelScope2022/status/2018925461649396202) 2026-02-04T05:51Z [----] followers, [----] engagements "๐ ACE-Step v1.5 is out: an open-source music foundation model that runs locally on consumer GPUs (4GB VRAM) and generates full songs in 2s (A100) or 10s (RTX 3090). โ Beats most commercial models in quality โ Train a personalized LoRA from just a few tracks โ Built on a novel LM+DiT hybrid architecture with internal RLno external reward bias โ Supports 50+ languages + editing (covers repainting vocalBGM) ๐ค https://modelscope.cn/models/ACE-Step/Ace-Step1.5 https://modelscope.cn/models/ACE-Step/Ace-Step1.5" [X Link](https://x.com/ModelScope2022/status/2018977724048777344) 2026-02-04T09:19Z [----] followers, [----] engagements "๐ Meet Intern-S1-Pro: A massive 1T parameter MoE model for Multimodal Science Reasoning โ [---] Experts (22B active) โ SOTA in AI4Science (Chemistry Materials Bio) โ FoPE + Time-series modeling (up to [--] points) โ Native "Thinking Mode" support Open-source science just leveled up. ๐งช๐ป Model: Github: https://github.com/InternLM/Intern-S1 https://www.modelscope.cn/models/Shanghai_AI_Laboratory/Intern-S1-Pro https://github.com/InternLM/Intern-S1 https://www.modelscope.cn/models/Shanghai_AI_Laboratory/Intern-S1-Pro" [X Link](https://x.com/ModelScope2022/status/2019068430041919653) 2026-02-04T15:20Z [----] followers, [----] engagements "Introducing Singularity Cinema (MS-Agent framework)๐an LLM-powered tool that transforms documents or prompts into storyboards generating stunning 5-minute+ videos via its Manim/Remotion dual-engine. By fusing nano-banana-pro๐ Qwen-Image๐จ Sora and edge-tts it delivers high-quality information-dense contentperfect for science economics and history creators looking to scale their social media presence. ๐: ๐ฆPrompt: Convert xx/deploy.md into a short video in a blue-themed style making sure to use the important images from the document. The short video must be in English. ๐ฌOutput Video๐" [X Link](https://x.com/ModelScope2022/status/2019443081461002520) 2026-02-05T16:08Z [----] followers, [----] engagements "From a local designer in Chengdu to a global AI sensation with 117k+ downloadsDaniels journey shows whats possible when open-source meets imagination. Now its your turn. Join the Qwen-Image LoRA Competition and start building your own legacy. The next top creator could be you. https://modelscope.ai/active/qwenimagelora LoRA Training Contest: Win iPhone/PS5 and other prizes Gaussian splash LoRA creation process shared. Competition information: https://t.co/MEEZHQcr83 Training address: https://t.co/uxwtMlJzoP @Ali_TongyiLab @ModelScope2022 #HappyQwensday #QwenImageLora https://t.co/5igMVZUO5i" [X Link](https://x.com/ModelScope2022/status/2019612106086314106) 2026-02-06T03:20Z [----] followers, [----] engagements "ChenkinNoob-XL v0.2 is here A high-fidelity SDXL-to-Rectified Flow conversion delivering vivid color range and superior lighting adherence. โ Performance: Eliminates "base model greyness"; Superior contrast/dark scene rendering; High stability across wide CFG ranges (3-6); Optimal convergence at 20-28 steps. โ Ecosystem: Native ComfyUI workflow included; ReForge-ready with built-in Flow support; LoRA-compatible via Pochi.toml (Refresh scripts). โ Architecture: Rectified Flow (RF) linear ODE trajectory; Trained on 47M samples using 8xH20 nodes; Leverages Cosine Optimal Transport and Protected" [X Link](https://x.com/ModelScope2022/status/2019637177471173090) 2026-02-06T05:00Z [----] followers, [----] engagements "StepFun open-sourced Step-3.5-Flashfeaturing powerful reasoning and Agent intelligence. Speeds up to [---] TPS Competes with closed-source models in Agent and Math tasks Excels at complex long-chain reasoning. Dont miss out https://x.com/i/broadcasts/1OyKAjZgZPaGb https://x.com/i/broadcasts/1OyKAjZgZPaGb" [X Link](https://x.com/ModelScope2022/status/2019744317792018474) 2026-02-06T12:05Z [----] followers, [----] engagements "Meet X-Coder ๐IIGroup drops a powerhouse suite for code LLMs โ X-Coder-SFT-376k: Massive synthetic dataset via GPT-o3-mini/DeepSeek-R1. โ X-Coder-RL-40k: Targeted RL alignment for logic & verification. โ X-Coder-8B Models: SOTA reasoning built on Qwen3. Experience the power of fully synthetic code evolution. ๐ป๐ฅ ๐ Models & Dataset: #LLM #OpenCode #Qwen3 #DeepSeek #Coding https://modelscope.cn/models/IIGroup/X-Coder-SFT-Qwen3-8B https://modelscope.cn/datasets/IIGroup/X-Coder-RL-40k https://modelscope.cn/datasets/IIGroup/X-Coder-SFT-376k" [X Link](https://x.com/ModelScope2022/status/2020745735059443936) 2026-02-09T06:25Z [----] followers, [----] engagements "YOLO26 is here๐ The most advanced and deployment-ready vision model from @ultralytics redefining real-time edge computing performance. โ Performance: 43% CPU inference speed boost vs YOLO11; Real-time performance on low-power hardware; [--] scalable sizes (n/s/m/l/x) for optimized speed-accuracy trade-offs. โ Ecosystem: Unified framework for Detection Segmentation Pose OBB and Tracking; Simplified export for seamless hardware integration. โ Logic/Architecture: Removed DFL for enhanced hardware compatibility; End-to-end NMS-free inference; ProgLoss + STAL for small target precision; MuSGD" [X Link](https://x.com/ModelScope2022/status/2020777189575467356) 2026-02-09T08:30Z [----] followers, [----] engagements "2/4 Key Technical Highlights: - Architecture: Native single-stream design integrating generation editing and segmentation. - Audio: First model to unify Speech Audio and Music generation via 12.5Hz continuous representation. - Video: Time-Interleaved VideoRoPE for enhanced temporal modeling. https://twitter.com/i/web/status/2021580867009671592 https://twitter.com/i/web/status/2021580867009671592" [X Link](https://x.com/ModelScope2022/status/2021580867009671592) 2026-02-11T13:43Z [----] followers, [---] engagements "๐ GLM-5 is now open-sourceand ranked #4 globally #1 among open models on Artificial Analysis Beyond "vibe coding" it delivers Agentic Engineering: autonomously architects debugs and ships production systems. Coding performance rivals Claude Opus [---] (SWE-bench: [----] ๐ SOTA) while dominating long-horizon agent tasks (BrowseComp MCP-Atlas ๐). 744B params MIT License ๐ ModelScope: https://modelscope.cn/models/ZhipuAI/GLM-5 https://modelscope.cn/models/ZhipuAI/GLM-5" [X Link](https://x.com/ModelScope2022/status/2021792323227820181) 2026-02-12T03:43Z [----] followers, [----] engagements "๐ฟ Meet Sirchmunk: The agentic search engine that operates directly on your raw data. Turn files into evolving intelligence in real-time Build information retrieval pipelines that require ZERO vector databases ZERO pre-indexing and ZERO chunking: ๐ฏ Embedding-Free Retrieval Drop files in and search instantly. No ETL pipelines no vector silos and no setup hours. ๐ฒ Monte Carlo Evidence Sampling It doesn't "read" full docsit explores focuses and synthesizes. 500-page manual Only the most relevant snippets hit the LLM. ๐ง Self-Evolving Knowledge Every search builds reusable knowledge clusters." [X Link](https://x.com/ModelScope2022/status/2022304582207238586) 2026-02-13T13:39Z [----] followers, [---] engagements "LongCat-Flash-LiteisNow Live Meituan's 68.5B MoE model pioneering "Embedding Expansion" for leading Agent and Coding performance. โ Performance: 2.9B4.5B activated params; 500-700 tokens/s throughput; 256K context via YARN; SOTA SWE-Bench (54.4%) and -Bench scores. โ Ecosystem: Optimized SGLang-FluentLLM engine with custom CUDA kernels; Open API with 50M free daily tokens. โ Architecture: N-gram embedding layer for localized semantic precision; Hash-collision mitigation via sub-table decomposition; Triple-step Speculative Decoding synergy. ๐ค ๐ง ๐ Tech" [X Link](https://x.com/ModelScope2022/status/2020383958819832285) 2026-02-08T06:27Z [----] followers, [----] engagements "๐1/4 Introducing Ming-flash-omni-2.0: An open-source Omni-MLLM breakthrough. Based on the Ling-2.0 MoE architecture (100B total / 6B active) it unifies vision speech and image editing into a single model matching specialist performance across all modalities. ๐งต๐ ๐ค ModelScope: ๐ฆ GitHub: https://github.com/inclusionAI/Ming https://www.modelscope.cn/models/inclusionAI/Ming-flash-omni-2.0 https://github.com/inclusionAI/Ming https://www.modelscope.cn/models/inclusionAI/Ming-flash-omni-2.0" [X Link](https://x.com/ModelScope2022/status/2021580863277084888) 2026-02-11T13:43Z [----] followers, [----] engagements "๐Introducing Z-Image the full non-distilled version of Z-Image for high-quality customizable image generation. โ Model: โ Demo: โจ Capabilities: Photorealistic + diverse art styles (anime illustration digital art) Full CFG & precise negative prompts for reliable prompt following Multi-subject generation with distinct identitiesno "clone faces" Community-friendly: ready for LoRA ControlNet and other fine-tuning workflows ๐ฅ Highlights: Full architecture (vs. distilled Turbo): better diversity & training stability ideal for customization High output variance across seeds = less homogenization" [X Link](https://x.com/ModelScope2022/status/2016178781522890948) 2026-01-27T15:57Z [----] followers, 59.5K engagements "Z-Image-Fun-Lora-Distill is here A high-efficiency Distill LoRA enabling CFG-free high-fidelity generation for Z-Image in just 4-8 steps. โ 4-step and 8-step variants; CFG-free inference (CFG=1.0); [----] update delivers superior skin texture and color consistency compared to previous iterations. โ Native ComfyUI weights available; full integration with the VideoX-Fun framework; compatible with Z-Image ControlNet Union [---] and third-party LoRAs. โ Dual-distillation of both sampling steps and CFG; trained from scratch without Z-Image-Turbo weights; optimized for rapid generation and derivative" [X Link](https://x.com/ModelScope2022/status/2021412342723207600) 2026-02-11T02:33Z [----] followers, [----] engagements "๐ Ring-2.5-1T is open-source The first hybrid linear 1T thinking model. ๐ Gold Tier reasoning IMO25 (35/42) & CMO25 (105/126) scores. ๐ Agent-native compatible with Claude Code OpenClaw and SGLang. โกHybrid Linear breakthrough 10x lower memory & 3x throughput for long context. ๐ง 1T MoE structure with 63B active parameters (1:7 MLA + Lightning Linear). ๐ป Long-Horizon mastery autonomously built a 32-bit OS from scratch in [--] hours. Model: GitHub: https://github.com/inclusionAI/Ring-V2.5 https://modelscope.cn/models/inclusionAI/Ring-2.5-1T https://github.com/inclusionAI/Ring-V2.5" [X Link](https://x.com/ModelScope2022/status/2022223579421839523) 2026-02-13T08:17Z [----] followers, 13K engagements "Introducing FireRed-Image-Edit-1.0 from FireRedTeam ๐ Its officially the new SOTA for general image editing. โ Better than Closed-Source: Outperforms Nano-Banana & Seedream4.0 on GEdit benchmarks. โ Native Evolution: Built from T2I foundations not just a "patch" on existing models. โ Style Mastery: Scored a record-breaking 4.97/5.0 in style transfer. โ High-Fidelity Text: Keeps original font styles perfectly. โ Virtual Try-on: Native support for multi-image joint editing. โ Bilingual: Native support for both English & Chinese prompts. Apache [---] license. Local deployment ready. ๐ค Model: ๐ " [X Link](https://x.com/ModelScope2022/status/2022687312442704320) 2026-02-14T15:00Z [----] followers, 69.4K engagements "How 0.9B Models Are Outperforming the Giants: A Comparison of Four New Chinese VLM-OCR Models 1/3 ๐ Early [----] has brought a massive surge of open-source models from DeepSeek Z-org PaddlePaddle and Tencent YouTu. A tiny 0.9B parameter model GLM-OCR topped the global OmniDocBench v1.5 leaderboard outscoring giants like Gemini-3 Pro and Qwen3-VL. We are moving away from rigid "detect and recognize" pipelines toward Vision-Language Models (VLMs) that actually "read" and understand document structures like humans do. https://twitter.com/i/web/status/2018615132134551565" [X Link](https://x.com/ModelScope2022/status/2018615132134551565) 2026-02-03T09:18Z [----] followers, 30.1K engagements "Z-Image-Fun-Controlnet-Union & Lora-Distill are now open-source ๐Expanding the Z-Image ecosystem with ultra-fast inference and high-precision spatial control. โ Performance: Achieve 8-step inference (CFG: 1.0) via Distill LoRA and professional-grade multi-condition control (Canny Depth Pose Inpaint) across 15+ layer blocks. โ Ecosystem: Full cross-compatibility between Distill LoRA and Union ControlNet; includes "Lite" versions for low-VRAM edge devices and seamless natural blending. โ Unified Control: Support for 7+ spatial conditions in a single model with adjustable control_context_scale" [X Link](https://x.com/ModelScope2022/status/2019339552650092830) 2026-02-05T09:17Z [----] followers, 11K engagements "Qwen3-ASR & Qwen3-ForcedAligner are now open-source ๐๐ High-performance streaming-ready speech AI for [--] languages and dialects. ๐: Demo: ๐คThe Models โ ASR-1.7B: Open-source SOTA. Outperforms Whisper-v3; rivals GPT-4o & Gemini. โ ASR-0.6B: Insane efficiency92ms latency; transcribes 2000s of audio in 1s. โ ForcedAligner-0.6B: Superior timestamp accuracy across [--] languages. ๐Key Features โ Universal: Supports [--] languages + [--] Chinese dialects. โ Robust: Expertly handles singing heavy BGM and low SNR. โ Unified: Single model for both streaming & offline (up to 20min audio). โกPerformance &" [X Link](https://x.com/ModelScope2022/status/2016857817597301174) 2026-01-29T12:55Z [----] followers, [----] engagements "๐ FantasyWorld is now open source Recently accepted to ICLR [----] and ranking #1 on the WorldScore leaderboard this framework offers a unified approach to video and 3D generation. โจKey Technical Highlights: โ Direct 3D Inference: Input an image and camera trajectory to get frame-consistent video dense depth maps and point clouds instantly. โ Dual-Branch IRG Architecture: An "Imagination" branch handles visual appearance while a "Geometry" branch ensures spatial accuracy. โ Wan [---] Integration: The released 14B models leverage the latest Wan-DiT architectures for superior temporal consistency and" [X Link](https://x.com/ModelScope2022/status/2022264617871757514) 2026-02-13T11:00Z [----] followers, [----] engagements "Say hello to Ling-2.5-1T: 1T params 63B active MIT licensed โก Hybrid Linear Attention: 1:7 MLA + Lightning Linear. Beats Kimi K2 on long-context throughput. ๐ง 4x token efficiency: Composite rewards match frontier thinking models using 4x fewer tokens. ๐ 1M context: YaRN to 1M. Beats Kimi K2.5 & DeepSeek V3.2 on RULER/MRCR. Perfect NIAH scores. ๐ SOTA tool use: Agentic RL trained. Leads BFCL-V4. Native support for Claude Code OpenCode OpenClaw. ๐ฏ Follows instructions: Bidirectional RL + agent verification. High-density zero fluff. Honest: Gap vs GPT-5.2/Gemini [--] Pro on long-horizon tasks." [X Link](https://x.com/ModelScope2022/status/2023240865649447320) 2026-02-16T03:39Z [----] followers, [----] engagements "New Z-Image-Turbo control model released ๐ More control blocks. Inpaint mode. Better details. - 6-block ControlNet - Trained from scratch on 1M images @1328 res - Canny / HED / Depth / Pose / MLSD - Best control_context_scale: 0.65-0.80 Model: Demo: Code: https://github.com/aigc-apps/VideoX-Fun https://modelscope.cn/studios/AI-ModelScope/ZIT-Controlnet https://modelscope.cn/models/PAI/Z-Image-Turbo-Fun-Controlnet-Union https://modelscope.cn/models/PAI/Z-Image-Turbo-Fun-Controlnet-Union https://github.com/aigc-apps/VideoX-Fun https://modelscope.cn/studios/AI-ModelScope/ZIT-Controlnet" [X Link](https://x.com/ModelScope2022/status/2000766764922401091) 2025-12-16T03:15Z [----] followers, 33.1K engagements "๐Model & Demo are now live on our international site Model: - - - Demo: https://modelscope.ai/studios/FunAudioLLM/Fun-CosyVoice3-0.5B https://modelscope.ai/models/FunAudioLLM/Fun-CosyVoice3-0.5B-2512 https://modelscope.ai/models/FunAudioLLM/Fun-ASR-Nano-2512 https://modelscope.ai/models/FunAudioLLM/Fun-ASR-MLT-Nano-2512 https://modelscope.ai/studios/FunAudioLLM/Fun-CosyVoice3-0.5B https://modelscope.ai/models/FunAudioLLM/Fun-CosyVoice3-0.5B-2512 https://modelscope.ai/models/FunAudioLLM/Fun-ASR-Nano-2512 https://modelscope.ai/models/FunAudioLLM/Fun-ASR-MLT-Nano-2512" [X Link](https://x.com/ModelScope2022/status/2000864318758511044) 2025-12-16T09:43Z [----] followers, [---] engagements "๐คMeet Nanbeige4-3B from Boss Zhipina 3B-parameter LLM that outperforms Qwen3-32B on math (AIME) science (GPQA) and tool calling (BFCL-V4) while matching Qwen3-30B-A3B on human preference alignment (Arena-Hard-V2). How โ 23T tokens of ultra-curated data โ Fine-grained WSD scheduler โ 30M+ high-quality SFT instructions โ Multi-stage RL + innovative distillation (DPD) โ Chain-of-thought reconstruction & deliberative generation It even ranks top [--] on WritingBench & EQ-Bench3beating models 100x larger like GLM-4.5 and Deepseek-R1. All models + tech report now open-source: ๐ Weights: ๐ Paper:" [X Link](https://x.com/ModelScope2022/status/2000890664486363362) 2025-12-16T11:28Z [----] followers, 25.8K engagements "๐ MiMo-V2-Flash FREE API is now live on ModelScope The first major release since Fuli Luo joined Xiaomiand its built for real-world agentic AI. โก MiMo-V2-Flash: an open high-performance MoE model with 309B total / 15B active parameters 256K context window 150+ tokens/s generation thanks to native Multi-Token Prediction ๐ฅ Key wins for developers: โ Hybrid attention (5:1 SWA + Global) [--] less KV cache full long-context recall โ 73.4% on SWE-Bench Verified new SOTA for open-source models โ Matches DeepSeek-V3.2 on reasoning but much faster in practice โจ API-readyperfect for building smart" [X Link](https://x.com/ModelScope2022/status/2001479829095620906) 2025-12-18T02:29Z [----] followers, 16K engagements "๐Qwen-Image-Layered is now live on ModelScopean incredible model that can intelligently decompose any image into multiple RGBA layers ๐ค Model: ๐ Try Demo ๐ Paper: https://modelscope.cn/papers/2512.15603 https://modelscope.cn/studios/Qwen/Qwen-Image-Layered https://modelscope.ai/models/Qwen/Qwen-Image-Layered https://modelscope.cn/papers/2512.15603 https://modelscope.cn/studios/Qwen/Qwen-Image-Layered https://modelscope.ai/models/Qwen/Qwen-Image-Layered ๐จ Qwen-Image-Layered is LIVE native image decomposition fully open-sourced โจ Why it stands out โ Photoshop-grade layering Physically" [X Link](https://x.com/ModelScope2022/status/2002038699878695146) 2025-12-19T15:29Z [----] followers, [----] engagements "Its still Sunday and Im already hyped for next week. ๐The mysterious Eastern power is dropping a trilogy of magic: ๐ฎ Next week: A secret image-generation model ๐ค Next week: A secret code-agent model ๐ The week after: A secret voice model Buckle upthings are about to get interesting" [X Link](https://x.com/ModelScope2022/status/2002679068203028809) 2025-12-21T09:54Z [----] followers, 10.1K engagements "๐ Meet GLM-4.7 your new coding partner supercharged Try it now on ModelScope free API access for a limited time โ +5.8% on SWE-bench (73.8%) โ +12.9% on Multilingual coding (66.7%) โ +16.5% on Terminal Bench [---] (41%) โ Smarter agent reasoning & tool use โ Better UI cleaner webpages sharper slides โ +12.4% on complex reasoning (HLE: 42.8%)From coding to creativity GLM-4.7 delivers across the board. ๐ค Model: https://modelscope.cn/models/ZhipuAI/GLM-4.7 https://modelscope.cn/models/ZhipuAI/GLM-4.7" [X Link](https://x.com/ModelScope2022/status/2003290346135187635) 2025-12-23T02:23Z [----] followers, [----] engagements "๐ New on ModelScope: QwenLong-L1.5 is now fully open-source A 30B model (3B active params) that matches GPT-5 & Gemini-2.5-Pro in long-context reasoning. ๐ฅ Key wins: โ +31.7 pts on OpenAIs MRCR (128K context SOTA across all models) โ Matches Gemini-2.5-Pro on [--] major long-QA benchmarks โ +9.69 on CorpusQA +6.16 on LongBench-V2 How Three breakthroughs: [--] Synthetic data at scale: 14.1K long-reasoning samples from 9.2B tokens no human labeling. Avg. length: 34K tokens (max: 119K). [--] Stable RL training: Task-balanced sampling + Adaptive Entropy-Controlled Policy Optimization (AEPO) for reliable" [X Link](https://x.com/ModelScope2022/status/2003370363590226313) 2025-12-23T07:41Z [----] followers, 44.2K engagements "@Sean60133791259 @bdsqlsz Could I ask which country you are in Well test the speed. You can DM me" [X Link](https://x.com/ModelScope2022/status/2003648027375075799) 2025-12-24T02:04Z [----] followers, [--] engagements "Merry Christmas everyone ๐โจ ๐Introducing BEYOND REALITY Z IMAGE 1.0: a community-tuned checkpoint merging LoRA enhancements into Z-Image optimized for film-style portrait aesthetics and high-frequency texture detail think skin pores fabric weaves and wall surfaces rendered with striking realism. โ Keeps Zs photorealism intact โ Adds cinematic color depth & tonal nuance โ Significantly upgrades portrait + environmental textures โ Runs on 8GB VRAM (FP8) tested on laptops Note: Training focused on high-quality synthetic portrait photography so complex poses or non-human textures may vary." [X Link](https://x.com/ModelScope2022/status/2003737630245343345) 2025-12-24T08:00Z [----] followers, 13.5K engagements "We shouldve shouted this louder: Youre officially missing out if you aren't using Civision ๐คซ ๐จ FREE Image & Video Gen & FREE LoRA Training โจ A UI so clean you dont need a PhD to master it. Weve got the heavy hitters ready: Z-Image Qwen-Image Qwen-Image-Edit Wan2.2 and more. All the power zero cost. Right here on ModelScope: ๐โจ https://modelscope.ai/civision/imageGeneration https://modelscope.ai/civision/imageGeneration" [X Link](https://x.com/ModelScope2022/status/2004110641687417306) 2025-12-25T08:43Z [----] followers, [----] engagements "๐New on ModelScope: ChenkinNoob-XL v0.2 is live Independently developed by the ChenkinNoob team as a fine-tuned extension of noobai-XL-1.1 v0.2 incorporates Danbooru data up to Nov [--] 2025enhancing performance while preserving the original aesthetic. โจ Key improvements over v0.1: โ Stronger character consistency โ Refined details (hands accessories layered outfits) โ Higher prompt adherence ๐จ Faithful to the noob visual DNA: clean composition solid anatomy and harmonious color palettes. โ Recommended config: CFG: [--] Steps: [----] Sampler: Euler a Resolution: [----] (e.g. [-------] 10241024)" [X Link](https://x.com/ModelScope2022/status/2004378356645814448) 2025-12-26T02:26Z [----] followers, 30.7K engagements "๐ New on ModelScope: MiniMax M2.1 is open-source โ SOTA in 8+ languages (Rust Go Java C++ TS Kotlin Obj-C JS) โ Full-stack Web & mobile dev: Android/iOS 3D visuals vibe coding that actually ships โ Smarter faster 30% fewer tokens with lightning mode (M2.1-lightning) for high-TPS workflows โ Top-tier on SWE-bench VIBE and custom coding/review benchmarks โ Works flawlessly in Cursor Cline Droid BlackBox and more Its not just better code its AI-native development end to end. ๐ Model: https://modelscope.cn/models/MiniMax/MiniMax-M2.1 https://modelscope.cn/models/MiniMax/MiniMax-M2.1" [X Link](https://x.com/ModelScope2022/status/2004462984698253701) 2025-12-26T08:03Z [----] followers, 17.3K engagements "@nnnnmailcom Hi there SDXL models should be supported. Could you let me know exactly what issue you're running into" [X Link](https://x.com/ModelScope2022/status/2005487313158279461) 2025-12-29T03:53Z [----] followers, [--] engagements "๐400+ Free APIs are now available on We currently support: โ DeepSeek series โ Qwen3 series & Qwen-Image โ Z-Image-Turbo and many more Check it out: Go to - Models - Filter by API-Inference. (Currently available on PC only.) ๐ Give them a spin and let us know what you think โDM us for support. https://modelscope.ai/modelsfilter=inference_type&page=1&tabKey=task http://modelscope.ai http://ModelScope.ai http://modelscope.ai http://ModelScope.ai https://modelscope.ai/modelsfilter=inference_type&page=1&tabKey=task http://modelscope.ai http://ModelScope.ai http://modelscope.ai" [X Link](https://x.com/ModelScope2022/status/2005549668886040939) 2025-12-29T08:01Z [----] followers, 17.8K engagements "@blankbraindead oh I see. May I know which region you are in" [X Link](https://x.com/ModelScope2022/status/2005557772797215085) 2025-12-29T08:33Z [----] followers, [---] engagements "๐ Tencent HunYuan open-sourced HY-MT1.5 A lightweight powerhouse for translation: โจ 1.8B model: 0.18s latency runs offline on mobile & outperforms most commercial APIs. Even rivals Gemini-3.0-Pro at 90% quality โก 7B model: Upgraded with WMT25-champion accuracy & fewer artifacts. ๐ [--] languages + [--] Chinese dialects ๐ Context-aware custom glossaries format-preserving output ๐ก Perfect for edge + cloud deployments Models: https://modelscope.cn/collections/Tencent-Hunyuan/HY-MT15 https://modelscope.cn/collections/Tencent-Hunyuan/HY-MT15" [X Link](https://x.com/ModelScope2022/status/2005944042853667257) 2025-12-30T10:08Z [----] followers, 17K engagements "One of your New Year open-source gifts ๐จ is on its way" [X Link](https://x.com/ModelScope2022/status/2005968451538759734) 2025-12-30T11:45Z [----] followers, 10.7K engagements "๐New on ModelScope: Qwen-Image-2512 is here ๐ This version delivers a massive leap in realism and control: โจ Photorealistic humans dramatically reduced AI look with accurate aging natural expressions and fine details (yes even individual hair strands) ๐ฟ Ultra-detailed nature fur foliage water and mist rendered with unprecedented texture fidelity ๐ค Reliable text & layout generates complex infographics timelines and multilingual slides with correct structure and legible text ๐ Evaluated in 10k+ blind tests: now the strongest open-source image modelrivaling closed alternatives. ๐ก Supports" [X Link](https://x.com/ModelScope2022/status/2006302672073507194) 2025-12-31T09:53Z [----] followers, [----] engagements "2026 is almost here Wishing you a very Happy New Year ๐ What open-source models or product features are at the top of your wishlist for the coming year ๐ Wed love to see if we can help make those wishes come true" [X Link](https://x.com/ModelScope2022/status/2006341719198163109) 2025-12-31T12:28Z [----] followers, [----] engagements "๐Qwen-Image-2512 is now live on Model: Try: https://modelscope.ai/civision/imageGeneration/tab=default https://modelscope.ai/models/Qwen/Qwen-Image-2512/summary http://ModelScope.ai https://modelscope.ai/civision/imageGeneration/tab=default https://modelscope.ai/models/Qwen/Qwen-Image-2512/summary http://ModelScope.ai" [X Link](https://x.com/ModelScope2022/status/2006419353856782394) 2025-12-31T17:37Z [----] followers, [----] engagements "๐ New on ModelScope: IQuest Research just dropped IQuest-Coder-V1 a new family of open-source code LLMs (7B to 40B) with 128K context GQA and two specialized variants: - Instruct: for coding assistance - Thinking: RL-tuned for agentic reasoning & error recovery - Plus: a novel Loop architecture (40B only) for efficient deployment via parameter sharing across [--] iterations. ๐ All models support repo-level understanding via Code-Flow training learning from commit sequences and code evolution not just static snippets. ๐ Benchmarks: - [----] on SWE-Bench Verified - [----] on LiveCodeBench v6" [X Link](https://x.com/ModelScope2022/status/2007712347373351418) 2026-01-04T07:14Z [----] followers, 10.9K engagements "Big news ๐ now officially supports FREE LoRA training for Qwen-Image-2512 Z-Image-Turbo and Qwen-Image-Edit-2511. Time to build ๐ ๐จModelScope Civision: https://modelscope.ai/civision/modelTraining http://ModelScope.ai https://modelscope.ai/civision/modelTraining http://ModelScope.ai https://modelscope.ai/civision/modelTraining http://ModelScope.ai https://modelscope.ai/civision/modelTraining http://ModelScope.ai" [X Link](https://x.com/ModelScope2022/status/2007787503802294735) 2026-01-04T12:13Z [----] followers, 21.5K engagements "Qwen-Image-2512-Fast is officially here โก ๐ Try the Demo on ModelScope now: Experience a massive 20x speedup generating high-quality images in just [--] to [--] steps Using CFG distillation weve achieved lightning-fast inference without compromising quality. https://modelscope.cn/studios/kelseye/Qwen-Image-2512-Turbo-LoRA-Demo/summary https://modelscope.cn/studios/kelseye/Qwen-Image-2512-Turbo-LoRA-Demo/summary https://modelscope.cn/studios/kelseye/Qwen-Image-2512-Turbo-LoRA-Demo/summary https://modelscope.cn/studios/kelseye/Qwen-Image-2512-Turbo-LoRA-Demo/summary" [X Link](https://x.com/ModelScope2022/status/2007792938248708207) 2026-01-04T12:35Z [----] followers, 30.5K engagements "@WolfyBlair Yes you'll get daily free quota (magicubes) and you can earn more by participating in community activities" [X Link](https://x.com/ModelScope2022/status/2008002207099293874) 2026-01-05T02:26Z [----] followers, [---] engagements "๐ New on ModelScope: The Wuli Team just released a Turbo LoRA that makes Qwen-Image-2512 inference 20x fasterโก without sacrificing quality. Key updates: โ 4-step turbo inference โ Enhanced texture & detail โ Native ComfyUI support Model: Demo: https://modelscope.cn/studios/kelseye/Qwen-Image-2512-Turbo-LoRA-Demo https://modelscope.ai/models/Wuli-Art/Qwen-Image-2512-Turbo-LoRA https://modelscope.ai/models/Wuli-Art/Qwen-Image-2512-Turbo-LoRA Qwen-Image-2512-Fast is officially here โก ๐ Try the Demo on ModelScope now: https://t.co/LRShwwbiNh Experience a massive 20x speedup generating" [X Link](https://x.com/ModelScope2022/status/2008020913309298809) 2026-01-05T03:41Z [----] followers, 11.3K engagements "๐ค Introducing InternVLA-A1 now fully open-sourced Many VLA models follow instructions well in static scenes but struggle in dynamic environments (conveyor belts rotating platforms multi-robot setups). Why They see the presentbut cant imagine the future. InternVLA-A1 solution: unify perception imagination and action in one model: โ Scene understanding: Image + text task parsing โ Task imagination: Predict future frames reason about dynamics โ Guided control: Execute actions steered by visual foresight Powered by InternData-A1 - Large-scale high-quality simulated dataset InternVLA-A1 stays" [X Link](https://x.com/ModelScope2022/status/2008137224575992238) 2026-01-05T11:23Z [----] followers, 38.2K engagements "๐ Small model BIG agency Meet Youtu-LLM a native agentic 1.96B LLM that thinks plans and acts by itself no external frameworks needed. โ Outperforms 4B8B models on math coding & agent benchmarks โ Trained on 200B+ agentic reasoning trajectories โ 128K context open-weight & ready to run anywhere Think lightweight cant be smart Think again. ๐ค Model: ๐ GitHub: ๐ Paper: https://www.modelscope.cn/papers/2512.24618 https://github.com/TencentCloudADP/youtu-tip/blob/master/youtu-llm https://modelscope.cn/collections/Tencent-YouTu-Research/Youtu-LLM https://www.modelscope.cn/papers/2512.24618" [X Link](https://x.com/ModelScope2022/status/2008447725600469060) 2026-01-06T07:57Z [----] followers, 11.8K engagements "๐ TeleChat3-105B-A4.7B-Thinking is now open source A 105B sparse MoE model with fine-grained routing: - [---] experts only [--] activated per token (4.7B active params) - Trained end-to-end on domestic compute - Strong across code math agents writing check HumanEval-X (92.7%) & SWE-Bench (51%) It can even build a working shooter game ๐น or a video site frontendfully autonomously. Also released: TeleChat3-36B-Thinking (dense 36B) optimized for reasoning & roleplay. โ Supports vLLM SGLang LLaMA-Factory โ Full Thinking mode with think./think reasoning traces โ Tuned for long-context and multi-turn" [X Link](https://x.com/ModelScope2022/status/2008499004158411187) 2026-01-06T11:20Z [----] followers, 16.8K engagements "The #EmbodiedAI data gap just got a lot smaller. ๐ @AGIBOTofficial released the Genie Sim [---] datasetthe largest open-source sim dataset in the field ๐ค 10000+ hours of high-quality sim demos ๐ฏ 200+ tasks & 100k scenarios ๐ Multi-sensor: RGB-D Stereo & Kinematics Standardized in LeRobot format. Ready to train โ Dataset: โ GitHub: https://github.com/AgibotTech/genie_sim https://modelscope.cn/datasets/agibot_world/GenieSim3.0-Dataset https://github.com/AgibotTech/genie_sim https://modelscope.cn/datasets/agibot_world/GenieSim3.0-Dataset" [X Link](https://x.com/ModelScope2022/status/2008795887666147377) 2026-01-07T07:00Z [----] followers, [----] engagements "AI for all ๐ Hosted by @alibaba_cloud the Alibaba Cloud AIGC Championship @ Milano Cortina [----] is officially live Create your Winter Olympics videos with the tools and resources powered by ModelScope as an official partner. โธโ ๐ Win Big ๐ Milan [----] Tickets: A trip to the Winter Olympics ๐ Legacy: Works archived in the Olympic Museum. ๐ฐ Cash Prizes: Reward your creative storytelling. ๐ฌ Pro Tech: High-speed AI video via Wan [---]. Global Entrance Creating #ModelScope #AlibabaCloud #MilanoCortina2026 #Wan26 #Olympics https://modelscope.ai/civision/videoGeneration" [X Link](https://x.com/ModelScope2022/status/2008824590517321874) 2026-01-07T08:54Z [----] followers, [----] engagements "๐ComfyUI now supports Z-Image-Turbo LoRAs trained with ModelScope Civision Huge thanks to our amazing community members for the incredibly fast integration ๐ Everyone is welcome to train LoRAs for free on Civision and thanks to your valuable feedback training speed has been significantly improved ๐ https://modelscope.ai/civision/modelTraining https://github.com/Comfy-Org/ComfyUI/pull/11805 https://github.com/Comfy-Org/ComfyUI/pull/11805 https://modelscope.ai/civision/modelTraining https://github.com/Comfy-Org/ComfyUI/pull/11805 https://github.com/Comfy-Org/ComfyUI/pull/11805" [X Link](https://x.com/ModelScope2022/status/2010551381233709197) 2026-01-12T03:16Z [----] followers, 11.9K engagements "GLM-Image is here Try it out and let us know what you think: https://x.com/ModelScope2022/status/2011262011997651194 ๐GLM-Image is now open-source a breakthrough in "Cognitive Generation." By combining Autoregressive models with Diffusion Decoders it finally solves the struggle of generating posters PPTs and knowledge-heavy infographics. ๐ Highlights: โ #1 on CVTG-2K & LongText-Bench https://t.co/cF3z0VFvc9 https://x.com/ModelScope2022/status/2011262011997651194 ๐GLM-Image is now open-source a breakthrough in "Cognitive Generation." By combining Autoregressive models with Diffusion" [X Link](https://x.com/ModelScope2022/status/2011267263043944540) 2026-01-14T02:40Z [----] followers, [----] engagements "Step-Audio-R1.1 by @StepFun_ai just set a new SOTA on the Artificial Analysis Speech Reasoning leaderboard ๐ It outperforms Grok Gemini and GPT-Realtime with a 96.4% accuracy rate. โ Native Audio Reasoning (End-to-End) โ Audio-native CoT (Chain of Thought) โ Real-time streaming inference โ FULLY OPEN SOURCE ๐ Demo: ๐ค Model: https://modelscope.cn/models/stepfun-ai/Step-Audio-R1.1 https://modelscope.cn/studios/stepfun-ai/Step-Audio-R1 https://modelscope.cn/models/stepfun-ai/Step-Audio-R1.1 https://modelscope.cn/studios/stepfun-ai/Step-Audio-R1" [X Link](https://x.com/ModelScope2022/status/2011687986338136089) 2026-01-15T06:32Z [----] followers, 15.6K engagements "@SteveWarnerFL Hi would you mind sharing more about how you use it in your work Are you trying to modify a specific layer or add/remove one Wed love to better understand your situation" [X Link](https://x.com/ModelScope2022/status/2012050686171889776) 2026-01-16T06:34Z [----] followers, [--] engagements "Step3-VL-10B Demo: https://modelscope.cn/studios/stepfun-ai/step3-vl-10b https://modelscope.cn/studios/stepfun-ai/step3-vl-10b" [X Link](https://x.com/ModelScope2022/status/2013213460721475715) 2026-01-19T11:34Z [----] followers, [----] engagements "๐ Introducing GLM-4.7-Flash the strongest 30B-class open model that balances performance & efficiency โ Top scores on AIME GPQA SWE-bench -Bench & more โ Optimized for agentic coding tool use and reasoning โ Runs locally with vLLM / SGLang / Transformers Perfect for developers who want Claude-level coding power at a fraction of the cost ๐ก Try it now: ๐ค ๐ ๐ ๐ https://github.com/zai-org/GLM-4.5 https://z.ai/blog/glm-4.7 https://chat.z.ai https://modelscope.cn/models/ZhipuAI/GLM-4.7-Flash https://github.com/zai-org/GLM-4.5 https://z.ai/blog/glm-4.7 https://chat.z.ai" [X Link](https://x.com/ModelScope2022/status/2013273444679590292) 2026-01-19T15:32Z [----] followers, 11.3K engagements "How reliable is your model's tool usage. REALLY ๐ง Inspired by @Kimi_Moonshot's K2 Vendor Verifier we are leveling up Agentic Evaluation with EvalScope. The standardizing Function Calling evaluation allows us to bridge the gap between "claims of support" and "RELIABLE execution" in the field. ๐ ๐ Docs: ๐ป Code: ๐น Standardized: Quantifies MoonshotAI K2 verification logic. ๐น Comprehensive: Measures decision-making + JSON schema compliance. ๐น Simple: Plug-and-play with your data or our benchmarks. Don't let your Agent go live without a proper check-up #LLM #AIAgents #FunctionCalling" [X Link](https://x.com/ModelScope2022/status/2014224614252253541) 2026-01-22T06:32Z [----] followers, [----] engagements "@YashasGunderia Hi you'll need to request to join the organization first and then you'll have access to the GPU resources. Here is the link: https://www.modelscope.ai/organization/xGPU-Explorers https://www.modelscope.ai/organization/xGPU-Explorers" [X Link](https://x.com/ModelScope2022/status/2014937070008205667) 2026-01-24T05:43Z [----] followers, [---] engagements "Thinking Machines Lab proved On-Policy Distillation slashes LLM training costs by 10x and we show you how to reproduce their research. Invest [--] minutes in this guideas we unpack the theory tech details experiment results and code to instantly transform your fine-tuning budget๐ Related Resources๐): โ Slash training compute by 10X. โ Achieve robust RL performance with zero forgetting. โ Get the ready-to-use ms-SWIFT + vLLM code for deployment. Related Resources - TML Blog: - (Open-source implementation for reproducing On-Policy Distillation) - On-Policy Distillation Documentation: - Example" [X Link](https://x.com/ModelScope2022/status/1983159127934488946) 2025-10-28T13:09Z [----] followers, [----] engagements "VibeThinker-1.5B is here ๐ and it flips the bigger = smarter myth on its head. โ Just 1.5B params โ Trained via novel Spectrum-to-Signal Principle (SSP) โ Beats models 400x larger (e.g. 671B DeepSeek-R1) on hard math benchmarks (AIME24/25 HMMT25) โ Matches 456B MiniMax-M1 and ties Mistrals Magistral-Medium on LiveCodeBench v6 โ Post-training cost: $8K (vs. $290K+ for others) ๐ก SSP first encourages divergent exploration of solution paths then uses RL to converge on optimal strategies smarter training not bigger models. ๐ Open-sourced to empower researchers & smaller teams whove been priced" [X Link](https://x.com/ModelScope2022/status/1988955029064278216) 2025-11-13T13:00Z [----] followers, [----] engagements "๐ GLM-4.6V is live @Zai_org 's new open-source multimodal models just dropped on ModelScope and we provide a FREE API for immediate accessโ GLM-4.6V (106B) for cloud & high-performance workloads GLM-4.6V-Flash (9B) lightweight fast great for local inference ๐ฅ Key dev perks: โ Native multimodal tool calling pass images/docs directly as function args no OCR detour โ 128K context handles 150-page docs or hour-long videos in one go โ Visual Action pipeline powers real multimodal agents (e.g. find this outfit online returns structured shopping list) โ 50% cheaper than GLM-4.5V $1/million input" [X Link](https://x.com/ModelScope2022/status/1997995482459865290) 2025-12-08T11:43Z [----] followers, [----] engagements "Weve prepared ModelScope and Tongyi swag for everyone attending the workshopready for a day full of great takeaways ๐ Join Us in Seoul: AI Innovation Meets Creativity @AMD @Alibaba_Qwen @ModelScope2022 @alibaba_cloud AMDs AI Developer Meetup in Seoul (Dec 10) is filling FAST. As a key partner were bringing you the future of generative AI. ๐ Dec [--] ๐ Seoul ๐ Free https://t.co/b5XcmEmbPf ๐ Join Us in Seoul: AI Innovation Meets Creativity @AMD @Alibaba_Qwen @ModelScope2022 @alibaba_cloud AMDs AI Developer Meetup in Seoul (Dec 10) is filling FAST. As a key partner were bringing you the" [X Link](https://x.com/ModelScope2022/status/1998565458912620942) 2025-12-10T01:28Z [----] followers, [----] engagements "Huge update for Z-Image-Turbo-Fun-Controlnet-Union [---] ๐ Key technical upgrades: โ New 1.9GB Lite model for low-VRAM & natural blending โ Fixed mask leakage in inpainting โ Multi-resolution dataset refactor (up to 1536px) โ 8-step distillation for crisp blur-free Turbo gen No more bright spot artifacts. Just pure control. ๐ค Model: https://modelscope.ai/models/PAI/Z-Image-Turbo-Fun-Controlnet-Union-2.1 https://modelscope.ai/models/PAI/Z-Image-Turbo-Fun-Controlnet-Union-2.1" [X Link](https://x.com/ModelScope2022/status/2011356116119334957) 2026-01-14T08:34Z [----] followers, 20.5K engagements "Qwen-Image-Layered-Control is now supported by ComfyUI Try it out: https://modelscope.ai/models/DiffSynth-Studio/Qwen-Image-Layered-Control/files DiffSynth-Studio has open-sourced Qwen-Image-Layered-Control ๐ It enables precise text-controlled layer extractionperfect for poster decomposition and graphic design workflows. Control image layers like never before: โ Single-layer output via text prompts โ Native https://t.co/4QwOKqTFLs https://modelscope.ai/models/DiffSynth-Studio/Qwen-Image-Layered-Control/files DiffSynth-Studio has open-sourced Qwen-Image-Layered-Control ๐ It enables precise" [X Link](https://x.com/ModelScope2022/status/2011733445324783764) 2026-01-15T09:33Z [----] followers, 22.9K engagements "Real-Qwen-Image-V2 is here ๐ธ Developed by wikeeyang this fine-tuned version of Qwen-Image-2512 is built for those who demand peak realism and production-grade quality. ๐ค Model: Key Highlights: โจ Sharper details & enhanced realism โจ Optimized for Asian facial aesthetics โจ Superior text & image generation quality โจ Highly compatible with LoRAs & custom workflows Tech Specs: โ CFG: [---] - [---] โ Steps: [--] - [--] โ Sampler: Euler / Simple (flexible) โ Model Shift: [---] - [---] https://www.modelscope.cn/models/wikeeyang/Real-Qwen-Image-V2 https://www.modelscope.cn/models/wikeeyang/Real-Qwen-Image-V2" [X Link](https://x.com/ModelScope2022/status/2012752195939168522) 2026-01-18T05:01Z [----] followers, 28.4K engagements "๐ Meet STEP3-VL-10Bit delivers SOTA-level visual perception complex reasoning and human-aligned intelligenceredefining efficiency in open multimodal AI. โ Beats or matches models [----] larger (like GLM-4.6V Qwen3-VL even Gemini [---] Pro) โ Achieves SOTA on MMMU MathVision OCRBench ScreenSpot and more โ Trained on 1.2T tokens + 1400+ RL rounds (RLHF + RLVR) โ Supports PaCoRe: parallel collaborative reasoning (128K context) Despite its compact size it leads the 10B class in: STEM reasoning (94.43% on AIME [----] w/ PaCoRe) Visual perception (92.05 on MMBench) GUI understanding & OCR Spatial" [X Link](https://x.com/ModelScope2022/status/2013213019690639431) 2026-01-19T11:32Z [----] followers, 217.6K engagements "Latency or intelligence Step-Audio-R1.1 ends the trade-off. Fei Tian Multimodal Large Model Researcher at StepFun introduces Mind-Paced Speakingenabling real-time voice AI that thinks while speaking. Fast responses. Deep reasoning. No compromise. https://x.com/i/broadcasts/1mrGmBabZavJy https://x.com/i/broadcasts/1mrGmBabZavJy" [X Link](https://x.com/ModelScope2022/status/2013219373948039628) 2026-01-19T11:57Z [----] followers, [----] engagements "Introducing AgentCPM-Report: An 8B on-device agent that rivals closed-source giants in deep research and professional report writing. The Performance: ๐ Ranked #1 in Insight Depth across DeepResearch Bench Deep Consult and DeepResearch Gymoutperforming top-tier closed models. Highlights: โ 100% Private: Zero data leakage. Your sensitive data stays on your hardware. โ UltraRAG Stack: Efficient local indexing for private PDFs and docs. โ Production Ready: One-click Docker deployment. How does an 8B model hit SOTA ๐ง It breaks the "one-shot" generation bottleneck with a "Writing as Reasoning"" [X Link](https://x.com/ModelScope2022/status/2013586921046344058) 2026-01-20T12:18Z [----] followers, [----] engagements "๐ Chroma [---] is here and its open From @flashlabsdotai : the worlds first open-source end-to-end real-time speech-to-speech dialogue model with personalized voice cloning. โจ 150ms end-to-end latency ๐งฌ High-fidelity voice cloning from just seconds of audio ๐ Speaker similarity (SIM) = [-----] +10.96% over human baseline ๐ง Strong reasoning in only 4B params ๐ Fully open weights + code A true open alternative to closed commercial systems and already optimized with @lmsysorgs SGLang for even faster inference ๐ Paper: ๐ค Model: ๐ป Code: Were excited to see what the community builds with" [X Link](https://x.com/ModelScope2022/status/2014006971855466640) 2026-01-21T16:07Z [----] followers, 14.3K engagements "๐พ Meet Sinong (): The first heavy-duty LLM suite for Agriculture ๐ ๐งฌ Expertise: Crop breeding animal medicine smart farming & ag-economics. ๐ค Models: 8B & 32B parameters ๐ Stack: 4B+ tokens (240k+ papers 8k+ books patents) ๐ง Tech: Native CoT reasoning + Multi-Agent RAG for zero-hallucination scientific retrieval. ๐ Open Source: https://modelscope.cn/models/NAULLM/Sinong1.0-32B https://modelscope.cn/models/NAULLM/Sinong1.0-8B https://modelscope.cn/models/NAULLM/Sinong1.0-32B https://modelscope.cn/models/NAULLM/Sinong1.0-8B" [X Link](https://x.com/ModelScope2022/status/2014545364347715871) 2026-01-23T03:46Z [----] followers, [----] engagements "@noctus91 Hi there This isn't actually an error. To use API inference you just need to bind your Alibaba Cloud account (no extra cost involved)" [X Link](https://x.com/ModelScope2022/status/2014754313999155689) 2026-01-23T17:37Z [----] followers, [---] engagements "โก๐จTomorrow. This week is going to be intriguing" [X Link](https://x.com/ModelScope2022/status/2015613317088522594) 2026-01-26T02:30Z [----] followers, 51.8K engagements "DeepSeek-OCR-2 is now on ModelScope โจ Meet DeepSeek-OCR-2 a vision-language model that reimagines document intelligence with human-like visual understanding. Key Highlights ๐ง Visual Causal Flow: Understands layout semantics and structurenot just pixels ๐ Dynamic Resolution: Processes documents at (0-6)768 + [----] with adaptive token generation โ Flexible Prompts: One model for multiple tasks: - Convert to markdown preserve tables & formatting - Parse the figure extract chart insights - Locate "text" pixel-level grounding - OCR this image 100+ language support โก Production Ready:" [X Link](https://x.com/ModelScope2022/status/2016035065298616539) 2026-01-27T06:26Z [----] followers, [----] engagements "๐ Meet Kimi K2.5 ๐ This is Kimis most intelligent and versatile model to date achieving SOTA performance across coding vision and agentic workflows. Model: Paper: Highlights: โ Native Multimodal Architecture: Seamlessly integrates vision & text. From screenshots to screen recordings K2.5 "sees" and reasons across images and videos to break the limits of text-only prompts. โ Agent Clusters: A world-first capability. K2.5 can autonomously spawn up to [---] "clones" to work in parallel on complex tasks slashing execution time by 4.5x. โ The Ultimate Office/Coding Assistant: * Office Pro: Masters" [X Link](https://x.com/ModelScope2022/status/2016043017610957071) 2026-01-27T06:58Z [----] followers, [----] engagements "The DiffSynth-Studio team also dropped Z-Image-i2L (Image-to-LoRA). ๐จโจ This model takes a single image as input and instantly outputs a custom LoRA tailored to that specific style or feature. Model: Demo: โ Zero Manual Effort: No captioning or training process required. โ One-Shot Magic: Get a style LoRA from just one reference image. โ Style Transfer: Seamlessly extract and apply unique features to new prompts. โ Lower Barrier: Personalized model creation is now accessible to everyone. https://modelscope.cn/studios/DiffSynth-Studio/Z-Image-i2L" [X Link](https://x.com/ModelScope2022/status/2016179961779126461) 2026-01-27T16:02Z [----] followers, [----] engagements "๐ Meet LingBot-VLA: A pragmatic Vision-Language-Action model designed to bridge the gap between perception and execution in robotics. ๐ค โ LingBot-VLA-4B: Lightweight & versatile. โ LingBot-VLA-4B-Depth: Enhanced for high-precision spatial tasks. Highlights: ๐ง Powerful Core: Built on the Qwen2.5-VL-3B foundation mastering multi-tasking and dual-arm coordination across 9+ robot configs. ๐ Elite Performance: Outperforms competitors like [---] and GR00T in success rates (SR) on both GM-100 (Real-world) and RoboTwin [---] (Sim). ๐ Hyper-Efficient: 1.52.8x faster training than existing VLA codebases" [X Link](https://x.com/ModelScope2022/status/2016209598471196691) 2026-01-27T18:00Z [----] followers, [----] engagements "HunyuanImage-3.0 is here: A high-performance multimodal model for unified text-to-image generation and creative editing. ๐๐งต๐ ๐คModel: 1/ The Model Suite ๐ฆA versatile lineup for any workflow: โ Base 3.0: The core T2I powerhouse. โ 3.0-Instruct: Advanced I2I and creative editing. โ Instruct-Distil: Optimized for 8-step sampling and rapid deployment. 2/ Core Capabilities ๐จBeyond renderingits a creative partner. Features Unified Generation (T2I I2I and 3-image fusion) with built-in CoT reasoning and Prompt Enhancement to turn sparse ideas into professional descriptions. 3/ Technical Highlights" [X Link](https://x.com/ModelScope2022/status/2016841562274066648) 2026-01-29T11:51Z [----] followers, [----] engagements "Meet LingBot-VA: The future of robot learning is visual. ๐ค๐ฅ Its an autoregressive diffusion framework that predicts future video frames and decodes actions simultaneously. Instead of just reacting it reasons: "I expect the world to look like this next so I should move like that." Key Highlights: ๐ง Infinite Memory: Unlike memoryless models that get stuck in loops LingBot-VA handles "recurrent states" (like opening the same box twice) with ease by remembering the full history. ๐ณ Generalist Skills: From folding clothes to making breakfast and high-precision screw pickingits a true" [X Link](https://x.com/ModelScope2022/status/2017067214919389569) 2026-01-30T02:47Z [----] followers, [----] engagements "PaddleOCR-VL-1.5 is open-sourced A 0.9B parameter multimodal model that masters the "curved & distorted" document challenge. ๐ ๐ ModelScope: ๐ Demo: Key Technical Breakthroughs: โ Polygon-based Localization: No more rigid rectangles. It natively supports irregular boxes to fit warped tilted or screen-captured text perfectly. ๐ โ Seal & Stamp Recognition: New dedicated capability to handle complex official document verification. ๐ฎ โ Cross-Page Logic: Automatically merges tables and identifies headings across physical page breakscrucial for full-document semantics. โ Performance Giant:" [X Link](https://x.com/ModelScope2022/status/2017113730866434213) 2026-01-30T05:52Z [----] followers, [----] engagements "Ready to make this February legendary ๐โจThe Qwen-Image LoRA Training Competition is officially LIVE Weve teamed up with @Ali_TongyiLab and @Alibaba_Qwen to bring you some absolutely grand prizes:๐ฑ iPhone [--] Pro Max (512GB) ๐ฎ Sony PS5 Pro ๐ $800 Shopping Cards (or equivalent prizes) Stop scrolling and start training ๐ ๐จJoin here: #HappyQwensday #QwenImageLoRA https://modelscope.ai/active/qwenimagelora https://modelscope.ai/active/qwenimagelora" [X Link](https://x.com/ModelScope2022/status/2017939891846512907) 2026-02-01T12:35Z [----] followers, [----] engagements "2/3 ๐ Four specialist models were released in just one week each mastering a specific real-world challenge: โ DeepSeek-OCR2: Uses "Visual Causal Flow" to handle reading order in complex multi-column layouts. โ GLM-OCR (Z-org): An efficiency powerhouse using Multi-Token Prediction (MTP) to reduce costs by 90% while excelling at formulas. โ PaddleOCR-VL-1.5 (PaddlePaddle): Built for "Real5" messy conditionsperfect for curved pages and camera glare. โ Youtu-Parsing (Tencent): Delivers 22x faster inference and converts flowcharts directly into Mermaid code." [X Link](https://x.com/ModelScope2022/status/2018615134479057183) 2026-02-03T09:18Z [----] followers, [----] engagements "Here they come They are now at or alumni of these elite AI labs: Alibaba Tongyi Meta (FAIR/Superintelligence) Google DeepMind Tencent AI. Academia: SJTU OSU UChicago NUS WashU. https://x.com/i/broadcasts/1ZkJzZRdQNoJv https://x.com/i/broadcasts/1ZkJzZRdQNoJv" [X Link](https://x.com/ModelScope2022/status/2019288799780770192) 2026-02-05T05:55Z [----] followers, [----] engagements "@pranaysuyash http://ModelScope.ai http://ModelScope.ai" [X Link](https://x.com/ModelScope2022/status/2019390286355656859) 2026-02-05T12:38Z [----] followers, [---] engagements "1/4๐งต MS-Agent is Rebuilding A massive functional update to ModelScopes agent frameworkintroducing autonomous skill scaling and industrial-grade toolkits. โ Agent Skills: Skill Discovery via local/Hub repo_id matching; โ Multi-skill Collaboration via DAG-based planning; โ Progressive Analysis to optimize context window for small models. โ Robust Executor: Secure Docker/Sandbox execution; autonomous self-reflection and bug-fix cycles. Github: https://github.com/modelscope/ms-agent/tree/main/ms_agent/skill https://github.com/modelscope/ms-agent/tree/main/ms_agent/skill" [X Link](https://x.com/ModelScope2022/status/2019600619443744930) 2026-02-06T02:34Z [----] followers, [---] engagements "2/4 ๐ป Code Genesis is Now Live A SOTA end-to-end synthesis engine specializing in zero-to-one development over standard "vibe-coding." โ Capabilities: Industrial-grade Code Genesis for full-stack synthesis; Language Server integration for stability and clean code aesthetics. โ Technical Edge: Superior benchmarks in end-to-end logic flow; minimalist architecture for rapid project scaffolding. Repo: Tech Report: https://github.com/modelscope/ms-agent/tree/main/projects/code_genesis/TECH_REPORT.md https://github.com/modelscope/ms-agent/tree/main/projects/code_genesis" [X Link](https://x.com/ModelScope2022/status/2019600621608022159) 2026-02-06T02:34Z [----] followers, [---] engagements "๐ฌ Singularity Cinema is Breaking Benchmarks The GLCC-winning workshop for high-density long-form automated video production. โ Generation: LLM-driven storyboards; Manim/Remotion dual-engine animation; 5min+ cinematic output for Science Econ and History. โ Model Fusion: Seamlessly integrates Qwen-Image Sora edge-tts and nano-banana-pro; transforms static docs into dynamic media. Demo: https://github.com/modelscope/ms-agent/tree/main/projects/singularity_cinema https://github.com/modelscope/ms-agent/tree/main/projects/singularity_cinema" [X Link](https://x.com/ModelScope2022/status/2019600623738765722) 2026-02-06T02:34Z [----] followers, [---] engagements "4/4 ๐ ms-agent WebUI is Live A self-bootstrapped interface generated entirely by the Code Genesis engine. โ Usage: pip install ms-agent -U - ms-agent ui; instant GUI access for agent orchestration. โ Roadmap: DeepResearch capability launching soonSOTA autonomous retrieval and synthesis incoming. Full Framework: https://github.com/modelscope/ms-agent https://github.com/modelscope/ms-agent" [X Link](https://x.com/ModelScope2022/status/2019600625835864392) 2026-02-06T02:34Z [----] followers, [---] engagements Limited data mode. Full metrics available with subscription: lunarcrush.com/pricing
@ModelScope2022 ModelScopeModelScope posts on X about native, agentic, ai, inference the most. They currently have [-----] followers and [---] posts still getting attention that total [------] engagements in the last [--] hours.
Social category influence technology brands 9% finance 2% travel destinations 2% stocks 2% products 1% gaming 1%
Social topic influence native #117, agentic #156, ai 10%, inference #190, realtime 7%, model 7%, math 6%, flow 5%, tencent 4%, science 4%
Top accounts mentioned or mentioned by @furaidosu @alitongyilab @alibabacloud @alibabaqwen @ali_tongyilab @barrakali @dx8152 @1328 @ultralytics @sean60133791259 @bdsqlsz @nnnnmailcom @blankbraindead @wolfyblair @agibotofficial @stepfunai @stevewarnerfl @kimimoonshots @yashasgunderia @zaiorg
Top posts by engagements in the last [--] hours
"๐ฟ Meet Sirchmunk: The agentic search engine that operates directly on your raw data. Turn files into evolving intelligence in real-time Build information retrieval pipelines that require ZERO vector databases ZERO pre-indexing and ZERO chunking: ๐ฏ Embedding-Free Retrieval Drop files in and search instantly. No ETL pipelines no vector silos and no setup hours. ๐ฒ Monte Carlo Evidence Sampling It doesn't "read" full docsit explores focuses and synthesizes. 500-page manual Only the most relevant snippets hit the LLM. ๐ง Self-Evolving Knowledge Every search builds reusable knowledge clusters."
X Link 2026-02-13T18:00Z [----] followers, [----] engagements
"Love seeing what the MiniMax team shipped with M2.5 ๐ Real-world productivity strong agentic workflows and impressive efficiency numbers this is serious engineering. SOTA across coding search and tool use while getting faster is a big deal. ๐ฅ Excited to share that the model will be coming to ModelScope soon for the open-source community. Stay tuned ๐โจ Introducing M2.5 an open-source frontier model designed for real-world productivity. - SOTA performance at coding (SWE-Bench Verified 80.2%) search (BrowseComp 76.3%) agentic tool-calling (BFCL 76.8%) & office work. - Optimized for efficient"
X Link 2026-02-12T17:09Z [----] followers, [----] engagements
"Nanbeige just dropped Nanbeige4.1-3B a tiny powerhouse that outclasses models 10x its size.๐ โ
256k Context Window โ
"Deep Search" Agent capabilities built-in โ
Two-stage Code RL (Correctness โก Efficiency) โ
Beats Qwen3-32B on key benchmarks The era of high-performance low-cost local deployment is here. ๐ค: https://modelscope.cn/models/nanbeige/Nanbeige4.1-3B https://modelscope.cn/models/nanbeige/Nanbeige4.1-3B"
X Link 2026-02-14T11:11Z [----] followers, 11.9K engagements
"Introducing LingBot-World: An open-source world simulator pushing the boundaries of video generation. ๐ ๐ High-Fidelity: Realistic scientific & stylized. ๐ง Long-Term Memory: Minute-level consistency. โก Real-Time: 1s latency at [--] FPS. ๐ Apache [---] Licensed. Model: Github: https://github.com/Robbyant/lingbot-world https://modelscope.cn/models/Robbyant/lingbot-world-base-cam https://github.com/Robbyant/lingbot-world https://modelscope.cn/models/Robbyant/lingbot-world-base-cam"
X Link 2026-01-28T17:30Z [----] followers, 29.3K engagements
"Tencent YouTu Research open-sourcedYoutu-VL-4B-Instruct a compact VLM redefining visual perception viaVLUAS. ๐โก ๐Highlights: โ
All-in-One Vision: SOTA performance in detection segmentation depth estimation and pose estimationwithouttask-specific heads. โ
OCR & Reasoning: Exceptional at complex document parsing and multimodal math reasoning. โ
GUI Agent Ready: Optimized for world-understanding and interface navigation. โ
Efficiency: 4B parameters make it perfect for edge deployment and fast inference. ๐งPerformance: Outperforms many larger models inOmniDocBenchandVision-Centrictasks while"
X Link 2026-01-30T11:00Z [----] followers, [----] engagements
"Stepfun open-sourcedStep-3.5-Flash a powerhouse model specifically architected for high-speed reasoning and complex Agentic workflows. ๐ Model: Key Technical Specs: โ
Sparse MoE Architecture: 196B total params but only11B activeper token. SOTA efficiency. โ
MTP-3 (Multi-Token Prediction): It predicts3 tokens at once hitting a blistering350 TPSfor code-heavy tasks. โก โ
Hybrid Attention (SWA + Full): A 3:1 mix that masters256K contextwindows while keeping compute costs low. โ
Parallel Thinking: Massively boosted performance for multi-step reasoning and deep search. Why Devs should care: - Built"
X Link 2026-02-02T06:52Z [----] followers, 34K engagements
"Beyond-Reality-Z-Image [---] is here.๐ Highlights: โ
Texture SOTA: Optimized high-frequency details for what might be the best out-of-the-box skin and fabric textures in the Z-model family. โ
Cinematic Lighting: Re-engineered color balance to achieve that authentic "film look" without heavy prompting. โ
Speed x Versatility: Generates at Turbo speeds (10-15 steps) but supports broader artistic styles and 2D/Anime aesthetics than the official release. โ
Efficiency: Optimized for low CFG (1.0) making it highly stable and responsive to LoRA integration. Available now on ModelScope: Recommended Stack:"
X Link 2026-02-02T13:03Z [----] followers, 16.4K engagements
"How do Agents "Grow from Experience" A dialogue on Key Issues and Technical Paths of Self-Evolution AI AMA Deep Roundtable Vol.1 [----] was about Reasoning. [----] is the "Mirror Moment"Agents are shifting from static tools to digital lifeforms that learn iterate and evolve from every failure. ๐ https://twitter.com/i/web/status/2018712297829732601 https://twitter.com/i/web/status/2018712297829732601"
X Link 2026-02-03T15:44Z [----] followers, [----] engagements
"Meetthe1stAuthorsbehindthemostinfluentialpapers:earlyexperienceDREAMGYMAgentEvolverSPICER-ZeroEvo-MemoryandSEAgent. TheyarenowatoralumnioftheseeliteAIlabs: ๐นAlibabaTongyiLabMetaFAIRMetaSuperintelligenceLabsGoogleDeepMindTencentAISeattleLab ๐นSJTUOSUUChicagoNUSWashU 6leadresearchers1stage.Unrivaledtechnicaldepth.๐ง https://twitter.com/i/web/status/2018712300744847796 https://twitter.com/i/web/status/2018712300744847796"
X Link 2026-02-03T15:44Z [----] followers, [---] engagements
"MiniCPM-o [---] brings true "Omni" capabilities (Vision + Audio + Text) into a single 9B end-to-end architecture (Qwen3-8B + SigLip2 + Whisper + CosyVoice2).๐ โ
Vision SOTA: OpenCompass score of 77.6surpassing GPT-4o and Gemini [---] Pro in VLM tasks with just 9B params. โ
Real-time Full Duplex: It "sees" "hears" and "speaks" simultaneously. Supports active interaction (it can interrupt or initiate conversation based on live video). โ
World-Class OCR: Beats DeepSeek-OCR [--] and GPT-4o in document parsing (OmniDocBench leader). โ
Voice Cloning: Clone any voice with just a short reference"
X Link 2026-02-04T05:51Z [----] followers, [----] engagements
"๐ ACE-Step v1.5 is out: an open-source music foundation model that runs locally on consumer GPUs (4GB VRAM) and generates full songs in 2s (A100) or 10s (RTX 3090). โ
Beats most commercial models in quality โ
Train a personalized LoRA from just a few tracks โ
Built on a novel LM+DiT hybrid architecture with internal RLno external reward bias โ
Supports 50+ languages + editing (covers repainting vocalBGM) ๐ค https://modelscope.cn/models/ACE-Step/Ace-Step1.5 https://modelscope.cn/models/ACE-Step/Ace-Step1.5"
X Link 2026-02-04T09:19Z [----] followers, [----] engagements
"๐ Meet Intern-S1-Pro: A massive 1T parameter MoE model for Multimodal Science Reasoning โ
[---] Experts (22B active) โ
SOTA in AI4Science (Chemistry Materials Bio) โ
FoPE + Time-series modeling (up to [--] points) โ
Native "Thinking Mode" support Open-source science just leveled up. ๐งช๐ป Model: Github: https://github.com/InternLM/Intern-S1 https://www.modelscope.cn/models/Shanghai_AI_Laboratory/Intern-S1-Pro https://github.com/InternLM/Intern-S1 https://www.modelscope.cn/models/Shanghai_AI_Laboratory/Intern-S1-Pro"
X Link 2026-02-04T15:20Z [----] followers, [----] engagements
"Introducing Singularity Cinema (MS-Agent framework)๐an LLM-powered tool that transforms documents or prompts into storyboards generating stunning 5-minute+ videos via its Manim/Remotion dual-engine. By fusing nano-banana-pro๐ Qwen-Image๐จ Sora and edge-tts it delivers high-quality information-dense contentperfect for science economics and history creators looking to scale their social media presence. ๐: ๐ฆPrompt: Convert xx/deploy.md into a short video in a blue-themed style making sure to use the important images from the document. The short video must be in English. ๐ฌOutput Video๐"
X Link 2026-02-05T16:08Z [----] followers, [----] engagements
"From a local designer in Chengdu to a global AI sensation with 117k+ downloadsDaniels journey shows whats possible when open-source meets imagination. Now its your turn. Join the Qwen-Image LoRA Competition and start building your own legacy. The next top creator could be you. https://modelscope.ai/active/qwenimagelora LoRA Training Contest: Win iPhone/PS5 and other prizes Gaussian splash LoRA creation process shared. Competition information: https://t.co/MEEZHQcr83 Training address: https://t.co/uxwtMlJzoP @Ali_TongyiLab @ModelScope2022 #HappyQwensday #QwenImageLora https://t.co/5igMVZUO5i"
X Link 2026-02-06T03:20Z [----] followers, [----] engagements
"ChenkinNoob-XL v0.2 is here A high-fidelity SDXL-to-Rectified Flow conversion delivering vivid color range and superior lighting adherence. โ
Performance: Eliminates "base model greyness"; Superior contrast/dark scene rendering; High stability across wide CFG ranges (3-6); Optimal convergence at 20-28 steps. โ
Ecosystem: Native ComfyUI workflow included; ReForge-ready with built-in Flow support; LoRA-compatible via Pochi.toml (Refresh scripts). โ
Architecture: Rectified Flow (RF) linear ODE trajectory; Trained on 47M samples using 8xH20 nodes; Leverages Cosine Optimal Transport and Protected"
X Link 2026-02-06T05:00Z [----] followers, [----] engagements
"StepFun open-sourced Step-3.5-Flashfeaturing powerful reasoning and Agent intelligence. Speeds up to [---] TPS Competes with closed-source models in Agent and Math tasks Excels at complex long-chain reasoning. Dont miss out https://x.com/i/broadcasts/1OyKAjZgZPaGb https://x.com/i/broadcasts/1OyKAjZgZPaGb"
X Link 2026-02-06T12:05Z [----] followers, [----] engagements
"Meet X-Coder ๐IIGroup drops a powerhouse suite for code LLMs โ
X-Coder-SFT-376k: Massive synthetic dataset via GPT-o3-mini/DeepSeek-R1. โ
X-Coder-RL-40k: Targeted RL alignment for logic & verification. โ
X-Coder-8B Models: SOTA reasoning built on Qwen3. Experience the power of fully synthetic code evolution. ๐ป๐ฅ ๐ Models & Dataset: #LLM #OpenCode #Qwen3 #DeepSeek #Coding https://modelscope.cn/models/IIGroup/X-Coder-SFT-Qwen3-8B https://modelscope.cn/datasets/IIGroup/X-Coder-RL-40k https://modelscope.cn/datasets/IIGroup/X-Coder-SFT-376k"
X Link 2026-02-09T06:25Z [----] followers, [----] engagements
"YOLO26 is here๐ The most advanced and deployment-ready vision model from @ultralytics redefining real-time edge computing performance. โ
Performance: 43% CPU inference speed boost vs YOLO11; Real-time performance on low-power hardware; [--] scalable sizes (n/s/m/l/x) for optimized speed-accuracy trade-offs. โ
Ecosystem: Unified framework for Detection Segmentation Pose OBB and Tracking; Simplified export for seamless hardware integration. โ
Logic/Architecture: Removed DFL for enhanced hardware compatibility; End-to-end NMS-free inference; ProgLoss + STAL for small target precision; MuSGD"
X Link 2026-02-09T08:30Z [----] followers, [----] engagements
"2/4 Key Technical Highlights: - Architecture: Native single-stream design integrating generation editing and segmentation. - Audio: First model to unify Speech Audio and Music generation via 12.5Hz continuous representation. - Video: Time-Interleaved VideoRoPE for enhanced temporal modeling. https://twitter.com/i/web/status/2021580867009671592 https://twitter.com/i/web/status/2021580867009671592"
X Link 2026-02-11T13:43Z [----] followers, [---] engagements
"๐ GLM-5 is now open-sourceand ranked #4 globally #1 among open models on Artificial Analysis Beyond "vibe coding" it delivers Agentic Engineering: autonomously architects debugs and ships production systems. Coding performance rivals Claude Opus [---] (SWE-bench: [----] ๐ SOTA) while dominating long-horizon agent tasks (BrowseComp MCP-Atlas ๐). 744B params MIT License ๐ ModelScope: https://modelscope.cn/models/ZhipuAI/GLM-5 https://modelscope.cn/models/ZhipuAI/GLM-5"
X Link 2026-02-12T03:43Z [----] followers, [----] engagements
"๐ฟ Meet Sirchmunk: The agentic search engine that operates directly on your raw data. Turn files into evolving intelligence in real-time Build information retrieval pipelines that require ZERO vector databases ZERO pre-indexing and ZERO chunking: ๐ฏ Embedding-Free Retrieval Drop files in and search instantly. No ETL pipelines no vector silos and no setup hours. ๐ฒ Monte Carlo Evidence Sampling It doesn't "read" full docsit explores focuses and synthesizes. 500-page manual Only the most relevant snippets hit the LLM. ๐ง Self-Evolving Knowledge Every search builds reusable knowledge clusters."
X Link 2026-02-13T13:39Z [----] followers, [---] engagements
"LongCat-Flash-LiteisNow Live Meituan's 68.5B MoE model pioneering "Embedding Expansion" for leading Agent and Coding performance. โ
Performance: 2.9B4.5B activated params; 500-700 tokens/s throughput; 256K context via YARN; SOTA SWE-Bench (54.4%) and -Bench scores. โ
Ecosystem: Optimized SGLang-FluentLLM engine with custom CUDA kernels; Open API with 50M free daily tokens. โ
Architecture: N-gram embedding layer for localized semantic precision; Hash-collision mitigation via sub-table decomposition; Triple-step Speculative Decoding synergy. ๐ค ๐ง ๐ Tech"
X Link 2026-02-08T06:27Z [----] followers, [----] engagements
"๐1/4 Introducing Ming-flash-omni-2.0: An open-source Omni-MLLM breakthrough. Based on the Ling-2.0 MoE architecture (100B total / 6B active) it unifies vision speech and image editing into a single model matching specialist performance across all modalities. ๐งต๐ ๐ค ModelScope: ๐ฆ GitHub: https://github.com/inclusionAI/Ming https://www.modelscope.cn/models/inclusionAI/Ming-flash-omni-2.0 https://github.com/inclusionAI/Ming https://www.modelscope.cn/models/inclusionAI/Ming-flash-omni-2.0"
X Link 2026-02-11T13:43Z [----] followers, [----] engagements
"๐Introducing Z-Image the full non-distilled version of Z-Image for high-quality customizable image generation. โ
Model: โ
Demo: โจ Capabilities: Photorealistic + diverse art styles (anime illustration digital art) Full CFG & precise negative prompts for reliable prompt following Multi-subject generation with distinct identitiesno "clone faces" Community-friendly: ready for LoRA ControlNet and other fine-tuning workflows ๐ฅ Highlights: Full architecture (vs. distilled Turbo): better diversity & training stability ideal for customization High output variance across seeds = less homogenization"
X Link 2026-01-27T15:57Z [----] followers, 59.5K engagements
"Z-Image-Fun-Lora-Distill is here A high-efficiency Distill LoRA enabling CFG-free high-fidelity generation for Z-Image in just 4-8 steps. โ
4-step and 8-step variants; CFG-free inference (CFG=1.0); [----] update delivers superior skin texture and color consistency compared to previous iterations. โ
Native ComfyUI weights available; full integration with the VideoX-Fun framework; compatible with Z-Image ControlNet Union [---] and third-party LoRAs. โ
Dual-distillation of both sampling steps and CFG; trained from scratch without Z-Image-Turbo weights; optimized for rapid generation and derivative"
X Link 2026-02-11T02:33Z [----] followers, [----] engagements
"๐ Ring-2.5-1T is open-source The first hybrid linear 1T thinking model. ๐ Gold Tier reasoning IMO25 (35/42) & CMO25 (105/126) scores. ๐ Agent-native compatible with Claude Code OpenClaw and SGLang. โกHybrid Linear breakthrough 10x lower memory & 3x throughput for long context. ๐ง 1T MoE structure with 63B active parameters (1:7 MLA + Lightning Linear). ๐ป Long-Horizon mastery autonomously built a 32-bit OS from scratch in [--] hours. Model: GitHub: https://github.com/inclusionAI/Ring-V2.5 https://modelscope.cn/models/inclusionAI/Ring-2.5-1T https://github.com/inclusionAI/Ring-V2.5"
X Link 2026-02-13T08:17Z [----] followers, 13K engagements
"Introducing FireRed-Image-Edit-1.0 from FireRedTeam ๐ Its officially the new SOTA for general image editing. โ
Better than Closed-Source: Outperforms Nano-Banana & Seedream4.0 on GEdit benchmarks. โ
Native Evolution: Built from T2I foundations not just a "patch" on existing models. โ
Style Mastery: Scored a record-breaking 4.97/5.0 in style transfer. โ
High-Fidelity Text: Keeps original font styles perfectly. โ
Virtual Try-on: Native support for multi-image joint editing. โ
Bilingual: Native support for both English & Chinese prompts. Apache [---] license. Local deployment ready. ๐ค Model: ๐ "
X Link 2026-02-14T15:00Z [----] followers, 69.4K engagements
"How 0.9B Models Are Outperforming the Giants: A Comparison of Four New Chinese VLM-OCR Models 1/3 ๐ Early [----] has brought a massive surge of open-source models from DeepSeek Z-org PaddlePaddle and Tencent YouTu. A tiny 0.9B parameter model GLM-OCR topped the global OmniDocBench v1.5 leaderboard outscoring giants like Gemini-3 Pro and Qwen3-VL. We are moving away from rigid "detect and recognize" pipelines toward Vision-Language Models (VLMs) that actually "read" and understand document structures like humans do. https://twitter.com/i/web/status/2018615132134551565"
X Link 2026-02-03T09:18Z [----] followers, 30.1K engagements
"Z-Image-Fun-Controlnet-Union & Lora-Distill are now open-source ๐Expanding the Z-Image ecosystem with ultra-fast inference and high-precision spatial control. โ
Performance: Achieve 8-step inference (CFG: 1.0) via Distill LoRA and professional-grade multi-condition control (Canny Depth Pose Inpaint) across 15+ layer blocks. โ
Ecosystem: Full cross-compatibility between Distill LoRA and Union ControlNet; includes "Lite" versions for low-VRAM edge devices and seamless natural blending. โ
Unified Control: Support for 7+ spatial conditions in a single model with adjustable control_context_scale"
X Link 2026-02-05T09:17Z [----] followers, 11K engagements
"Qwen3-ASR & Qwen3-ForcedAligner are now open-source ๐๐ High-performance streaming-ready speech AI for [--] languages and dialects. ๐: Demo: ๐คThe Models โ
ASR-1.7B: Open-source SOTA. Outperforms Whisper-v3; rivals GPT-4o & Gemini. โ
ASR-0.6B: Insane efficiency92ms latency; transcribes 2000s of audio in 1s. โ
ForcedAligner-0.6B: Superior timestamp accuracy across [--] languages. ๐Key Features โ
Universal: Supports [--] languages + [--] Chinese dialects. โ
Robust: Expertly handles singing heavy BGM and low SNR. โ
Unified: Single model for both streaming & offline (up to 20min audio). โกPerformance &"
X Link 2026-01-29T12:55Z [----] followers, [----] engagements
"๐ FantasyWorld is now open source Recently accepted to ICLR [----] and ranking #1 on the WorldScore leaderboard this framework offers a unified approach to video and 3D generation. โจKey Technical Highlights: โ
Direct 3D Inference: Input an image and camera trajectory to get frame-consistent video dense depth maps and point clouds instantly. โ
Dual-Branch IRG Architecture: An "Imagination" branch handles visual appearance while a "Geometry" branch ensures spatial accuracy. โ
Wan [---] Integration: The released 14B models leverage the latest Wan-DiT architectures for superior temporal consistency and"
X Link 2026-02-13T11:00Z [----] followers, [----] engagements
"Say hello to Ling-2.5-1T: 1T params 63B active MIT licensed โก Hybrid Linear Attention: 1:7 MLA + Lightning Linear. Beats Kimi K2 on long-context throughput. ๐ง 4x token efficiency: Composite rewards match frontier thinking models using 4x fewer tokens. ๐ 1M context: YaRN to 1M. Beats Kimi K2.5 & DeepSeek V3.2 on RULER/MRCR. Perfect NIAH scores. ๐ SOTA tool use: Agentic RL trained. Leads BFCL-V4. Native support for Claude Code OpenCode OpenClaw. ๐ฏ Follows instructions: Bidirectional RL + agent verification. High-density zero fluff. Honest: Gap vs GPT-5.2/Gemini [--] Pro on long-horizon tasks."
X Link 2026-02-16T03:39Z [----] followers, [----] engagements
"New Z-Image-Turbo control model released ๐ More control blocks. Inpaint mode. Better details. - 6-block ControlNet - Trained from scratch on 1M images @1328 res - Canny / HED / Depth / Pose / MLSD - Best control_context_scale: 0.65-0.80 Model: Demo: Code: https://github.com/aigc-apps/VideoX-Fun https://modelscope.cn/studios/AI-ModelScope/ZIT-Controlnet https://modelscope.cn/models/PAI/Z-Image-Turbo-Fun-Controlnet-Union https://modelscope.cn/models/PAI/Z-Image-Turbo-Fun-Controlnet-Union https://github.com/aigc-apps/VideoX-Fun https://modelscope.cn/studios/AI-ModelScope/ZIT-Controlnet"
X Link 2025-12-16T03:15Z [----] followers, 33.1K engagements
"๐Model & Demo are now live on our international site Model: - - - Demo: https://modelscope.ai/studios/FunAudioLLM/Fun-CosyVoice3-0.5B https://modelscope.ai/models/FunAudioLLM/Fun-CosyVoice3-0.5B-2512 https://modelscope.ai/models/FunAudioLLM/Fun-ASR-Nano-2512 https://modelscope.ai/models/FunAudioLLM/Fun-ASR-MLT-Nano-2512 https://modelscope.ai/studios/FunAudioLLM/Fun-CosyVoice3-0.5B https://modelscope.ai/models/FunAudioLLM/Fun-CosyVoice3-0.5B-2512 https://modelscope.ai/models/FunAudioLLM/Fun-ASR-Nano-2512 https://modelscope.ai/models/FunAudioLLM/Fun-ASR-MLT-Nano-2512"
X Link 2025-12-16T09:43Z [----] followers, [---] engagements
"๐คMeet Nanbeige4-3B from Boss Zhipina 3B-parameter LLM that outperforms Qwen3-32B on math (AIME) science (GPQA) and tool calling (BFCL-V4) while matching Qwen3-30B-A3B on human preference alignment (Arena-Hard-V2). How โ
23T tokens of ultra-curated data โ
Fine-grained WSD scheduler โ
30M+ high-quality SFT instructions โ
Multi-stage RL + innovative distillation (DPD) โ
Chain-of-thought reconstruction & deliberative generation It even ranks top [--] on WritingBench & EQ-Bench3beating models 100x larger like GLM-4.5 and Deepseek-R1. All models + tech report now open-source: ๐ Weights: ๐ Paper:"
X Link 2025-12-16T11:28Z [----] followers, 25.8K engagements
"๐ MiMo-V2-Flash FREE API is now live on ModelScope The first major release since Fuli Luo joined Xiaomiand its built for real-world agentic AI. โก MiMo-V2-Flash: an open high-performance MoE model with 309B total / 15B active parameters 256K context window 150+ tokens/s generation thanks to native Multi-Token Prediction ๐ฅ Key wins for developers: โ
Hybrid attention (5:1 SWA + Global) [--] less KV cache full long-context recall โ
73.4% on SWE-Bench Verified new SOTA for open-source models โ
Matches DeepSeek-V3.2 on reasoning but much faster in practice โจ API-readyperfect for building smart"
X Link 2025-12-18T02:29Z [----] followers, 16K engagements
"๐Qwen-Image-Layered is now live on ModelScopean incredible model that can intelligently decompose any image into multiple RGBA layers ๐ค Model: ๐ Try Demo ๐ Paper: https://modelscope.cn/papers/2512.15603 https://modelscope.cn/studios/Qwen/Qwen-Image-Layered https://modelscope.ai/models/Qwen/Qwen-Image-Layered https://modelscope.cn/papers/2512.15603 https://modelscope.cn/studios/Qwen/Qwen-Image-Layered https://modelscope.ai/models/Qwen/Qwen-Image-Layered ๐จ Qwen-Image-Layered is LIVE native image decomposition fully open-sourced โจ Why it stands out โ
Photoshop-grade layering Physically"
X Link 2025-12-19T15:29Z [----] followers, [----] engagements
"Its still Sunday and Im already hyped for next week. ๐The mysterious Eastern power is dropping a trilogy of magic: ๐ฎ Next week: A secret image-generation model ๐ค Next week: A secret code-agent model ๐ The week after: A secret voice model Buckle upthings are about to get interesting"
X Link 2025-12-21T09:54Z [----] followers, 10.1K engagements
"๐ Meet GLM-4.7 your new coding partner supercharged Try it now on ModelScope free API access for a limited time โ
+5.8% on SWE-bench (73.8%) โ
+12.9% on Multilingual coding (66.7%) โ
+16.5% on Terminal Bench [---] (41%) โ
Smarter agent reasoning & tool use โ
Better UI cleaner webpages sharper slides โ
+12.4% on complex reasoning (HLE: 42.8%)From coding to creativity GLM-4.7 delivers across the board. ๐ค Model: https://modelscope.cn/models/ZhipuAI/GLM-4.7 https://modelscope.cn/models/ZhipuAI/GLM-4.7"
X Link 2025-12-23T02:23Z [----] followers, [----] engagements
"๐ New on ModelScope: QwenLong-L1.5 is now fully open-source A 30B model (3B active params) that matches GPT-5 & Gemini-2.5-Pro in long-context reasoning. ๐ฅ Key wins: โ
+31.7 pts on OpenAIs MRCR (128K context SOTA across all models) โ
Matches Gemini-2.5-Pro on [--] major long-QA benchmarks โ
+9.69 on CorpusQA +6.16 on LongBench-V2 How Three breakthroughs: [--] Synthetic data at scale: 14.1K long-reasoning samples from 9.2B tokens no human labeling. Avg. length: 34K tokens (max: 119K). [--] Stable RL training: Task-balanced sampling + Adaptive Entropy-Controlled Policy Optimization (AEPO) for reliable"
X Link 2025-12-23T07:41Z [----] followers, 44.2K engagements
"@Sean60133791259 @bdsqlsz Could I ask which country you are in Well test the speed. You can DM me"
X Link 2025-12-24T02:04Z [----] followers, [--] engagements
"Merry Christmas everyone ๐โจ ๐Introducing BEYOND REALITY Z IMAGE 1.0: a community-tuned checkpoint merging LoRA enhancements into Z-Image optimized for film-style portrait aesthetics and high-frequency texture detail think skin pores fabric weaves and wall surfaces rendered with striking realism. โ
Keeps Zs photorealism intact โ
Adds cinematic color depth & tonal nuance โ
Significantly upgrades portrait + environmental textures โ
Runs on 8GB VRAM (FP8) tested on laptops Note: Training focused on high-quality synthetic portrait photography so complex poses or non-human textures may vary."
X Link 2025-12-24T08:00Z [----] followers, 13.5K engagements
"We shouldve shouted this louder: Youre officially missing out if you aren't using Civision ๐คซ ๐จ FREE Image & Video Gen & FREE LoRA Training โจ A UI so clean you dont need a PhD to master it. Weve got the heavy hitters ready: Z-Image Qwen-Image Qwen-Image-Edit Wan2.2 and more. All the power zero cost. Right here on ModelScope: ๐โจ https://modelscope.ai/civision/imageGeneration https://modelscope.ai/civision/imageGeneration"
X Link 2025-12-25T08:43Z [----] followers, [----] engagements
"๐New on ModelScope: ChenkinNoob-XL v0.2 is live Independently developed by the ChenkinNoob team as a fine-tuned extension of noobai-XL-1.1 v0.2 incorporates Danbooru data up to Nov [--] 2025enhancing performance while preserving the original aesthetic. โจ Key improvements over v0.1: โ
Stronger character consistency โ
Refined details (hands accessories layered outfits) โ
Higher prompt adherence ๐จ Faithful to the noob visual DNA: clean composition solid anatomy and harmonious color palettes. โ Recommended config: CFG: [--] Steps: [----] Sampler: Euler a Resolution: [----] (e.g. [-------] 10241024)"
X Link 2025-12-26T02:26Z [----] followers, 30.7K engagements
"๐ New on ModelScope: MiniMax M2.1 is open-source โ
SOTA in 8+ languages (Rust Go Java C++ TS Kotlin Obj-C JS) โ
Full-stack Web & mobile dev: Android/iOS 3D visuals vibe coding that actually ships โ
Smarter faster 30% fewer tokens with lightning mode (M2.1-lightning) for high-TPS workflows โ
Top-tier on SWE-bench VIBE and custom coding/review benchmarks โ
Works flawlessly in Cursor Cline Droid BlackBox and more Its not just better code its AI-native development end to end. ๐ Model: https://modelscope.cn/models/MiniMax/MiniMax-M2.1 https://modelscope.cn/models/MiniMax/MiniMax-M2.1"
X Link 2025-12-26T08:03Z [----] followers, 17.3K engagements
"@nnnnmailcom Hi there SDXL models should be supported. Could you let me know exactly what issue you're running into"
X Link 2025-12-29T03:53Z [----] followers, [--] engagements
"๐400+ Free APIs are now available on We currently support: โ
DeepSeek series โ
Qwen3 series & Qwen-Image โ
Z-Image-Turbo and many more Check it out: Go to - Models - Filter by API-Inference. (Currently available on PC only.) ๐ Give them a spin and let us know what you think โDM us for support. https://modelscope.ai/modelsfilter=inference_type&page=1&tabKey=task http://modelscope.ai http://ModelScope.ai http://modelscope.ai http://ModelScope.ai https://modelscope.ai/modelsfilter=inference_type&page=1&tabKey=task http://modelscope.ai http://ModelScope.ai http://modelscope.ai"
X Link 2025-12-29T08:01Z [----] followers, 17.8K engagements
"@blankbraindead oh I see. May I know which region you are in"
X Link 2025-12-29T08:33Z [----] followers, [---] engagements
"๐ Tencent HunYuan open-sourced HY-MT1.5 A lightweight powerhouse for translation: โจ 1.8B model: 0.18s latency runs offline on mobile & outperforms most commercial APIs. Even rivals Gemini-3.0-Pro at 90% quality โก 7B model: Upgraded with WMT25-champion accuracy & fewer artifacts. ๐ [--] languages + [--] Chinese dialects ๐ Context-aware custom glossaries format-preserving output ๐ก Perfect for edge + cloud deployments Models: https://modelscope.cn/collections/Tencent-Hunyuan/HY-MT15 https://modelscope.cn/collections/Tencent-Hunyuan/HY-MT15"
X Link 2025-12-30T10:08Z [----] followers, 17K engagements
"One of your New Year open-source gifts ๐จ is on its way"
X Link 2025-12-30T11:45Z [----] followers, 10.7K engagements
"๐New on ModelScope: Qwen-Image-2512 is here ๐ This version delivers a massive leap in realism and control: โจ Photorealistic humans dramatically reduced AI look with accurate aging natural expressions and fine details (yes even individual hair strands) ๐ฟ Ultra-detailed nature fur foliage water and mist rendered with unprecedented texture fidelity ๐ค Reliable text & layout generates complex infographics timelines and multilingual slides with correct structure and legible text ๐ Evaluated in 10k+ blind tests: now the strongest open-source image modelrivaling closed alternatives. ๐ก Supports"
X Link 2025-12-31T09:53Z [----] followers, [----] engagements
"2026 is almost here Wishing you a very Happy New Year ๐ What open-source models or product features are at the top of your wishlist for the coming year ๐ Wed love to see if we can help make those wishes come true"
X Link 2025-12-31T12:28Z [----] followers, [----] engagements
"๐Qwen-Image-2512 is now live on Model: Try: https://modelscope.ai/civision/imageGeneration/tab=default https://modelscope.ai/models/Qwen/Qwen-Image-2512/summary http://ModelScope.ai https://modelscope.ai/civision/imageGeneration/tab=default https://modelscope.ai/models/Qwen/Qwen-Image-2512/summary http://ModelScope.ai"
X Link 2025-12-31T17:37Z [----] followers, [----] engagements
"๐ New on ModelScope: IQuest Research just dropped IQuest-Coder-V1 a new family of open-source code LLMs (7B to 40B) with 128K context GQA and two specialized variants: - Instruct: for coding assistance - Thinking: RL-tuned for agentic reasoning & error recovery - Plus: a novel Loop architecture (40B only) for efficient deployment via parameter sharing across [--] iterations. ๐ All models support repo-level understanding via Code-Flow training learning from commit sequences and code evolution not just static snippets. ๐ Benchmarks: - [----] on SWE-Bench Verified - [----] on LiveCodeBench v6"
X Link 2026-01-04T07:14Z [----] followers, 10.9K engagements
"Big news ๐ now officially supports FREE LoRA training for Qwen-Image-2512 Z-Image-Turbo and Qwen-Image-Edit-2511. Time to build ๐ ๐จModelScope Civision: https://modelscope.ai/civision/modelTraining http://ModelScope.ai https://modelscope.ai/civision/modelTraining http://ModelScope.ai https://modelscope.ai/civision/modelTraining http://ModelScope.ai https://modelscope.ai/civision/modelTraining http://ModelScope.ai"
X Link 2026-01-04T12:13Z [----] followers, 21.5K engagements
"Qwen-Image-2512-Fast is officially here โก ๐ Try the Demo on ModelScope now: Experience a massive 20x speedup generating high-quality images in just [--] to [--] steps Using CFG distillation weve achieved lightning-fast inference without compromising quality. https://modelscope.cn/studios/kelseye/Qwen-Image-2512-Turbo-LoRA-Demo/summary https://modelscope.cn/studios/kelseye/Qwen-Image-2512-Turbo-LoRA-Demo/summary https://modelscope.cn/studios/kelseye/Qwen-Image-2512-Turbo-LoRA-Demo/summary https://modelscope.cn/studios/kelseye/Qwen-Image-2512-Turbo-LoRA-Demo/summary"
X Link 2026-01-04T12:35Z [----] followers, 30.5K engagements
"@WolfyBlair Yes you'll get daily free quota (magicubes) and you can earn more by participating in community activities"
X Link 2026-01-05T02:26Z [----] followers, [---] engagements
"๐ New on ModelScope: The Wuli Team just released a Turbo LoRA that makes Qwen-Image-2512 inference 20x fasterโก without sacrificing quality. Key updates: โ
4-step turbo inference โ
Enhanced texture & detail โ
Native ComfyUI support Model: Demo: https://modelscope.cn/studios/kelseye/Qwen-Image-2512-Turbo-LoRA-Demo https://modelscope.ai/models/Wuli-Art/Qwen-Image-2512-Turbo-LoRA https://modelscope.ai/models/Wuli-Art/Qwen-Image-2512-Turbo-LoRA Qwen-Image-2512-Fast is officially here โก ๐ Try the Demo on ModelScope now: https://t.co/LRShwwbiNh Experience a massive 20x speedup generating"
X Link 2026-01-05T03:41Z [----] followers, 11.3K engagements
"๐ค Introducing InternVLA-A1 now fully open-sourced Many VLA models follow instructions well in static scenes but struggle in dynamic environments (conveyor belts rotating platforms multi-robot setups). Why They see the presentbut cant imagine the future. InternVLA-A1 solution: unify perception imagination and action in one model: โ
Scene understanding: Image + text task parsing โ
Task imagination: Predict future frames reason about dynamics โ
Guided control: Execute actions steered by visual foresight Powered by InternData-A1 - Large-scale high-quality simulated dataset InternVLA-A1 stays"
X Link 2026-01-05T11:23Z [----] followers, 38.2K engagements
"๐ Small model BIG agency Meet Youtu-LLM a native agentic 1.96B LLM that thinks plans and acts by itself no external frameworks needed. โ
Outperforms 4B8B models on math coding & agent benchmarks โ
Trained on 200B+ agentic reasoning trajectories โ
128K context open-weight & ready to run anywhere Think lightweight cant be smart Think again. ๐ค Model: ๐ GitHub: ๐ Paper: https://www.modelscope.cn/papers/2512.24618 https://github.com/TencentCloudADP/youtu-tip/blob/master/youtu-llm https://modelscope.cn/collections/Tencent-YouTu-Research/Youtu-LLM https://www.modelscope.cn/papers/2512.24618"
X Link 2026-01-06T07:57Z [----] followers, 11.8K engagements
"๐ TeleChat3-105B-A4.7B-Thinking is now open source A 105B sparse MoE model with fine-grained routing: - [---] experts only [--] activated per token (4.7B active params) - Trained end-to-end on domestic compute - Strong across code math agents writing check HumanEval-X (92.7%) & SWE-Bench (51%) It can even build a working shooter game ๐น or a video site frontendfully autonomously. Also released: TeleChat3-36B-Thinking (dense 36B) optimized for reasoning & roleplay. โ
Supports vLLM SGLang LLaMA-Factory โ
Full Thinking mode with think./think reasoning traces โ
Tuned for long-context and multi-turn"
X Link 2026-01-06T11:20Z [----] followers, 16.8K engagements
"The #EmbodiedAI data gap just got a lot smaller. ๐ @AGIBOTofficial released the Genie Sim [---] datasetthe largest open-source sim dataset in the field ๐ค 10000+ hours of high-quality sim demos ๐ฏ 200+ tasks & 100k scenarios ๐ Multi-sensor: RGB-D Stereo & Kinematics Standardized in LeRobot format. Ready to train โ
Dataset: โ
GitHub: https://github.com/AgibotTech/genie_sim https://modelscope.cn/datasets/agibot_world/GenieSim3.0-Dataset https://github.com/AgibotTech/genie_sim https://modelscope.cn/datasets/agibot_world/GenieSim3.0-Dataset"
X Link 2026-01-07T07:00Z [----] followers, [----] engagements
"AI for all ๐ Hosted by @alibaba_cloud the Alibaba Cloud AIGC Championship @ Milano Cortina [----] is officially live Create your Winter Olympics videos with the tools and resources powered by ModelScope as an official partner. โธโ ๐ Win Big ๐ Milan [----] Tickets: A trip to the Winter Olympics ๐ Legacy: Works archived in the Olympic Museum. ๐ฐ Cash Prizes: Reward your creative storytelling. ๐ฌ Pro Tech: High-speed AI video via Wan [---]. Global Entrance Creating #ModelScope #AlibabaCloud #MilanoCortina2026 #Wan26 #Olympics https://modelscope.ai/civision/videoGeneration"
X Link 2026-01-07T08:54Z [----] followers, [----] engagements
"๐ComfyUI now supports Z-Image-Turbo LoRAs trained with ModelScope Civision Huge thanks to our amazing community members for the incredibly fast integration ๐ Everyone is welcome to train LoRAs for free on Civision and thanks to your valuable feedback training speed has been significantly improved ๐ https://modelscope.ai/civision/modelTraining https://github.com/Comfy-Org/ComfyUI/pull/11805 https://github.com/Comfy-Org/ComfyUI/pull/11805 https://modelscope.ai/civision/modelTraining https://github.com/Comfy-Org/ComfyUI/pull/11805 https://github.com/Comfy-Org/ComfyUI/pull/11805"
X Link 2026-01-12T03:16Z [----] followers, 11.9K engagements
"GLM-Image is here Try it out and let us know what you think: https://x.com/ModelScope2022/status/2011262011997651194 ๐GLM-Image is now open-source a breakthrough in "Cognitive Generation." By combining Autoregressive models with Diffusion Decoders it finally solves the struggle of generating posters PPTs and knowledge-heavy infographics. ๐ Highlights: โ
#1 on CVTG-2K & LongText-Bench https://t.co/cF3z0VFvc9 https://x.com/ModelScope2022/status/2011262011997651194 ๐GLM-Image is now open-source a breakthrough in "Cognitive Generation." By combining Autoregressive models with Diffusion"
X Link 2026-01-14T02:40Z [----] followers, [----] engagements
"Step-Audio-R1.1 by @StepFun_ai just set a new SOTA on the Artificial Analysis Speech Reasoning leaderboard ๐ It outperforms Grok Gemini and GPT-Realtime with a 96.4% accuracy rate. โ
Native Audio Reasoning (End-to-End) โ
Audio-native CoT (Chain of Thought) โ
Real-time streaming inference โ
FULLY OPEN SOURCE ๐ Demo: ๐ค Model: https://modelscope.cn/models/stepfun-ai/Step-Audio-R1.1 https://modelscope.cn/studios/stepfun-ai/Step-Audio-R1 https://modelscope.cn/models/stepfun-ai/Step-Audio-R1.1 https://modelscope.cn/studios/stepfun-ai/Step-Audio-R1"
X Link 2026-01-15T06:32Z [----] followers, 15.6K engagements
"@SteveWarnerFL Hi would you mind sharing more about how you use it in your work Are you trying to modify a specific layer or add/remove one Wed love to better understand your situation"
X Link 2026-01-16T06:34Z [----] followers, [--] engagements
"Step3-VL-10B Demo: https://modelscope.cn/studios/stepfun-ai/step3-vl-10b https://modelscope.cn/studios/stepfun-ai/step3-vl-10b"
X Link 2026-01-19T11:34Z [----] followers, [----] engagements
"๐ Introducing GLM-4.7-Flash the strongest 30B-class open model that balances performance & efficiency โ
Top scores on AIME GPQA SWE-bench -Bench & more โ
Optimized for agentic coding tool use and reasoning โ
Runs locally with vLLM / SGLang / Transformers Perfect for developers who want Claude-level coding power at a fraction of the cost ๐ก Try it now: ๐ค ๐ ๐ ๐ https://github.com/zai-org/GLM-4.5 https://z.ai/blog/glm-4.7 https://chat.z.ai https://modelscope.cn/models/ZhipuAI/GLM-4.7-Flash https://github.com/zai-org/GLM-4.5 https://z.ai/blog/glm-4.7 https://chat.z.ai"
X Link 2026-01-19T15:32Z [----] followers, 11.3K engagements
"How reliable is your model's tool usage. REALLY ๐ง Inspired by @Kimi_Moonshot's K2 Vendor Verifier we are leveling up Agentic Evaluation with EvalScope. The standardizing Function Calling evaluation allows us to bridge the gap between "claims of support" and "RELIABLE execution" in the field. ๐ ๐ Docs: ๐ป Code: ๐น Standardized: Quantifies MoonshotAI K2 verification logic. ๐น Comprehensive: Measures decision-making + JSON schema compliance. ๐น Simple: Plug-and-play with your data or our benchmarks. Don't let your Agent go live without a proper check-up #LLM #AIAgents #FunctionCalling"
X Link 2026-01-22T06:32Z [----] followers, [----] engagements
"@YashasGunderia Hi you'll need to request to join the organization first and then you'll have access to the GPU resources. Here is the link: https://www.modelscope.ai/organization/xGPU-Explorers https://www.modelscope.ai/organization/xGPU-Explorers"
X Link 2026-01-24T05:43Z [----] followers, [---] engagements
"Thinking Machines Lab proved On-Policy Distillation slashes LLM training costs by 10x and we show you how to reproduce their research. Invest [--] minutes in this guideas we unpack the theory tech details experiment results and code to instantly transform your fine-tuning budget๐ Related Resources๐): โ
Slash training compute by 10X. โ
Achieve robust RL performance with zero forgetting. โ
Get the ready-to-use ms-SWIFT + vLLM code for deployment. Related Resources - TML Blog: - (Open-source implementation for reproducing On-Policy Distillation) - On-Policy Distillation Documentation: - Example"
X Link 2025-10-28T13:09Z [----] followers, [----] engagements
"VibeThinker-1.5B is here ๐ and it flips the bigger = smarter myth on its head. โ
Just 1.5B params โ
Trained via novel Spectrum-to-Signal Principle (SSP) โ
Beats models 400x larger (e.g. 671B DeepSeek-R1) on hard math benchmarks (AIME24/25 HMMT25) โ
Matches 456B MiniMax-M1 and ties Mistrals Magistral-Medium on LiveCodeBench v6 โ
Post-training cost: $8K (vs. $290K+ for others) ๐ก SSP first encourages divergent exploration of solution paths then uses RL to converge on optimal strategies smarter training not bigger models. ๐ Open-sourced to empower researchers & smaller teams whove been priced"
X Link 2025-11-13T13:00Z [----] followers, [----] engagements
"๐ GLM-4.6V is live @Zai_org 's new open-source multimodal models just dropped on ModelScope and we provide a FREE API for immediate accessโ GLM-4.6V (106B) for cloud & high-performance workloads GLM-4.6V-Flash (9B) lightweight fast great for local inference ๐ฅ Key dev perks: โ
Native multimodal tool calling pass images/docs directly as function args no OCR detour โ
128K context handles 150-page docs or hour-long videos in one go โ
Visual Action pipeline powers real multimodal agents (e.g. find this outfit online returns structured shopping list) โ
50% cheaper than GLM-4.5V $1/million input"
X Link 2025-12-08T11:43Z [----] followers, [----] engagements
"Weve prepared ModelScope and Tongyi swag for everyone attending the workshopready for a day full of great takeaways ๐ Join Us in Seoul: AI Innovation Meets Creativity @AMD @Alibaba_Qwen @ModelScope2022 @alibaba_cloud AMDs AI Developer Meetup in Seoul (Dec 10) is filling FAST. As a key partner were bringing you the future of generative AI. ๐
Dec [--] ๐ Seoul ๐ Free https://t.co/b5XcmEmbPf ๐ Join Us in Seoul: AI Innovation Meets Creativity @AMD @Alibaba_Qwen @ModelScope2022 @alibaba_cloud AMDs AI Developer Meetup in Seoul (Dec 10) is filling FAST. As a key partner were bringing you the"
X Link 2025-12-10T01:28Z [----] followers, [----] engagements
"Huge update for Z-Image-Turbo-Fun-Controlnet-Union [---] ๐ Key technical upgrades: โ
New 1.9GB Lite model for low-VRAM & natural blending โ
Fixed mask leakage in inpainting โ
Multi-resolution dataset refactor (up to 1536px) โ
8-step distillation for crisp blur-free Turbo gen No more bright spot artifacts. Just pure control. ๐ค Model: https://modelscope.ai/models/PAI/Z-Image-Turbo-Fun-Controlnet-Union-2.1 https://modelscope.ai/models/PAI/Z-Image-Turbo-Fun-Controlnet-Union-2.1"
X Link 2026-01-14T08:34Z [----] followers, 20.5K engagements
"Qwen-Image-Layered-Control is now supported by ComfyUI Try it out: https://modelscope.ai/models/DiffSynth-Studio/Qwen-Image-Layered-Control/files DiffSynth-Studio has open-sourced Qwen-Image-Layered-Control ๐ It enables precise text-controlled layer extractionperfect for poster decomposition and graphic design workflows. Control image layers like never before: โ
Single-layer output via text prompts โ
Native https://t.co/4QwOKqTFLs https://modelscope.ai/models/DiffSynth-Studio/Qwen-Image-Layered-Control/files DiffSynth-Studio has open-sourced Qwen-Image-Layered-Control ๐ It enables precise"
X Link 2026-01-15T09:33Z [----] followers, 22.9K engagements
"Real-Qwen-Image-V2 is here ๐ธ Developed by wikeeyang this fine-tuned version of Qwen-Image-2512 is built for those who demand peak realism and production-grade quality. ๐ค Model: Key Highlights: โจ Sharper details & enhanced realism โจ Optimized for Asian facial aesthetics โจ Superior text & image generation quality โจ Highly compatible with LoRAs & custom workflows Tech Specs: โ
CFG: [---] - [---] โ
Steps: [--] - [--] โ
Sampler: Euler / Simple (flexible) โ
Model Shift: [---] - [---] https://www.modelscope.cn/models/wikeeyang/Real-Qwen-Image-V2 https://www.modelscope.cn/models/wikeeyang/Real-Qwen-Image-V2"
X Link 2026-01-18T05:01Z [----] followers, 28.4K engagements
"๐ Meet STEP3-VL-10Bit delivers SOTA-level visual perception complex reasoning and human-aligned intelligenceredefining efficiency in open multimodal AI. โ
Beats or matches models [----] larger (like GLM-4.6V Qwen3-VL even Gemini [---] Pro) โ
Achieves SOTA on MMMU MathVision OCRBench ScreenSpot and more โ
Trained on 1.2T tokens + 1400+ RL rounds (RLHF + RLVR) โ
Supports PaCoRe: parallel collaborative reasoning (128K context) Despite its compact size it leads the 10B class in: STEM reasoning (94.43% on AIME [----] w/ PaCoRe) Visual perception (92.05 on MMBench) GUI understanding & OCR Spatial"
X Link 2026-01-19T11:32Z [----] followers, 217.6K engagements
"Latency or intelligence Step-Audio-R1.1 ends the trade-off. Fei Tian Multimodal Large Model Researcher at StepFun introduces Mind-Paced Speakingenabling real-time voice AI that thinks while speaking. Fast responses. Deep reasoning. No compromise. https://x.com/i/broadcasts/1mrGmBabZavJy https://x.com/i/broadcasts/1mrGmBabZavJy"
X Link 2026-01-19T11:57Z [----] followers, [----] engagements
"Introducing AgentCPM-Report: An 8B on-device agent that rivals closed-source giants in deep research and professional report writing. The Performance: ๐ Ranked #1 in Insight Depth across DeepResearch Bench Deep Consult and DeepResearch Gymoutperforming top-tier closed models. Highlights: โ
100% Private: Zero data leakage. Your sensitive data stays on your hardware. โ
UltraRAG Stack: Efficient local indexing for private PDFs and docs. โ
Production Ready: One-click Docker deployment. How does an 8B model hit SOTA ๐ง It breaks the "one-shot" generation bottleneck with a "Writing as Reasoning""
X Link 2026-01-20T12:18Z [----] followers, [----] engagements
"๐ Chroma [---] is here and its open From @flashlabsdotai : the worlds first open-source end-to-end real-time speech-to-speech dialogue model with personalized voice cloning. โจ 150ms end-to-end latency ๐งฌ High-fidelity voice cloning from just seconds of audio ๐ Speaker similarity (SIM) = [-----] +10.96% over human baseline ๐ง Strong reasoning in only 4B params ๐ Fully open weights + code A true open alternative to closed commercial systems and already optimized with @lmsysorgs SGLang for even faster inference ๐ Paper: ๐ค Model: ๐ป Code: Were excited to see what the community builds with"
X Link 2026-01-21T16:07Z [----] followers, 14.3K engagements
"๐พ Meet Sinong (): The first heavy-duty LLM suite for Agriculture ๐ ๐งฌ Expertise: Crop breeding animal medicine smart farming & ag-economics. ๐ค Models: 8B & 32B parameters ๐ Stack: 4B+ tokens (240k+ papers 8k+ books patents) ๐ง Tech: Native CoT reasoning + Multi-Agent RAG for zero-hallucination scientific retrieval. ๐ Open Source: https://modelscope.cn/models/NAULLM/Sinong1.0-32B https://modelscope.cn/models/NAULLM/Sinong1.0-8B https://modelscope.cn/models/NAULLM/Sinong1.0-32B https://modelscope.cn/models/NAULLM/Sinong1.0-8B"
X Link 2026-01-23T03:46Z [----] followers, [----] engagements
"@noctus91 Hi there This isn't actually an error. To use API inference you just need to bind your Alibaba Cloud account (no extra cost involved)"
X Link 2026-01-23T17:37Z [----] followers, [---] engagements
"โก๐จTomorrow. This week is going to be intriguing"
X Link 2026-01-26T02:30Z [----] followers, 51.8K engagements
"DeepSeek-OCR-2 is now on ModelScope โจ Meet DeepSeek-OCR-2 a vision-language model that reimagines document intelligence with human-like visual understanding. Key Highlights ๐ง Visual Causal Flow: Understands layout semantics and structurenot just pixels ๐ Dynamic Resolution: Processes documents at (0-6)768 + [----] with adaptive token generation โ Flexible Prompts: One model for multiple tasks: - Convert to markdown preserve tables & formatting - Parse the figure extract chart insights - Locate "text" pixel-level grounding - OCR this image 100+ language support โก Production Ready:"
X Link 2026-01-27T06:26Z [----] followers, [----] engagements
"๐ Meet Kimi K2.5 ๐ This is Kimis most intelligent and versatile model to date achieving SOTA performance across coding vision and agentic workflows. Model: Paper: Highlights: โ
Native Multimodal Architecture: Seamlessly integrates vision & text. From screenshots to screen recordings K2.5 "sees" and reasons across images and videos to break the limits of text-only prompts. โ
Agent Clusters: A world-first capability. K2.5 can autonomously spawn up to [---] "clones" to work in parallel on complex tasks slashing execution time by 4.5x. โ
The Ultimate Office/Coding Assistant: * Office Pro: Masters"
X Link 2026-01-27T06:58Z [----] followers, [----] engagements
"The DiffSynth-Studio team also dropped Z-Image-i2L (Image-to-LoRA). ๐จโจ This model takes a single image as input and instantly outputs a custom LoRA tailored to that specific style or feature. Model: Demo: โ
Zero Manual Effort: No captioning or training process required. โ
One-Shot Magic: Get a style LoRA from just one reference image. โ
Style Transfer: Seamlessly extract and apply unique features to new prompts. โ
Lower Barrier: Personalized model creation is now accessible to everyone. https://modelscope.cn/studios/DiffSynth-Studio/Z-Image-i2L"
X Link 2026-01-27T16:02Z [----] followers, [----] engagements
"๐ Meet LingBot-VLA: A pragmatic Vision-Language-Action model designed to bridge the gap between perception and execution in robotics. ๐ค โ
LingBot-VLA-4B: Lightweight & versatile. โ
LingBot-VLA-4B-Depth: Enhanced for high-precision spatial tasks. Highlights: ๐ง Powerful Core: Built on the Qwen2.5-VL-3B foundation mastering multi-tasking and dual-arm coordination across 9+ robot configs. ๐ Elite Performance: Outperforms competitors like [---] and GR00T in success rates (SR) on both GM-100 (Real-world) and RoboTwin [---] (Sim). ๐ Hyper-Efficient: 1.52.8x faster training than existing VLA codebases"
X Link 2026-01-27T18:00Z [----] followers, [----] engagements
"HunyuanImage-3.0 is here: A high-performance multimodal model for unified text-to-image generation and creative editing. ๐๐งต๐ ๐คModel: 1/ The Model Suite ๐ฆA versatile lineup for any workflow: โ
Base 3.0: The core T2I powerhouse. โ
3.0-Instruct: Advanced I2I and creative editing. โ
Instruct-Distil: Optimized for 8-step sampling and rapid deployment. 2/ Core Capabilities ๐จBeyond renderingits a creative partner. Features Unified Generation (T2I I2I and 3-image fusion) with built-in CoT reasoning and Prompt Enhancement to turn sparse ideas into professional descriptions. 3/ Technical Highlights"
X Link 2026-01-29T11:51Z [----] followers, [----] engagements
"Meet LingBot-VA: The future of robot learning is visual. ๐ค๐ฅ Its an autoregressive diffusion framework that predicts future video frames and decodes actions simultaneously. Instead of just reacting it reasons: "I expect the world to look like this next so I should move like that." Key Highlights: ๐ง Infinite Memory: Unlike memoryless models that get stuck in loops LingBot-VA handles "recurrent states" (like opening the same box twice) with ease by remembering the full history. ๐ณ Generalist Skills: From folding clothes to making breakfast and high-precision screw pickingits a true"
X Link 2026-01-30T02:47Z [----] followers, [----] engagements
"PaddleOCR-VL-1.5 is open-sourced A 0.9B parameter multimodal model that masters the "curved & distorted" document challenge. ๐ ๐ ModelScope: ๐ Demo: Key Technical Breakthroughs: โ
Polygon-based Localization: No more rigid rectangles. It natively supports irregular boxes to fit warped tilted or screen-captured text perfectly. ๐ โ
Seal & Stamp Recognition: New dedicated capability to handle complex official document verification. ๐ฎ โ
Cross-Page Logic: Automatically merges tables and identifies headings across physical page breakscrucial for full-document semantics. โ
Performance Giant:"
X Link 2026-01-30T05:52Z [----] followers, [----] engagements
"Ready to make this February legendary ๐โจThe Qwen-Image LoRA Training Competition is officially LIVE Weve teamed up with @Ali_TongyiLab and @Alibaba_Qwen to bring you some absolutely grand prizes:๐ฑ iPhone [--] Pro Max (512GB) ๐ฎ Sony PS5 Pro ๐ $800 Shopping Cards (or equivalent prizes) Stop scrolling and start training ๐ ๐จJoin here: #HappyQwensday #QwenImageLoRA https://modelscope.ai/active/qwenimagelora https://modelscope.ai/active/qwenimagelora"
X Link 2026-02-01T12:35Z [----] followers, [----] engagements
"2/3 ๐ Four specialist models were released in just one week each mastering a specific real-world challenge: โ
DeepSeek-OCR2: Uses "Visual Causal Flow" to handle reading order in complex multi-column layouts. โ
GLM-OCR (Z-org): An efficiency powerhouse using Multi-Token Prediction (MTP) to reduce costs by 90% while excelling at formulas. โ
PaddleOCR-VL-1.5 (PaddlePaddle): Built for "Real5" messy conditionsperfect for curved pages and camera glare. โ
Youtu-Parsing (Tencent): Delivers 22x faster inference and converts flowcharts directly into Mermaid code."
X Link 2026-02-03T09:18Z [----] followers, [----] engagements
"Here they come They are now at or alumni of these elite AI labs: Alibaba Tongyi Meta (FAIR/Superintelligence) Google DeepMind Tencent AI. Academia: SJTU OSU UChicago NUS WashU. https://x.com/i/broadcasts/1ZkJzZRdQNoJv https://x.com/i/broadcasts/1ZkJzZRdQNoJv"
X Link 2026-02-05T05:55Z [----] followers, [----] engagements
"@pranaysuyash http://ModelScope.ai http://ModelScope.ai"
X Link 2026-02-05T12:38Z [----] followers, [---] engagements
"1/4๐งต MS-Agent is Rebuilding A massive functional update to ModelScopes agent frameworkintroducing autonomous skill scaling and industrial-grade toolkits. โ
Agent Skills: Skill Discovery via local/Hub repo_id matching; โ
Multi-skill Collaboration via DAG-based planning; โ
Progressive Analysis to optimize context window for small models. โ
Robust Executor: Secure Docker/Sandbox execution; autonomous self-reflection and bug-fix cycles. Github: https://github.com/modelscope/ms-agent/tree/main/ms_agent/skill https://github.com/modelscope/ms-agent/tree/main/ms_agent/skill"
X Link 2026-02-06T02:34Z [----] followers, [---] engagements
"2/4 ๐ป Code Genesis is Now Live A SOTA end-to-end synthesis engine specializing in zero-to-one development over standard "vibe-coding." โ
Capabilities: Industrial-grade Code Genesis for full-stack synthesis; Language Server integration for stability and clean code aesthetics. โ
Technical Edge: Superior benchmarks in end-to-end logic flow; minimalist architecture for rapid project scaffolding. Repo: Tech Report: https://github.com/modelscope/ms-agent/tree/main/projects/code_genesis/TECH_REPORT.md https://github.com/modelscope/ms-agent/tree/main/projects/code_genesis"
X Link 2026-02-06T02:34Z [----] followers, [---] engagements
"๐ฌ Singularity Cinema is Breaking Benchmarks The GLCC-winning workshop for high-density long-form automated video production. โ
Generation: LLM-driven storyboards; Manim/Remotion dual-engine animation; 5min+ cinematic output for Science Econ and History. โ
Model Fusion: Seamlessly integrates Qwen-Image Sora edge-tts and nano-banana-pro; transforms static docs into dynamic media. Demo: https://github.com/modelscope/ms-agent/tree/main/projects/singularity_cinema https://github.com/modelscope/ms-agent/tree/main/projects/singularity_cinema"
X Link 2026-02-06T02:34Z [----] followers, [---] engagements
"4/4 ๐ ms-agent WebUI is Live A self-bootstrapped interface generated entirely by the Code Genesis engine. โ
Usage: pip install ms-agent -U - ms-agent ui; instant GUI access for agent orchestration. โ
Roadmap: DeepResearch capability launching soonSOTA autonomous retrieval and synthesis incoming. Full Framework: https://github.com/modelscope/ms-agent https://github.com/modelscope/ms-agent"
X Link 2026-02-06T02:34Z [----] followers, [---] engagements
Limited data mode. Full metrics available with subscription: lunarcrush.com/pricing
/creator/twitter::ModelScope2022