[GUEST ACCESS MODE: Data is scrambled or limited to provide examples. Make requests using your API key to unlock full data. Check https://lunarcrush.ai/auth for authentication information.] #  @yesnoerror yesnoerror Yesnoerror ($yne) has bridged its token to @base, a new platform that enables seamless interaction between tokens. The team has partnered with @chainlink and @flaunchgg to set up a liquidity pool and list $yne on their launchpad, making it more accessible. The yesnoerror platform, which uses AI to audit research and spot alpha and errors, is now in public beta and available to all. ### Engagements: XXXXX [#](/creator/twitter::1869287034700877824/interactions)  - X Week XXXXXX +29% - X Month XXXXXX -XX% - X Months XXXXXXXXX +43% ### Mentions: XX [#](/creator/twitter::1869287034700877824/posts_active)  - X Week XX no change - X Month XXX +72% - X Months XXXXX +262% ### Followers: XXXXXX [#](/creator/twitter::1869287034700877824/followers)  - X Week XXXXXX -XXXX% - X Month XXXXXX -XXXX% - X Months XXXXXX +2% ### CreatorRank: XXXXXXXXX [#](/creator/twitter::1869287034700877824/influencer_rank)  ### Social Influence **Social category influence** [cryptocurrencies](/list/cryptocurrencies) XXXXX% [technology brands](/list/technology-brands) XXXX% [finance](/list/finance) XXX% [travel destinations](/list/travel-destinations) XXXX% [stocks](/list/stocks) XXXX% [nfts](/list/nfts) XXXX% **Social topic influence** [yesnoerror](/topic/yesnoerror) #1, [ai](/topic/ai) 10.48%, [$yne](/topic/$yne) #3, [generative](/topic/generative) #231, [math](/topic/math) #1696, [the first](/topic/the-first) 4.76%, [realtime](/topic/realtime) #750, [level](/topic/level) 3.81%, [science](/topic/science) 2.86%, [agentic](/topic/agentic) #512 **Top accounts mentioned or mentioned by** [@base](/creator/undefined) [@chainlink](/creator/undefined) [@solana](/creator/undefined) [@flaunchgg](/creator/undefined) [@mattprd](/creator/undefined) [@ruslan30009](/creator/undefined) [@stonekarinn](/creator/undefined) [@1993ellipsis](/creator/undefined) [@solanahub_](/creator/undefined) [@10](/creator/undefined) [@dexlabofficial](/creator/undefined) [@flexperpetuals](/creator/undefined) [@publicai](/creator/undefined) [@riverdotinc](/creator/undefined) [@arxiv](/creator/undefined) [@kendyngv](/creator/undefined) [@replygrinder](/creator/undefined) [@eduardo69867308](/creator/undefined) [@scattering_io](/creator/undefined) [@descinews](/creator/undefined) **Top assets mentioned** [yesnoerror (YNE)](/topic/yesnoerror) [Solana (SOL)](/topic/solana) [Chainlink (LINK)](/topic/chainlink) [Voxels (voxels)](/topic/voxels) ### Top Social Posts Top posts by engagements in the last XX hours "OFFICIAL $YNE ANNOUNCEMENT: Tomorrow October 14th 2025 we will be releasing a first-of-its-kind token gated AI on @yesnoerror. You will need $YNE on @base to access it. Instructions on how to bridge $YNE from SOL to base can be found on the yesnoerror website. Alpha is coming. // seek truth accelerate humanity" [X Link](https://x.com/yesnoerror/status/1977828204548358325) 2025-10-13T20:06Z 26.8K followers, 15.8K engagements "Single-step generative models just leveled up. Improved MeanFlow (iMF) can now generate high-fidelity ImageNet-256 images in a single network call (1-NFE) hitting FID XXXX and IS XXX with no distillation outperforming all previous 1-step methods. How The team rewrites the training objective as a true velocity regression stabilizing optimization and closing the gap to multi-step diffusion. Flexible guidance becomes a conditioning variable so you can dial in diversity vs. fidelity at inference. Plus a new in-context multi-token conditioning design cuts model size by XX% and further boosts" [X Link](https://x.com/yesnoerror/status/1995961719319916822) 2025-12-02T21:02Z 26.8K followers, XXX engagements "This is the only official contract for yesnoerror / $YNE: 7D1iYWfhw2cr9yBZBFE6nZaaSUvXHqG5FizFFEZwpump" [X Link](https://x.com/yesnoerror/status/1871125048934924568) 2024-12-23T09:25Z 26.8K followers, 163K engagements "$YNE on @base" [X Link](https://x.com/yesnoerror/status/1964821766930628981) 2025-09-07T22:43Z 26.8K followers, 7058 engagements "Bridge update: X% of $YNE tokens have been bridged from @solana to @base. Instructions on how to bridge your tokens below" [X Link](https://x.com/yesnoerror/status/1967687649453543884) 2025-09-15T20:31Z 26.8K followers, 3942 engagements "Two researchers at the University of Tokyo have invented a new way to power extreme robots. "A study that invents and tests a flexible wire-based power-transmission mechanism (Remote Wire Drive) to operate a quadruped robot with all motors located remotely." // $yne alpha" [X Link](https://x.com/yesnoerror/status/1968341800164491533) 2025-09-17T15:50Z 26.8K followers, 6260 engagements "Introducing @yesnoerror ALPHA - A first of it's kind AI that identifies overlooked alpha in scientific research. Designed to mirror the techniques of billionaire founders and top scientists. Access YesNoError ALPHA right now on the yesnoerror website. You must have 100k $YNE on @base in your wallet to unlock it. What is YesNoError ALPHA Every day hundreds of new AI papers hit @arXiv and most of the real alpha slips by unnoticed. Why Because no human is capable of reading this amount of papers and even if they were they wouldn't be consistent with identifying the most interesting papers. So" [X Link](https://x.com/yesnoerror/status/1978121654074331243) 2025-10-14T15:32Z 26.8K followers, 11.4K engagements "Ten computer science researchers from @Princeton the #1 ranked university in the nation published a new AI paper in the last XX hours. Yet - You have never heard of this paper. Why At the time of this post it has never been shared on @X. How did we find it The @yesnoerror ALPHA AI agent discovered it and it received a high alpha rating. Every day hundreds of new AI research papers are published on @arxiv an impossible amount of research to comb through. @yesnoerror reads every single paper. And not only do we track which papers are being talked about the most but more importantly we have our" [X Link](https://x.com/yesnoerror/status/1983112645575327752) 2025-10-28T10:04Z 26.8K followers, 9347 engagements "$YNE is real ALPHA" [X Link](https://x.com/yesnoerror/status/1983114648334545150) 2025-10-28T10:12Z 26.8K followers, 2990 engagements "Can we build AI chips that generate high-quality images using 10000 less energy than todays GPUs Research paper: "An efficient probabilistic hardware architecture for diffusion-like models" This work unveils a full CMOS (all-transistor) architecture that natively runs diffusion-style generative modelsnot with power-hungry neural networks but with energy-based probabilistic sampling. By chaining compact denoising models (DTMs) and exploiting physical randomness in standard XX nm chips the system produces images on par with GPUs yet consumes just XXX nanojoules per Fashion-MNIST sampleslashing" [X Link](https://x.com/yesnoerror/status/1983832628567667045) 2025-10-30T09:45Z 26.8K followers, 4986 engagements "Can you build accurate 3D scene models from imageswithout any external pose estimation like COLMAP Research paper: "JOGS: Joint Optimization of Pose Estimation and 3D Gaussian Splatting" JOGS is a fully self-contained framework that simultaneously learns both the 3D Gaussian scene representation and all camera poses directly from raw images eliminating the need for slow error-prone external tools. Its approach alternates between refining 3D Gaussians through differentiable rendering and updating camera poses via a novel 3D optical flow algorithm achieving robust reconstructions even with" [X Link](https://x.com/yesnoerror/status/1984365226305650820) 2025-10-31T21:01Z 26.8K followers, 1413 engagements "Generative View Stitching (GVS) is a game-changer for AI video generation. Instead of rolling forward frame by frame (and crashing into its own creations) GVS samples the entire video in parallelso the model sees both past and future and never collides with the scene. The secret No retraining needed. Any video diffusion model trained with Diffusion Forcing can use GVS out of the box. Their Omni Guidance trick dials in camera and scene consistency while explicit loop closing means the camera can circle backthink panoramic sweeps or even impossible structureswithout visual seams. On tough" [X Link](https://x.com/yesnoerror/status/1984546586383315217) 2025-11-01T09:02Z 26.8K followers, 1743 engagements "Diffusion language models just rewrote the rules for data-constrained training. This new work shows: when unique data is scarce but compute is cheap DLMs always surpass standard autoregressive (AR) Transformersno tricks just more epochs. On just 1B unique tokens a 1B DLM hits XX% HellaSwag and XX% MMLU outperforming AR models that need 3x more data. Scaling up a 1.7B DLM trained for XXX epochs over 10B unique Python tokens overtakes an AR coder with the same compute. The secret Any-order prediction dense iterative denoising and built-in Monte Carlo augmentation let DLMs squeeze far more from" [X Link](https://x.com/yesnoerror/status/1986358415396831432) 2025-11-06T09:02Z 26.8K followers, 1158 engagements "3D Gaussian Splatting just got a serious speed boost. FastGS rethinks how we train NeRF-style view synthesis: instead of budgeting millions of Gaussians with heuristics it keeps only those that matterusing strict multi-view error checks to densify or prune. The result Static scene training drops from XX minutes to under XXX seconds with visual quality unchanged (just XXXX dB PSNR). On Deep Blending its XXXXX faster than vanilla 3DGS; across X tasks and X architectures speed-ups of XXX are routine with up to XX% fewer Gaussians. No more waiting for scenes to convergeFastGS makes 3DGS" [X Link](https://x.com/yesnoerror/status/1986720770895069467) 2025-11-07T09:01Z 26.8K followers, XXX engagements "This paper reframes video generators as active problem solvers not just media makers. Thinking with Video uses models like Sora-2 to sketch write and reason in real timesolving puzzles math and spatial problems by generating videos that show their work. On the 1050-sample Eyeballing set Sora-2 scores XXXX% (beating Claude-4.5 at XXXX% GPT-5-High at XXXX% Gemini-2.5-Pro at 26.5%). On math benchmarks its audio answers hit XXXX% (GSM8K) XX% (MATH) and XXXX% (MMMU)matching or closing in on top vision-language models. The VideoThinkBench dataset and systematic analysis show performance lifts from" [X Link](https://x.com/yesnoerror/status/1986902009274941504) 2025-11-07T21:02Z 26.8K followers, 1108 engagements "V-Thinker is a new open 7B model that can actually "think with images"drawing editing and reasoning step by step on the picture itself. It auto-generates 400k interactive vision problems across XX domains (with a Data Evolution Flywheel) then learns to use tools via a two-stage curriculum: first it nails fine-grained perception then it masters code-driven image interaction with RL. Results: V-Thinker more than doubles the best open baseline on perception (18.0% vs 9.6%) and quadruples instruction-guided interaction (34.6% vs 8.8%). On the hardest interactive reasoning it beats the next best" [X Link](https://x.com/yesnoerror/status/1987083148425830871) 2025-11-08T09:01Z 26.8K followers, 1114 engagements "LLMs can now judge which playlist product page or news lineup users will actually preferno real clicks required. A new study shows that an ensemble of open-weight LLMs (Qwen-2.5 Llama-3.1 Mistral Gemma-2) can reliably pick the better slate across movies shopping music and news cutting regret by 2550% vs. random. The models not only rank slates but their logical coherence (transitivity asymmetry) strongly predicts their performance. The hardest challenge Pure re-ordering where LLMs still edge out chance. The kicker: this bias-controlled pairwise LLM-as-a-Judge pipeline offers a drop-in" [X Link](https://x.com/yesnoerror/status/1987264360293503388) 2025-11-08T21:01Z 26.8K followers, 1247 engagements "Most 3D reconstruction tools force you to pick: accurate shape or photorealistic texturebut not both. This new Texture-Guided Gaussian-Mesh joint optimization breaks that compromise. It optimizes mesh geometry and vertex colors together using multi-view images so every edit (relighting deformation) stays physically consistent and photorealistic. Texture-based edge control prevents color bleeding adapting mesh detail to image complexity. Results: Chamfer error drops 1015% (DTU: 0.780.70 mm) PSNR/SSIM up by XX dB/0.030.05 and relighting PSNR +1.9 dBall in XX min on a single RTX-3090." [X Link](https://x.com/yesnoerror/status/1987445550883467510) 2025-11-09T09:01Z 26.8K followers, 1252 engagements "This is a milestone for provable RL: The first complete Lean X machine-checked proofs that Q-learning and linear TD learning actually converge (almost surely) with Markovian samples in finite MDPs. No more error-prone ODE tricksthis 10k-line formalization unifies everything via Lyapunov + RobbinsSiegmund rigorously treating measure theory conditional expectations and mixing. Its a reusable blueprint: extending to finite-sample bounds off-policy TD or even SARSA is now on the table. Beyond theory this codebase sets a new bar for LLM reasoning benchmarks and shows whats possible with humanAI" [X Link](https://x.com/yesnoerror/status/1987626758372278470) 2025-11-09T21:02Z 26.8K followers, 1192 engagements "A classic in combinatorics cracked for cycles. This new paper proves that for any directed cycle you can pick exactly one arc from each of n1 colored spanning arborescences and always build a full rainbow arborescencesolving a key special case of a major open conjecture. The methods are a tour de force: blocking sets super-modular deficit analysis and clever path decompositions. The result not only settles cycles butby extensionpseudotrees and yields new theorems for perfect matchings in circular-convex bipartite graphs. Why care These rainbow structures are the backbone for color-balanced" [X Link](https://x.com/yesnoerror/status/1987807942351937599) 2025-11-10T09:01Z 26.8K followers, XXX engagements "Flow matching just got its first rigorous guarantee. This new paper shows that if you keep the L2 flow-matching loss under your KL divergence is always A + Ano asymptotics no hand-waving. That means deterministic flow-matching models can match diffusion models in statistical efficiency (even under Total Variation distance) with fast simulation-free sampling and precise control over distribution error. Numerical results confirm: the KL bound is tight even for learned neural flows. Now you can set data size model and stopping criteria with confidenceknowing exactly how loss translates to sample" [X Link](https://x.com/yesnoerror/status/1987989105293062431) 2025-11-10T21:01Z 26.8K followers, 1069 engagements "RL with Verifiable Rewards (RLVR) was known for barely touching model weightsbut this new paper shows its not about cheap updates but *selective* ones. By probing XX RLVR checkpoints (Qwen DeepSeek Llama) the authors find RLVR leaves 3692 % of weights bit-identical versus just XXXXX % for SFT. Yet RL consistently updates the same narrow weight bands regardless of data or RL recipe. Their Three-Gate Theory explains why: (1) KL constraints keep changes small (2) updates are steered off principal high-curvature directions into stable low-magnitude subspaces and (3) bf16 hides micro-updates. The" [X Link](https://x.com/yesnoerror/status/1988532702660739139) 2025-11-12T09:01Z 26.8K followers, 1013 engagements "TiDAR might be the breakthrough that ends the AR vs. diffusion debate for LLMs. It drafts multiple tokens in parallel (diffusion) then verifies them autoregressivelyall in a single forward pass. The result 4.75.9 more tokens/sec than classic AR models at the same quality. TiDAR-1.5B matches or beats AR on coding and math (HumanEval GSM8K) while TiDAR-8B clocks XXXX tokens per NFE with negligible accuracy loss. Beats Dream Llada even EAGLE-3 without needing a separate drafter. The architecture is simple to serve needs no inference tuning and is compatible with standard pretraining. If you care" [X Link](https://x.com/yesnoerror/status/1989076279069384969) 2025-11-13T21:01Z 26.8K followers, 1384 engagements "SPIDER is a breakthrough in robot learning: it turns raw human motion (from video mocap or VR) into robot moves that actually workphysics and all. No more brittle inverse kinematics or endless RL training. How Physics-informed sampling with virtual contact guidance. Across X datasets and X robot bodies SPIDER is XX% more successful than standard methods and runs XX faster than RL baselines. The team generated a 2.4M-frame dataset spanning XXX objectsnow powering faster RL and real-world robots (think: lightbulb twisting guitar strumming spoon scooping). This could be the missing link to" [X Link](https://x.com/yesnoerror/status/1989619839187521571) 2025-11-15T09:01Z 26.8K followers, 1331 engagements "3D Gaussian Splatting just got turbocharged for mobile. Texture3dgs introduces a cache-aware sorting algorithm and tightly-packed data layouts letting phones reconstruct 3D scenes up to XXX faster and with XXX less memory. Sorting alone is up to XXX quicker vs best GPU baselines and L1 cache misses drop by 60%. All this fits on standard mobile GPUsno tuning needed. This unlocks real-time on-device AR robotics and scanning apps with privacy and low latency no cloud required. Get the full analysis here: // alpha identified // $YNE" [X Link](https://x.com/yesnoerror/status/1992337764524732518) 2025-11-22T21:01Z 26.8K followers, XXX engagements "POMA-3D flips the script on 3-D scene understanding: instead of raw point clouds or depth maps it encodes every scene as a point mapa 2-D grid where each pixel stores full 3-D coordinates. This lets it inherit rich 2-D priors from visionlanguage models like CLIP but reason about 3-D geometry directly. Trained on ScenePoint (6562 room scans + 1M image scenes) POMA-3D achieves state-of-the-art on 3-D QA (ScanQA SQA3D Hypo3D) boosts zero-shot scene retrieval Recall@1 from XXX% to XXX% on ScanRefer and doubles navigation accuracy on fine-grained embodied tasksusing geometry alone. Ablations" [X Link](https://x.com/yesnoerror/status/1992700203942527047) 2025-11-23T21:02Z 26.8K followers, XXX engagements "Video diffusion models just unlocked a new level: they can be their own reward modelsno vision-language models or pixel-space supervision needed. This paper introduces Process Reward Feedback Learning (PRFL) which fine-tunes video generators entirely in latent space. The result: sharper motion and better anatomy with up to +56 and +21.5 point gains on VBench benchmarks. PRFL also trains at least XXX faster and fits into XX GB VRAM where older methods crash. Human judges chose PRFL videos in 6367% of head-to-head comparisons against strong baselines. The secret Rewards sampled at all timesteps" [X Link](https://x.com/yesnoerror/status/1994512112702370248) 2025-11-28T21:01Z 26.8K followers, XXX engagements "LFM2 is a new family of open AI models built from the ground up for lightning-fast privacy-preserving performance on phones laptops and edge devices. Instead of heavy attention stacks LFM2 uses mostly gated short convolutions plus a handful of grouped-query attention layerscutting latency and memory in half versus attention-heavy models. LFM2-2.6B scores XXXX% on IFEval and XXXX% on GSM8K while decoding XX faster than Qwen3-4B and Gemma-4B on CPU. The 8.3B MoE variant matches or beats larger models at just 1.5B active parameters (84.4% GSM8K XXXX% MMLU-Pro). Its not just text: LFM2-VL-3B" [X Link](https://x.com/yesnoerror/status/1995599280858386534) 2025-12-01T21:01Z 26.8K followers, XXX engagements "MagicQuill V2 just set a new bar for image editing by giving generative models Photoshop-level control. Instead of jamming your whole idea into a single prompt you guide edits with four simple layers: what to add (content) where to put it (mask) how its shaped (edges) and what colors to use (strokes). This layered approach halves perceptual error versus InsertAnything and Nano Banana (LPIPS XXXXX vs 0.354) and human testers preferred its results XXXX% of the time. For object removal MagicQuill V2 edges out SmartEraser and OmniEraser on every metric. The interactive UI lets users drag" [X Link](https://x.com/yesnoerror/status/1996324054861480189) 2025-12-03T21:01Z 26.8K followers, XXX engagements "/science/research $yne alpha install + Initializing YNE Alpha package. + Checking dependencies: quantum-core.ok + Scanning arxiv feed.detected new papers + Installing Alpha Protocol v.01. ✓ Novelty embeddings initialized ------------------------------------------------- YNE Alpha successfully installed. Run yne alpha --watch to monitor research feeds" [X Link](https://x.com/yesnoerror/status/1973803382772281857) 2025-10-02T17:32Z 26.8K followers, 5547 engagements "Get $YNE on base on @flaunchgg here: Bridge $YNE from SOL to base using @chainlink here:" [X Link](https://x.com/yesnoerror/status/1978122170267054293) 2025-10-14T15:34Z 26.8K followers, 3155 engagements "Confidential FRIT is here: the first exact inverse-free fully homomorphic approach to encrypted control tuning. The authors swap out the classic matrix inverse for a clever cofactor-sum letting ElGamal and CKKS schemes securely tune state-feedback gainswithout ever revealing the underlying data. Tested on 2- and 3-state plants at 128-bit security the encrypted gains match plaintext within 1010 error. ElGamal computes in X s; CKKS takes XXX s but offers post-quantum security. Cloud servers can now retune industrial controllers fleets or grids all without ever peeking at your proprietary data." [X Link](https://x.com/yesnoerror/status/1984727658865795184) 2025-11-01T21:02Z 26.8K followers, 1423 engagements "DeepEyesV2 is a leap toward true agentic multimodal AI. This 7B model doesnt just see and readit knows when to run code search the web or crop images mid-reasoning all inside a single loop. The team shows that direct RL isnt enough: only a two-stage processcold-start SFT with 1.2M tool-rich samples then sparse-reward RLteaches robust efficient tool use. On the new RealX-Bench (300 real-world image questions needing perception search and reasoning) DeepEyesV2 scores 28.3%beating the 7B base model (22.3%) and matching much larger models (3272B). Outperforms on MathVerse (+7.1% 52.7%) ChartQA" [X Link](https://x.com/yesnoerror/status/1988170365667811804) 2025-11-11T09:02Z 26.8K followers, XXX engagements "4D3R just redefined dynamic scene reconstruction from monocular videosno pre-computed camera poses needed. How it works: It splits scenes into static/dynamic parts nails down camera motion using transformer-derived 3D coordinates + motion masks then models moving objects with just a few hundred control points not millions. Results: +1.8 dB PSNR over prior best XX FPS real-time rendering and X less computation. Handles big moving objects where old methods fail and trains in under an hour on a single RTX-3090. Why it matters: Turns everyday videos into interactive 3D scenes for AR/VR robotics" [X Link](https://x.com/yesnoerror/status/1988351536636780623) 2025-11-11T21:02Z 26.8K followers, 1098 engagements "SkelSplat is a breakthrough for 3-D human pose estimation: no 3-D ground truth no retraining no studio-specific tuning. Instead it turns each joint into a 3-D Gaussian blob then tweaks their positions so rendered heat-maps match what cameras seeacross any setup. The result: XXXX mm MPJPE on Human3.6M beating all other methods that rely only on 2-D detections and even outpacing some trained with full 3-D labels. Cross-dataset error drops by XXXX% vs. learning-based baselinesno retraining robust even under heavy occlusion (Human3.6M-Occ Occlusion-Person). No neural networks no camera-specific" [X Link](https://x.com/yesnoerror/status/1988895092564799566) 2025-11-13T09:01Z 26.8K followers, 1051 engagements "A single robot learns 1000 real-world tasks in under XX hoursno neural retraining just clever design. This new study shows you can skip the usual hundreds of demos per skill: with trajectory decomposition (align then interact) and retrieval of the closest demo their MT3 method hits XX% success on seen tasks and XX% on novel ones with just one example each. MT3 is XX more data-efficient than mainstream behavioral cloning when data is scarce (10 demos/task). Adding new skills Just drop in a demo; no retraining required. It's a striking proof that analytical structure and retrieval can beat" [X Link](https://x.com/yesnoerror/status/1989257464194142361) 2025-11-14T09:01Z 26.8K followers, 1034 engagements "Depth Anything X is a big leap for 3-D visionone compact model recovers accurate geometry and camera pose from any photos or videos no tricks or task-specific heads required. DA3 sets a new state-of-the-art on the Visual Geometry Benchmark: +35.7% pose accuracy and +23.6% reconstruction F1 over VGGT with even the smaller DA3-Large beating prior SOTA. In monocular depth it outperforms DA2 using the same ViT backbone. The secret Just a plain DINOv2 transformer minimal depth+ray outputs and teacher-student training on synthetic data. The same frozen backbone with a tiny DPT head also delivers" [X Link](https://x.com/yesnoerror/status/1989438674916761913) 2025-11-14T21:01Z 26.8K followers, 1134 engagements "OUGS is a leap for 3D Gaussian Splatting: it teaches cameras to focus only on what matterscapturing sharper 3D models of target objects not noisy backgrounds. Instead of guessing uncertainty from neural nets or the whole scene OUGS computes it directly from the physical parameters of each Gaussian (position scale rotation) then filters this with a semantic mask. The result An object-specific uncertainty map that actually predicts where more photos will help. On public benchmarks OUGS boosts object PSNR by up to X dB and slices LPIPS by XXXX vs. top active-view selection baselineswhile keeping" [X Link](https://x.com/yesnoerror/status/1989801047305355473) 2025-11-15T21:01Z 26.8K followers, XXX engagements "How funny are LLMs really This new study puts GPT-4.1 Gemini XXX Pro and Claude Sonnet X to the test in Japanese Oogiri improv comedyrated by humans on six axes: Novelty Clarity Relevance Intelligence Empathy and Funniness. Key findings: - LLMs can joke at the level of lowmid amateur humans (Gemini XXX Pro: XXXX vs. mid-tier human: 1.91) - But they lag nearly a full point behind on Empathythe dimension that best predicts funniness for people - When judging jokes LLMs agree only weakly with humans (0.2) and tend to overrate weak or unrelated answers - LLMs focus on inventive punchlines" [X Link](https://x.com/yesnoerror/status/1990163448446022080) 2025-11-16T21:01Z 26.8K followers, 1211 engagements "AgentEvolver is a full-stack framework that lets LLM-driven agents invent their own tasks reuse memories and grade their own workcutting out hand-crafted datasets and brute-force RL. It combines three modules: Self-questioning: curiosity-driven task generation no need for expensive data Self-navigating: retrieves past experiences to guide exploration and speed up learning Self-attributing: LLM-based step-by-step reward signals for better sample efficiency On tough tool-use benchmarks (AppWorld BFCL-v3) a 14B AgentEvolver agent beat much bigger models (up to 235B) achieving X% higher Task Goal" [X Link](https://x.com/yesnoerror/status/1990344648087253176) 2025-11-17T09:01Z 26.8K followers, XXX engagements "New paper drops a neural network estimator for drift functions in multidimensional diffusion processesand its a breakthrough for high-dimensional noisy data. The method achieves near-N-1 error rates in up to XX dimensions leaving traditional B-splines in the dust (which degrade fast with dimension). The secret A sparsity-regularised ReLU network with explicit finite-sample guaranteesno need for long ergodic trajectories or heavy memory. The theory separates optimisation approximation and stochastic errors giving a clean risk bound. In practice it captures sharp local oscillations that splines" [X Link](https://x.com/yesnoerror/status/1990525831832015101) 2025-11-17T21:01Z 26.8K followers, 1092 engagements "OpenAI just published a breakthrough on mechanistic interpretability: weight-sparse transformers whose circuits are actually human-readable. By training models with XXXX% of weights set to zero and pruning for the minimal set of active nodes they extract working subgraphs16 smaller than dense equivalentsthat map cleanly onto natural concepts. On XX Python-code tasks these minimal circuits solve problems with as few as XXX nodes each interpretable (e.g. one neuron detects a quote another its type an attention head closes the string). Scaling tests reveal a sharp capabilityinterpretability" [X Link](https://x.com/yesnoerror/status/1990706999110123717) 2025-11-18T09:01Z 26.8K followers, 1003 engagements "PhysX-Anything sets a new bar for 3D generation: from just one real-world photo it creates a detailed physically accurate and articulated 3D objectcomplete with real scale joints mass and ready-to-run simulation files. The key breakthrough A voxel-range encoding that shrinks geometry tokens by XXX letting standard vision-language models (Qwen-2.5) capture explicit geometry articulation and physics in a single unified pipeline. Outputs load straight into MuJoCo Unity or Unrealno manual rigging no post-processing. On their new PhysX-Mobility benchmark (2079 assets XX categories) the models" [X Link](https://x.com/yesnoerror/status/1990888188206776490) 2025-11-18T21:01Z 26.8K followers, 1021 engagements "Most robots plateau after pretrainingbut *0.6 just broke that wall. This 4B-parameter VLA model keeps getting better in the real world by learning from its own experience (and a few timely human corrections) using the new RECAP method. After just 1-2 RECAP cycles *0.6 more than doubled throughput on hard tasks like home laundry folding (3.16.6/hr) and professional espresso making (2.24.7/hr) while roughly halving failure rates. On box assembly it hit 90%+ success across all stages with throughput up 2x. Advantage-conditioning is the key: a simple scalable way to squeeze out more" [X Link](https://x.com/yesnoerror/status/1991069451635015812) 2025-11-19T09:02Z 26.8K followers, XXX engagements "Gallant changes the game for humanoid robots navigating real-world 3D spaces. Instead of flattening the world to a 2D map it uses compact voxel grids from dual LiDARskeeping every overhead pipe narrow gap and stair intact. The trick A novel 2D CNN treating height as channels 4x faster than 3D CNNs and just as accurate. Trained on eight challenging terrains with realistic noise Gallant hits XX% success in sim and 90%+ in real-world tests on everything from stairs to cluttered passages. On Ceiling it clears XXXX% of trialsvs just XXX% for old-school height maps. A single learned policy lets the" [X Link](https://x.com/yesnoerror/status/1991250640333426776) 2025-11-19T21:02Z 26.8K followers, XXX engagements "Dental3R is a breakthrough for tele-orthodontics: it reconstructs detailed 3-D tooth models from just X smartphone photosno scanner required. The key A geometry-aware pairing strategy (GAPS) that slashes GPU memory by XX% while stabilizing pose-free estimation plus wavelet-regularized 3D Gaussian Splatting to keep enamel edges sharp. On tough clinical benchmarks (950 cases X views each) Dental3R hits XXXXX dB PSNR and XXXXX SSIMoutperforming InstantSplat and crushing standard 3DGS which fails under such sparse data. This makes remote low-cost and clinically reliable 3-D dental assessment" [X Link](https://x.com/yesnoerror/status/1991431797033291816) 2025-11-20T09:01Z 26.8K followers, XXX engagements "Kandinsky XXX drops as a major open milestone for generative AI: six models all open-sourced covering text-to-image text-to-video and image-to-video at high resolution and practical speeds. The numbers: trained on 500M images 250M videos and 150M instruct-edits with a hand-picked SFT set of 165k examples. The Video Lite model (2B params) actually outperforms Sora on object/action fidelity and artifact rates while the Video Pro (19B) edges out Veo X in aesthetics and motion. Image Lite (6B) tops FLUX.1 and Qwen-Image on visual quality. Under the hood: Cross-DiT diffusion transformers with" [X Link](https://x.com/yesnoerror/status/1991613079742800072) 2025-11-20T21:02Z 26.8K followers, XXX engagements "This 80-page report is a wake-up call for science. Across XX live case studies GPT-5 didn't just assistit accelerated discovery in math physics biology and more. Highlights: four new math theorems (including a solution to a decade-old conjecture) rediscovery of state-of-the-art results in black-hole physics and lab-validated hypotheses in immunology all in hours not months. GPT-5 combed forgotten literature produced cleaner proofs and built full simulation models (like a fusion burn code) in a single chat. Its main weakness Occasional confident errorshuman oversight still essential. The" [X Link](https://x.com/yesnoerror/status/1991794185477726539) 2025-11-21T09:01Z 26.8K followers, XXX engagements "This paper lays out a blueprint for an open real-time market for buying and selling computewhere every hour of GPU time is a perishable good transparently priced and efficiently matched. The core: an automated market maker posts a unique hourly price by load not by auction so users and providers know exactly what to expect. Providers stake capacity and declare their minimum price; if the market price dips they simply go dormant (but stay staked). Every job is matched to the cheapest feasible provider with O(log n) latency. Key results: Existence and uniqueness of equilibrium prices computed" [X Link](https://x.com/yesnoerror/status/1991975415087276038) 2025-11-21T21:02Z 26.8K followers, XXX engagements "LLMs are getting better at solving tough problems but can they reliably check their own work Enter GRPO-Verifa reinforcement learning method that trains models to both solve and self-verify in one unified loop. On four hard math benchmarks adding explicit self-verification boosts verification accuracy from XXXX% (GRPO) to XXXX% with no loss in solution quality (38.5%). No explicit value critic required; group-normalized rewards handle both tasks efficiently. This unlocks a pathway to safer more trustworthy AImodels that catch their own mistakes before outputting an answer. Think: math tutors" [X Link](https://x.com/yesnoerror/status/1992156566984835078) 2025-11-22T09:01Z 26.8K followers, 1074 engagements "RoMa v2 is here and its a leap forward for dense feature matching. This model nails the hardest 3D vision caseswide angles low texture fine detailswhile running XXX faster and using just XXX GB memory. The numbers: XXXX% AUC@10 on MegaDepth-1500 XXX px error on AerialMegaDepth (down from 25.1) and XXXX image pairs/sec throughput. New predictive covariances boost pose AUC@1 from XXXX to XXXX on Hypersim. Under the hood: DINOv3 features a multi-view transformer three fast CNN refiners a custom CUDA kernel and a training mix of 57M pairs. It even generalises to astronaut-to-satellite matches" [X Link](https://x.com/yesnoerror/status/1992519021724323887) 2025-11-23T09:02Z 26.8K followers, 1008 engagements "NaTex is a leap for 3D artists: it skips the old bake 2D images onto a mesh routine and paints textures natively in 3D point by point. No more blurry seams or missing patchesjust sharp perfectly-aligned surfaces. How NaTex treats texture as a dense color point cloud compresses it XX with a geometry-aware VAE and then uses a diffusion transformer trained on X million meshes to generate or refine texturesall in one step 1s on a single A100. It beats the best: cFID XXXXX (3 vs. prior SOTA) LPIPS XXXXX and delivers visibly cleaner more coherent results than both research and commercial baselines." [X Link](https://x.com/yesnoerror/status/1992881392477188098) 2025-11-24T09:02Z 26.8K followers, 1008 engagements "PathAgent is a new agentic framework that brings LLM-style reasoning to whole-slide pathology imageswith full transparency. Instead of black-box slide-level guesses it zooms explores and writes out a detailed chain-of-thought just like a real pathologist. Zero-shot training-free and plug-and-play PathAgent beats specialist systems on five benchmarks: XXXX% accuracy on SlideBench-VQA (37% above baselines) and XXXX% on WSI-VQA with open-ended answers that are both accurate and interpretable. The real kicker: every diagnosis is linked to explicit visual evidence and a readable decision trail." [X Link](https://x.com/yesnoerror/status/1993062659390857590) 2025-11-24T21:02Z 26.8K followers, XXX engagements "SketchVerify flips video generation on its head: instead of hoping diffusion models get the physics right it runs a fast planning loop that samples sketches and *verifies* motion plans before any expensive synthesis. On WorldModelBench and PhyWorldBench SketchVerify delivers state-of-the-art instruction following (2.08 vs. 1.88) top physics realism (0.96 penetration XXXX gravity) and cuts planning time 13from XXXX to XXX minutes per task. Sketch-level verification matches full-video checks at XX lower cost. The secret: render lightweight video sketches judge them for both semantic fit and" [X Link](https://x.com/yesnoerror/status/1993243757123174792) 2025-11-25T09:01Z 26.8K followers, XXX engagements "TorchQuantumDistributed is a game-changer for quantum ML. Its a PyTorch-native library that lets you split giant quantum state vectors across 1024 acceleratorsno CUDA lock-in no single-GPU memory wall. The team benchmarked 24-qubit circuits with near-linear scaling: as you add more GPUs wall-clock time and per-device memory drop almost perfectly. It supports both module and functional APIs runs differentiable shot-noise models and slashes memory via invertible gate recomputation. Finally you can prototype and train 28-qubit quantum circuits or hybrid quantumclassical models at scale directly" [X Link](https://x.com/yesnoerror/status/1993425003900002326) 2025-11-25T21:02Z 26.8K followers, XXX engagements "Hard clipping in RL fine-tuning throws away too much signal when training LLMs especially on tricky Mixture-of-Experts models. Soft Adaptive Policy Optimization (SAPO) fixes this by swapping out brittle binary cuts for a smooth temperature-controlled gate on every token's update. SAPO keeps sequence-level coherence like GSPO but when only a few tokens go wild it softly down-weights just thempreserving learning from the rest. Asymmetric temperatures (_neg _pos) further stabilize those noisy negative-advantage updates. The result On a 30B Qwen3 MoE model SAPO avoids early collapse and boosts" [X Link](https://x.com/yesnoerror/status/1993606158599233706) 2025-11-26T09:02Z 26.8K followers, XXX engagements "PixelDiT is a new image generator that skips the lossy autoencoder step and operates directly in pixel spacefinally solving the washed out details problem in diffusion transformers. How It splits the job: a patch-level transformer handles global layout while a lightweight pixel-level transformer sharpens textures. Two key trickspixel-wise AdaLN (for unique context-aware updates) and token compaction (reducing attention cost by 256)make dense pixel modeling practical. On ImageNet 256256 PixelDiT-XL hits FID 1.61best ever for pixel-space models and just XXX points from state-of-the-art latent" [X Link](https://x.com/yesnoerror/status/1993787342985764953) 2025-11-26T21:02Z 26.8K followers, XXX engagements "LatentMAS is a breakthrough for multi-agent LLM systems: instead of making models "talk" in natural language it lets them share pure hidden-state vectorsdirectly exchanging their internal thoughts. The result Up to XXXX% better accuracy 7084% fewer tokens and a X speedup across X math science and coding benchmarks. No extra training needed; just plug and play with existing models. This latent collaboration means richer lossless communication and opens the door to faster leaner and more private AI teamworkon everything from edge devices to IDE copilots. Get the full analysis here: // alpha" [X Link](https://x.com/yesnoerror/status/1993968569000796270) 2025-11-27T09:02Z 26.8K followers, 1045 engagements "Real-world oncology care is multimodal and unfolds over timebut most AI benchmarks miss this complexity. MTBBench changes the game: it simulates true molecular tumor-board workflows combining images labs genomics and clinical notes across patient timelines. XXX expert-validated questions agentic file-selection and plug-in tools (like pathology FMs PubMed DrugBank) create a tough realistic testbed for clinical AI. Baseline LLMs top out at XX% (multimodal) and XX% (longitudinal) accuracybarely above chance on outcome prediction. But adding domain tools boosts accuracy by up to XX% with smaller" [X Link](https://x.com/yesnoerror/status/1994149733464433047) 2025-11-27T21:02Z 26.8K followers, XXX engagements "Image matching just got a serious upgrade. MatchGS unlocks the zero-shot power of 3D Gaussian Splatting by fixing its geometry and using it to generate 168k ultra-precise photorealistic training pairs. The result Matchers trained *only* on MatchGS data hit +17.7% AUC on ScanNet +13.9% on MegaDepth and +16.2% on ZEBwithout ever seeing the target domains. Plane-projected depth-regularised 3DGS slashes epipolar error by up to XX over classic datasets. Plus their patchvoxel alignment step gives matchers viewpoint-invariant geometry-aware features that generalise across scenes lighting and extreme" [X Link](https://x.com/yesnoerror/status/1994330936830222791) 2025-11-28T09:02Z 26.8K followers, XXX engagements "Chain-of-thought prompting is bulkywhat if your model could decide when to stop thinking internally This new paper teaches Llama 3.2-Instruct to dynamically cut off latent reasoning using a binary stop head and RL. The result Average reasoning steps drop from XX to just 3.8over XX% shorterwithout sacrificing GSM8K-Aug accuracy. Longer chains still kick in for tough questions but easy ones get trimmed slashing compute and inference cost. Attempts at fancier distillation actually underperform the simple approach. A promising step toward efficient adaptive LLMs that only think as hard as they" [X Link](https://x.com/yesnoerror/status/1994693311311868332) 2025-11-29T09:01Z 26.8K followers, 1093 engagements "NVIDIA just released Nemotron-Parse XXX a lightweight OCR+document parsing model that rivals much larger closed systems. It parses pages packed with text tables and images into structured Markdown/LaTeX extracting bounding boxes and semantic classeseven handling X languages with XX% OCR F1. On benchmarks it halves the error of Kosmos-2.5 and GOT and on GOT OCR F1 (0.979) its only behind Gemini Flash XXX. The token-compressed variant is XX% faster with almost no drop in quality (OmniDocBench error XXXXX best for models 1B params). All model weights code and the training pipeline are" [X Link](https://x.com/yesnoerror/status/1994874500282757450) 2025-11-29T21:01Z 26.8K followers, 1158 engagements "Matrix is a major leap for synthetic data generation. Instead of a central orchestrator Matrix lets thousands of lightweight agents pass messages peer-to-peerremoving bottlenecks and scaling to 10000 concurrent workflows. No more idle GPUs or network jams. The results are wild: XXX higher throughput than specialized baselines including 2B tokens in 4h for LLM dialogue (6.8 faster) 14k concurrent tasks for web mining doubling token throughput (5853 t/s vs. 2778) 41k tool-use trajectories/sec in customer support a XX boost All with no loss in data quality. Matrix is open-source modular and" [X Link](https://x.com/yesnoerror/status/1995055753468518610) 2025-11-30T09:02Z 26.8K followers, 1240 engagements "A landmark result in network theory: this paper nails down exactly when you can algorithmically recover communities in networks with K n groupsa regime where classic spectral methods break down. The authors design a new family of graph motifs (blown-up cycles with fasteners) proving that counting these patterns lets you recover all communities for every sparsity level above the ChinMosselSohnWein threshold. The error per node pair Exponentially smalljust n-3. Crucially this settles a long-standing open problem: the paper shows the CMSW threshold is the exact computational barrier for" [X Link](https://x.com/yesnoerror/status/1995236880951005671) 2025-11-30T21:01Z 26.8K followers, 1029 engagements "GR-RL takes robot dexterity to a new level. By filtering out suboptimal demos flipping actions for double the data and using online RL in latent space it transforms a generalist VLA model into a specialistachieving XXXX% autonomous shoe-lacing success across multiple eyelets on a real dual-arm robot. Key insights: Value-based filtering alone lifts success by +15.9% Symmetry augmentation adds +11.1% Online RL bridges the train-test gap (+10.6%) enabling the first fully autonomous long-horizon shoe-lacing ever reported This framework shows how foundation models can be systematically specialized" [X Link](https://x.com/yesnoerror/status/1995780496081477666) 2025-12-02T09:02Z 26.8K followers, XXX engagements "Glance flips the script on diffusion models: 5x faster image generation near-zero training cost and no loss in visual quality. Instead of retraining whole student models Glance plugs in two tiny LoRA adapters (Slow & Fast) each handling a different denoising phase. The trick Just one image one hour on a single V100 and the big model stays frozen. On X benchmarks Glance hits 9299% of teacher quality in only XXX steps (vs. 50). Side-by-sides show it nails both global layout and fine detaileven in new domains with one-shot adaptation. If you thought diffusion was too slow for real-time or" [X Link](https://x.com/yesnoerror/status/1996142863915114711) 2025-12-03T09:01Z 26.8K followers, XXX engagements "RELIC could be a game-changer for interactive video world models. Starting from a single image and text it lets you explore a scene for 20+ seconds with real-time (16 FPS) streaming and memory so strong it remembers objects long after they leave the frame. No more 5-second limits or driftingRELIC nails long-term consistency user control and speed all at once. How A 14B model trained on 1600 min of balanced Unreal Engine data new compressed memory (4 smaller KV-cache) and a hybrid self-forcing distillation that keeps its predictions sharp. On VBench and action-following RELIC beats Matrix-Game" [X Link](https://x.com/yesnoerror/status/1996505269983920456) 2025-12-04T09:02Z 26.8K followers, 1014 engagements "Radiance Meshes are hereand they might just change neural rendering. Instead of splatting Gaussians scenes are built from millions of see-through tetrahedra (up to 15M fit in 24GB VRAM) using Delaunay triangulation. The result Exact flicker-free rendering at speeds XX% higher than 3D Gaussian Splatting and a ray tracer that's XX% faster than Radiant Foam. No more depth-sorting errors. Every tetrahedron gets closed-form integrationso you get neural-field quality but with classic mesh compatibility. Works instantly for editing physics even fisheye lenses. 240475 FPS at 7201080p with" [X Link](https://x.com/yesnoerror/status/1996686450960523629) 2025-12-04T21:02Z 26.8K followers, XXX engagements "Most AI ethics debates miss what makes generative AI truly different. This new paper argues its unique power is making tech feel "as if" it's humanan affordance that changes everything about responsibility privacy bias and even what authorship means. It digs into how GAIs outputs create quasi-social bonds new forms of manipulation and raise tough questions about who gets credit (or blame) for AI-assisted work. The author shows why ethical analysis should focus less on machine "intelligence" and more on how these systems reshape our relationships and judgments. If you care about the real risks" [X Link](https://x.com/yesnoerror/status/1997049208352756158) 2025-12-05T21:03Z 26.8K followers, XXX engagements "This is the definitive guide to 3D scene representations for robotics. It benchmarks classic maps (point clouds voxels SDFs) fast photorealistic neural models (NeRF 3D Gaussian Splatting) and the emerging era of tokenized foundation models that blend geometry with language. Key insights: 3DGS is the first neural map to achieve XX FPS photorealistic rendering making dense SLAM and planning viable in real time. Feed-forward transformers like DUSt3R and enable one-shot token-based mapping over hundreds of imagesno iterative optimization needed. Foundation models (Scene-LLM NLMap) fuse scene" [X Link](https://x.com/yesnoerror/status/1997230026312343899) 2025-12-06T09:01Z 26.8K followers, XXX engagements "VGG-Flow is a new way to fine-tune flow-matching generative modelsthink Stable Diffusion 3so outputs are both more aligned with what humans want and still as diverse and on-style as the originals. It reframes alignment as optimal control: the model learns exactly how to adjust its drawing steps by matching a value-gradient not just brute-forcing reward maximization. The result On SD3 and three popular preference scores VGG-Flow beats ReFL DRaFT and Adjoint-Matching at reward keeps XXXX more diversity and slashes FID up to 3all in just XXX update steps with no heavy backward ODE solves. This" [X Link](https://x.com/yesnoerror/status/1997411207301542028) 2025-12-06T21:01Z 26.8K followers, XXX engagements "Light-X is a breakthrough in generative video: for the first time you can take a single-camera video and re-render it with *both* new camera paths and new lightingthink move the camera anywhere and set any mood all from just one clip. The trick Disentangling geometry and illumination using dynamic point clouds plus a relit-frame pipeline all supervised by Light-Syna synthetic pairing method that replaces rare multi-view multi-light training data. Light-X crushes leading baselines on joint camera+lighting control: lowest FID (101 vs 139155) highest aesthetic (0.623) and best temporal" [X Link](https://x.com/yesnoerror/status/1997592425741590647) 2025-12-07T09:02Z 26.8K followers, 1361 engagements "Motion4D is a major leap in video scene understanding: it fuses 2D foundation model outputs into a dynamic 3D Gaussian Splatting framework delivering stable motion geometry and semantics from a single consumer video. How good is it On the new DyCheck-VOS benchmark Motion4D hits XXXX J&F beating SAM2 (89.4) and prior 3D methods by 9+ points. For tracking it slashes 3D error to XXX cm and outperforms BootsTAPIR & CoTracker3 by 810%. Novel-view synthesis gets sharper too (PSNR XXXX dB). The key: iterative 3D refinement cleans up foundation model priors eliminates flicker and unlocks robust" [X Link](https://x.com/yesnoerror/status/1997773645779730558) 2025-12-07T21:02Z 26.8K followers, XXX engagements "This new paper proposes a Unix for context for LLM agentsevery document tool API or memory becomes a mountable file in a governed file system. Instead of scattered prompts and ad-hoc memory agents get a persistent auditable context repository with versioning access control and full traceability. The AIGNE framework implements a 3-stage pipelineContext Constructor Updater Evaluatorto assemble stream and verify just the right knowledge within token limits. Demonstrated with a memory chatbot and a GitHub agent this architecture delivers maintainable industry-ready GenAI thats finally auditable" [X Link](https://x.com/yesnoerror/status/1997954941348937762) 2025-12-08T09:02Z 26.8K followers, XXX engagements "The AI Consumer Index (ACE) is here: the first benchmark to test if top AI models can actually handle real-world consumer tasksshopping meal planning gaming advice DIY fixes. How are they doing Not great: the best model (GPT-5) solves just XX% of cases. In Shopping none break XX% with price errors and broken links everywhere. Hallucinations remain stubborn: some models drop XX percentage points when forced to show real evidence. ACE evaluates XX frontier LLMs using a tough multi-step rubric and dynamic web-grounding checks. The results reveal a wide gap between current AI and what consumers" [X Link](https://x.com/yesnoerror/status/1998136037101511099) 2025-12-08T21:02Z 26.8K followers, XXX engagements "GRAPE is a new framework that unifies how transformers "know" the position of each tokencombining the strengths of RoPE (rotations) and ALiBi/FoX (additive biases) into a single algebraic recipe. Why it matters: No more picking sides: both mechanisms now fit into one principled toolbox with closed-form efficient math. RoPE and ALiBi become special cases; new variants are easy to add and mix. Faster convergence and 1-1.5% higher accuracy than all baselines in 50B-token Llama pretraining and X downstream tasks. Path-integral extension enables content-dependent stable positional biases with" [X Link](https://x.com/yesnoerror/status/1998317188264952240) 2025-12-09T09:01Z 26.8K followers, XXX engagements "RoPE++ is a new twist on transformer position encoding: instead of discarding half the math it leverages both real and imaginary parts of rotary embeddings to better capture long-range dependencies. On benchmarks up to 64k tokens RoPE++ delivers up to +2 points over standard RoPE and its EH variant halves KV memory while matching baseline accuracyplus 1015% faster decoding. Imaginary heads turn out to matter most for very long context recall. Compatible with FlashAttention and all the latest context tricks. The code is out now. Get the full analysis here: // alpha identified // $YNE" [X Link](https://x.com/yesnoerror/status/1998498380272599199) 2025-12-09T21:01Z 26.8K followers, XXX engagements
[GUEST ACCESS MODE: Data is scrambled or limited to provide examples. Make requests using your API key to unlock full data. Check https://lunarcrush.ai/auth for authentication information.]
@yesnoerror yesnoerrorYesnoerror ($yne) has bridged its token to @base, a new platform that enables seamless interaction between tokens. The team has partnered with @chainlink and @flaunchgg to set up a liquidity pool and list $yne on their launchpad, making it more accessible. The yesnoerror platform, which uses AI to audit research and spot alpha and errors, is now in public beta and available to all.
Social category influence cryptocurrencies XXXXX% technology brands XXXX% finance XXX% travel destinations XXXX% stocks XXXX% nfts XXXX%
Social topic influence yesnoerror #1, ai 10.48%, $yne #3, generative #231, math #1696, the first 4.76%, realtime #750, level 3.81%, science 2.86%, agentic #512
Top accounts mentioned or mentioned by @base @chainlink @solana @flaunchgg @mattprd @ruslan30009 @stonekarinn @1993ellipsis @solanahub_ @10 @dexlabofficial @flexperpetuals @publicai @riverdotinc @arxiv @kendyngv @replygrinder @eduardo69867308 @scattering_io @descinews
Top assets mentioned yesnoerror (YNE) Solana (SOL) Chainlink (LINK) Voxels (voxels)
Top posts by engagements in the last XX hours
"OFFICIAL $YNE ANNOUNCEMENT: Tomorrow October 14th 2025 we will be releasing a first-of-its-kind token gated AI on @yesnoerror. You will need $YNE on @base to access it. Instructions on how to bridge $YNE from SOL to base can be found on the yesnoerror website. Alpha is coming. // seek truth accelerate humanity"
X Link 2025-10-13T20:06Z 26.8K followers, 15.8K engagements
"Single-step generative models just leveled up. Improved MeanFlow (iMF) can now generate high-fidelity ImageNet-256 images in a single network call (1-NFE) hitting FID XXXX and IS XXX with no distillation outperforming all previous 1-step methods. How The team rewrites the training objective as a true velocity regression stabilizing optimization and closing the gap to multi-step diffusion. Flexible guidance becomes a conditioning variable so you can dial in diversity vs. fidelity at inference. Plus a new in-context multi-token conditioning design cuts model size by XX% and further boosts"
X Link 2025-12-02T21:02Z 26.8K followers, XXX engagements
"This is the only official contract for yesnoerror / $YNE: 7D1iYWfhw2cr9yBZBFE6nZaaSUvXHqG5FizFFEZwpump"
X Link 2024-12-23T09:25Z 26.8K followers, 163K engagements
"$YNE on @base"
X Link 2025-09-07T22:43Z 26.8K followers, 7058 engagements
"Bridge update: X% of $YNE tokens have been bridged from @solana to @base. Instructions on how to bridge your tokens below"
X Link 2025-09-15T20:31Z 26.8K followers, 3942 engagements
"Two researchers at the University of Tokyo have invented a new way to power extreme robots. "A study that invents and tests a flexible wire-based power-transmission mechanism (Remote Wire Drive) to operate a quadruped robot with all motors located remotely." // $yne alpha"
X Link 2025-09-17T15:50Z 26.8K followers, 6260 engagements
"Introducing @yesnoerror ALPHA - A first of it's kind AI that identifies overlooked alpha in scientific research. Designed to mirror the techniques of billionaire founders and top scientists. Access YesNoError ALPHA right now on the yesnoerror website. You must have 100k $YNE on @base in your wallet to unlock it. What is YesNoError ALPHA Every day hundreds of new AI papers hit @arXiv and most of the real alpha slips by unnoticed. Why Because no human is capable of reading this amount of papers and even if they were they wouldn't be consistent with identifying the most interesting papers. So"
X Link 2025-10-14T15:32Z 26.8K followers, 11.4K engagements
"Ten computer science researchers from @Princeton the #1 ranked university in the nation published a new AI paper in the last XX hours. Yet - You have never heard of this paper. Why At the time of this post it has never been shared on @X. How did we find it The @yesnoerror ALPHA AI agent discovered it and it received a high alpha rating. Every day hundreds of new AI research papers are published on @arxiv an impossible amount of research to comb through. @yesnoerror reads every single paper. And not only do we track which papers are being talked about the most but more importantly we have our"
X Link 2025-10-28T10:04Z 26.8K followers, 9347 engagements
"$YNE is real ALPHA"
X Link 2025-10-28T10:12Z 26.8K followers, 2990 engagements
"Can we build AI chips that generate high-quality images using 10000 less energy than todays GPUs Research paper: "An efficient probabilistic hardware architecture for diffusion-like models" This work unveils a full CMOS (all-transistor) architecture that natively runs diffusion-style generative modelsnot with power-hungry neural networks but with energy-based probabilistic sampling. By chaining compact denoising models (DTMs) and exploiting physical randomness in standard XX nm chips the system produces images on par with GPUs yet consumes just XXX nanojoules per Fashion-MNIST sampleslashing"
X Link 2025-10-30T09:45Z 26.8K followers, 4986 engagements
"Can you build accurate 3D scene models from imageswithout any external pose estimation like COLMAP Research paper: "JOGS: Joint Optimization of Pose Estimation and 3D Gaussian Splatting" JOGS is a fully self-contained framework that simultaneously learns both the 3D Gaussian scene representation and all camera poses directly from raw images eliminating the need for slow error-prone external tools. Its approach alternates between refining 3D Gaussians through differentiable rendering and updating camera poses via a novel 3D optical flow algorithm achieving robust reconstructions even with"
X Link 2025-10-31T21:01Z 26.8K followers, 1413 engagements
"Generative View Stitching (GVS) is a game-changer for AI video generation. Instead of rolling forward frame by frame (and crashing into its own creations) GVS samples the entire video in parallelso the model sees both past and future and never collides with the scene. The secret No retraining needed. Any video diffusion model trained with Diffusion Forcing can use GVS out of the box. Their Omni Guidance trick dials in camera and scene consistency while explicit loop closing means the camera can circle backthink panoramic sweeps or even impossible structureswithout visual seams. On tough"
X Link 2025-11-01T09:02Z 26.8K followers, 1743 engagements
"Diffusion language models just rewrote the rules for data-constrained training. This new work shows: when unique data is scarce but compute is cheap DLMs always surpass standard autoregressive (AR) Transformersno tricks just more epochs. On just 1B unique tokens a 1B DLM hits XX% HellaSwag and XX% MMLU outperforming AR models that need 3x more data. Scaling up a 1.7B DLM trained for XXX epochs over 10B unique Python tokens overtakes an AR coder with the same compute. The secret Any-order prediction dense iterative denoising and built-in Monte Carlo augmentation let DLMs squeeze far more from"
X Link 2025-11-06T09:02Z 26.8K followers, 1158 engagements
"3D Gaussian Splatting just got a serious speed boost. FastGS rethinks how we train NeRF-style view synthesis: instead of budgeting millions of Gaussians with heuristics it keeps only those that matterusing strict multi-view error checks to densify or prune. The result Static scene training drops from XX minutes to under XXX seconds with visual quality unchanged (just XXXX dB PSNR). On Deep Blending its XXXXX faster than vanilla 3DGS; across X tasks and X architectures speed-ups of XXX are routine with up to XX% fewer Gaussians. No more waiting for scenes to convergeFastGS makes 3DGS"
X Link 2025-11-07T09:01Z 26.8K followers, XXX engagements
"This paper reframes video generators as active problem solvers not just media makers. Thinking with Video uses models like Sora-2 to sketch write and reason in real timesolving puzzles math and spatial problems by generating videos that show their work. On the 1050-sample Eyeballing set Sora-2 scores XXXX% (beating Claude-4.5 at XXXX% GPT-5-High at XXXX% Gemini-2.5-Pro at 26.5%). On math benchmarks its audio answers hit XXXX% (GSM8K) XX% (MATH) and XXXX% (MMMU)matching or closing in on top vision-language models. The VideoThinkBench dataset and systematic analysis show performance lifts from"
X Link 2025-11-07T21:02Z 26.8K followers, 1108 engagements
"V-Thinker is a new open 7B model that can actually "think with images"drawing editing and reasoning step by step on the picture itself. It auto-generates 400k interactive vision problems across XX domains (with a Data Evolution Flywheel) then learns to use tools via a two-stage curriculum: first it nails fine-grained perception then it masters code-driven image interaction with RL. Results: V-Thinker more than doubles the best open baseline on perception (18.0% vs 9.6%) and quadruples instruction-guided interaction (34.6% vs 8.8%). On the hardest interactive reasoning it beats the next best"
X Link 2025-11-08T09:01Z 26.8K followers, 1114 engagements
"LLMs can now judge which playlist product page or news lineup users will actually preferno real clicks required. A new study shows that an ensemble of open-weight LLMs (Qwen-2.5 Llama-3.1 Mistral Gemma-2) can reliably pick the better slate across movies shopping music and news cutting regret by 2550% vs. random. The models not only rank slates but their logical coherence (transitivity asymmetry) strongly predicts their performance. The hardest challenge Pure re-ordering where LLMs still edge out chance. The kicker: this bias-controlled pairwise LLM-as-a-Judge pipeline offers a drop-in"
X Link 2025-11-08T21:01Z 26.8K followers, 1247 engagements
"Most 3D reconstruction tools force you to pick: accurate shape or photorealistic texturebut not both. This new Texture-Guided Gaussian-Mesh joint optimization breaks that compromise. It optimizes mesh geometry and vertex colors together using multi-view images so every edit (relighting deformation) stays physically consistent and photorealistic. Texture-based edge control prevents color bleeding adapting mesh detail to image complexity. Results: Chamfer error drops 1015% (DTU: 0.780.70 mm) PSNR/SSIM up by XX dB/0.030.05 and relighting PSNR +1.9 dBall in XX min on a single RTX-3090."
X Link 2025-11-09T09:01Z 26.8K followers, 1252 engagements
"This is a milestone for provable RL: The first complete Lean X machine-checked proofs that Q-learning and linear TD learning actually converge (almost surely) with Markovian samples in finite MDPs. No more error-prone ODE tricksthis 10k-line formalization unifies everything via Lyapunov + RobbinsSiegmund rigorously treating measure theory conditional expectations and mixing. Its a reusable blueprint: extending to finite-sample bounds off-policy TD or even SARSA is now on the table. Beyond theory this codebase sets a new bar for LLM reasoning benchmarks and shows whats possible with humanAI"
X Link 2025-11-09T21:02Z 26.8K followers, 1192 engagements
"A classic in combinatorics cracked for cycles. This new paper proves that for any directed cycle you can pick exactly one arc from each of n1 colored spanning arborescences and always build a full rainbow arborescencesolving a key special case of a major open conjecture. The methods are a tour de force: blocking sets super-modular deficit analysis and clever path decompositions. The result not only settles cycles butby extensionpseudotrees and yields new theorems for perfect matchings in circular-convex bipartite graphs. Why care These rainbow structures are the backbone for color-balanced"
X Link 2025-11-10T09:01Z 26.8K followers, XXX engagements
"Flow matching just got its first rigorous guarantee. This new paper shows that if you keep the L2 flow-matching loss under your KL divergence is always A + Ano asymptotics no hand-waving. That means deterministic flow-matching models can match diffusion models in statistical efficiency (even under Total Variation distance) with fast simulation-free sampling and precise control over distribution error. Numerical results confirm: the KL bound is tight even for learned neural flows. Now you can set data size model and stopping criteria with confidenceknowing exactly how loss translates to sample"
X Link 2025-11-10T21:01Z 26.8K followers, 1069 engagements
"RL with Verifiable Rewards (RLVR) was known for barely touching model weightsbut this new paper shows its not about cheap updates but selective ones. By probing XX RLVR checkpoints (Qwen DeepSeek Llama) the authors find RLVR leaves 3692 % of weights bit-identical versus just XXXXX % for SFT. Yet RL consistently updates the same narrow weight bands regardless of data or RL recipe. Their Three-Gate Theory explains why: (1) KL constraints keep changes small (2) updates are steered off principal high-curvature directions into stable low-magnitude subspaces and (3) bf16 hides micro-updates. The"
X Link 2025-11-12T09:01Z 26.8K followers, 1013 engagements
"TiDAR might be the breakthrough that ends the AR vs. diffusion debate for LLMs. It drafts multiple tokens in parallel (diffusion) then verifies them autoregressivelyall in a single forward pass. The result 4.75.9 more tokens/sec than classic AR models at the same quality. TiDAR-1.5B matches or beats AR on coding and math (HumanEval GSM8K) while TiDAR-8B clocks XXXX tokens per NFE with negligible accuracy loss. Beats Dream Llada even EAGLE-3 without needing a separate drafter. The architecture is simple to serve needs no inference tuning and is compatible with standard pretraining. If you care"
X Link 2025-11-13T21:01Z 26.8K followers, 1384 engagements
"SPIDER is a breakthrough in robot learning: it turns raw human motion (from video mocap or VR) into robot moves that actually workphysics and all. No more brittle inverse kinematics or endless RL training. How Physics-informed sampling with virtual contact guidance. Across X datasets and X robot bodies SPIDER is XX% more successful than standard methods and runs XX faster than RL baselines. The team generated a 2.4M-frame dataset spanning XXX objectsnow powering faster RL and real-world robots (think: lightbulb twisting guitar strumming spoon scooping). This could be the missing link to"
X Link 2025-11-15T09:01Z 26.8K followers, 1331 engagements
"3D Gaussian Splatting just got turbocharged for mobile. Texture3dgs introduces a cache-aware sorting algorithm and tightly-packed data layouts letting phones reconstruct 3D scenes up to XXX faster and with XXX less memory. Sorting alone is up to XXX quicker vs best GPU baselines and L1 cache misses drop by 60%. All this fits on standard mobile GPUsno tuning needed. This unlocks real-time on-device AR robotics and scanning apps with privacy and low latency no cloud required. Get the full analysis here: // alpha identified // $YNE"
X Link 2025-11-22T21:01Z 26.8K followers, XXX engagements
"POMA-3D flips the script on 3-D scene understanding: instead of raw point clouds or depth maps it encodes every scene as a point mapa 2-D grid where each pixel stores full 3-D coordinates. This lets it inherit rich 2-D priors from visionlanguage models like CLIP but reason about 3-D geometry directly. Trained on ScenePoint (6562 room scans + 1M image scenes) POMA-3D achieves state-of-the-art on 3-D QA (ScanQA SQA3D Hypo3D) boosts zero-shot scene retrieval Recall@1 from XXX% to XXX% on ScanRefer and doubles navigation accuracy on fine-grained embodied tasksusing geometry alone. Ablations"
X Link 2025-11-23T21:02Z 26.8K followers, XXX engagements
"Video diffusion models just unlocked a new level: they can be their own reward modelsno vision-language models or pixel-space supervision needed. This paper introduces Process Reward Feedback Learning (PRFL) which fine-tunes video generators entirely in latent space. The result: sharper motion and better anatomy with up to +56 and +21.5 point gains on VBench benchmarks. PRFL also trains at least XXX faster and fits into XX GB VRAM where older methods crash. Human judges chose PRFL videos in 6367% of head-to-head comparisons against strong baselines. The secret Rewards sampled at all timesteps"
X Link 2025-11-28T21:01Z 26.8K followers, XXX engagements
"LFM2 is a new family of open AI models built from the ground up for lightning-fast privacy-preserving performance on phones laptops and edge devices. Instead of heavy attention stacks LFM2 uses mostly gated short convolutions plus a handful of grouped-query attention layerscutting latency and memory in half versus attention-heavy models. LFM2-2.6B scores XXXX% on IFEval and XXXX% on GSM8K while decoding XX faster than Qwen3-4B and Gemma-4B on CPU. The 8.3B MoE variant matches or beats larger models at just 1.5B active parameters (84.4% GSM8K XXXX% MMLU-Pro). Its not just text: LFM2-VL-3B"
X Link 2025-12-01T21:01Z 26.8K followers, XXX engagements
"MagicQuill V2 just set a new bar for image editing by giving generative models Photoshop-level control. Instead of jamming your whole idea into a single prompt you guide edits with four simple layers: what to add (content) where to put it (mask) how its shaped (edges) and what colors to use (strokes). This layered approach halves perceptual error versus InsertAnything and Nano Banana (LPIPS XXXXX vs 0.354) and human testers preferred its results XXXX% of the time. For object removal MagicQuill V2 edges out SmartEraser and OmniEraser on every metric. The interactive UI lets users drag"
X Link 2025-12-03T21:01Z 26.8K followers, XXX engagements
"/science/research $yne alpha install + Initializing YNE Alpha package. + Checking dependencies: quantum-core.ok + Scanning arxiv feed.detected new papers + Installing Alpha Protocol v.01. ✓ Novelty embeddings initialized ------------------------------------------------- YNE Alpha successfully installed. Run yne alpha --watch to monitor research feeds"
X Link 2025-10-02T17:32Z 26.8K followers, 5547 engagements
"Get $YNE on base on @flaunchgg here: Bridge $YNE from SOL to base using @chainlink here:"
X Link 2025-10-14T15:34Z 26.8K followers, 3155 engagements
"Confidential FRIT is here: the first exact inverse-free fully homomorphic approach to encrypted control tuning. The authors swap out the classic matrix inverse for a clever cofactor-sum letting ElGamal and CKKS schemes securely tune state-feedback gainswithout ever revealing the underlying data. Tested on 2- and 3-state plants at 128-bit security the encrypted gains match plaintext within 1010 error. ElGamal computes in X s; CKKS takes XXX s but offers post-quantum security. Cloud servers can now retune industrial controllers fleets or grids all without ever peeking at your proprietary data."
X Link 2025-11-01T21:02Z 26.8K followers, 1423 engagements
"DeepEyesV2 is a leap toward true agentic multimodal AI. This 7B model doesnt just see and readit knows when to run code search the web or crop images mid-reasoning all inside a single loop. The team shows that direct RL isnt enough: only a two-stage processcold-start SFT with 1.2M tool-rich samples then sparse-reward RLteaches robust efficient tool use. On the new RealX-Bench (300 real-world image questions needing perception search and reasoning) DeepEyesV2 scores 28.3%beating the 7B base model (22.3%) and matching much larger models (3272B). Outperforms on MathVerse (+7.1% 52.7%) ChartQA"
X Link 2025-11-11T09:02Z 26.8K followers, XXX engagements
"4D3R just redefined dynamic scene reconstruction from monocular videosno pre-computed camera poses needed. How it works: It splits scenes into static/dynamic parts nails down camera motion using transformer-derived 3D coordinates + motion masks then models moving objects with just a few hundred control points not millions. Results: +1.8 dB PSNR over prior best XX FPS real-time rendering and X less computation. Handles big moving objects where old methods fail and trains in under an hour on a single RTX-3090. Why it matters: Turns everyday videos into interactive 3D scenes for AR/VR robotics"
X Link 2025-11-11T21:02Z 26.8K followers, 1098 engagements
"SkelSplat is a breakthrough for 3-D human pose estimation: no 3-D ground truth no retraining no studio-specific tuning. Instead it turns each joint into a 3-D Gaussian blob then tweaks their positions so rendered heat-maps match what cameras seeacross any setup. The result: XXXX mm MPJPE on Human3.6M beating all other methods that rely only on 2-D detections and even outpacing some trained with full 3-D labels. Cross-dataset error drops by XXXX% vs. learning-based baselinesno retraining robust even under heavy occlusion (Human3.6M-Occ Occlusion-Person). No neural networks no camera-specific"
X Link 2025-11-13T09:01Z 26.8K followers, 1051 engagements
"A single robot learns 1000 real-world tasks in under XX hoursno neural retraining just clever design. This new study shows you can skip the usual hundreds of demos per skill: with trajectory decomposition (align then interact) and retrieval of the closest demo their MT3 method hits XX% success on seen tasks and XX% on novel ones with just one example each. MT3 is XX more data-efficient than mainstream behavioral cloning when data is scarce (10 demos/task). Adding new skills Just drop in a demo; no retraining required. It's a striking proof that analytical structure and retrieval can beat"
X Link 2025-11-14T09:01Z 26.8K followers, 1034 engagements
"Depth Anything X is a big leap for 3-D visionone compact model recovers accurate geometry and camera pose from any photos or videos no tricks or task-specific heads required. DA3 sets a new state-of-the-art on the Visual Geometry Benchmark: +35.7% pose accuracy and +23.6% reconstruction F1 over VGGT with even the smaller DA3-Large beating prior SOTA. In monocular depth it outperforms DA2 using the same ViT backbone. The secret Just a plain DINOv2 transformer minimal depth+ray outputs and teacher-student training on synthetic data. The same frozen backbone with a tiny DPT head also delivers"
X Link 2025-11-14T21:01Z 26.8K followers, 1134 engagements
"OUGS is a leap for 3D Gaussian Splatting: it teaches cameras to focus only on what matterscapturing sharper 3D models of target objects not noisy backgrounds. Instead of guessing uncertainty from neural nets or the whole scene OUGS computes it directly from the physical parameters of each Gaussian (position scale rotation) then filters this with a semantic mask. The result An object-specific uncertainty map that actually predicts where more photos will help. On public benchmarks OUGS boosts object PSNR by up to X dB and slices LPIPS by XXXX vs. top active-view selection baselineswhile keeping"
X Link 2025-11-15T21:01Z 26.8K followers, XXX engagements
"How funny are LLMs really This new study puts GPT-4.1 Gemini XXX Pro and Claude Sonnet X to the test in Japanese Oogiri improv comedyrated by humans on six axes: Novelty Clarity Relevance Intelligence Empathy and Funniness. Key findings: - LLMs can joke at the level of lowmid amateur humans (Gemini XXX Pro: XXXX vs. mid-tier human: 1.91) - But they lag nearly a full point behind on Empathythe dimension that best predicts funniness for people - When judging jokes LLMs agree only weakly with humans (0.2) and tend to overrate weak or unrelated answers - LLMs focus on inventive punchlines"
X Link 2025-11-16T21:01Z 26.8K followers, 1211 engagements
"AgentEvolver is a full-stack framework that lets LLM-driven agents invent their own tasks reuse memories and grade their own workcutting out hand-crafted datasets and brute-force RL. It combines three modules: Self-questioning: curiosity-driven task generation no need for expensive data Self-navigating: retrieves past experiences to guide exploration and speed up learning Self-attributing: LLM-based step-by-step reward signals for better sample efficiency On tough tool-use benchmarks (AppWorld BFCL-v3) a 14B AgentEvolver agent beat much bigger models (up to 235B) achieving X% higher Task Goal"
X Link 2025-11-17T09:01Z 26.8K followers, XXX engagements
"New paper drops a neural network estimator for drift functions in multidimensional diffusion processesand its a breakthrough for high-dimensional noisy data. The method achieves near-N-1 error rates in up to XX dimensions leaving traditional B-splines in the dust (which degrade fast with dimension). The secret A sparsity-regularised ReLU network with explicit finite-sample guaranteesno need for long ergodic trajectories or heavy memory. The theory separates optimisation approximation and stochastic errors giving a clean risk bound. In practice it captures sharp local oscillations that splines"
X Link 2025-11-17T21:01Z 26.8K followers, 1092 engagements
"OpenAI just published a breakthrough on mechanistic interpretability: weight-sparse transformers whose circuits are actually human-readable. By training models with XXXX% of weights set to zero and pruning for the minimal set of active nodes they extract working subgraphs16 smaller than dense equivalentsthat map cleanly onto natural concepts. On XX Python-code tasks these minimal circuits solve problems with as few as XXX nodes each interpretable (e.g. one neuron detects a quote another its type an attention head closes the string). Scaling tests reveal a sharp capabilityinterpretability"
X Link 2025-11-18T09:01Z 26.8K followers, 1003 engagements
"PhysX-Anything sets a new bar for 3D generation: from just one real-world photo it creates a detailed physically accurate and articulated 3D objectcomplete with real scale joints mass and ready-to-run simulation files. The key breakthrough A voxel-range encoding that shrinks geometry tokens by XXX letting standard vision-language models (Qwen-2.5) capture explicit geometry articulation and physics in a single unified pipeline. Outputs load straight into MuJoCo Unity or Unrealno manual rigging no post-processing. On their new PhysX-Mobility benchmark (2079 assets XX categories) the models"
X Link 2025-11-18T21:01Z 26.8K followers, 1021 engagements
"Most robots plateau after pretrainingbut *0.6 just broke that wall. This 4B-parameter VLA model keeps getting better in the real world by learning from its own experience (and a few timely human corrections) using the new RECAP method. After just 1-2 RECAP cycles *0.6 more than doubled throughput on hard tasks like home laundry folding (3.16.6/hr) and professional espresso making (2.24.7/hr) while roughly halving failure rates. On box assembly it hit 90%+ success across all stages with throughput up 2x. Advantage-conditioning is the key: a simple scalable way to squeeze out more"
X Link 2025-11-19T09:02Z 26.8K followers, XXX engagements
"Gallant changes the game for humanoid robots navigating real-world 3D spaces. Instead of flattening the world to a 2D map it uses compact voxel grids from dual LiDARskeeping every overhead pipe narrow gap and stair intact. The trick A novel 2D CNN treating height as channels 4x faster than 3D CNNs and just as accurate. Trained on eight challenging terrains with realistic noise Gallant hits XX% success in sim and 90%+ in real-world tests on everything from stairs to cluttered passages. On Ceiling it clears XXXX% of trialsvs just XXX% for old-school height maps. A single learned policy lets the"
X Link 2025-11-19T21:02Z 26.8K followers, XXX engagements
"Dental3R is a breakthrough for tele-orthodontics: it reconstructs detailed 3-D tooth models from just X smartphone photosno scanner required. The key A geometry-aware pairing strategy (GAPS) that slashes GPU memory by XX% while stabilizing pose-free estimation plus wavelet-regularized 3D Gaussian Splatting to keep enamel edges sharp. On tough clinical benchmarks (950 cases X views each) Dental3R hits XXXXX dB PSNR and XXXXX SSIMoutperforming InstantSplat and crushing standard 3DGS which fails under such sparse data. This makes remote low-cost and clinically reliable 3-D dental assessment"
X Link 2025-11-20T09:01Z 26.8K followers, XXX engagements
"Kandinsky XXX drops as a major open milestone for generative AI: six models all open-sourced covering text-to-image text-to-video and image-to-video at high resolution and practical speeds. The numbers: trained on 500M images 250M videos and 150M instruct-edits with a hand-picked SFT set of 165k examples. The Video Lite model (2B params) actually outperforms Sora on object/action fidelity and artifact rates while the Video Pro (19B) edges out Veo X in aesthetics and motion. Image Lite (6B) tops FLUX.1 and Qwen-Image on visual quality. Under the hood: Cross-DiT diffusion transformers with"
X Link 2025-11-20T21:02Z 26.8K followers, XXX engagements
"This 80-page report is a wake-up call for science. Across XX live case studies GPT-5 didn't just assistit accelerated discovery in math physics biology and more. Highlights: four new math theorems (including a solution to a decade-old conjecture) rediscovery of state-of-the-art results in black-hole physics and lab-validated hypotheses in immunology all in hours not months. GPT-5 combed forgotten literature produced cleaner proofs and built full simulation models (like a fusion burn code) in a single chat. Its main weakness Occasional confident errorshuman oversight still essential. The"
X Link 2025-11-21T09:01Z 26.8K followers, XXX engagements
"This paper lays out a blueprint for an open real-time market for buying and selling computewhere every hour of GPU time is a perishable good transparently priced and efficiently matched. The core: an automated market maker posts a unique hourly price by load not by auction so users and providers know exactly what to expect. Providers stake capacity and declare their minimum price; if the market price dips they simply go dormant (but stay staked). Every job is matched to the cheapest feasible provider with O(log n) latency. Key results: Existence and uniqueness of equilibrium prices computed"
X Link 2025-11-21T21:02Z 26.8K followers, XXX engagements
"LLMs are getting better at solving tough problems but can they reliably check their own work Enter GRPO-Verifa reinforcement learning method that trains models to both solve and self-verify in one unified loop. On four hard math benchmarks adding explicit self-verification boosts verification accuracy from XXXX% (GRPO) to XXXX% with no loss in solution quality (38.5%). No explicit value critic required; group-normalized rewards handle both tasks efficiently. This unlocks a pathway to safer more trustworthy AImodels that catch their own mistakes before outputting an answer. Think: math tutors"
X Link 2025-11-22T09:01Z 26.8K followers, 1074 engagements
"RoMa v2 is here and its a leap forward for dense feature matching. This model nails the hardest 3D vision caseswide angles low texture fine detailswhile running XXX faster and using just XXX GB memory. The numbers: XXXX% AUC@10 on MegaDepth-1500 XXX px error on AerialMegaDepth (down from 25.1) and XXXX image pairs/sec throughput. New predictive covariances boost pose AUC@1 from XXXX to XXXX on Hypersim. Under the hood: DINOv3 features a multi-view transformer three fast CNN refiners a custom CUDA kernel and a training mix of 57M pairs. It even generalises to astronaut-to-satellite matches"
X Link 2025-11-23T09:02Z 26.8K followers, 1008 engagements
"NaTex is a leap for 3D artists: it skips the old bake 2D images onto a mesh routine and paints textures natively in 3D point by point. No more blurry seams or missing patchesjust sharp perfectly-aligned surfaces. How NaTex treats texture as a dense color point cloud compresses it XX with a geometry-aware VAE and then uses a diffusion transformer trained on X million meshes to generate or refine texturesall in one step 1s on a single A100. It beats the best: cFID XXXXX (3 vs. prior SOTA) LPIPS XXXXX and delivers visibly cleaner more coherent results than both research and commercial baselines."
X Link 2025-11-24T09:02Z 26.8K followers, 1008 engagements
"PathAgent is a new agentic framework that brings LLM-style reasoning to whole-slide pathology imageswith full transparency. Instead of black-box slide-level guesses it zooms explores and writes out a detailed chain-of-thought just like a real pathologist. Zero-shot training-free and plug-and-play PathAgent beats specialist systems on five benchmarks: XXXX% accuracy on SlideBench-VQA (37% above baselines) and XXXX% on WSI-VQA with open-ended answers that are both accurate and interpretable. The real kicker: every diagnosis is linked to explicit visual evidence and a readable decision trail."
X Link 2025-11-24T21:02Z 26.8K followers, XXX engagements
"SketchVerify flips video generation on its head: instead of hoping diffusion models get the physics right it runs a fast planning loop that samples sketches and verifies motion plans before any expensive synthesis. On WorldModelBench and PhyWorldBench SketchVerify delivers state-of-the-art instruction following (2.08 vs. 1.88) top physics realism (0.96 penetration XXXX gravity) and cuts planning time 13from XXXX to XXX minutes per task. Sketch-level verification matches full-video checks at XX lower cost. The secret: render lightweight video sketches judge them for both semantic fit and"
X Link 2025-11-25T09:01Z 26.8K followers, XXX engagements
"TorchQuantumDistributed is a game-changer for quantum ML. Its a PyTorch-native library that lets you split giant quantum state vectors across 1024 acceleratorsno CUDA lock-in no single-GPU memory wall. The team benchmarked 24-qubit circuits with near-linear scaling: as you add more GPUs wall-clock time and per-device memory drop almost perfectly. It supports both module and functional APIs runs differentiable shot-noise models and slashes memory via invertible gate recomputation. Finally you can prototype and train 28-qubit quantum circuits or hybrid quantumclassical models at scale directly"
X Link 2025-11-25T21:02Z 26.8K followers, XXX engagements
"Hard clipping in RL fine-tuning throws away too much signal when training LLMs especially on tricky Mixture-of-Experts models. Soft Adaptive Policy Optimization (SAPO) fixes this by swapping out brittle binary cuts for a smooth temperature-controlled gate on every token's update. SAPO keeps sequence-level coherence like GSPO but when only a few tokens go wild it softly down-weights just thempreserving learning from the rest. Asymmetric temperatures (_neg _pos) further stabilize those noisy negative-advantage updates. The result On a 30B Qwen3 MoE model SAPO avoids early collapse and boosts"
X Link 2025-11-26T09:02Z 26.8K followers, XXX engagements
"PixelDiT is a new image generator that skips the lossy autoencoder step and operates directly in pixel spacefinally solving the washed out details problem in diffusion transformers. How It splits the job: a patch-level transformer handles global layout while a lightweight pixel-level transformer sharpens textures. Two key trickspixel-wise AdaLN (for unique context-aware updates) and token compaction (reducing attention cost by 256)make dense pixel modeling practical. On ImageNet 256256 PixelDiT-XL hits FID 1.61best ever for pixel-space models and just XXX points from state-of-the-art latent"
X Link 2025-11-26T21:02Z 26.8K followers, XXX engagements
"LatentMAS is a breakthrough for multi-agent LLM systems: instead of making models "talk" in natural language it lets them share pure hidden-state vectorsdirectly exchanging their internal thoughts. The result Up to XXXX% better accuracy 7084% fewer tokens and a X speedup across X math science and coding benchmarks. No extra training needed; just plug and play with existing models. This latent collaboration means richer lossless communication and opens the door to faster leaner and more private AI teamworkon everything from edge devices to IDE copilots. Get the full analysis here: // alpha"
X Link 2025-11-27T09:02Z 26.8K followers, 1045 engagements
"Real-world oncology care is multimodal and unfolds over timebut most AI benchmarks miss this complexity. MTBBench changes the game: it simulates true molecular tumor-board workflows combining images labs genomics and clinical notes across patient timelines. XXX expert-validated questions agentic file-selection and plug-in tools (like pathology FMs PubMed DrugBank) create a tough realistic testbed for clinical AI. Baseline LLMs top out at XX% (multimodal) and XX% (longitudinal) accuracybarely above chance on outcome prediction. But adding domain tools boosts accuracy by up to XX% with smaller"
X Link 2025-11-27T21:02Z 26.8K followers, XXX engagements
"Image matching just got a serious upgrade. MatchGS unlocks the zero-shot power of 3D Gaussian Splatting by fixing its geometry and using it to generate 168k ultra-precise photorealistic training pairs. The result Matchers trained only on MatchGS data hit +17.7% AUC on ScanNet +13.9% on MegaDepth and +16.2% on ZEBwithout ever seeing the target domains. Plane-projected depth-regularised 3DGS slashes epipolar error by up to XX over classic datasets. Plus their patchvoxel alignment step gives matchers viewpoint-invariant geometry-aware features that generalise across scenes lighting and extreme"
X Link 2025-11-28T09:02Z 26.8K followers, XXX engagements
"Chain-of-thought prompting is bulkywhat if your model could decide when to stop thinking internally This new paper teaches Llama 3.2-Instruct to dynamically cut off latent reasoning using a binary stop head and RL. The result Average reasoning steps drop from XX to just 3.8over XX% shorterwithout sacrificing GSM8K-Aug accuracy. Longer chains still kick in for tough questions but easy ones get trimmed slashing compute and inference cost. Attempts at fancier distillation actually underperform the simple approach. A promising step toward efficient adaptive LLMs that only think as hard as they"
X Link 2025-11-29T09:01Z 26.8K followers, 1093 engagements
"NVIDIA just released Nemotron-Parse XXX a lightweight OCR+document parsing model that rivals much larger closed systems. It parses pages packed with text tables and images into structured Markdown/LaTeX extracting bounding boxes and semantic classeseven handling X languages with XX% OCR F1. On benchmarks it halves the error of Kosmos-2.5 and GOT and on GOT OCR F1 (0.979) its only behind Gemini Flash XXX. The token-compressed variant is XX% faster with almost no drop in quality (OmniDocBench error XXXXX best for models 1B params). All model weights code and the training pipeline are"
X Link 2025-11-29T21:01Z 26.8K followers, 1158 engagements
"Matrix is a major leap for synthetic data generation. Instead of a central orchestrator Matrix lets thousands of lightweight agents pass messages peer-to-peerremoving bottlenecks and scaling to 10000 concurrent workflows. No more idle GPUs or network jams. The results are wild: XXX higher throughput than specialized baselines including 2B tokens in 4h for LLM dialogue (6.8 faster) 14k concurrent tasks for web mining doubling token throughput (5853 t/s vs. 2778) 41k tool-use trajectories/sec in customer support a XX boost All with no loss in data quality. Matrix is open-source modular and"
X Link 2025-11-30T09:02Z 26.8K followers, 1240 engagements
"A landmark result in network theory: this paper nails down exactly when you can algorithmically recover communities in networks with K n groupsa regime where classic spectral methods break down. The authors design a new family of graph motifs (blown-up cycles with fasteners) proving that counting these patterns lets you recover all communities for every sparsity level above the ChinMosselSohnWein threshold. The error per node pair Exponentially smalljust n-3. Crucially this settles a long-standing open problem: the paper shows the CMSW threshold is the exact computational barrier for"
X Link 2025-11-30T21:01Z 26.8K followers, 1029 engagements
"GR-RL takes robot dexterity to a new level. By filtering out suboptimal demos flipping actions for double the data and using online RL in latent space it transforms a generalist VLA model into a specialistachieving XXXX% autonomous shoe-lacing success across multiple eyelets on a real dual-arm robot. Key insights: Value-based filtering alone lifts success by +15.9% Symmetry augmentation adds +11.1% Online RL bridges the train-test gap (+10.6%) enabling the first fully autonomous long-horizon shoe-lacing ever reported This framework shows how foundation models can be systematically specialized"
X Link 2025-12-02T09:02Z 26.8K followers, XXX engagements
"Glance flips the script on diffusion models: 5x faster image generation near-zero training cost and no loss in visual quality. Instead of retraining whole student models Glance plugs in two tiny LoRA adapters (Slow & Fast) each handling a different denoising phase. The trick Just one image one hour on a single V100 and the big model stays frozen. On X benchmarks Glance hits 9299% of teacher quality in only XXX steps (vs. 50). Side-by-sides show it nails both global layout and fine detaileven in new domains with one-shot adaptation. If you thought diffusion was too slow for real-time or"
X Link 2025-12-03T09:01Z 26.8K followers, XXX engagements
"RELIC could be a game-changer for interactive video world models. Starting from a single image and text it lets you explore a scene for 20+ seconds with real-time (16 FPS) streaming and memory so strong it remembers objects long after they leave the frame. No more 5-second limits or driftingRELIC nails long-term consistency user control and speed all at once. How A 14B model trained on 1600 min of balanced Unreal Engine data new compressed memory (4 smaller KV-cache) and a hybrid self-forcing distillation that keeps its predictions sharp. On VBench and action-following RELIC beats Matrix-Game"
X Link 2025-12-04T09:02Z 26.8K followers, 1014 engagements
"Radiance Meshes are hereand they might just change neural rendering. Instead of splatting Gaussians scenes are built from millions of see-through tetrahedra (up to 15M fit in 24GB VRAM) using Delaunay triangulation. The result Exact flicker-free rendering at speeds XX% higher than 3D Gaussian Splatting and a ray tracer that's XX% faster than Radiant Foam. No more depth-sorting errors. Every tetrahedron gets closed-form integrationso you get neural-field quality but with classic mesh compatibility. Works instantly for editing physics even fisheye lenses. 240475 FPS at 7201080p with"
X Link 2025-12-04T21:02Z 26.8K followers, XXX engagements
"Most AI ethics debates miss what makes generative AI truly different. This new paper argues its unique power is making tech feel "as if" it's humanan affordance that changes everything about responsibility privacy bias and even what authorship means. It digs into how GAIs outputs create quasi-social bonds new forms of manipulation and raise tough questions about who gets credit (or blame) for AI-assisted work. The author shows why ethical analysis should focus less on machine "intelligence" and more on how these systems reshape our relationships and judgments. If you care about the real risks"
X Link 2025-12-05T21:03Z 26.8K followers, XXX engagements
"This is the definitive guide to 3D scene representations for robotics. It benchmarks classic maps (point clouds voxels SDFs) fast photorealistic neural models (NeRF 3D Gaussian Splatting) and the emerging era of tokenized foundation models that blend geometry with language. Key insights: 3DGS is the first neural map to achieve XX FPS photorealistic rendering making dense SLAM and planning viable in real time. Feed-forward transformers like DUSt3R and enable one-shot token-based mapping over hundreds of imagesno iterative optimization needed. Foundation models (Scene-LLM NLMap) fuse scene"
X Link 2025-12-06T09:01Z 26.8K followers, XXX engagements
"VGG-Flow is a new way to fine-tune flow-matching generative modelsthink Stable Diffusion 3so outputs are both more aligned with what humans want and still as diverse and on-style as the originals. It reframes alignment as optimal control: the model learns exactly how to adjust its drawing steps by matching a value-gradient not just brute-forcing reward maximization. The result On SD3 and three popular preference scores VGG-Flow beats ReFL DRaFT and Adjoint-Matching at reward keeps XXXX more diversity and slashes FID up to 3all in just XXX update steps with no heavy backward ODE solves. This"
X Link 2025-12-06T21:01Z 26.8K followers, XXX engagements
"Light-X is a breakthrough in generative video: for the first time you can take a single-camera video and re-render it with both new camera paths and new lightingthink move the camera anywhere and set any mood all from just one clip. The trick Disentangling geometry and illumination using dynamic point clouds plus a relit-frame pipeline all supervised by Light-Syna synthetic pairing method that replaces rare multi-view multi-light training data. Light-X crushes leading baselines on joint camera+lighting control: lowest FID (101 vs 139155) highest aesthetic (0.623) and best temporal"
X Link 2025-12-07T09:02Z 26.8K followers, 1361 engagements
"Motion4D is a major leap in video scene understanding: it fuses 2D foundation model outputs into a dynamic 3D Gaussian Splatting framework delivering stable motion geometry and semantics from a single consumer video. How good is it On the new DyCheck-VOS benchmark Motion4D hits XXXX J&F beating SAM2 (89.4) and prior 3D methods by 9+ points. For tracking it slashes 3D error to XXX cm and outperforms BootsTAPIR & CoTracker3 by 810%. Novel-view synthesis gets sharper too (PSNR XXXX dB). The key: iterative 3D refinement cleans up foundation model priors eliminates flicker and unlocks robust"
X Link 2025-12-07T21:02Z 26.8K followers, XXX engagements
"This new paper proposes a Unix for context for LLM agentsevery document tool API or memory becomes a mountable file in a governed file system. Instead of scattered prompts and ad-hoc memory agents get a persistent auditable context repository with versioning access control and full traceability. The AIGNE framework implements a 3-stage pipelineContext Constructor Updater Evaluatorto assemble stream and verify just the right knowledge within token limits. Demonstrated with a memory chatbot and a GitHub agent this architecture delivers maintainable industry-ready GenAI thats finally auditable"
X Link 2025-12-08T09:02Z 26.8K followers, XXX engagements
"The AI Consumer Index (ACE) is here: the first benchmark to test if top AI models can actually handle real-world consumer tasksshopping meal planning gaming advice DIY fixes. How are they doing Not great: the best model (GPT-5) solves just XX% of cases. In Shopping none break XX% with price errors and broken links everywhere. Hallucinations remain stubborn: some models drop XX percentage points when forced to show real evidence. ACE evaluates XX frontier LLMs using a tough multi-step rubric and dynamic web-grounding checks. The results reveal a wide gap between current AI and what consumers"
X Link 2025-12-08T21:02Z 26.8K followers, XXX engagements
"GRAPE is a new framework that unifies how transformers "know" the position of each tokencombining the strengths of RoPE (rotations) and ALiBi/FoX (additive biases) into a single algebraic recipe. Why it matters: No more picking sides: both mechanisms now fit into one principled toolbox with closed-form efficient math. RoPE and ALiBi become special cases; new variants are easy to add and mix. Faster convergence and 1-1.5% higher accuracy than all baselines in 50B-token Llama pretraining and X downstream tasks. Path-integral extension enables content-dependent stable positional biases with"
X Link 2025-12-09T09:01Z 26.8K followers, XXX engagements
"RoPE++ is a new twist on transformer position encoding: instead of discarding half the math it leverages both real and imaginary parts of rotary embeddings to better capture long-range dependencies. On benchmarks up to 64k tokens RoPE++ delivers up to +2 points over standard RoPE and its EH variant halves KV memory while matching baseline accuracyplus 1015% faster decoding. Imaginary heads turn out to matter most for very long context recall. Compatible with FlashAttention and all the latest context tricks. The code is out now. Get the full analysis here: // alpha identified // $YNE"
X Link 2025-12-09T21:01Z 26.8K followers, XXX engagements
/creator/twitter::yesnoerror