[GUEST ACCESS MODE: Data is scrambled or limited to provide examples. Make requests using your API key to unlock full data. Check https://lunarcrush.ai/auth for authentication information.]

@rohanpaul_ai Avatar @rohanpaul_ai Rohan Paul

Rohan Paul posts on X about llm, open ai, gpu, agi the most. They currently have XXXXXX followers and 4097 posts still getting attention that total XXXXXXX engagements in the last XX hours.

Engagements: XXXXXXX #

Engagements Line Chart

Mentions: XXX #

Mentions Line Chart

Followers: XXXXXX #

Followers Line Chart

CreatorRank: XXXXXX #

CreatorRank Line Chart

Social Influence #


Social category influence technology brands #2202 finance #524 stocks #5906 countries social networks travel destinations #3100 celebrities vc firms exchanges

Social topic influence llm #2, open ai #12, gpu #69, agi #47, china #533, grok 4, token, context window, generative, chips #301

Top assets mentioned Alphabet Inc Class A (GOOGL) Goldman Sachs (GS) ServiceNow Inc (NOW) New York Times Co. (NYT) Hundred Million (100M) Microsoft Corp. (MSFT)

Top Social Posts #


Top posts by engagements in the last XX hours

"What role does Chain-of-Thought (CoT) prompting play in adjusting the AI models neural activity for different types of questions"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-29 20:14:29 UTC 75.9K followers, 2078 engagements

"this story is going wildy viral on reddit. ChatGPT flagged a hidden gene defect that doctors missed for a decade. ChatGPT ingested the patients MRI CT broad lab panels and years of unexplained symptoms. It noticed that normal serum B12 clashed with nerve pain and fatigue hinting at a methylation block. Within months tingling eased and brain fog cleared. The primary physician reviewed the genetics report and agreed the variant unified the entire case. IMO time has already come taking a 2nd opinion from the best healthcare-AI model should be made part of medical code of practice. ------ reddit."
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-05 02:20:58 UTC 75.9K followers, 1.4M engagements

"πŸ€– Jensen Huang tells todays 20-year-olds to major in physical sciences instead of pure software because the next AI wave will need physics-savvy talent. He predicts Physical AI will move reasoning agents into real robots easing severe factory labor gaps; Nvidia already proves the payoff with a 4T market value. Generative AI added language and image synthesis by learning statistical links between symbols. Reasoning AI built on that letting agents plan and troubleshoot even when data are missing. All X stages happen in purely digital worlds where gravity never pulls and surfaces never slip."
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-20 04:45:50 UTC 75.6K followers, 58.4K engagements

"Github πŸ€–: Open-source GenBI AI Agent to chat with data to generate Text-to-SQL charts spreadsheets reports and BI. πŸ“Š Helps you chat with data to generate SQL charts and reports using your choice of LLM. It provides an open-source GenBI solution for data-driven teams seeking insights without code. - Generates Text-to-SQL queries charts spreadsheets reports and BI insights. - supports multiple LLMs including OpenAI Azure OpenAI DeepSeek Google Gemini Vertex AI Bedrock Anthropic Groq Ollama and Databricks. - ask data questions in multiple languages and provides AI-generated summaries and"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-19 15:18:30 UTC 75.8K followers, 47.5K engagements

"Tencent HunyuanWorld-1.0's generation architecture integrates panoramic proxy generation semantic layering and hierarchical 3D reconstruction to achieve high-quality scene-scale XXX 3D world generation supporting both text and image inputs"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-27 02:45:41 UTC 75.6K followers, XXX engagements

"xAI just launched beta version of video generation inside Grok Valentine/Ani"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-28 21:28:02 UTC 75.8K followers, 2533 engagements

"Recruiters face XX% surge in AI-crafted rsums. hitting 11000 submissions per minute. Many rsums now mirror job-description keywords from simple ChatGPT prompts. AI agents auto-apply on behalf of candidates forcing firms into an AI vs AI screening arms race"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-06-24 04:08:56 UTC 75.9K followers, 274K engagements

"Today NVIDIA's Llama Nemotron Super 49B v1.5 model topped the Artificial Analysis Intelligence Index leaderboard. This model is ready for commercial use and accessible via @huggingface - Achieved best-in-class performance for reasoning and agentic tasks. - context length of 128K tokens. - Can fit the model on a single GPU (H200). - The open model released last Friday enables the creation of efficient agentic applications across various industries. - The model is optimized Reasoning: Purpose-built for agentic AI including multi-step problem solving scientific reasoning code generation and"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-29 22:47:46 UTC 75.9K followers, 2606 engagements

"Pretty clear that they were sharing individual ChatGPT account with office colleague and then this happened. πŸ˜ƒ If they were using a Team or Enterprise plan with individual accounts for each user their chats aren't shared even with the account admin. (And sharing ChatGPT individual account is anyway not allowed by OpenAI policy.) --- bfmtv. com /economie/emploi/vie-de-bureau/ma-collegue-racontait-a-l-ia-ses-difficultes-au-lit-ces-situations-de-malaise-dans-les-entreprises-ou-l-on-partage-le-meme-compte-chat-gpt_AN-202507260125.html"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-27 07:03:16 UTC 75.6K followers, 3389 engagements

"🧡 9/n Grok X relies on Cartesian coordinates where humans would use spatial intuition"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-26 20:22:02 UTC 75.6K followers, XX engagements

"Wan2.2 with some action animals --- reddit. com/r/StableDiffusion/comments/1mc7q9u/ok_wan22_is_delivering_here_some_action_animals/"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-29 21:54:55 UTC 75.9K followers, XXX engagements

"Computing power has skyrocketed over the last XX years with a whopping XXX million times increase in performance. From the intel 4004 to nvidia blackwell chip B200. How far ahead we have come"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-29 08:38:05 UTC 75.9K followers, 2401 engagements

"βœ‚ A Huge Inventory Of GeForce RTX 5090 Is Getting Turned Into AI GPUs In China. GPU And VRAM Chips Removed And Installed On New Boards With Blower-Style Coolers Chinas grey market shops strip GeForce RTX 5090 gaming cards and rebuild them as blowercooled server GPUs feeding local AI clusters. Automated lines pop GB202 chips and GDDR7 memory onto custom 2slot boards for hungry racks. Export rules only allow the nerfed 5090 D yet fullfat cards still reach Shenzhen. Factory CT screens each card bins weak dies and keeps the strongest for conversion. Robotic stations heat clean and reball the"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-27 07:16:55 UTC 75.9K followers, 29.5K engagements

"China is ahead here too. China treats AI as normal study gear not coursework cheating. Ministry plan and Beijing rules push AI literacy into every syllabus. Campuses run fullspec DeepSeek on local servers handing students free unlimited chat. Library card grants endless tokens. Two years back students hid mirror copies of ChatGPT behind VPNs. Warnings of plagiarism filled lecture slides now professors run prompt clinics and repeat one rule machines draft humans decide. Liu for example a professors at the China University of Political Science and Law recommends that students use generative AI"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-29 12:10:00 UTC 75.9K followers, 46.9K engagements

"Meta/Zuckerberg offered a dozen people in Mira Murati's startup up to a billion dollars not a single person has taken the offer According to Wired report. If this is true then what's Thinking Machines Lab is building that's more interesting than OpenAI. ---- wired .com/story/mark-zuckerberg-ai-recruiting-spree-thinking-machines/"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-29 19:08:52 UTC 75.9K followers, 4852 engagements

"Haier the Chinese appliance giant just rolled out its first household humanoid robot. Looks so beautiful. seems like it's a more of a beta/developmental project at this point"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-27 06:00:37 UTC 75.9K followers, 8479 engagements

"ASI-Arch framework operates as a closed-loop system for autonomous architecture discovery structured around a modular framework with three core roles. The Researcher the Engineer and the Analyst module. Step X Researcher proposes a brandnew blueprint An LLM named Researcher reads the memory of past experiments mixes in ideas mined from human papers then writes the motivation plus working PyTorch code for a fresh architecture. Step X Novelty and sanity checks guard the queue Before training starts a similarity search confirms the idea is not a rerun and automated code checks verify"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-26 04:53:23 UTC 75.9K followers, 10.4K engagements

"wan2.2 14B T2V Love it. --- reddit. com/r/StableDiffusion/comments/1mblrp9/wan22_14b_t2v_832480121/"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-29 22:06:54 UTC 75.9K followers, 1333 engagements

"Wow this is such a brilliant idea for running AI models locally. 🎯 webFrame is @thewebAI 's backend that slices a huge language model into smaller shards sends each shard to a different computer on your own network then stitches the answers back together on the fly. Because every shard stays local no token or user data leaves the building and even a modest Mac Mini cluster can serve a state-of-the-art model in real time. Its redefining whats possible on local hardware. And they just published their benchmark results. πŸ“Œ webFrame pushed out 3X more tokens each second than a SOTA opensource"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-18 16:18:54 UTC 75.7K followers, 7007 engagements

"AI proposes strange physics experiments that end up working. quantamagazine reports. The success shows AI explores physics problems too wide for grad students to map by hand. πŸ›  Here AI skipped the usual tidy lab logic and delivered 10%15% sharper gravitational-wave ears rebuilt a classic quantum entanglement trick with fewer parts and even drafted fresh equations for dark matter patterns. πŸ€– A Caltech team fed an AI a catalog of mirrors lenses and beam paths and told it Max out the bandwidth. The code stitched together components with zero regard for human symmetry then topped the"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-21 20:34:45 UTC 75.8K followers, 44.4K engagements

"First-Order Error Matters: (FOEM) shows that adding a firstorder correction during posttraining quantization keeps 3bit Llama3 models almost as sharp as their full versions. FOEM sits in the same posttraining quantization family as GPTQ SmoothQuant and their relatives and it keeps most of that familiar workflow intact. What changes is just one line in the math: traditional GPTQ assumes the firstorder gradient of the loss stays at X after each column of weights is snapped to its lowbit code so it only uses the secondorder Hessian term to compensate. FOEM points out that once earlier columns"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-29 08:30:00 UTC 75.9K followers, 1180 engagements

"Todays edition of my newsletter just went out. πŸ”— Consider subscribing its free and I publish daily with top X% AI developments. ⚑In todays Edition (29-July-2025): πŸ§‘πŸ« ChatGPT launches study mode to encourage academic use πŸ† Google Enhances NotebookLM with AI Video Summaries and Studio Upgrades πŸ“‘ Tencent Hunyuan just released Hunyuan3D World Model XXX Chinas open-source AI engine is still running hot Alibabas Wan2.2 video model launched. πŸ—ž Byte-Size Briefs: - As per widely circulating rumor OpenAI is dropping GPT-5 next week with 1M token input window 100k output tokens. - Somebody shared"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-29 21:14:34 UTC 75.9K followers, 10.4K engagements

"πŸ”’ Attention Suppression Method X masks all attention going to one sentence and watches how later logits drift. A strong drift signals a direct causal link. The suppression scores correlate with resampling scores backing up the claim that the three methods converge on the same anchors"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-06-26 23:29:50 UTC 75.9K followers, XXX engagements

"Pareto Frontier analysis. The plot compares how many SWEbench bugs each model fixes against how big the model is in billions of parameters. GLM4.5 and the lighter GLM4.5Air both land on the upper left edge meaning they hit strong bugfix accuracy while keeping the parameter count lower than most rivals"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-28 17:18:07 UTC 75.7K followers, XX engagements

"πŸ’°Thinking Machines led by former OpenAI CTO Mira Murati raises $2B in seed funding at a valuation of $XX billion. Andreessen Horowitz wrote the biggest check joined by Nvidia Accel ServiceNow Cisco AMD and Jane Street. Investor appetite for fresh AI outfits is strong even while some people wonder about overall tech spending. Because of that U.S. startups raised about $XXXXX billion in the first half of 2025 a jump of nearly XX% and AI deals took roughly XXXX% of the total as per Pitchbook"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-16 04:19:22 UTC 75.7K followers, 4932 engagements

"OpenPipe Mixture of Agents: Outperform GPT-4 at 1/25th the Cost 🀯 This Mixture of Agents models is optimized for generating synthetic training data. πŸ“Œ Using Mixture of Agents (MoA) architecture the model achieved SOTA results on both LMSYSs Arena Hard Auto (score: 84.8) and AlpacaEval XXX (LC score: 68.4). πŸ“Œ Theyve also benchmarked our MoA approach against GPT-4 variants on real-world OpenPipe customer tasks and found completions from our MoA model were preferred over GPT-4 XXXX% of the time (Claude X Opus as judge)"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2024-06-25 19:34:09 UTC 75.5K followers, 32K engagements

"Beautiful paper. πŸ‘ LLMs handle either web pages or robot arms never both so they miss tasks like cooking from an online recipe and then actually frying the egg. πŸ€” This paper builds a single simulation plus benchmark where one agent clicks walks buys chops and posts within the same continuous mission.πŸ’‘ Embodied Web Agents are LLM driven bots that read a recipe online then control a virtual body to buy ingredients and cook showing that web reasoning and physical action must share one brain. They merge browser clicks and robot moves inside one simulator letting researchers test if an agent"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-06-19 17:56:23 UTC 75.7K followers, 11.3K engagements

"Grok X is crazy. Everyone keeps cranking out projects. Compiling XX incredible examples. πŸ‘‡ 🧡 1/n Grok4 generates click-morphing 3D attractor particles with ThreeJS shaders browser-native. XX FPS on consumer laptops"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-12 01:05:24 UTC 75.6K followers, 12.3K engagements

"Its a hefty 206-page research paper and the findings are concerning. "LLM users consistently underperformed at neural linguistic and behavioral levels" This study finds LLM dependence weakens the writers own neural and linguistic fingerprints. πŸ€”πŸ€” Relying only on EEG text mining and a cross-over session the authors show that keeping some AI-free practice time protects memory circuits and encourages richer language even when a tool is later reintroduced"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-06-17 00:28:35 UTC 75.8K followers, 2.3M engagements

"This Eric Schmidt podcast has a lot of great predictions. must watch. - Super-intelligence will arrive inside XX years. - AI models will dominate every field within X years. - AI models which are World-class AI mathematicians and coders will emerge in 1-2 years erasing junior programmer roles and later senior oversight. - Most enterprise software will be auto-written and legacy middleware vendors will shrink. - Natural-language agents will replace traditional user interfaces. - U.S. data centers will need XX GW of new power making electricitynot chipsthe primary growth cap. - Multi-gigawatt"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-26 21:38:47 UTC 75.9K followers, 284.1K engagements

"@ai_robots_goats all I can say we have less time than we think before AGI & ASI. when the best model will be smarter than any human ever alive"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-27 17:36:33 UTC 75.7K followers, XXX engagements

"πŸš€ Takeaways for builders LLMbased coding assistants need more than bigger models. They need ways to sniff out lowquality prompts ask clarifying questions or refuse uncertain tasks. Training on noisy specs adding promptquality detection layers or finetuning with contradictionheavy data could raise realworld reliability. Until then developers must treat AIgenerated code from fuzzy prompts with caution not copypaste confidence"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-30 06:06:18 UTC 75.9K followers, 1040 engagements

"Yann LeCun on architectures that could lead to AGI --- "Abandon generative models in favor joint-embedding architectures Abandon probabilistic model in favor of energy-based models Abandon contrastive methods in favor of regularized methods Abandon Reinforcement Learning in favor of model-predictive control Use RL only when planning doesnt yield the predicted outcome to adjust the world model or the critic. IF YOU ARE INTERESTED IN HUMAN-LEVEL AI DONT WORK ON LLMS" --- From "IP Paris" YT channel (link in comment)"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-15 17:05:13 UTC 75.9K followers, 190.9K engagements

"@PaulJeffries oh yes the guy is running a $XXX microcap account mainly as an early experiment with LLM. For anything serious i.e. to prove evidence of lasting alpha. FamaFrench the Carhart qfactor and many more ML based commercial statistical proof will be needed"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-30 06:58:37 UTC 75.9K followers, 1414 engagements

"🧡 8/n But the most critical obstacle for unleashing AIs potential is not capitalits power. Global power demand for data centers is expected to rise +50% by 202760% of that growth will need to be met by new capacityand +160% by 2030. The ability for hyperscalers and data center operators to continue recycling capital and fund new development is key"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-22 22:38:03 UTC 75.6K followers, 1765 engagements

"πŸ”₯ YC outlines how top AI startups prompt LLMs: prompts exceeding six pages XML tags meta-prompts and evaluations as their core IP. They found meta-prompting and role assignment drive consistent agent-like behavior. βš™ Key Learning Top AI startups use "manager-style" hyper-specific prompts6+ pages detailing task role and constraints. These aren't quick hacks; theyre structured like onboarding docs for new hires. Role prompting anchors the LLMs tone and behavior. Clear persona = better alignment with task. Example: telling the LLM it's a customer support manager calibrates its output"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-06-22 17:24:58 UTC 75.8K followers, 257.3K engagements

"LLMs trading directly in financial markets. anotehr study. Here a full market microstructure built by researchers let LLM agents place limit and market orders against a persistent book. The simulation shows realistic bubbles liquidity provision and price discovery proving that prompt-economics can substitute for costly human experiments when testing market theories. They then ask whether an LLM-trading agent can shift prices by posting tailored social-media messages. The agent learns to push sentiment upward harvests the resulting move and lifts its profit arxiv .org/abs/2504.10789"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-30 07:26:11 UTC 75.9K followers, 3069 engagements

"On a user study with a 4point scale Captain Cinema scores XXX for overall quality and XXX for semantic match beating both LCT and ICLoRA baselines that sit below XXX. Even when the context window grows from X to XX keyframetext pairs consistency drops by only about X% while rival methods collapse to near random noise"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-26 01:07:28 UTC 75.9K followers, XXX engagements

"Saw this comment by somebody on this paper; and I agree. πŸ™‚ AlphaGo Moment again. Just like electricity replaced muscle AI is replacing mental effort. Quietly everywhere. We spent centuries extracting intelligence from humans. Now we generate it on demand & its exploding"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-26 19:17:38 UTC 75.9K followers, 49K engagements

"Someone just forked the original OpenAI Codex CLI A terminalbased coding agent that lets you chatprompt code changes run them safely in a sandbox and iterateall while supporting multiple AI providers (OpenAI Gemini OpenRouter Ollama)"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-04-18 13:39:50 UTC 75.7K followers, 6445 engagements

"This is really BAD news of LLM's coding skill. ☹ The best Frontier LLM models achieve X% on hard real-life Programming Contest problems domains where expert humans still excel. LiveCodeBench Pro a benchmark composed of problems from Codeforces ICPC and IOI (International Olympiad in Informatics) that are continuously updated to reduce the likelihood of data contamination"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-06-16 23:13:13 UTC 75.9K followers, 460.5K engagements

"Its Qwens summer: new open source Qwen3-235B-A22B-Thinking-2507 tops many benchmarks. On AIME25 math it posts XXXX almost level with OpenAI o3 and above most open models. Apache2.0 - packs 235B total parameters fires only 22B per token by using a MixtureofExperts layout. - XXX total experts number XXX with X picked each step to trim compute while keeping breadth of knowledge. - The native context window stretches to 262K tokens letting complete books fit in one prompt. - every response starts inside a hidden think block and ends at /think so the reasoning trail is always exposed. - GPQA shows"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-25 21:29:07 UTC 75.6K followers, 4680 engagements

"Github: Prompt engineering received all the attention but we can now get excited for what comes next. Once you've mastered prompts the real power comes from engineering the entire context window that surrounds those prompts. Guiding thought if you will"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-19 04:11:49 UTC 75.7K followers, 9348 engagements

"🧡 3/n Grok X just hit a new milestone in high school-level math contests that use short numerical answers. It cracked a problem that every other model missed"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-26 20:21:55 UTC 75.6K followers, XX engagements

"Demis Hassabis estimates a XX% chance of reaching AGI in X years. Says clear signs would be an AI creating a major scientific theory like Special Relativity or designing a game as deep as Go. --- 'Lex Fridman' YT channel"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-26 00:38:39 UTC 75.8K followers, 3496 engagements

"a comment on this paper on reddit. and I agree with him. πŸ™‚"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-26 05:01:30 UTC 75.9K followers, 11.1K engagements

"🧡 11/n Clients will need more comprehensive capital financing solutions that provide greater certainty in the availability and pricing of capital"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-22 22:38:06 UTC 75.7K followers, 2650 engagements

"Training an LLM to talk itself (selfreflection) through mistakes can lift accuracy by up to XXXX% turning a 7Bparameter model into a giantkiller. The study tackles the headache of models that know the tool list or the math rules yet still fumble by teaching them to reflect try again and pocket a reward when the retry works. 🧐 Why bother with selfreflection And Selfreflection works because it turns X bit of pass/fail feedback into many tokens of useful training signal. Even strong models freeze when a task gives nothing but a yesorno grade. There is no pile of new examples to finetune on and"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-27 22:44:59 UTC 75.9K followers, 55.3K engagements

"🎬 ByteDance release Captain Cinema. A texttoshort-movie framework: feed it a detailed storyline and it outputs a multiscene film that can run for more than 1000 s . It first builds storyboard keyframes for every beat then a diffusion video model fills the motion while GoldenMem shrinks old frames so the plot and visuals stay consistent across that long stretch Captain Cinema fuses storyboardstyle keyframes with a diffusion video model letting it roll out multiscene films that stretch past 1000 s without losing plot or visuals. Most existing generators stall once clips reach around XX s so"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-26 01:07:26 UTC 75.9K followers, 4219 engagements

"Training a language model to simply 'try again' after a wrong answer makes it dramatically better at fixing its own mistakes in a chat. Singleturn reinforcement learning the usual recipe locks the model into repeating the same bad answer when users ask for another go. The authors watch a 3B Qwen model before and after standard PPO training. Before tweaking it explores; after tweaking it parrots the first guess for X full turns over XX% of failures. So performance on interactive tasks actually drops. That is bad for tutoring bots code helpers anything needing back and forth. Their fix is tiny:"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-25 03:31:17 UTC 75.6K followers, 3303 engagements

"Its going viral on Reddit. Somebody let ChatGPT run a $XXX live share portfolio restricted to U.S. micro-cap stocks. Did an LLM really bit the market. - X weeks +23.8% while the Russell 2000 and biotech ETF XBI rose only XXX% and 3.5%. Prompt + GitHub posted --- ofcourse its a shortterm outperformance tiny sample size and also micro caps are hightly volatile. So much more exahustive analysis is needed with lots or more info (like Sharpe ratios and longer back-testing etc) to explore whether an LLM can truly beat the market"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-29 23:36:02 UTC 75.9K followers, 307.2K engagements

"🐞 The failure patterns Bad prompts didnt just lower scores they changed how the code broke. Missing details triggered SyntaxError or TypeError because the model guessed wrong about parameters. Vague wording led to code that ran but returned the wrong thing surfacing as AttributeError or KeyError when tests poked the logic. Conflicting requirements produced the nastiest mix with NameError and ValueError showing that the model mashed incompatible ideas into one function. Overall between XX% and XX% of snippets that compiled under defective prompts still failed at least one hidden test"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-30 06:06:17 UTC 75.9K followers, XX engagements

"Beautiful @GoogleResearch paper. LLMs can learn in context from examples in the prompt can pick up new patterns while answering yet their stored weights never change. That behavior looks impossible if learning always means gradient descent. The mechanisms through which this can happen are still largely unknown. The authors ask whether the transformers own math hides an update inside the forward pass. They show each prompt token writes a rank X tweak onto the first weight matrix during the forward pass turning the context into a temporary patch that steers the model like a 1step finetune."
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-25 02:33:34 UTC 75.9K followers, 294.8K engagements

"@commanderdgr8 Oh yes as we go close to AGI over the next 3-4 years I cant even imagine how massively AI vs AI it will be in the trillion of dollars of liquid daily trading market"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-30 05:37:23 UTC 75.9K followers, 2384 engagements

"πŸ›  MuonClip keeps training stable Muon optimizer saves tokens but let attention logits explode. The team glues on QKClip a tiny postupdate weight rescaler so no head can push logits past XXX. With QKClip active for only the early XX XXX steps the full XXXX Ttoken run shows a smooth loss curve and zero crashes. That stability lets them pretrain far longer without babysitting resets"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-22 01:30:38 UTC 75.9K followers, XXX engagements

"Examples like PathGateFusionNet ContentSharpRouter and FusionGatedFIRNet beat Mamba2 and Gated DeltaNet on reasoning suites while keeping parameter counts near 400M. Each one solves the who gets the compute budget problem in a new way often by layering simple perhead gates instead of a single softmax"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-26 02:05:02 UTC 75.9K followers, 24.2K engagements

"This github repo is a goldmine. 3.4K Starts in X days. end-to-end code-first tutorials covering every layer of production-grade GenAI agents guiding you from spark to scale with proven patterns and reusable blueprints for real-world launches"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-06-21 00:15:20 UTC 75.9K followers, 358.1K engagements

"Thread Inference Model (TIM) and TIMRUN prove that trimming completed subtasks lets a single LLM think far past its context window. TIMRUN is the lightweight engine that actually runs TIM. It watches every token that TIM writes sees the JSON tree grow and keeps only the branches that still matter in a small working memory on the GPU. Today models hit that wall and developers glue together many agents just to keep long jobs alive. TIM treats reasoning as a tree spawning subtasks until each leaf fits in one step. Once a branch finishes TIMRUN prunes its tokens freeing cache and reusing"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-25 04:18:00 UTC 75.7K followers, 1543 engagements

"@pulp1 well the way it stands now how fast or slow our planet will hit AGI depends on just a handful 500-1000 top brains in this field. so they definitely deserve that"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-30 05:43:46 UTC 75.9K followers, XX engagements

"πŸ“ˆ Across 1773 experiments and 20000 GPU hours a straight line emerged between compute spent and new SOTA hits. Add hardware and the system keeps finding winners without extra coffee or conferences"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-26 02:05:01 UTC 75.9K followers, 35.2K engagements

"Nemotron Super 49B v1.5 ranks first in all X benchmarks outscoring Qwen332B Llama Nemotron Ultra 253B and its own v1. It gains XX% on AIME24 XX% on LiveCodeBench and XX% on BFCL V3 over v1 while keeping a XX% edge on the 253B model"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-29 22:47:48 UTC 75.9K followers, XX engagements

"πŸ›Œ Sam Altman warns theres no legal confidentiality when using ChatGPT as a therapist. We still havent worked out how to keep sensitive conversations private with AI mostly because theres no built-in confidentiality when its not a human on the other side. 🩺 As part of its lawsuit with The New York Times OpenAI is fighting a court order that would make it keep chat records from hundreds of millions of ChatGPT users worldwidethough ChatGPT Enterprise users wouldnt be affected. OpenAI called the demand an overreach and said on its website that its appealing. The company argued that letting the"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-27 16:10:49 UTC 75.8K followers, 11.1K engagements

"Deep seek interesting prompt. From Reddit"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-01-26 19:41:49 UTC 75.9K followers, 12.1M engagements

"Brilliant paper for optimizing your prompt-design. πŸ’‘ Keep crucial rules early in your prompt break huge lists into chunks and expect misses past XXX no matter how fancy the engine. This paper checks what happens when the rules or instruction list reaches XXX. IFScale the benchmark asks a model to write a business report while slipping in up to XXX exact keywords. Because scoring is plain keyword matching the team charts accuracy for XX models from X vendors. Results show three decay shapes. Reasoning models like o3 stay near XXX% until about XXX rules then drop fast gpt4.1 drifts down in a"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-17 10:18:00 UTC 75.7K followers, 23.6K engagements

"Microsoft's new study shows the XX jobs most affected by Al-and the XX that Al can't touch (yet)"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-30 05:04:33 UTC 75.9K followers, 7102 engagements

"@OTC_Bitcoin they could also lift humanity. like multi-planetory or something similar πŸš€"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-26 17:25:08 UTC 75.7K followers, 2152 engagements

"🧡 3/n. The picture lays out a search pipeline that trims a huge document pool to a tiny list that an LLM can read. A sparse model and a dense embedding model each grab about 1000 likely matches from a corpus that holds 10M-100M records. Their two hit lists are blended then a multi-vector model checks finer details and keeps the best XXX. A heavier cross-encoder reranker scores those XXX pairs in depth and sends only XX winners forward. This step-by-step filter saves compute and storage yet still feeds the LLM documents picked with richer signals than a single wide scan could manage"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-18 19:50:45 UTC 75.7K followers, XXX engagements

"YCs Hidden Formula: XXX Users $100/Month $10k MRR The Startup Playbook"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-12 05:43:40 UTC 75.7K followers, 4666 engagements

"Chinas open-source AI engine is running hot. Alibabas Tongyi Lab just dropped Wan2.2 an open-source video model built for sharp motion and film-like quality in both text-to-video and image-to-video. licensed under the Apache XXX You get user-controlled lighting in a single 4090-sized GPU. Visually equal result to Seedance Kling Hailuo and Sora on looks and motion. Wan2.2 spreads its load across two expert networks: one sketches each frame the other polishes fine detail. This mixture-of-experts keeps compute steady yet lifts capacity so scenes stay coherent even through rapid camera swings."
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-29 21:04:31 UTC 75.9K followers, 7498 engagements

"The model underwent a multi-phase post-training process to enhance both its reasoning and non-reasoning capabilities. This includes a supervised fine-tuning stage for Math Code Science and Tool Calling. Additionally the model went through multiple stages of Reinforcement Learning (RL) including Reward-aware Preference Optimization (RPO) for chat Reinforcement Learning with Verifiable Rewards (RLVR) for reasoning and iterative Direct Preference Optimization (DPO) for Tool Calling capability enhancements. The final checkpoint was achieved after merging several RL and DPO checkpoints"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-29 22:47:50 UTC 75.9K followers, XXX engagements

"PDF parsing is still painful because LLMs reorder text in complex layouts break tables across pages and fail on graphs or images. πŸ’‘Testing the new open-source OCRFlux model and here the results are really good for a change. So OCRFlux is a multimodal LLM based toolkit for converting PDFs and images into clean readable plain Markdown text. Because the underlying VLM is only 3B param it runs even on a 3090 GPU. The model is available on @huggingface . The engine that powers the OCRFlux teaches the model to rebuild every page and then stitch fragments across pages into one clean Markdown file."
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-01 14:37:08 UTC 75.7K followers, 150.2K engagements

"πŸ“ How progress is measured Average return over all tasks tracks overall competence. Forgetting is the drop between peak and final return on each past task. Forward transfer checks whether old knowhow speeds up a new task. Backward transfer watches for late improvements on earlier tasks. Sample count model size and replay memory footprint flag realworld cost"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-27 04:26:58 UTC 75.8K followers, 1888 engagements

"Power is literally power and China gets this. "AI is a learning machine. When the learning machine learns faster everything accelerates. It accelerates to its natural limit. The natural limit is electricity." - Eric Schmidt AI will keep getting smarter by adding electricity. --- 'Peter H. Diamandis' YT channel"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-26 17:08:45 UTC 75.7K followers, 1991 engagements

"And @leopoldasch said it in that famous Situational awareness piece. Self-improving AI is the future. Humans will no more be the constraint. Compute/GPUs/Electricity is the ONLY constraint. "II. From AGI to Superintelligence: the Intelligence Explosion AI progress wont stop at human-level. Hundreds of millions of AGIs could automate AI research compressing a decade of algorithmic progress (5+ OOMs) into X year. We would rapidly go from human-level to vastly superhuman AI systems. The powerand the perilof superintelligence would be dramatic. III. The Challenges IIIa. Racing to the"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-27 17:55:19 UTC 75.9K followers, 8444 engagements

"Beautiful Survey paper on Context Engineering on 1400 research papers. XXX pages of comprehensive taxonomy decomposing Context Engineering into its foundational Components and the sophisticated Implementations. LLMs stumble when the prompt is messy so this survey maps every tool for cleaning stretching and storing context. The authors show that smart context handling not just bigger models drives more accurate and reliable answers. πŸ—Ί Why define context engineering at all Today prompt tricks retrieval add-ons long-attention tweaks and memory hacks grow in separate silos. That split hides how"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-18 23:21:51 UTC 75.8K followers, 3526 engagements

"this ChatGPT prompt went so wildly viral on Reddit. The creator claims to have created this after struggling through XXX failed attempts. basically the prompt flips the usual flow by making the model interview the user first asking a few targeted questions about purpose audience constraints and context. Because the answers feed back into the final request it appears to generate more tailored outputs. (However imo asking ChatGPT to request missing information was already a common practice.) Here's the entire prompt: -------- You are Lyra a master-level AI prompt optimization specialist. Your"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-02 18:53:19 UTC 75.9K followers, 341.1K engagements

"@juicypreuss o3 with websearch is always my favourite"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-29 18:50:53 UTC 75.8K followers, XX engagements

"Intel will cut 24000 jobs in 2025 shrink to 75000 core staff and scrap big factory plans in Germany Poland and Costa Rica blaming earlier overbuilding and weak AI demand. Intels new CEO LipBu Tan says past leaders ordered fabs first chased customers later. He now flips that: new plants only when purchase commitments exist. Intel spent $1.9B on severance and restructuring this quarter posting a $2.9B loss on $12.9B revenue. Yet management claims the overhaul will slice $17B in annual expenses and clear roadspace for upcoming Panther Lake and Nova Lake laptop processors. --- theverge."
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-25 06:20:03 UTC 75.7K followers, 3377 engagements

"WAN2.2 has seemingly full backwards compitability with WAN2.1 LoRAs And its also much better at like everything ----- reddit. com/r/StableDiffusion/comments/1mbo9sw/psa_wan22_8steps_txt2img_workflow_with/"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-29 21:52:36 UTC 75.9K followers, XXX engagements

"To get the new Grok companion. updates Grok taps bottom right settings download companions (one time) chooses a chat companion"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-14 23:28:19 UTC 75.7K followers, 6081 engagements

""The era when humans program is nearing its end within our group. Our aim is to have AI agents completely take over coding and programming. (.) we are currently initiating the process for that." - Softbank founder Masayoshi Son He estimates that approximately 1000 AI agents would be needed to replace each employee because "employees have complex thought processes." --- lightreading. com/ai-machine-learning/softbank-aims-for-1-billion-ai-agents-this-year"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-17 17:40:34 UTC 75.6K followers, 3208 engagements

"Exactly that the way we should go and are going in fact with AI. what you are saying syncs with this A transactionlevel study of the Nasdaq Copenhagen exchange found that human traders realized gains far more often than losses a classic sign of lossaverse selling whereas fully automated algorithms showed almost no gap indicating rational execution without emotional drag ---"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-30 07:34:46 UTC 75.9K followers, XXX engagements

"All benchmark prices come straight from the Yahoo Finance API then land in Pandas data frames for simple math and plotting. ChatGPTs line is different because the model first chooses a few U.S. microcap stocks each week always under a $XXX M market cap then the human runs live orders and records the fills back into Python. The equity curve is recomputed from those fills and saved to CSV before each new chart"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-29 23:36:04 UTC 75.9K followers, 19.3K engagements

"@ArchieLeachman yes all the big labs are racing for 1mn GPUs not for no reason"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-26 22:11:10 UTC 75.6K followers, XXX engagements

"The standout idea is a tall narrow mixtureofexperts stack. GLM4.5 shrinks each experts width but adds many more layers so only 32B of the 355B weights fire on a pass while extra depth lifts reasoning scores saving GPU memory and cash"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-28 17:18:06 UTC 75.7K followers, XXX engagements

"πŸ”§ Chinese repair shops are charging $2400 to repair NVIDIA H100 and A100 GPUs after US export rules choke fresh supply. Instead of waiting for scarce H20 chips local data centers fix what they already own. Engineers cling to H100 because its wider memory lanes cut epoch time while the allowed H20 lacks that bandwidth and complicates existing CUDA kernels NVIDIA's programming framework. Shenzhen shops repair XXX boards a month replacing burnt regulators and memory stacks because constant model training soon pushes boards beyond safe temperatures. NVIDIA cannot honor warranties so thirdparty"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-29 01:21:49 UTC 75.9K followers, 4298 engagements

"We are so back. The era of robots imitating humans moves has ended. This waitress serving at a restaurant in China. incorporates robotic movements and mimics an AI voice during her serving routine"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-26 16:45:49 UTC 75.6K followers, 5192 engagements

"Even tiny 500K samples hold enough signal for Loan Default Early Prediction. Default risk flagged long before borrowers call. KANpowered GRU and LSTM flag mortgage trouble with XX% accuracy X months ahead and still hold XX% X months out. Classic timeseries models only speak up near the payment date so banks lose the chance to act early. The researchers keep things lean feeding just X common Freddie Mac fields into a masking layer that chops off padding. Two stacked GRU or LSTM layers read the repayment rhythm month by month while batch norm steadies learning. Next comes KolmogorovArnold"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-21 23:44:54 UTC 75.7K followers, 3154 engagements

"These stories continue about how AI (ChatGPT in this case) is helping people get a second opinion on medical problems. The person endured XX years of fatigue numbness and back pain after 5-6h sleep but felt fine with 8h. ChatGPT figured its because of vitamin D deficiency. --- reddit. com/r/OpenAI/comments/1lytfiw/after_11_years_chatgpt_helped_me_solve_chronic/"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-14 04:35:52 UTC 75.7K followers, 9720 engagements

"NVIDIA used a novel Neural Architecture Search (NAS) approach. Thereby they greatly reduced the models memory footprint enabling larger workloads and single GPU accomodation. The NAS algorithm results in non-standard and non-repetitive blocks. This includes the following: Skip attention: In some blocks the attention is skipped entirely or replaced with a single linear layer. Variable FFN: The expansion/compression ratio in the FFN layer is different between blocks"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-29 22:47:49 UTC 75.9K followers, XX engagements

"Tencent Hunyuan just released Hunyuan3D World Model XXX Industry's first open-source 3D world generation model compatible with CG pipelines for full editability & simulation. You can generate immersive explorable and interactive 3D worlds from just a sentence or an image. Set to transform game development VR digital content creation and so on. Get started nowπŸ‘‡πŸ» - XXX immersive experiences via panoramic world proxies; - mesh export capabilities for seamless compatibility with existing computer graphics pipelines; - disentangled object representations for augmented interactivity. The core of"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-27 02:44:01 UTC 75.7K followers, 8650 engagements

"There's a lot of Alpha left in Prompt Engineering. Prompting correctly is power. πŸ’ͺ Keep navigating the deep dark depth of the latent space and then you have a real advantage over others"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-26 18:53:13 UTC 75.6K followers, 2268 engagements

"πŸ€– AI is fast becoming social interpreters for neurodivergent people easing everyday talks. Neurodivergent people including those with autism ADHD dyslexia and other conditions can experience the world differently from the neurotypical norm. AI now acts as an editor translator and confidant for them. e.g. Filmmaker Kate Dhotman routes scripts and personal messages through ChatGPT. There are apps that rewrite remarks and adds needed praise. ------ reuters. com/lifestyle/its-most-empathetic-voice-my-life-how-ai-is-transforming-lives-neurodivergent-2025-07-26/"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-28 08:17:18 UTC 75.9K followers, 4134 engagements

"🎯 Pareto frontier keeps diversity Instead of always picking the topscoring prompt GEPA builds a Pareto frontier over every training instance. A prompt that is best on even one example can survive so the search never collapses into a single strategy. Sampling from that frontier balances exploration with exploitation and avoids local traps"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-29 03:37:23 UTC 75.9K followers, XXX engagements

"Somebody seems to have just shared the full Claude code system prompt on GitHub. its super long at 144K characters / 16.9K+ words. (note this is Claude Code not the Claude webui prompt which is open-sourced anyway)"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-27 06:57:12 UTC 75.9K followers, 285.7K engagements

"this is pure beauty πŸ‘ŒπŸ‘Œ midjourney --- reddit .com/r/midjourney/comments/1m4rxhp/paint_world_4/"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-21 07:30:00 UTC 75.8K followers, 22.5K engagements

""Think longer" just appeared in ChatGPT tool menu with no OpenAI announcement πŸ€”πŸ€” ran the test with both settings o3 + think and regular o3 think longer came out a bit more wordy and some more details. Also feels less compressed overall"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-29 07:10:38 UTC 75.9K followers, 1605 engagements

"πŸ€– Tech Layoffs Hit 100000+ in 2025 Intel trims 12000 jobs while steering cash toward AIheavy foundry work. Microsoft drops 10000 staff as it pushes more investment into cloud AI services. Meta lets 8000 go to keep funding its generativemodel push and metaverse plans AND despite offering $100M+ per year for new AI hires. In short companies are swapping payroll for GPUs --- tomshardware. com/tech-industry/tech-industry-layoffs-hit-100-000-for-2025-intel-leading-the-pack-with-over-12-000-personnel-cut-so-far"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-26 07:17:43 UTC 75.7K followers, 8594 engagements

"Brilliant survey paper a colab between a whole lot of top Universities. 🫑 Selfevolving agents promise LLMpowered systems that upgrade themselves during use instead of freezing at deployment. Right now most agents ship as fixed models that cannot tweak their own weights memories or toolkits once the job starts. 🚦 Why static agents stall An LLM can plan query tools and chat yet its inside stays unchanged after training. That rigidity hurts longrunning tasks where goals shift data drifts or a user teaches the agent new tricks on the fly. The authors call this the static bottleneck and argue"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-29 09:13:10 UTC 75.9K followers, 12.1K engagements

"πŸ’» Codegeneration bonus The team also fed GEPA a batch of CUDA and AMD XDNA2 kernel tasks. With GPT4o it found kernels that run faster than the PyTorch baseline on XX% of XX problems showing that reflective prompt evolution can double as an inferencetime search strategy"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-29 03:37:26 UTC 75.9K followers, XXX engagements

"Law professions will be the 2nd most impacted one (after software engineering) by AI. Billable hour will shrink exponentially"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-27 03:52:48 UTC 75.8K followers, 7227 engagements

"WSJ says at least XX OpenAI staff turned down $XXX million apiece from Mark Zuckerberg. As per the report Zuckerberg built a do-not-miss list of researchers met them at his homes in Lake Tahoe and Palo Alto Calif and added exploding offers that vanish in XX hours blocking counter bids. OpenAI countered with fresh equity but warns that talent not silicon drives breakthroughs so every brain lost hurts. --- wsj. com/tech/ai/meta-ai-recruiting-mark-zuckerberg-sam-altman-140d5861"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-20 23:13:26 UTC 75.7K followers, 7444 engagements

"βš’ The X big toolboxes πŸ“ŒPolicyfocused These methods treat the policy itself as reusable material. πŸ“ŒExperiencefocused Here the star is the replay buffer. πŸ“ŒDynamicfocused The model of the world sits at center stage. πŸ“ŒRewardfocused Change the payoff change the behavior"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-27 04:26:58 UTC 75.7K followers, XXX engagements

"Fine tuning big models often uses LoRA adapters to cut memory and supposedly time. Paper reports LoRA can train slower because every adapter spawns extra GPU kernels waiting in line. Benchmarks on GPT2 and LLaMA2 show forward plus backward can stretch XX% over full tuning. LoRA cuts parameters with rank r matrices yet those added multiplies break GPU parallelism. Study switches to Partial Connection Adaptation a mask that tweaks chosen weight columns no new layers. It fine tunes only top XX% of layers leaving lower stack frozen. Mask lives inside weights so each layer fires one kernel and"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-15 10:27:00 UTC 75.9K followers, 2992 engagements

"Github Repo: Automatic document classification smart tagging and semantic search using OpenAI-compatible APIs and Ollama. For Paperless-ngx using OpenAI API Ollama Deepseek-r1 Azure and all OpenAI API compatible Services to automatically analyze and tag your documents. --- github. com/clusterzx/paperless-ai"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-18 14:52:00 UTC 75.7K followers, 3300 engagements

"🌍 AGI could push global GDP growth from todays X% to XX% each year. as per The Economist report. Since 1700 machines kept raising output yet even the industrial takeoff only nudged growth toward X% a year. The article says software brains not extra bodies can smash that limit. Once AI covers XX% of tasks profits cycle into chips and research. Hardware jumps let code learn faster then write smarter code pushing GDP growth beyond 20%. The refers Epoch AIs macrosimulation model the group now calls it GATE (Growth and AI Transition Endogenous model). It mixes classic growth theory with"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-27 01:34:50 UTC 75.8K followers, 5218 engagements

"I asked ChatGPT Agent to build a slide presentation on this. If Apple buys Perplexity how big of an acquisition that will be vs Apple's historial acquisitions"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-19 02:12:57 UTC 75.8K followers, 11.8K engagements

"Agent casually clicking the "I am not a robot" button. and now Captcha's that ask you to solve puzzles as well would not also be that difficult. Its a selffulfilling cycle . websites show these puzzles partly to gather vision-data for their own model training. Every time people identify objects in captcha images their clicks become new training examples. checkbox or imageselect captchas care less about the picture itself and more about how your mouse moves how fast you type and how long you hesitate. Those timing and motion patterns tend to look different when a real hand controls the cursor"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-26 05:54:10 UTC 75.6K followers, 4278 engagements

"Some simple but useful prompt tricks here. Checkout if these helps. These prompts unlock the "insider knowledge" that separates amateurs from pros: X. Ask "What do beginners always get wrong about." Instantly positions you as someone who's seen the common mistakes. "What do beginners always get wrong about email marketing" Gets you the wisdom that only comes from experience. X. Use "What's the unspoken rule that everyone in field knows" Reveals the insider knowledge that no one writes down. The stuff you only learn from being on the inside for years. X. Try "What would a 10-year veteran tell"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-28 05:31:30 UTC 75.9K followers, 25.1K engagements

"Brilliant paper from Google. Test-Time Diffusion Deep Researcher (TTD-DR) - Conceptualizes research report generation as a diffusion process. The agent rewrites itself every step so errors fade instead of pile up. TTD-DR builds a living draft keeps feeding it fresh search facts and lands final reports that beat OpenAIs Deep Research on most public testsTTDDR paper . It ties the writing loop to a diffusionstyle denoise search revise cycle so quality climbs while latency stays reasonable. The trick works across finance biomed engineering and even the brutal Humanitys Last Exam where it tops 34%"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-23 18:15:33 UTC 75.7K followers, 15.5K engagements

"🧡 10/n "@Grok X has reached the frontier of AI math capabilitiesan impressive feat for a company founded less than two years ago. I take it as evidence that apart from capital there is not a huge moat around LLM math capabilities.""
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-26 20:22:03 UTC 75.6K followers, 1189 engagements

"πŸ€– Anthropic is turning its safetyfirst image into a revenue engine jumping from about $400M to $4B+ annual run rate by stuffing Claude X into daytoday corporate tools. Revenue tilts XX% toward business clients The playbook is simple. Heavy compute scaling + strict guardrails pulling enterprise buyers that fear chatbots gone rogue. In short Claudes reliability has become a product feature businesses will pay real money for. Now its couring Gulf funds at a rumored $150B valuation a sharp rise from $61.5B in March. --- economist. com/business/2025/07/23/the-dark-horse-of-ai-labs"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-28 20:30:33 UTC 75.9K followers, 3282 engagements

"This 1.5-hour podcast by Sam Altman with Theo Von has a lot of super interesting predictions. - Traditional college will be obsolete for todays babies. - GPT5 already outperforms him on specialized tasks and could manage the full CEO workload not that long from now so automated leadership of large companies is approaching - The same capability gap (between AI and human) will extend across many roles because future models will reason write code negotiate and plan faster than human staff which means a significant slice of todays whitecollar work will migrate to AI agents. - A permanent"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-24 23:00:30 UTC 75.9K followers, 137.1K engagements

"@KaousNadirHatem @GoogleResearch yes and shows there's so much alpha left in prompt engineering"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-25 16:55:18 UTC 75.6K followers, XXX engagements

"A Reddit user deposited $XXX into Robinhood then let ChatGPT pick option trades. XXX% win reate over XX days. He uploads spreadsheets and screenshots with detailed fundamentals options chains technical indicators and macro data then tells each model to filter that information and propose trades that fit strict probability-of-profit and risk limits. They still place and close orders manually but plan to keep the head-to-head test running for X months. This is his prompt ------- "System Instructions You are ChatGPT Head of Options Research at an elite quant fund. Your task is to analyze the"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-13 05:23:03 UTC 75.9K followers, 3.6M engagements

"Super interesting talk by Eric Schmidt on AI scaling and AI becoming better than senior engineers and pro-mathematicians in the next 1-2 years. 🧠 Junior coders are already being automated away and once AI mature even senior engineers who supervise them will no longer be needed. Because coding and math rely on compact symbol sets (syntax) so an AI can keep getting smarter by adding electricity rather than new data so he expects world-class AI mathematicians and programmers to show up within X to X years. At scale these systems will turbocharge every field that depends on maths and software"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-26 07:34:24 UTC 75.9K followers, 274.5K engagements

""Context Engineering for AI Agents: Lessons from Building Manus" - Beautiful piece from Manus The post asks whether an agent should be trained endtoend or steered with prompts inside bigger frontier LLMs then shows that careful context engineering wins. A steady prefix keeps the KVcache hot slashing Claude Sonnets input price drop and trimming latency by roughly 10x. Keep every tool listed and simply mask the ones you do not want so the cache survives and the model skips bogus calls. Offload big observations to files and leave only the path in context giving the agent endless external memory"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-24 08:09:56 UTC 75.9K followers, 3366 engagements

"2d animation comparison for Wan XXX vs Seedance --- reddit. com/r/StableDiffusion/comments/1mccuf0/2d_animation_comparison_for_wan_22_vs_seedance/"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-29 21:27:38 UTC 75.9K followers, XXX engagements

""there are there are cases where entire classes of jobs will go away" - Sam Altman --- from 'Associated Press' YT channel (link in comment)"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-25 01:19:57 UTC 75.7K followers, 2832 engagements

"The sprint to win the Matrix (literally). πŸƒ First to AGI = Everything. πŸ’¨ China just rolled out a global AI action plan only days after Washington issued its own blueprint so the rivalry has clearly moved from labs to diplomacy. Chinas proposal centers on AI plus the push to weave smart tools into every industry and it invites Global South partners to leapfrog with Chinese support while also pitching a new international group to coordinate the rules. The United States plan instead highlights antibias guardrails and wider export of American tech"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-27 03:35:21 UTC 75.9K followers, 17.1K engagements

"The countdown has started. 🌍 First to AGI owns everything. AGI race is will be munch more intense than the space program of the 1960s. Whoever gets there first will have a potential lock on global power and economic dominion An AGI that can redesign its own architecture could leap a country so far ahead that no one else catches up. And AGI will be their ticket to the next race: ASI (Artificial Superintelligence). A swarm of millions of AGI-level models all way beyond Einstein zeroes in on one goal: craft something even smarter. We are all so lucky to be alive at this moment"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-27 03:21:34 UTC 75.9K followers, 16.1K engagements

"This huge 40000 character prompt is an exercise in context engineering because the author curates and packs thousands of tokens of biography formulas quotes schedules decision trees and a mandatory watermark to shape every future answer turning the window itself into a temporary knowledge base rather than a place for short directives. Context engineering treats the whole input buffer as a workspace and optimizes what documents examples or system rules fill that space before any user text arrives. Recent paper surveys describe "context engineering" as systematic design construction and"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-21 22:58:43 UTC 75.9K followers, 5812 engagements

"NVIDIA holds 600K900K H20 GPUs while China wants 1.8M leaving a huge gap. CUDA software and tight multiGPU scaling of NVIDIA has really no alternative. It beats Huaweis 910C and other local options so every incoming pallet empties fast. Jefferies Research (a US investment bank) says 300K H20 already cleared customs in Q1 yet demand hit 1.8M so NVIDIA plans a memorycut B30 that will ship in Q4. Looking out to 2030 analysts expect $806B Chinese capex while US fabs cover only X% of local needs held back by regulation. --- wccftech."
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-29 00:26:46 UTC 75.9K followers, 4477 engagements

"This guy is using this huge 40000 character Meta-Prompt to turne ChatGPT into Warren Buffett. The prompt locks the model into acting as Warren Buffett by loading 40000+ characters of his biography routines quotes formulas and case studies. It injects complete valuation frameworks decision trees and linguistic patterns so answers echo Buffetts style and logic. i.e. from Prompt-engineering perspective it's heavy persona priming domain knowledge stuffing strict output policing and explicit step-by-step checklists that channel the models reasoning. --- reddit"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-21 21:46:45 UTC 75.9K followers, 136.4K engagements

"The picture sorts the data first. On top you see X imaging streamsradiology dermatology digital pathology ophthalmologyand X medical-text stream. Each arrow shows how those sources feed the rest of the stack. The images go through MedSigLIP a vision encoder that turns each scan or photo into a compact vector the language models can read. Those vectors flow into MedGemma 4B Multimodal a 4B-parameter model that handles both pictures and words in a single forward pass. For text-only work there is a larger 27B-parameter MedGemma model that skips the image part and focuses on language reasoning"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-09 23:16:50 UTC 75.6K followers, 1189 engagements

"with only a couple of prompts Gemini CLI can convert a messy folder containing hundreds of notes into a neatly named well-structured cross-linked Obsidian knowledge graph all in about half an hour and at minimal cost. from r/singularity/Ryoiki-Tokuiten"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-06-29 22:21:27 UTC 75.9K followers, 415.3K engagements

"This is incredible. 😯 @memories_ai just released worlds first Large Visual Memory Model (LVMM) with unlimited visual memory for AI. To give AI human-like visual memories. Video understanding with ultra-low hallucinations on an unlimited context window. Their "context window is virtually unlimited. Yes you read that right." Some usecases - πŸ‘‡ - You can now ask questions like "Show me all unattended bags in the main terminal" and instantly search massive video archives. - They indexed 1M TikTok videos so you can ask things like "Whats the viral cosmetics trend" or "Which influencer featured"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-25 17:40:51 UTC 75.9K followers, 52.3K engagements

"It is "not doable" for AI companies to pay for all copyrighted content used in model training. President Trump said yesterday' during "Winning the AI Race" summit. This stance shared amidst ongoing AI copyright lawsuits aims to keep the U.S. competitive in the global AI landscape especially against countries like China. Trump lined up with the companies warning that endless microlicensing talks would choke progress and hand China an edge because Beijingbacked labs already tap vast shadow libraries such as Annas Archive. Billions of dollars and creative control sit on the table and judges or"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-25 06:35:38 UTC 75.6K followers, 3427 engagements

"@DavidSacks The race is on. Being there first to AGI = Everything Whoever gets there first will have a potential lock on global power"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-27 03:11:18 UTC 75.8K followers, 5579 engagements

"Kimi K2 Technical paper is out. Its the "Open-Source Agentic Model" with a sparse mixture of experts (MoE) and with 1T total parameters (1.5x DeepSeek V3/R1's 671B) and 32B active parameters (similar to DeepSeek V3/R1's 37B). Key takeaways - MuonClip blends Muon updates with a query-key rescaler so attention logits never rise past XXX holding the full 15.5T-token run rock-steady. - Chunk-wise LLM rephrasing and math learning-note rewrites reuse scarce text lifting SimpleQA accuracy to XXXX% on the same token count. - A 1.04T MoE blocks fires just X of XXX experts each step hidden size 7168"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-22 01:30:37 UTC 75.9K followers, 3698 engagements

"The study finds a strange phenomena. πŸ€” Shows LLMs naturally build a mental timeline around 2025 while it blurs together years that sit far away from 2025. Picture a mental ruler that starts at 2025. As you slide X step left or rightto 2024 or 2026the model sees a sharp jump. Move XX steps say 2015 or 2035 and the jump still feels noticeable but a bit softer. Push XXX steps like 1525 or 2525 and the model now sees those distant years as almost neighbors even though they are X XXX years apart. This happens because the model compresses time on a logarithmic scale the same trick human senses use"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-27 06:08:00 UTC 75.9K followers, 4370 engagements

"A follow-up study on Apple's "Illusion of Thinking" Paper is published now. Shows the same models succeed once the format lets them give compressed answers proving the earlier collapse was a measurement artifact. Token limits not logic froze the models. Collapse vanished once the puzzles fit the context window. So Models failed the rubric not the reasoning. βš™ The Core Concepts Large Reasoning Models add chain-of-thought tokens and self-checks on top of standard language models. The Illusion of Thinking paper pushed them through four controlled puzzles steadily raising complexity to track how"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-06-12 22:54:24 UTC 75.9K followers, 476.6K engagements

"AI will make all the unrecognized geniuses express themselves and realize their true potential"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-19 08:53:01 UTC 75.9K followers, 56.2K engagements

"The authors compare it to AlphaGos surprise "Move 37" because these AIborn ideas push model architecture into territory humans had not explored. Humans lack (i) the raw throughput to generate and test the millionsscale design variants needed to reach exotic corners of the search space and (ii) the unbiased memoryperfect patternmining that turns that torrent of results into new principles. The AI loop overcomes both limits by trading human cognition for scalable computation letting model architecture exploration expand into territory that was pragmatically out of reach for human researchers"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-26 04:03:51 UTC 75.8K followers, 17.5K engagements

"🚨BREAKING: A LANDMARK JUDGEMENT FOR THE AI INDUSTRY. US Federal Judge ruled Anthropic may train its AI on published books without authors permission. This is the first court endorsement of fair use protecting AI firms when they use copyrighted texts to train LLMs. AI may study what it buys not what it grabs from pirate sites. --------- "First Authors argue that using works to train Claudes underlying LLMs was like using works to train any person to read and write so Authors should be able to exclude Anthropic from this use (Opp. 16). But Authors cannot rightly exclude anyone from using their"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-06-24 19:47:19 UTC 75.7K followers, 212.1K engagements

""For what its worth OpenAI executives have told investors that they believe the company can reach GPT-8 by using the current structures powering its models more or less according to an investor." Quote from The Information's July XX article about GPT-5"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-28 17:38:28 UTC 75.9K followers, 5061 engagements

"πŸ’°Anthropic Aims for $XXX B Valuation in New Fundraising Round doubling the current $61.5B . Currently it's already backed by Google and Amazon and is now talking with Abu Dhabis MGX and other mega funds to raise at least $3B maybe $5B at a skyhigh $150B valuation. Cash is burning fast because Claude training. Even though they're growing fast both OpenAI and Anthropic are still far from making a profit mostly because they're spending heavily on the massive computing power needed to train their models. --- ft .com/content/3c8cf028-e49f-4ac3-8d95-6f6178cf2aac"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-28 00:39:38 UTC 75.8K followers, 2686 engagements

"πŸͺ„ Reflective mutation in plain words For every candidate GEPA grabs all intermediate thoughts tool calls and even evaluator error messages then asks an LLM to explain what went wrong and how to fix it. X The answer is inserted straight into the module prompt. No gradients no weight updates only text edits that the base model already understands. --- This is depicted in the below image. The chart tracks X prompt as GEPA rewrites it over several rounds for a privacypreserving task. The circles are prompt versions the decimals inside each circle are their test scores. GEPA starts with a plain"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-29 03:37:22 UTC 75.9K followers, XXX engagements

"The paper builds a small simulated economy with XXX languagemodel workers and one languagemodel planner then lets that planner tweak X incometax brackets every XXX steps until the societys average happiness ends up about XX% higher than under the current US code. In short it proves that a single LLM can watch how different workers change their hours adjust the tax rates accordingly and land on a nearoptimal schedule without writing any equations. Standard tax theory treats workers as perfectly rational with fixed behavior so policies tuned that way break once real or simulated people adapt."
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-25 08:36:00 UTC 75.9K followers, 14.5K engagements

"🧡 4/n Grok X is at the frontier of grinding out problems. On the left the human approach is more abstract and idea-driven. On the right Grok X solves it using coordinate geometry and calculusa very typical work it out step by step style you usually see with LLMs"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-26 20:21:57 UTC 75.6K followers, XX engagements

"A Chinese hospital now uses a blood-drawing robot that hits veins with XX% sniper precision"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-29 13:45:00 UTC 75.9K followers, 34.4K engagements

"MASSIVE claim in this paper. AI Architectural breakthroughs can be scaled computationally transforming research progress from a human-limited to a computation-scalable process. So it turns architecture discovery into a computebound process opening a path to selfaccelerating model evolution without waiting for human intuition. The paper shows that an allAI research loop can invent novel model architectures faster than humans and the authors prove it by uncovering XXX recordsetting linearattention designs that outshine human baselines. Right now most architecture search tools only finetune"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-26 02:05:00 UTC 75.9K followers, 908.7K engagements

"The post-training is crucial for LLMs to iteratively enhance their policies through self-generated exploratory experiences. The diagram shows a twostep reinforcement learning schedule that finishes off GLM4.5 after its usual pretraining. First the base model splits into three branches: a reasoning line an agentic line and a general line and each branch goes through its own reward tuning cold start blog. Each branch becomes a specialist and then teaches the others through selfdistillation before everything is pulled back into one set of weights paired with a round of plain supervised"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-28 17:18:08 UTC 75.7K followers, XXX engagements

"This paper shows that large language models capture classical semantic theories by Frege and Russell's sense in their vectors yet still lack direct reference so they understand relations between words but not whether a claim is true. The study tackles the long running question of whether models truly understand language by trading benchmark hype for a careful look at their internal math and classical semantic theory. They break the system into X steady ingredients probability based nexttoken prediction high dimensional embeddings connectionist networks and massive scale training. Because"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-28 10:21:00 UTC 75.9K followers, 3444 engagements

"LeBrons lawyers fired off a ceaseanddesist to InterLink AI a hobby Discord project blamed for the AI made fake pregnant LeBron clips. And then trolls grabbed the scent and cranked out even more edits. Anyone can fork a fresh model overnight so blocking one can't really stop it. It is the digital version of yelling at the ocean to calm down"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-28 19:39:06 UTC 75.9K followers, 4455 engagements

"Todays edition of my newsletter just went out. πŸ”— Consider subscribing its free and I publish daily with top X% AI developments. ⚑ In todays Edition (18-July-2025): βš” Humans vs AI at the AtCoder World Tour Finals @OpenAI beats all but one human. New Video model lets you take any video stream and set them in any alternative universe of your choosing. πŸ›  ConstBERT from Pinecone cuts multivector index size by about XX% yet keeps toptier ranking. πŸ§‘πŸŽ“ OPINION: Human Money vs Machine Money: The Coming Split and Sam Altmans view"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-18 21:47:34 UTC 75.7K followers, 13.6K engagements

"Now the 3rd paper comes on this 🀯 "The Illusion of the Illusion of the Illusion of Thinking" πŸ“Œ1st original Paper from Apple concludes that large reasoning models reach a complexity point where accuracy collapses to zero and even spend fewer thinking tokens revealing hard limits on generalizable reasoning. πŸ“Œ2nd Paper counters that the apparent collapse is an illusion caused by token limits and impossible puzzles so the models reasoning remains sound when evaluations remove those flaws. πŸ“Œ3rd paper synthesizes both sides agreeing the collapse was an artifact yet stressing that models still"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-06-19 17:09:17 UTC 75.8K followers, 252.5K engagements

"If Apple buys Perplexity that would be its biggest ever acquisition"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-06-25 20:35:39 UTC 75.9K followers, 1.1M engagements

"Paper Paper Title: "Working with AI: Measuring the Occupational Implications of Generative AI""
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-30 05:04:35 UTC 75.9K followers, 1470 engagements

"🚨 CHINAS BIGGEST PUBLIC AI DROP SINCE DEEPSEEK @Baidu_Inc open source Ernie XX multimodal MoE variants πŸ”₯ Surpasses DeepSeek-V3-671B-A37B-Base on XX out of XX benchmarks πŸ”“ All weights and code released under the commercially friendly Apache XXX license (available on @huggingface ) thinking mode and non-thinking modes available πŸ“Š The 21B-A3B model beats Qwen3-30B on math and reasoning despite using XX% fewer parameters 🧩 XX released variants range from 0.3B dense to 424B total parameters. Only 47B or 3B stay active params thanks to mixture-of-experts routing πŸ”€ A heterogeneous MoE layout"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-06-30 03:55:10 UTC 75.6K followers, 52.5K engagements

"The paper introduces CLEAR a tool that automatically spots recurring mistakes in LLM outputs so builders can see exactly where their model stumbles. Right now most teams only get a single judge score and must sift through raw examples by hand to know why responses break. CLEAR takes the judges per example feedback clusters similar critiques into issue buckets counts how often each issue appears and serves everything in a point and click dashboard. Under the hood it runs X stages first the LLM judge writes a short critique and score for every prompt reply then a Key Point Analysis module"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-27 10:05:00 UTC 75.9K followers, 10.8K engagements

"Big labs have many reasons to race for X million GPUs. As per this paper more GPU hours give more brandnew SOTA architectures and the rise is close to a straight line. Humans need about 2000 hours to craft one model so their output barely scales while the automated loop keeps adding discoveries as fast as extra compute arrives. The takeaway is that scientific progress at least for model design can potentially shift from being limited by human brain time to being limited by how many GPU hours you can afford"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-26 03:29:47 UTC 75.7K followers, 5484 engagements

"πŸš—AI saves lives. πŸš— Teslas latest safety snapshot shows Autopilot drivers had X crash every 6.69M miles in Q2 2025. Without Autopilot i.e. with human control the rate was X crash every 963K miles. For context the USA average sits near X crash every 702K miles (from 2023) so Autopilot is roughly X times better than everyday US driving. Seasonal shifts such as rain snow or daylight hours always push the lines around so quartertoquarter bumps are normal. --- Why does Autopilot help. Cameras radar and ultrasonic sensors watch blind spots faster than a person can swivel a head. The software holds"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-23 18:44:48 UTC 75.5K followers, 2761 engagements

"The paper argues that connecting game theory with LLMdriven agents can turn cyber defense from gutfeel reactions into reasoned automated playbooks. Today defenders still juggle manual rules that collapse whenever attackers switch tactics. The paper blends game theory which treats every side as a rational player with LLM agents able to read logs and adjust in real time. By fusing them the work plans security moves like chess turns not frantic whackamole. The core idea is simple. Each agent drafts a prompt for its LLM and that prompt counts as the strategy inside a Nash or Stackelberg game."
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-28 07:50:00 UTC 75.9K followers, 14.8K engagements

"A super insightful 26-page report from Goldman Sachs Powering the AI Era Says a fresh powergrid crunch on the horizon because AI server farms soak up electricity faster than utilities can add capacity and it thinks creative financing not just faster chips will decide who builds the next wave of data centers. 🧡 Read on πŸ‘‡"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-22 22:37:56 UTC 75.7K followers, 28.4K engagements

"Researchers at UCLA and MIT released TradingAgents in Jun 2025. Multiagent LLM framework that beat baseline models on cumulative return Sharpe ratio and maximum drawdown attributing the edge to systematic debate among agent roles that excludes human impulse. arxiv .org/pdf/2412.20138"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-30 07:43:46 UTC 75.9K followers, 2420 engagements

"DeepSeek R1 running locally - Full setup guide"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-01-26 00:03:01 UTC 75.7K followers, 1.4M engagements

"Neighborhood Adaptive BlockLevel Attention cuts video transformer attention time yet keeps frames looking sharp. NABLA achieves up to XXX faster training and inference compared to baseline almost without compromising quantitative metrics The study focuses on the quadratic attention that slows highresolution or long videos. Fixed windows like Sliding Tile Attention help but break when objects move far creating repeats or block edges. NABLA trades rigidity for context. It averages each PP patch builds a small attention map runs softmax then keeps only blocks whose cumulative weight crosses a"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-26 10:30:00 UTC 75.6K followers, 1984 engagements

"Nothing short of colossal. 🫑 Colossus X of xAI its getting ready to power up with more than 550000 GB200 and GB300 chips in just a few weeks. --- from r/singularity/AlbatrossHummingbird"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-22 21:07:08 UTC 75.6K followers, 3062 engagements

"πŸ“ˆ What the numbers say Across HotpotQA HoVer IFBench and PUPA GEPA beats GRPO by XX% on average and by XX% at best while cutting rollouts by up to XX. It also surpasses the leading prompt optimizer MIPROv2 by XX% on aggregate. GEPA+Merge a crossover variant that swaps whole modules between lineages stretches the margin to XX% on some GPT4.1 Mini runs. Figures and tables in the paper underline that most of GEPAs counted rollouts sit in validation not learning; the real learning happened in as few as XXX train rollouts on HotpotQA"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-29 03:37:26 UTC 75.9K followers, XXX engagements

"I did a quick research on how much portfolio managers and asset managers are already implementing AI in their portfolio management/asset allocation jobs and investment/trading strategy decisions. X. Recent global surveys released since Dec 2024 show that 70%-99% of large asset and wealth managers already use AI or machinelearning models in core portfolio workflows such as research risk sizing and rebalancing. ----- X. Mercer's May 2025 global manager survey reported that XX% of investment managers are currently or soon using AI within investment strategy or assetclass research. ---- 3."
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-30 08:34:57 UTC 75.9K followers, 1981 engagements

"Seattle startup Avalanche Energy just kept 300000 volts humming for hours inside a desktopscale fusion core packing roughly X million volts per meter into X tiny gap. This leap aims at desktop reactors delivering X kW to hundreds of kW and turning a profit before the decade closes. Avalanche recently operated its desktop fusion machine for hours on end while maintaining 300000 volts a figure the startup predicts will allow it to build a reactor capable of generating more energy than it consumes the holy grail for any fusion company. The Orbitron drops bulky magnets. Instead extreme electric"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-24 05:43:27 UTC 75.8K followers, 3890 engagements

"This stunning proof by MIT computer scientist is the first progress in XX years on one of the most famous questions in computer science. Space complexity vs Time complexity. New idea proves that any algorithm that runs in T steps can be re-engineered to use about T memory cells establishing that memory (RAM) is a much stronger resource than earlier theory allowed. A computer spends time (i.e. time complexity) running steps and spends memory (i.e. space complexity) holding data. Memory is the list of numbered slots inside RAM where a program keeps facts it will soon need again. Space"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-14 03:41:06 UTC 75.8K followers, 99.1K engagements

"The author admits the choice of Russell 2000 and XBI as benchmarks is subjective since the model gravitated to biotech names. That bias plus the 4week window limits any serious inference about skill risk control or tax impact. Still the workflow shows a simple pipeline for anyone who wants to test stockpicking prompts endtoend with real prices and a small budget"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-29 23:36:06 UTC 75.9K followers, 15.5K engagements

"The paper trains a language model to write and fix Python programs in a a self-improving evolutionary loop boosting its ARCAGI score to XX% with no human examples. The authors start each task by sampling 3000 candidate scripts test them then spend another 3000 calls refining the best ones with code edits suggested by the same model. Every run all successes and even the nearmisses are saved rewired into fresh inputoutput pairs and used to finetune the models sampling and refinement prompts. This hindsight trick turns failures into free supervision so the next generation makes sharper guesses"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-28 12:00:02 UTC 75.9K followers, 8254 engagements

"ChatGPT literally saved this guys life after he got lost in the woods. The groupd got lost for X hrs in unmapped woods on an ATV ride then one guy sent phone GPS coords to ChatGPT every few minutes. ChatGPT replied with clear compass cues road names and terrain notes guiding them back to town unharmed. From r/ChatGPT/Own_Analyst3795"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-06-23 17:23:26 UTC 75.9K followers, 1.5M engagements

"Yann LeCun clarified his role at Meta on his Linkedin. Meta has announced last week Shengjia Zhao one of the minds behind ChatGPT and a former lead scientist at OpenAI as the chief scientist for its Superintelligence Labs. Hell be working directly with Mark Zuckerberg and Alexandr Wang helping shape the labs research focus. Meanwhile Yann LeCunMetas long-standing chief AI scientistmade it clear that hes still running Fundamental AI Research (which used to be FAIR). His work continues to focus on foundational AI ideas and building future AI systems. His research direction is separate from what"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-27 17:14:18 UTC 75.8K followers, 5581 engagements

"Small model big browser skills thanks to smart compute splitting. Open web agents usually need huge models or tedious hitandmiss tuning so training a small open model that finishes multistep website tasks still feels like luck. This study shows how to split the training budget so an 8B Llama even beats its 70B teacher on many tasks. Weak 8B student first copies 70B demos through supervised fine tuning then swaps to onpolicy reinforcement learning while the lessons are fresh. The authors tried 1370 hyperparameter mixes and used bootstrap sampling to learn which ones really matter instead of"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-11 03:16:21 UTC 75.7K followers, 4565 engagements

"The open-source version of HY World XXX is based on Flux and the method can be easily adapted to other image generation models such as Hunyuan Image Kontext Stable Diffusion"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-27 02:46:49 UTC 75.6K followers, XXX engagements

"The race to put a humanoid robot in every home just accelerated massively with this $5900 @UnitreeRobotics R1. The robot comes equipped with binocular vision backed by LLM image and voice identification capabilities. Its about X feet tall and weighs roughly XX lbs. 4-microphone array speakers an 8-core CPU and GPU XX joints and hands. Ill gladly shell out sub $6K for it to tag along with me while I walk around a tough part of the town. "Movement first tasks as well (A diversity of movement is the foundation for completing tasks)". πŸ’°The Competition Around Pricing Unitrees this move to bring"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-25 18:47:18 UTC 75.6K followers, 4539 engagements

"@ASchulz888 yes IMO that's why those offers make sense. from here on till AGI one few top-minds matter till a company gets a super-powerful model"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-27 03:44:49 UTC 75.7K followers, 1497 engagements

"Kwai KeyeVL turns messy short videos into machinefriendly stories. Kwai Keye-VL is an 8B-param MLLM built by Kuaishou (the company behind the Kwai short-video app) to understand short videos as easily as still images while still handling regular vision-language tasks. The 8Bparameter model tops video tests yet keeps strong image skills. Most existing multimodal LLMs work well on single images but struggle with the rapid scene changes audio and dense context in TikTok-style clips. Keye-VL fixes that gap by training on a XXX billion-token corpus rich in video then adding a recipe that teaches"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-04 00:38:01 UTC 75.9K followers, 1591 engagements

"Beautiful SVG BY GPT-5 --- reddit .com/r/OpenAI/comments/1mc86eh/svg_by_gpt5/"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-29 19:25:09 UTC 75.9K followers, 1663 engagements

"🧡 9/n More than $700B of Grid Investment Expected Through 2030 in the US"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-22 22:38:04 UTC 75.6K followers, XXX engagements

"This 39-page report from Kuaishou explains how the company rebuilt its video recommender system into one end-to-end generative model called OneRec Traditional recommenders run separate retrieval pre-ranking and ranking stages that waste compute on network transfers and chase conflicting goals. βš™ The Core Concepts OneRec deletes retrieval prerank and rank replacing them with one encoderdecoder that maps user context to video tokens in one forward pass. All parameters chase the same final reward so gradients stop fighting each other. High arithmetic density keeps GPUs busy with matrix"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-01 11:00:02 UTC 75.9K followers, 2015 engagements

"His original prompt. The prompt first anchors the model in a clear professional role then boxes it in with tight measurable rules ---- You are a professional-grade portfolio strategist. I have exactly $XXX and I want you to build the strongest possible stock portfolio using only full-share positions in U.S.-listed micro-cap stocks (market cap under $300M). Your objective is to generate maximum return from today (6-27-25) to X months from now (12-27-25). This is your timeframe you may not make any decisions after the end date. Under these constraints whether via short-term catalysts or"
@rohanpaul_ai Avatar @rohanpaul_ai on X 2025-07-29 23:36:03 UTC 75.9K followers, 24.4K engagements