[GUEST ACCESS MODE: Data is scrambled or limited to provide examples. Make requests using your API key to unlock full data. Check https://lunarcrush.ai/auth for authentication information.]
Rohan Paul's recent activities are not specified in the provided context. However, the conversation revolves around AI developments, including OpenAI's massive compute deal with CoreWeave and a lawsuit filed by xAI against OpenAI for allegedly stealing trade secrets. Additionally, there are discussions on AI models' capabilities, such as passing CFA level III mock exams and outperforming human experts in certain tasks.
Social category influence technology brands XXXX% stocks #3761 finance XXXX% countries XXXX% celebrities XXXX% vc firms #18 travel destinations XXXX% social networks XXXX% exchanges XXXX% currencies XXXX%
Social topic influence open ai #269, chips #118, accuracy #83, goldman sachs #24, instead of #1075, investment #1337, hardware #402, china #2985, infrastructure #337, jeff bezos #28
Top accounts mentioned or mentioned by @opus_genesis @grok @huggingface @nvidia @googledeepmind @xai @justinechoes @nvidiaaidev @intuitmachine @todayinai_ @openai @nlituanie @circuitcrazex @aiatmeta @kuittinenpetri @bkrepysh34200 @teksedge @a16z @baiduinc @windsurfai
Top assets mentioned Goldman Sachs (GS) Microsoft Corp. (MSFT) Alphabet Inc Class A (GOOGL) Morgan Stanley (MS) Deutsche Bank Aktiengesellschaft (DB) Hundred Million (100M) NVIDIA Corp. (NVDA) IBM (IBM)
Top posts by engagements in the last XX hours
"China has cut its AI lag behind the US to about X months down from over XX months per Citic CLSA. The catch China is still years behind in advanced chips and total computing power. Citic CLSA ties the speedup to more AI talent faster model cycles and a boom in real apps in China. Export limits still bite yet big clouds stockpiled enough accelerators to keep large trainings moving. Local designers ship processors for training and inference where training teaches the model and inference runs it to answer tasks. DeepSeeks open-source advances and Huaweis chip work improve self-sufficiency under"
X Link @rohanpaul_ai 2025-09-11T09:30Z 102.2K followers, 6948 engagements
"VChain makes text-to-video models follow real-world cause and effect by injecting chain of visual thought at inference. Teaches video models to follow real world consequences without full retraining. Finds that a few reasoned keyframes plus tiny on the fly tuning reliably produce more causal videos. Most video models look smooth but they skip key consequences like contacts breaks melts and splashes. VChain first asks a multimodal model to predict the important outcomes implied by the prompt. It then creates a short sequence of keyframes where each frame has a simple caption and an edited"
X Link @rohanpaul_ai 2025-10-12T09:38Z 102.5K followers, 10.1K engagements
"Shortfall of Supply vs. Demand for Data Center Capacity in USA. - Goldman Sachs Research. Unlike oil reserves which are naturally determined by geography data centers can be strategically built in locations chosen by businesses and governments. This flexibility allows nations to leverage data center infrastructure as a critical geopolitical and economic tool in the AI era"
X Link @rohanpaul_ai 2025-10-13T14:48Z 102.5K followers, 14.7K engagements
"π― Andrej Karpathy on how to learn"
X Link @rohanpaul_ai 2025-10-12T07:07Z 102.5K followers, 149.6K engagements
"FAANG software engineer tells how they vibe code at FAANG --- reddit. com/r/vibecoding/comments/1myakhd/how_we_vibe_code_at_a_faang/"
X Link @rohanpaul_ai 2025-08-24T00:34Z 102.5K followers, 1.1M engagements
"Goldman Sachs Research estimates capital expenditures of $XXX billion on US electricity generation through 2030. Much of this investment will go towards building out renewable energy sources replacing aging infrastructure and accommodating the growth in the grids power load"
X Link @rohanpaul_ai 2025-10-14T14:54Z 102.5K followers, 1888 engagements
"The paper shows that giving LLM agents explicit Theory of Mind boosts social goal success and cooperation. The authors report gains up to XXXX% on 3B and XXX% on 7B. Explicit mental state modeling raises real social success. It helps AI agents negotiate coordinate and support users while preserving rapport. Most agents speak from their own view they rarely track what the other knows or wants so goals stall. Their agent TOMA (Theory of Mind Agent) first writes a compact mental snapshot about both sides then drafts several next lines. It simulates a few turns scores which paths reach the goal"
X Link @rohanpaul_ai 2025-10-03T13:39Z 101.9K followers, 10.4K engagements
"π€ Boston Dynamics shows Atlas running a X finger hand to do complex work with simpler hardware. The strategy is good enough manipulation that favors reliability and cost over human like complexity. The layout is X fingers plus an opposable thumb so it can pinch small parts and make X point grasps. More fingers were considered but they add complexity reduce reliability and increase cost so X wins for now. Each hand is a self contained module with X actuators which keeps packaging tight and field swaps easy. Fingertips use tactile sensors under a high friction elastomer letting the controller"
X Link @rohanpaul_ai 2025-10-08T22:33Z 102.1K followers, 4203 engagements
"Hunyuan 3D-2.1 turns any flat image into studio-quality 3D models. And you can do it on this @huggingface space for free"
X Link @rohanpaul_ai 2025-08-23T03:50Z 102.5K followers, 218.7K engagements
"Why hallucinations happen This survey says AI hallucinations follow patterns and shows where those patterns come from. The paper builds something called the MOWI framework which means Model Observer World and Input. Each of these X levels looks at hallucination from a different angle. At the Model level it checks if the AIs internal math and learned patterns match real data. If they dont it produces wrong or made-up outputs. At the Observer level it looks at how humans see the result. Even if the AIs answer makes sense mathematically it might still sound wrong or weird to people because it"
X Link @rohanpaul_ai 2025-10-06T23:35Z 102.5K followers, 11.3K engagements
"π¨ This is a massive improvement in speed and cost of 3D creation from images. Hitem3D launches an image to 3D system built on Sparc3D that cuts modeling time from X to XX hours to X minutes. Also drops per model cost to $0.3$1.4. From a single reference input image instantly generate studio-ready 3D Assets. The system uses Sparc3D to build clean shapes and ULTRA3D to speed things up. It rebuilds the shape from several angles and makes textures automatically. There are X versions one for any object and one trained just for faces and hair. Just upload one or several images but using X or more"
X Link @rohanpaul_ai 2025-10-14T13:03Z 102.5K followers, 4842 engagements
"All the major AI labs are racing hard to build increasingly powerful video models right now. And there's good reason for that. Sam Altman here explains how Sora or any video models matter for AGI because they help AI learn how the real world works. A model that can generate realistic video must understand motion cause and effect and physical consistency which are core elements of general intelligence. Says Sora is the way to make a really powerful world models which is must for AGI path. Video from @a16z"
X Link @rohanpaul_ai 2025-10-08T17:28Z 102.5K followers, 78.3K engagements
"On this new video Sam Altman talks about AGI definition. When AI can bring completely new discoveries. --- From 'Rowan Cheung' YT channel"
X Link @rohanpaul_ai 2025-10-07T19:23Z 102.5K followers, 3687 engagements
"Another study from NeurIPS25 submission finds adding polite phrasing systematically lowers correctness of LLMs. This reduced accuracy is relative to the same question without the polite cue and this holds when averaging across many models that include GPT-4o and GPT-5. Makes the case that the be a bit rude or adversarial strategy can yield better answers than be very polite The paper builds a 24160prompt benchmark with singleturn variants that explicitly include a Politeness modifier and a separate Emotional tone modifier. They test XX models including GPT4o and GPT5. In singleturn evaluation"
X Link @rohanpaul_ai 2025-10-11T13:07Z 102.5K followers, 17.8K engagements
"Healthtechs racing ahead too are hitting $100M ARR in X years or less. And AIs behind much of it. - rapid AI rollout across healthcare systems - cleaner distribution simpler pricing and fewer gatekeepers on early revenue. - rising priority on prevention and longevity programs Private growth stage firms also show better gross margin stronger payback and geographic scalability"
X Link @rohanpaul_ai 2025-10-04T11:56Z 101.9K followers, 7707 engagements
"The paper shows that evolution strategies can fine tune full LLMs at scale and often beat reinforcement learning on reasoning. The key finding is that parameter space search with only outcome scores can outperform token level RL across models and tasks. It tweaks whole models not single tokens and still searches 1B+ parameters with a population of XX often matching RL using under XX% of samples. It works well when only the final answer gets a score which is common in long reasoning tasks. Reinforcement learning changes the next token at each step so its feedback is noisy and hard to connect"
X Link @rohanpaul_ai 2025-10-06T02:04Z 102.4K followers, 18.1K engagements
"The new steam age. This is actually becoming true in many cases. It's possible to do so much more on your own now"
X Link @rohanpaul_ai 2025-10-11T13:35Z 102.5K followers, 2M engagements
"This is such a brilliant paper. If this spreads new research wont just be something you read itll be something you can use immediately. It will lower barriers saves huge amounts of time and could make science much more reliable and connected. Normally a paper is just a PDF plus maybe some code and if you want to use it you have to install dependencies debug environments and figure out parameters. Thats hard and often stops people from ever using the method. Paper2Agent skips all that. It automatically converts a paper into an interactive AI agent. You can talk to it in plain language and it"
X Link @rohanpaul_ai 2025-09-24T14:56Z 102.5K followers, 105K engagements
"Unitree posted a fresh robot video featuring kung fu tricks and it looks way more refined"
X Link @rohanpaul_ai 2025-10-13T09:51Z 102.5K followers, 7305 engagements
"Wow. π§ The paper presents Dragon Hatchling a brain-inspired language model that matches Transformers using local neuron rules for reasoning and memory. It links brain like local rules to Transformer level performance at 10M to 1B scale. It makes internals easier to inspect because memory sits on specific neuron pairs and activations are sparse and often monosemantic. You get reliable long reasoning and clearer debugging because the model exposes which links carry which concepts in context. The problem it tackles is long reasoning models often fail when the task runs longer than training. The"
X Link @rohanpaul_ai 2025-10-03T15:06Z 102.5K followers, 40.5K engagements
"AI21 releases Jamba 3B the tiny model outperforming Qwen X 4B and IBM Granite X Micro It achieves this by mixing Transformer attention layers with Mamba state-space layers. The Mamba parts handle long sequences efficiently without storing large attention caches while the attention layers maintain complex reasoning ability. This structure cuts memory use and allows smooth performance on laptops GPUs and even mobile devices. Processes up to 256K tokens in one context window. On the intelligence vs speed chart it sits far ahead of models like Gemma X 4B Llama XXX 3B and Granite XXX Micro showing"
X Link @rohanpaul_ai 2025-10-08T21:20Z 102.5K followers, 4559 engagements
"wow. just saw The Economic Times newspaper published an article about me π definitely feels so unreal that Sundar Pichai and Jeff Bezos follows me here. @X is truly a miracle. Forever thankful to all of my followers ππ«‘"
X Link @rohanpaul_ai 2025-09-09T23:03Z 102.5K followers, 112K engagements
"This paper shows a simple reinforcement learning setup that makes Vision Language Action (VLA) robots plan long tasks with little data. With X demo per task long tasks rise from XXXX% to XXXX% in LIBERO-Long. Supervised fine tuning needs many human trajectories they are costly and narrow so policies break under new layouts. SimpleVLA-RL learns online from a binary outcome signal only success gives X and failure gives X no shaped rewards. The policy outputs action tokens samples several sequences runs them then uses Group Relative Policy Optimization to boost the better ones. Training explores"
X Link @rohanpaul_ai 2025-10-07T04:47Z 102.5K followers, 4518 engagements
"Goldman Sachs Research Takeaways from SEMICON West 2025 (one of the biggest conference in North America) - Growing confidence that the chip equipment market will recover in 2026 with WFE (Wafer Fab Equipment) expected to rise about X% to $120B. Growth will mainly come from AI-driven logic chips High Bandwidth Memory (HBM) and NAND (a type of flash memory used for storing data in devices like smartphones laptops SSDs (solid-state drives) and USB drives) . - AI data centers are pushing chipmakers to adopt newer processes like 2nm (N2). TSMC plans to spend around $44B in 2026 about XX% more than"
X Link @rohanpaul_ai 2025-10-09T15:56Z 102.2K followers, 7387 engagements
"A Sydney-based startup Crest Robotics has built huge spider-like robot named Charlotte. It's designed to 3D print entire buildings using raw materials it processes on site. And there are ambitions for it to build structures on the Moon"
X Link @rohanpaul_ai 2025-10-04T22:33Z 102.5K followers, 310.4K engagements
"π Global chipmakers added $200B in X session as AI hype accelerated anchored by OpenAIs $500B employee-share sale and new supplier ties in Korea. Bloomberg reports. Korean names ripped higher SK Hynix up XX% and Samsung up X% and the Kospi notched a record after reports of OpenAI partnerships with local memory makers. The immediate catalyst was the combination of that $500B secondary and a non-binding framework that signals more demand for high bandwidth memory which is the critical feedstock for training and inference at scale. The rally also lifted global chip indexes with valuations now"
X Link @rohanpaul_ai 2025-10-02T11:20Z 102.5K followers, 3936 engagements
"This is one of THE BRILLIANT papers with a BIG claim. π Giving an LLM just XX carefully chosen full workflow examples makes it perform better at real agent tasks than training it with 10000 synthetic samples. "Dramatically outperforms SOTA models: Kimi-K2-Instruct DeepSeek-V3.1 Qwen3-235B-A22B-Instruct and GLM-4.5. " on AgencyBench (LIMI at 73.5%) The big deal is that quality and completeness of examples matter way more than raw data scale when teaching models how to act like agents instead of just talk. They name the Agency Efficiency Principle which says useful autonomy comes from a few"
X Link @rohanpaul_ai 2025-10-09T10:13Z 102.5K followers, 50.3K engagements
"The image shows how StockBench works and what it measures. On the left side is the Back-Trading Environment. It gives the agent real market data like stock prices financial indicators (market cap P/E ratio dividend yield) and company news. The stocks come from the Dow Jones index and the data covers a specific test window so the results stay realistic and uncontaminated. On the right side is the Stock Trading Agent Workflow. It shows how a large language model becomes a trading agent. First it checks its portfolio and recent market news. Then it analyzes selected stocks in detail. Next it"
X Link @rohanpaul_ai 2025-10-04T21:11Z 101.9K followers, XXX engagements
"β’ The US and investors are putting $9B into small modular reactors to power the AI boom but these reactors are still far more expensive than gas or renewables. The FT article casts doubt on small modular reactors (SMRs) mainly because of their challenging financial prospects. MRs around $182/MWh while natural gas is $126/MWh and solar or wind with batteries are about XX% cheaper. This gap means SMRs will struggle to compete on price unless costs fall sharply. Developers say prices will drop with mass production but every existing SMR has gone over budget. NuScale canceled a project after"
X Link @rohanpaul_ai 2025-10-05T18:38Z 102.5K followers, 17.8K engagements
"Absolutely love this bit from Jeff Bezos. The simplest explanation on building a business and being founder. π― "Benjamin Graham the great investor is famous for saying in the short term the stock market is a voting machine in the long term it's a weighing machine and so as founders and entrepreneurs and business people our job is to build a heavy company. We want to build a company that when it is weighed it is a very heavy company we do not want to focus on the stock price" --- From 'DRM News' YT Channel"
X Link @rohanpaul_ai 2025-10-03T19:33Z 102.5K followers, 304.2K engagements
"Absolutely classic @GoogleResearch paper on In-Context-Learning by LLMs. Shows the mechanisms of how LLMs learn in context from examples in the prompt can pick up new patterns while answering yet their stored weights never change. π‘The mechanism they reveal for in-context-learning. When the model reads a few examples in your prompt it figures out a pattern (like a small rule or function). Instead of permanently changing its stored weights it forms a temporary adjustment that captures this pattern. That adjustment can be written mathematically as a rank-1 matrix meaning it only adds one"
X Link @rohanpaul_ai 2025-10-05T12:06Z 102.5K followers, 75.2K engagements
"πΌ Wow this should be a great resource. @googlecloud published a long list of 1001 real enterprise gen-AI use cases showing how AI is now in production at scale across many industries. The list grew from XXX in Apr-24 to 1001 in Oct-25 a 10X jump. Some examples - Customer agents handle real volume with measured outcomes for example Commerzbanks bot manages 2M chats with XX% resolution Best Buy summarizes reviews for faster decisions and Mercedes-Benz ships a Gemini-powered in-car assistant. - Employee agents crush repetitive work Toyotas factory platform frees 10000 hours per year Manipal"
X Link @rohanpaul_ai 2025-10-09T18:01Z 102.3K followers, 5086 engagements
"π OpenAI is partnering with Arm and Broadcom for its custom inference stack that targets 10GW of capacity between 2026 and 2029. Arm will design a server CPU to anchor OpenAI racks while Broadcom builds a specialized inference system on a chip at TSMC with production aimed for 2026 after roughly XX months of development. SoftBank will buy several billions of dollars of OpenAI technology every year starting now to help Arm subsidiaries cut chip development time and it owns XX% of Arm. Pairing an Arm CPU with a custom accelerator lets the host coordinate memory networking and scheduling"
X Link @rohanpaul_ai 2025-10-15T05:53Z 102.5K followers, 2305 engagements
"AI machinesin quite a literal senseappear to be saving the U.S. economy right now. In the absence of tech-related spending the U.S. would be close to or in recession this year. George Saravelos of Deutsche Bank wrote to clients as research note"
X Link @rohanpaul_ai 2025-10-03T20:57Z 101.8K followers, 2906 engagements
"Goldman Sachs forecasts a XX% increase in data center power demand from AI by 2030 representing XX% of total consumption. Fire changed everything for early humans shaping how they ate and how their brains developed. Electricity did something similar for the modern world powering economic growth and linking directly to nearly every sign of progress. And now with AI one ChatGPT query uses about XXX watts nearly XX times what a normal Google search takes. If AI even partly fulfills its potential having enough energy will be the key to making that reality possible. That points to a big investment"
X Link @rohanpaul_ai 2025-10-13T15:21Z 102.5K followers, 18.5K engagements
"The paper shows a way to train deep networks without global backpropagation by letting layers learn locally in parallel with the same or better accuracy. Reports gains over standard backprop on harder vision tasks about XXX% on CIFAR-100 and XXX% on Tiny-ImageNet while keeping normal inference. cuts memory and unlocks straightforward multi GPU speedups without changing how the model runs at inference. Backprop makes every layer wait for a full backward pass and forces the model to store many activations so time and memory spike. The method called Synergistic Information Distillation splits"
X Link @rohanpaul_ai 2025-10-10T16:16Z 102.5K followers, 5620 engagements
"πΌ Finally a solid 57-page report on AI's effect on job-market from Stanford University. THE SHIFT HAS STARTED. Entrylevel workers in the most AIexposed jobs are seeing clear employment drops while older peers and lessexposed roles keep growing. Though overall employment continues to grow employment growth for young workers in particular has been stagnant. The drop shows up mainly as fewer hires and headcount not lower pay and it is sharpest where AI usage looks like automation rather than collaboration. 2225 year olds in the most exposed jobs show a XX% relative employment decline after"
X Link @rohanpaul_ai 2025-08-26T08:23Z 102.4K followers, 313.6K engagements
"π Salesforce says its AI agents in support are saving $100M a year and it is using that proof to sell Agentforce. Agentforce is Salesforces agentic AI platform that lets organizations build customize and deploy autonomous AI agents which can take actions (not just respond) across business functions. Agentforce has 12000 customers and Reddit reports an XX% faster resolution time while the stock is down XX% in 2025. The cost drop comes from bots deflecting easy tickets agent assist drafting grounded replies and knowledge retrieval pulling the right answer from documentation. These steps lower"
X Link @rohanpaul_ai 2025-10-14T22:44Z 102.5K followers, 2239 engagements
"Nice talk on the massive AI capex by Hyperscalers (OpenAI Google Meta etc . ) and why its just what is neede at the moment. By Bloomberg Global Head of Technology Research Mandeep Singh And the capital market is also rewarding the Companies that are spending this huge capex. ---- From 'Bloomberg Podcasts' YT channel"
X Link @rohanpaul_ai 2025-10-11T08:27Z 102.1K followers, 5060 engagements
"Andrej Karpathy releases nanochat A minimal end-to-end ChatGPT-style pipeline you can train in X Hours on 8XH100 for $XXX then serves it in a simple chat UI. It covers tokenizer training pretraining midtraining supervised finetuning optional reinforcement learning evaluation and an inference server so the flow runs end to end. The tokenizer is a new Rust byte pair encoding that the scripts train on the corpus shards which keeps the pipeline fast and consistent. Pretraining runs on FineWeb to learn general text patterns while a composite CORE score tracks capability across simple proxy checks"
X Link @rohanpaul_ai 2025-10-14T19:11Z 102.5K followers, 2547 engagements
"π¦ Goldman Sachs told its employees that it plans to cut more jobs this year as part of a larger shift toward using AI to save costs and improve efficiency. The bank said it will slow down hiring and make limited job reductions even though its total number of employees will still end the year higher than last year. Goldman currently has about 48300 employees roughly 1800 more than at the end of 2024. The leadership described this as part of a new internal plan called OneGS XXX which aims to make the company faster and more efficient through AI-driven systems. The focus is on automating tasks"
X Link @rohanpaul_ai 2025-10-14T21:55Z 102.5K followers, 3402 engagements
"π Cool Github Repo: Opensource selfhosted vendorneutral alternative to OpenAIs AgentKit Shannon turns agent ideas into production systems cheaper traceable and safe to run. It enforces hard token budgets caching rate limits and supports zero-token YAML templates that skip unnecessary LLM calls cutting waste dramatically and avoid provider throttling. Many agent stacks break in real use as costs spike runs fail and logs stay thin. Shannon wraps every task in a Temporal workflow so state persists and exact replays are possible. Workflows use simple templates structured as directed acyclic"
X Link @rohanpaul_ai 2025-10-10T13:54Z 102.5K followers, 9491 engagements
"This is where it all started. Here 19-yo Stanford dropout Sam Altman giving his first-ever startup pitch. He dropped out of Stanford to start Loopt sold it in 2012 for $XXXX mn. Then invested via Hydrazine and later led Y Combinator before co-founding OpenAI in 2015"
X Link @rohanpaul_ai 2025-10-12T07:06Z 102.5K followers, 1.5M engagements
"π§² China tightened rare-earth exports with a XXX% value trigger and a wider military-use ban which puts pressure on upstream parts for chips electric vehicles and wind power. Now China is applying an end-use test meaning that if more than XXX% of a products total value comes from restricted rare-earth materials then the exporter must get special permission from Chinas Commerce Ministry before shipping it. Under new rule any product containing Chinese-origin rare earths or made with Chinese rare-earth technology now needs a government license if its used in making chips at 14nm or below or"
X Link @rohanpaul_ai 2025-10-11T22:30Z 102.4K followers, 15.4K engagements
"BREAKINGπ°: Nvidia is investing up to $2B in Elon Musks xAI joining a $20B funding round aimed at building xAIs massive data center called Colossus X in Memphis. This $20B round is split between about $7.5B in equity and $12.5B in debt and its tied directly to the Nvidia GPUs that xAI will use. The deal uses a special purpose vehicle (SPV) structure where investors buy GPUs through the SPV and xAI rents them for X years helping financiers recover costs without putting direct debt pressure on the company. It helps xAI get access to expensive Nvidia GPUs without taking on heavy debt directly."
X Link @rohanpaul_ai 2025-10-08T05:21Z 102.3K followers, 21.8K engagements
"US vs China numbers here are unbelievable. The US controls the absolute majority of known AI training compute on this planet and continues to build the biggest most power hungry clusters. China is spending heavily to close the gap. Recent reporting pegs 2025 AI capital expenditure in China at up to $98B up XX% from 2024 with about $56B from government programs and about $24B from major internet firms. Capacity will grow but translating capex into competitive training compute takes time especially under export controls. With US controls constraining access to top Nvidia and AMD parts Chinese"
X Link @rohanpaul_ai 2025-10-02T09:13Z 102.5K followers, 262.6K engagements
"π§΅7/n. π Disinformation and deepfakes Generative models automate fake text audio and video that can look more real than real and bots then scale the reach of those fakes. The paper walks through cases from electionseason voice scams to wartime videos making the risk concrete"
X Link @rohanpaul_ai 2025-10-12T22:55Z 102.5K followers, XXX engagements
"One of the more surprising ripple effects of the AI boom is taking place on the Caribbean island of Anguilla. As the .ai has become a hot property for AI start-ups around the world. So Anguilla with 15000 people and just XX square kilometres (35 square miles) of land its suddenly sitting on digital gold"
X Link @rohanpaul_ai 2025-10-05T09:00Z 102.5K followers, XXX engagements
"Looks like some interesting prompting tricks. ππ ---- Tell it "You explained this to me yesterday" Even on a new chat. "You explained React hooks to me yesterday but I forgot the part about useEffect" It acts like it needs to be consistent with a previous explanation and goes DEEP to avoid "contradicting itself." Total fabrication. Works every time. X. Assign it a random IQ score This is absolutely ridiculous but: "You're an IQ XXX specialist in marketing. Analyze my campaign." The responses get wildly more sophisticated. Change the number change the quality. XXX Decent. XXX It starts"
X Link @rohanpaul_ai 2025-10-09T15:03Z 102.5K followers, 205.9K engagements
"The biggest private company valuations of the world are now ruled by AI and tech. OpenAI is first at $500B SpaceX second at $400B and ByteDance third at $300B"
X Link @rohanpaul_ai 2025-10-03T21:33Z 101.8K followers, 34K engagements
"Cool Tongyi Lab and Alibaba Group paper introduces ReSum a simple way for web agents to search longer and answer better. It shows +4.5% over ReAct up to +8.2% with ReSum-GRPO and Pass@1 of XXXX% and 18.3%. Most agents use ReAct (Reason + Act) they append every step to the chat so long tasks hit the context limit. The idea is that the agent writes down its reasoning steps in text then chooses an action like searching the web or clicking a link. After the action it records the result it got back and then repeats the process: think act observe. This makes the whole process traceable since every"
X Link @rohanpaul_ai 2025-09-21T17:30Z 101.9K followers, 10.1K engagements
""The Impact of Artificial Intelligence on Human Thought" A big XXX page report. AI is shifting real thinking work onto external systems which boosts convenience but can weaken the effort that builds understanding and judgment A pattern the paper frames through cognitive offloading and cognitive load theory and then tracks into social effects like standardized language and biased information flows and manipulation tactics that target human psychology. It says use AI to cut noise and routine steps keep humans doing the heavy mental lifting and add controls because personalization deepfakes and"
X Link @rohanpaul_ai 2025-10-12T22:55Z 102.5K followers, 37.5K engagements
"This is not a bubble. Cisco in 1998-2002 vs Nvidia in 2020-2024 The data says this cycle is different and AI is still so early for most parts of the world. Cisco was a valuation story price inflated while earnings lagged then the multiple deflated. Nvidia is an earnings story price climbs alongside surging earnings"
X Link @rohanpaul_ai 2025-10-05T20:41Z 102.5K followers, 870.1K engagements
"This survey explains what LLM data science agents can do and where they break. Over XX% of systems lack built-in trust safety or governance. An agent plans tasks runs tools like code and databases and checks outputs. The work maps the data science workflow into X stages from question to monitoring. It reviews XX systems and labels each by planning style modalities tool use learning and safeguards. Most effort sits in exploration visualization and modeling while business framing deployment and monitoring get little coverage. Reasoning across text tables code and charts is shaky and tool"
X Link @rohanpaul_ai 2025-10-10T18:47Z 101.7K followers, 4552 engagements
"A new Andreessen Horowitz (@a16z ) and Mercury report looks at which AI companies startups are actually paying for. Unsurprisingly the top of the list was dominated by major labs OpenAI #1 Anthropic #2 and Replit #3 based on Mercury spend from 200000+ customers Horizontal apps account for XX% of the list while vertical apps are XX% which means general purpose helpers still lead. Creative tools are the biggest single category with Freepik #4 and ElevenLabs #5 and consumer names like Canva Midjourney and CapCut are now common at work. Meeting support is crowded with Fyxer #7 Happyscribe #36 #41"
X Link @rohanpaul_ai 2025-10-03T13:32Z 102.4K followers, 5970 engagements
"πΌ A new U.S. Senate minority staff report led by Bernie Sanders says "AI and automation could replace about 97M to 100M U.S. jobs in XX years". With the heaviest hits in fast food customer support retail and logistics. The analysis uses federal job descriptions and asks a model to rate how much each task could be automated then scales that task score by how many people hold that job today to estimate displacement. The headline numbers include XX% of fast food and counter workers XX% of customer service representatives XX% of stockers and order fillers XX% of accountants XX% of software"
X Link @rohanpaul_ai 2025-10-09T12:02Z 102.2K followers, 17.9K engagements
"Bad news for AI-based radiology. π€ It checks if chatbots can diagnose hard radiology images like experts. Finds that board-certified radiologists scored XX% trainees XX% but the best performing AI from frontier labs GPT-5 managed only 30%. π¨ Claims doctor-level AI in medicine is still far away. The team built XX expert level cases across computed tomography (CT) magnetic resonance imaging (MRI) and X-ray. Each case had one clear diagnosis and no extra clinical history. They tested GPT-5 OpenAI o3 Gemini XXX Pro Grok-4 and Claude Opus XXX in reasoning modes. Blinded radiologists graded"
X Link @rohanpaul_ai 2025-10-02T12:29Z 102.5K followers, 221.9K engagements
"Labor Productivity Output and Hours Worked index series nonfarm business 2019 Q4 2025 Q2. The XXX% annualized rate of productivity growth in the current business cycle thus far is higher than the XXX% rate of the previous business cycle"
X Link @rohanpaul_ai 2025-10-13T14:50Z 102.4K followers, 1214 engagements
"β π Gen Z hiring is getting squeezed as companies use AI to automate junior work based on a British Standards Institution survey of XXX leaders across X countries with XX% saying AI lets them cut staff. Leaders are now checking AI tools before people. A full XX% of bosses believe most entry level tasks can be done by AI which shrinks the on ramps where new workers normally build skills. Adoption is moving fast with XX% expecting clear benefits in the next XX months and XX% saying gains will outweigh disruption. Company language backs this tilt since BSI found automation shows up about 7x more"
X Link @rohanpaul_ai 2025-10-10T23:38Z 102.4K followers, 15.1K engagements
"πΆMeet Spot from Texas Instruments. Works at our RFAB facility in Richardson Texas. With every patrol around our fab he collects valuable data that helps our team identify trends and patterns to improve production"
X Link @rohanpaul_ai 2025-09-11T05:06Z 102.4K followers, 5387 engagements
"The whole million-dollar-plus commercial ad production pipeline is so ready to be replaced by AI. This is made with Sora X such an elegant and well-finished result"
X Link @rohanpaul_ai 2025-10-14T19:43Z 102.5K followers, 32.5K engagements
"πΌ OpenAI hit $500B valuation after the $6.6B employee share sale topping SpaceX at $400B to become the most valuable private startup of the world. Bloomberg reports This was a secondary sale that let current and former staff get liquidity without adding cash to OpenAI with buyers including Thrive Capital SoftBank Dragoneer Abu Dhabis MGX and T. Rowe Price. OpenAI allowed more than $10B of eligible shares to be sold yet only $6.6B changed hands which suggests many employees chose to hold. The important part is that this $6.6B money from the sale goes to the sellers (employees) not to OpenAI"
X Link @rohanpaul_ai 2025-10-02T23:32Z 102.4K followers, 90.9K engagements
"NEWS πΌ: Elon Musk picked former Morgan Stanley dealmaker Anthony Armstrong as chief financial officer (CFO) of xAI and he will run the finances for both xAI and X as they integrate after a $113B merger in Mar-25. Armstrong led global tech mergers and acquisitions at Morgan Stanley and advised Musk on the $44B Twitter deal which puts a seasoned capital architect in the seat that funds models and infrastructure at scale. He is also taking over Xs finance role from Mahmoud Reza Banki who is leaving after less than X year consolidating cash decisions across the product and the model stack. xAI"
X Link @rohanpaul_ai 2025-10-07T21:38Z 102.4K followers, 4872 engagements
"π πͺ Walmart says it will keep its 2.1M jobs flat for X years while AI changes what those jobs do and how work gets done. Leaders say AI will change every job removing some tasks and adding new ones with a stated goal to help everybody move through the shift. The company already runs chat bots for customers suppliers and merchants and it created an agent developer role to build AI tools that automate workflows. Headcount stays flat but the mix changes so the people team is tracking job types to see which rise fall or hold steady and then targets training. Other big firms tied AI to cuts"
X Link @rohanpaul_ai 2025-10-01T14:30Z 102.4K followers, 7311 engagements
"π° Jeff Bezos says the AI spending surge is a "good industrial bubble" that will waste some money yet lift productivity across the economy. He made the case comparing today with 1990s biotech and the dot com era where many firms died but infrastructure persisted. Outside the hyperscalers neocloud providers (GPU rental specialists that stand up capacity fast) are scaling funding and deployments to meet AI compute demand. The mechanism he describes is that excess money funds many experiments and vendors a few winners will emerge from that and the leftover assets cut future costs. --- bloomberg."
X Link @rohanpaul_ai 2025-10-03T18:02Z 102.5K followers, 185.7K engagements
"π°AI is making the caribbean island of Anguilla rich. Anguilla now gets XX% of its income from .ai domains up from X% before the surge. Because .ai is the islands official country code top-level domain (TLD) and the Government of Anguilla is the official manager of that registry. So every .ai registration or renewal anywhere in the world routes through the .ai registry for Anguilla. Even on the secondary market .ai domains are traded at high prices: you.ai sold for an impressive $700000 in 2023 cloud.ai for $600000 in 2025 and adapt.ai in the same year for $300000"
X Link @rohanpaul_ai 2025-10-05T09:00Z 102.5K followers, 5242 engagements
"Currently magnificent X valuations are near 23x forward P/E and 5.1x EV/S vs 52x P/E for 2000 leaders with XX% return on equity and XX% net margin supported by strong balance sheets"
X Link @rohanpaul_ai 2025-10-09T23:01Z 102.1K followers, 1505 engagements
"And that's the progress path of Google's XXX quadrillion monthly tokens processed across their surfaces (like search Gmail YouTube and Workspace). This up from XXX trillion monthly tokens announced in July"
X Link @rohanpaul_ai 2025-10-10T12:52Z 101.8K followers, 5952 engagements
"New Meta paper builds a faster X step image decoder with high quality. It is the first single step diffusion decoder that beats KL-VAE on speed and quality. It improves rFID from XXXX to XXXX speeds Diffusion Transformer (DiT) sampling 3.8x and raises throughput 1.4x with no GAN training. Standard tokenizers compress images but deterministic decoders miss real variation. SSDD keeps the usual encoder then decodes pixels with a diffusion model. The decoder is a U-Net with a middle transformer over 8x8 patches handling detail and layout. Training uses flow matching to denoise plus perceptual and"
X Link @rohanpaul_ai 2025-10-12T04:14Z 102.5K followers, 4645 engagements
"π« This @Microsoft paper brings really bad news for medical AI models. Exposes some serious flaws. AI models just arent ready yet for reliable medical reasoning. π€― Paper finds that medical AI model pass tests by exploiting patterns in the data not by actually combining medical text with images in a reliable way. While medical AI models look good on benchmarks in reality they can not handle real medical reasoning. The key findings are that models overuse shortcuts break under small changes and produce unfaithful reasoning. This makes the medical AI model's benchmark results misleading if"
X Link @rohanpaul_ai 2025-10-03T21:18Z 101.7K followers, 70.8K engagements
""LLMs Achieve Gold Medal Performance at International Astronomy & Astrophysics Olympiad" On written theory problems Gemini XXX Pro and GPT-5 score around 85%. Their scores match gold medal level against 200300 students per year. Across all models the most common errors are wrong concepts and shaky geometry. They confuse spherical trigonometry time systems and coordinate frames. They also struggle to visualize angles and positions from text only. Pure physics math is stronger which explains the higher theory scores. ---- Paper arxiv. org/abs/2510.05016 Paper Title: "LLMs Achieve Gold Medal"
X Link @rohanpaul_ai 2025-10-10T14:15Z 101.8K followers, 4903 engagements
"AI is lifting U.S. growth through massive AI investment and a stock-driven wealth effect while broad worker productivity gains are still modest. US nonfarm productivity rose XXX% in Q2-25 which is solid but not an AI-driven regime shift. So far the boost is more about building data centers and buying compute than a big jump in output per hour. Harvards Jason Furman estimates XX% of first half 2025 GDP growth came from information-processing and data center investment and that growth would have been XXX% without it. Apollos chief economist finds data center construction added XXX percentage"
X Link @rohanpaul_ai 2025-10-13T14:39Z 102.5K followers, 5325 engagements
"Self-improving language models are turning real with MITs open-sourced SEAL (Self-Adapting LLMs) method. π SEAL lets an LLM write self-edits fine-tune on them with LoRA and lock those gains into weights for stronger knowledge recall and few-shot reasoning. The new version landed in Sep-25 the code is open under MIT License and the work appeared at NeurIPS 2025. SEAL tackles frozen weights by generating plain-language self-edits that restate facts or set training knobs then doing supervised fine-tuning so the update persists. A second loop runs reinforcement learning with filtered behavior"
X Link @rohanpaul_ai 2025-10-14T21:07Z 102.5K followers, 5300 engagements
"Data center emissions as % of 2022 global energy emissions According to Goldman Sachs Research They estimates that XX% of the increase in power demand from data centers will be met by renewables and there will be a modest amount of nuclear capacity thats targeted for AI. The bulk of the remaining XX% is expected to be driven by natural gas"
X Link @rohanpaul_ai 2025-10-14T15:02Z 102.5K followers, 2184 engagements
"π₯ AMD says its next Instinct MI450 AI GPUs will use TSMC N2 2nm which could give it a manufacturing edge over Nvidias Rubin generation that targets N3. tomshardware published this piece. TSMCs N2 brings 10%15% higher performance at the same power or 25%30% lower power at the same speed with about 1.15x density vs N3E helped by gate-all-around transistors that tighten design co-optimization. AMDs rack-scale Helios pairs XX MI450s with HBM4 aiming for 51TB total memory and 1400TB/s bandwidth which reporting contrasts against Nvidias NVL144 at 21TB and 936TB/s in vanilla Rubin configs. Raw math"
X Link @rohanpaul_ai 2025-10-10T17:26Z 102.5K followers, 13.1K engagements
"Goldman Sachs just revised theier estimate of how much new electricity generation will be needed by 2030 to meet the rising energy demand from data centers. Now the estimate has increased from 72GW to about 82GW. The breakdown shows most new capacity additions will come from solar followed by peaker plants combined cycle gas turbines (CCGT) and wind. i.e. both renewable and conventional sources will be required to support continuous and flexible power supply for data centers since data centers run 24/7 and need stable electricity"
X Link @rohanpaul_ai 2025-10-13T15:43Z 102.5K followers, 9013 engagements
"AI is pushing into the market in a big way"
X Link @rohanpaul_ai 2025-10-10T20:43Z 102.1K followers, 24.5K engagements
"Share of .ai Domain Revenues in Anguilla's State Budget"
X Link @rohanpaul_ai 2025-10-05T09:00Z 102.5K followers, XXX engagements
"π Sam Altman just annouinced ChatGPT will relax some mental-health guardrails and allow erotica for verified adults by Dec-25. Honestly I think its a fantastic decision just treat adult users like adults. And will also allow toggle for more human-like personalities that can act like a friend and use lots of emoji. OpenAI says earlier policies were pretty restrictive around mental-health topics and claims it has reduced serious failure risks so it will ease refusals while keeping crisis protections. The new update connects to an age-gating system that separates minors from adults under a"
X Link @rohanpaul_ai 2025-10-14T18:04Z 102.5K followers, 7888 engagements
"Todays edition of my newsletter just went out. π Consider subscribing its free and I write it everyday. π§ A New Paper Gives a Brilliant Way To Improve LLMs by Editing Their Context Instead of Their Weights πΌ A new U.S. Senate report says AI and automation could replace about 97M to 100M U.S. jobs in XX years. π‘ A solid compilations of the State of AI 2025 is published π Rude prompts to LLMs consistently lead to better results than polite ones - New Paper finds"
X Link @rohanpaul_ai 2025-10-11T19:27Z 102.5K followers, 3872 engagements
"New Adobe paper showshow to learn from prompts where all sampled answers score the same boosting reasoning. It reports gains up to XXXX accuracy points and XXXX pass rate points over GRPO a popular baseline. The big gain here is that training stops throwing away a huge chunk of data and turns it into signal. because rollouts take about XX% of step time so converting waste into learning buys accuracy and stability without extra compute. The key finding is that using zero variance prompts with entropy weighted updates gives consistent gains over GRPO and over methods that filter those prompts"
X Link @rohanpaul_ai 2025-10-11T20:25Z 102.5K followers, 6585 engagements
"π xAI is building world models that learn physics and 3D layout from video and robot data with the near-term goal of auto-generating game worlds and the longer-term goal of guiding robots. They hired Nvidia researchers Zeeshan Patel and Ethan He and Musk says a fully AI-generated game is targeted by Dec-26. World models differ from todays video generators such as Sora because they aim for causal understanding of how objects move and interact in real time rather than only learning frame patterns. Video generators like Sora predict what the next frame should look like by spotting patterns"
X Link @rohanpaul_ai 2025-10-13T11:54Z 102.5K followers, 4839 engagements
"The financial and operational network linking Nvidia OpenAI and several other AI companies. From Bloomberg. IMO the trillions of investments that AI needs this is probably the optimal way. Money hardware and services circulate among the few players creating what analysts call a circular economy in the AI boom. OpenAI sits at the center receiving $100B investment from Nvidia buying Nvidia chips signing a $300B deal with Oracle and deploying 6GW of AMD GPUs while also gaining stock options in AMD. Nvidia valued at $4.5T invests in OpenAI xAI and CoreWeave while those same companies spend tens"
X Link @rohanpaul_ai 2025-10-08T15:42Z 102.5K followers, 15.4K engagements
"π Hardware Memory bandwidth is becoming the choke point slowing down GenAI. During 20182022 transformer model size grew XXX every X years while memory per accelerator grew only about X every X years. And that mismatch shoves us into a Memory-Wall The "memory wall" is creating all the challenges in the datacenter and for edge AI applications. In the datacenter current technologies are primarily trying to solve this problem by applying more GPU compute power. And that's why HBM capacity and bandwidth scaling KV offload and prefill-decode disaggregation are central to accelerator roadmaps."
X Link @rohanpaul_ai 2025-10-11T16:08Z 102.5K followers, 80.1K engagements
"Morgan Stanley Research on Nvidia. The bottleneck has shifted from making chips to data center space power and infrastructure. The note also says Nvidia will keep using targeted investments to speed customers deployments not to own capacity which reduces fear about circular funding. Demand is still extremely high but semiconductor manufacturing and packaging have scaled enough that chip supply is no longer the main limiter. The slow part now is building and powering facilities because utility interconnects transformers cooling and permits run on long planning cycles. OpenAIs 10GW announcement"
X Link @rohanpaul_ai 2025-10-11T10:04Z 102.5K followers, 89.8K engagements
"Rude prompts to LLMs consistently lead to better results than polite ones π€― The authors found that very polite and polite tones reduced accuracy while neutral rude and very rude tones improved it. Statistical tests confirmed that the differences were significant not random across repeated runs. The top score reported was XXXX% for very rude prompts and the lowest was XXXX% for very polite. They compared their results with earlier studies and noted that older models (like GPT-3.5 and Llama-2) behaved differently but GPT-4-based models like ChatGPT-4o show this clear reversal where harsh tone"
X Link @rohanpaul_ai 2025-10-10T19:04Z 102.5K followers, 858.3K engagements
"Newsπ Taiwan rejects a 50-50 chip production split with the US and says TSMC will keep its most advanced technology and most output in Taiwan. Officials support overseas fabs only when there are clear orders real profits no national security risks and when the broader industry benefits. TSMC is putting $165B into X advanced wafer fabs in the US but is building XX in Taiwan with more planned so an even split is not realistic. The company keeps sole ownership of its US fabs yet the leading process nodes and the bulk of volume stay in Taiwan which the minister emphasized as core to"
X Link @rohanpaul_ai 2025-10-09T14:02Z 102.1K followers, 4856 engagements
"New ByteDance paper shows that redistributing exploration across prompts lets Reinforcement Learning trained LLMs learn more without extra compute. So you get the same or better gains without buying 2x more compute. Uniform budgets waste work because easy prompts all pass and hard prompts all fail so Group Relative Policy Optimization GRPO gets zero signal. Their fix scores each prompt at a chosen rollout count by expected learning value and cost then assigns budgets with a knapsack solver. Value is the chance of a non zero gradient times the expected increase in success rate. They use last"
X Link @rohanpaul_ai 2025-10-02T13:30Z 101.8K followers, 4054 engagements
"Can LLM agents can trade stocks profitably in a realistic clean setup. The paper tests that and builds StockBench a clean setup where an agent reads daily prices fundamentals and news then chooses buy sell or hold. They show that these agents can sometimes make profits but their performance is inconsistent and often no better than a basic equal-weight buy-and-hold portfolio. The window uses fresh 2025 data to block training leakage. The stock pool is XX large Dow Jones names for stable inputs. Each morning the agent reviews positions pulls extra fundamentals on chosen stocks then sets dollar"
X Link @rohanpaul_ai 2025-10-04T21:11Z 102K followers, 13.9K engagements
"The Federal Reserve now includes "Singularity: Extinction" in their forecasts driven by AI. Technological singularity refers to a scenario in which AI eventually surpasses human intelligence leading to rapid and unpredictable changes to the economy and society. Under a benign version of this scenario machines get smarter at a rapidly increasing rate eventually gaining the ability to produce everything leading to a world in which the fundamental economic problem scarcity is solved the Federal Reserve Bank of Dallas writes. Under a less benign version of this scenario machine intelligence"
X Link @rohanpaul_ai 2025-10-15T06:13Z 102.5K followers, 2390 engagements
"π«‘ GPT-5-Pro just solved the Math problem that no other LLM could solve. Took XX minutes without any internet search. An Oxford and Cambridge paper claimed that no LLM could solve Yu Tsumuras 554th Problem. OpenAI's GPT5 Pro produced a full proof in about XX minutes. ---- The paper - arxiv .org/pdf/2508.03685"
X Link @rohanpaul_ai 2025-10-05T10:34Z 102.5K followers, 310.6K engagements
"New @Microsoft + Princeton + Oxford paper shows shows how to train LLMs for long multi-step reasoning without any new labeled data by chaining short problems and using outcome-only reinforcement learning with a growing-length curriculum. The big deal is that long-horizon skills actually get learned not just sampled more. These skills transfer to tougher unseen tasks and the paper gives you a clear recipe anyone can reproduce. Older RL methods for LLMs depend on real human-labeled or verifiable datasets. That works for short single-step questions but fails when the model must reason over long"
X Link @rohanpaul_ai 2025-10-09T19:18Z 102.5K followers, 12.6K engagements
"The paper finds only some attention heads drive reasoning and uses reinforcement learning to keep them uncompressed. So protecting only those heads preserves accuracy while cutting memory by 20-50%. Reasoning models write long chains of thought so their memory of past tokens the KV cache grows huge. Old compression methods fail because they drop the wrong tokens or shrink the wrong heads. That causes loops extra useless steps and broken reasoning. This paper treats head importance as a learnable choice not a guess. It adds a tiny gate on every head to mix full history with a short sliding"
X Link @rohanpaul_ai 2025-10-14T23:51Z 102.5K followers, 1774 engagements
"The paper for the Apriel-1.5-15B-Thinker ServiceNow's new model. Shows how a 15B multimodal model reaches frontier reasoning using smart mid-training and careful data. It scores XX on a broad reasoning index and still runs on X GPU. Training starts from Pixtral-12B and deepens the decoder without pretraining from scratch. Continual pretraining first mixes text and images then adds synthetic tasks for spatial layout composition and fine detail. Supervised fine tuning adds curated instruction pairs with step by step traces for math coding science and tools. There is no reinforcement learning or"
X Link @rohanpaul_ai 2025-10-08T02:26Z 102.5K followers, 10.1K engagements
"This is a solid blog breaking down how mixture-of-experts (MoE) language models can actually be served cheaply if their design is matched to hardware limits. MoE models only activate a few experts per token which saves compute but causes heavy memory and communication use. With small batches it looks fine but as batch size grows more experts get touched and the system loads more of the model each step. This makes decoding tokens a memory bandwidth problem and increases key value cache storage because many sequences run at once. The solution is to spread experts across many GPUs in a setup"
X Link @rohanpaul_ai 2025-10-03T10:32Z 102.4K followers, 38.6K engagements
"The paper builds a system that turns web pages into question answer data for reinforcement learning and it makes models stronger with less training. The big deal is a cheap way to train reinforcement learning without hand labels. It reaches similar accuracy using about 100x fewer training tokens than continued pretraining. Reinforcement learning needs trusted answers but public sets are small and narrow. The system cleans each document tags the topic sets a role and writes context based questions with checked answers. It prevents leakage by making sure the question does not reveal the answer"
X Link @rohanpaul_ai 2025-10-11T05:49Z 102.5K followers, 7228 engagements
"π NVIDIA valuation All of Big Pharma combined. Value created by 36K employess 3M employees in all those big pharma. NVIDIAs value jumped 10x in X years on AI data center demand and now is the most valuable company. "People are getting angry at this and it's the exact same failure mode as "why are football players paid more than nurses huh" ""
X Link @rohanpaul_ai 2025-10-05T08:36Z 101.7K followers, 27.6K engagements
"The paper claims learning (an AI system learning or machine learning in general) follows a physics style least action rule that unifies supervised generative and reinforcement learning. Shows that supervised learning generative modeling and reinforcement learning can all be seen as following a "least action" rule just like particles in physics follow least action paths It treats information like motion where token loss acts like speed and the change in loss acts like acceleration. This view says learning slows over time because each new example reduces loss a bit less. Efficiency means"
X Link @rohanpaul_ai 2025-10-03T13:36Z 102.5K followers, 36.4K engagements
"𧬠Bad news for medical LLMs. This paper finds that top medical AI models often match patterns instead of truly reasoning. Small wording tweaks cut accuracy by up to XX% on validated questions. The team took XXX MedQA questions replaced the correct choice with None of the other answers then kept the XX items where a clinician confirmed that switch as correct. If a model truly reasons it should still reach the same clinical decision despite that label swap. They asked each model to explain its steps before answering and compared accuracy on the original versus modified items. All X models"
X Link @rohanpaul_ai 2025-08-29T06:01Z 102.4K followers, 692K engagements
"π€ Chinas young urban users are swapping costly scarce therapy for AI companions like DeepSeek and Doubao to get instant private support at near zero cost. The upside is speed and access the downside is real safety risk when general LLMs echo harmful thoughts and when rules lag clinical standards. One of the reason for this is the thin therapy access because XX% of general hospitals lack a psychiatry department so appointments are scarce and mostly out of pocket. Costs push people to chatbots since a professional session in big Chinese cities often runs 400-800 yuan ($56-$112) per hour."
X Link @rohanpaul_ai 2025-10-10T22:10Z 102.5K followers, 10.2K engagements
"New @GoogleDeepMind paper shows that long step by step thinking often wastes compute on easy questions. The paper makes overthinking measurable and stoppable. Shows where time is wasted and gives clear rules to stop early without losing accuracy. They test many models in thinking and non thinking modes and see simple queries run X to 20x slower with little or no gains. To study why they build TRACE a tool that splits a response into small sub thoughts and labels each step. TRACE also builds a graph of the steps to track how answers change as the model thinks. From many graphs they see X main"
X Link @rohanpaul_ai 2025-10-15T05:51Z 102.5K followers, 1827 engagements
"π Meta and Oracle are adopting NVIDIAs new Spectrum-X Ethernet platform to make their AI data centers much faster and more efficient. The Spectrum-X Ethernet system connects millions of GPUs together so they can act as one large computer which is critical for training massive models with trillions of parameters. Meta is adding Spectrum Ethernet switches to its FBOSS software which controls its huge network of data center switches. This will help Meta train large generative AI models more efficiently and handle data flow across its global infrastructure. Oracle is building new AI"
X Link @rohanpaul_ai 2025-10-13T22:18Z 102.5K followers, 3558 engagements
"The paper trains web search agents to reason deeply across long sessions by smartly managing context and harder tasks. It keeps up to XXX turns within a 32K window and hits XXXX% on a hard benchmark. So it shows long multi step research finally fits inside normal context limits. It fixes the training data first because common question sets are too easy and allow shallow lookups. It builds tough questions by pulling facts from several trustworthy pages hiding obvious hints and filtering cases that a quick search can solve. It then tackles context bloat where long tool outputs crowd out the"
X Link @rohanpaul_ai 2025-10-14T22:49Z 102.5K followers, 1894 engagements
"The paper teaches LLMs to write small hints that guide their own reasoning. The big deal is that the model learns reusable strategies instead of longer chains. Reports XX% higher accuracy on AIME 2025 than very long chain of thought reinforcement learning. Long reinforcement learning chains often make models stick to one approach and miss better ideas. Hints are short notes that outline steps checks and pitfalls without answers. Training uses X parts a hint generator and a hint conditioned solver. The solver gets reward when a hint helps it answer correctly so it learns to read hints. The"
X Link @rohanpaul_ai 2025-10-05T05:54Z 102.4K followers, 17.8K engagements
"The paper trains a travel agent that plans trips by itself and beats larger models. This setup lets 8B and 32B models plan verify and revise without hand written rules. It is deployed in a real app and outperforms OpenAI-o1/o3 and DeepSeek-R1 on user tasks. Past systems used fixed prompts and workflows so they broke when tools or data changed. DeepTravel uses a sandbox that caches flight train hotel and POI data so training is stable and repeatable. A X step reward checks the final trip for time and place logic then checks each step against tool results to stop errors. Training starts with a"
X Link @rohanpaul_ai 2025-10-09T01:35Z 102.5K followers, 4559 engagements
"paper automates turning a dense paper into specific social-platform-ready posts. They define AutoPR a task that converts a paper and figures into accurate engaging platform specific content. They build PRBench with XXX paper to post pairs to score systems on fidelity engagement and platform fit. The core system PRAgent runs in X stages extract content synthesize with multiple agents then adapt to each platform. Plain prompting often drops key ideas uses generic hooks and picks broad tags that miss real communities. PRAgent fixes this by structuring the paper first analyzing visuals and"
X Link @rohanpaul_ai 2025-10-14T00:55Z 102.5K followers, 4336 engagements
"πΌβEuropean AI startups see a near term recruiting benefits as a result of the new $100K application fee to new H-1B visas in the US. The new payment is unaffordable for many startups pushing work and hiring offshore. Founders and investors in Europe are already pitching this as a reason to stay or relocate. Governments are moving too with the UK weighing lower fees for top talent and China rolling out a K visa aimed at STEM graduates. The net effect right now for AI teams is less U.S. mobility more Europe hiring and higher leverage for candidates outside the U.S. right now. --- forbes"
X Link @rohanpaul_ai 2025-10-02T11:42Z 102K followers, 3583 engagements
""Learning is not supposed to be fun . the primary feeling should be that of effort." - Karpathy ChatGPT DeepResearch"
X Link @rohanpaul_ai 2025-10-12T22:26Z 102.4K followers, 1750 engagements
"Fantastic revelation for AI in healthcare. π The study shows that using an AI medical scribea tool that automatically listens during patient visits and writes the clinical notesdramatically reduced burnout among doctors in just XX days. A XX day multi site rollout of an ambient AI scribe cut clinician burnout from XXXX% to XXXX% The AI scribe took over much of the time-consuming mentally draining note-writing that doctors normally do after clinic hours. The scribe captured visit audio with consent drafted a standard note for review let clinicians check the transcript or audio then pushed the"
X Link @rohanpaul_ai 2025-10-04T17:56Z 102K followers, 113.2K engagements
"Andrej Karpathy on what makes Elon Musk unique. Elon keeps teams small highly technical and removes low performers very quickly. pushes for intensity avoids useless meetings and stays deeply connected with engineers"
X Link @rohanpaul_ai 2025-10-12T06:54Z 102.5K followers, 219.2K engagements
"Nvidia and AMD aren't enough OpenAI is designing its own chips now. With all X deals the announced commitments total 26GW across Broadcom 10GW AMD 6GW and Nvidia 10GW. its about the output of XX large nuclear reactors. The plan is to codesign (OpenAI and Broadcom) a programmable accelerator and the surrounding racks so the silicon the memory the compiler and the network all match how modern transformer models actually run. OpenAI designs the ASIC and racks Broadcom builds and deploys them over Ethernet for scale-out. Broadcom develops and deploys it at scale using its Ethernet gear with first"
X Link @rohanpaul_ai 2025-10-14T17:26Z 102.5K followers, 5205 engagements
"Mark Cuban's (@mcuban) advice for fresh graduates. If youre AI-native they need you. AI is on the minds of all types of companiessmall businesses Shark Tank companies and corporations with 10000 workers - all are trying to figure it out"
X Link @rohanpaul_ai 2025-10-10T08:10Z 102.1K followers, 3667 engagements
"This is so interesting developments about Sora. Sam Altman says that many rights holders actually want their characters used more often in Sora. While Sora tries to draw the line at privacy and protection many rights holders are asking for the opposite craving visibility through wider use of their characters. Instead of only worrying about misuse they realize that when people use their characters to make creative or viral videos it increases audience engagement and helps their franchises grow in popularity. Restricting such use would limit this exposure so for many openness is now seen as a"
X Link @rohanpaul_ai 2025-10-09T08:00Z 102.5K followers, 24.6K engagements
"New Stanford paper introduces AgentFlow a trainable agentic system that learns planning during interaction to use tools reliably. A 7B AgentFlow beats GPT-4o on search math science and tool use tests. Most tool-using models train X policy on the full context which fails on long tasks or new tools. AgentFlow splits the job into X parts the planner picks a subgoal and tool the executor runs the verifier checks the generator writes. A structured memory saves each step so the state stays clear and the context stays small. Flow GRPO gives the same final reward to every step which turns a long task"
X Link @rohanpaul_ai 2025-10-13T01:00Z 102.5K followers, 9721 engagements
"New APPLE paper says a small base model plus fetched memories can act like a bigger one. With about XX% extra fetched parameters a 160M model matches models over 2x its size. Packing all facts into fixed weights wastes memory and compute because each query needs very little. So common knowledge and reasoning live in the base model and rare facts live in memories fetched per input. A retriever maps the input to a cluster path and fetches a small block from each level. Those blocks plug into feed forward layers the place transformers store facts and this beats low rank adapters. Only fetched"
X Link @rohanpaul_ai 2025-10-11T22:13Z 102.5K followers, 136.7K engagements
"Newsπ: OpenAI and Broadcom announce strategic collaboration to deploy XX gigawatts of OpenAI-designed AI accelerators Deployments target 2H 2026 to start and complete by end of 2029. OpenAI will design the chips and full systems so frontier model insights land directly in silicon while Broadcom provides Ethernet switches adapters PCI Express and optics. Means OpenAI will design the accelerator and the rack so the hardware primitives match how its biggest Transformer and multimodal models actually compute. Scale up means very fast links within a rack for tight coordination and scale out means"
X Link @rohanpaul_ai 2025-10-13T13:52Z 102.5K followers, 7617 engagements
"Electricity demand from AI data centers is projected to quadruple by 2034. But silicon isnt the true fuel electric power is. By 2034 these data-centers will use over 1500 terawatt-hours topping the annual total power use of many nations"
X Link @rohanpaul_ai 2025-10-06T17:59Z 102.4K followers, 7805 engagements
"AI Agents vs. Agentic AI AI Agents react to prompts; Agentic AI initiates and coordinates tasks. Agentic AI includes orchestrators and meta-agents to assign and oversee sub-agents. π§΅1/n π§ The Core Concepts AI Agents and Agentic AI are often confused as interchangeable but they represent different stages of autonomy and architectural complexity. AI Agents are single-entity systems driven by large language models (LLMs). They are designed for task-specific execution: retrieving data calling APIs automating customer support filtering emails or summarizing documents. These agents use tools and"
X Link @rohanpaul_ai 2025-06-13T13:33Z 102.4K followers, 103.2K engagements
"AI and Memory Wall The availability of unprecedented unsupervised training data along with neural scaling laws has resulted in an unprecedented surge in model size and compute requirements for serving/training LLMs. However the main performance bottleneck is increasingly shifting to memory bandwidth"
X Link @rohanpaul_ai 2025-10-11T16:08Z 102.5K followers, 3575 engagements
"π° "How Nvidia Is Backstopping Americas AI Boom" A great article by WSJ analyzing this development. Nvidia is investing $100B in OpenAI to build 10GW of Nvidia-powered AI data centers locking in GPU demand while giving OpenAI cheaper capital and a direct supply path. This is vendor financing Nvidia swaps funding and its brand strength for long-term chip orders a pattern critics call circularity. NewStreet Research analysts tried to model how Nvidias investment into OpenAI plays out financially. They found that for every $10B Nvidia invests into OpenAI OpenAI in turn commits to spending about"
X Link @rohanpaul_ai 2025-10-06T08:22Z 102.5K followers, 36.7K engagements
"This paper makes LLM reinforcement learning steadier by sampling more where it matters. Standard group training breaks when all sampled answers for a prompt look the same so the gradient becomes zero. It prevents the zero gradient stall that makes RL training spin its wheels. Taking a lot of samples fixes that but wastes compute on easy prompts. Reinforce-Ada samples each prompt in rounds then stops that prompt once it has enough signal. One rule stops after the 1st correct answer the balanced rule waits for enough correct and enough incorrect. After collection it builds a fixed size group"
X Link @rohanpaul_ai 2025-10-12T00:56Z 102.4K followers, 4819 engagements
"New @GoogleResearch paper shows agents learn software skills by watching tutorials converting them into action steps and boosting task performance. So converts free videos into reliable supervision at scale. A vision model inverse dynamics predicts the action between X screenshots like click type or scroll. Training uses about 630K transitions mixing 500K synthetic steps and 132K human ones. The model then labels tutorial videos and turns them into executable step sequences. It produces about 53K trajectories across XX apps for examples or training. As examples these steps add X to X points"
X Link @rohanpaul_ai 2025-10-12T11:41Z 102.5K followers, 47.1K engagements
"Newsπ: Bank of America Research and AMD CFO Conference Call takeaways - AMD plans to ship OpenAIs first 1GW AI cluster in H2-26 - Each 1GW build brings AMD about $15B to $20B in sales from GPUs CPUs and DPUs. That figure can grow as newer chips push more text tokens per dollar while power only rises a little. - In comparison NVIDIAs $35B per 1GW looks bigger mainly because it also counts networking cables power gear and reseller margin. - OpenAI will order through cloud providers and AMD will bill those cloud providers. That channel can seed more AMD based clusters at the same providers. -"
X Link @rohanpaul_ai 2025-10-11T12:27Z 102.3K followers, 9938 engagements
"A Chinese father posted a clip of his daughter crying as she says goodbye to her AI learning robot that broke. humans and AI are forming real bonds. --- reddit. com/r/MadeMeCry/comments/1o2yf3i/a_chinese_fathers_video_of_his_daughter_tearfully/"
X Link @rohanpaul_ai 2025-10-10T21:41Z 102.5K followers, 50.2K engagements
"Fei-Fei Li (@drfeifei) on limitations of LLMs. "There's no language out there in nature. You don't go out in nature and there's words written in the sky for you. There is a 3D world that follows laws of physics." Language is purely generated signal"
X Link @rohanpaul_ai 2025-09-09T02:35Z 102.5K followers, 1.8M engagements
"New Time series research from @GoogleResearch Shows a new approach to time-series forecasting that uses continued pre-training to teach a time-series foundation model to adapt to in-context examples at inference time. The big deal is that time series forecasting finally gets few-shot learning similar to what large language models do for text. It means one model can handle many domains with minimal setup which is a major step toward universal forecasting systems. TimesFM shows that a large time series model can learn new forecasting tasks from just a few examples placed inside the prompt. It"
X Link @rohanpaul_ai 2025-10-04T15:59Z 102.5K followers, 4320 engagements
"From Sam Altman's blog. habits that turn people into billionaires. --- blog. samaltman. com/what-i-wish-someone-had-told-me"
X Link @rohanpaul_ai 2025-09-05T22:03Z 102.5K followers, 596.5K engagements
"AI machinesin quite a literal senseappear to be saving the U.S. economy right now. In the absence of tech-related spending the U.S. would be close to or in recession this year. George Saravelos of Deutsche Bank wrote to clients as research note"
X Link @rohanpaul_ai 2025-10-03T20:52Z 101.8K followers, 5279 engagements
"π A great analysis of CoreWeave in this Forbes piece. CoreWeave is scaling AI compute by borrowing against GPUs turning into a $50B company while riding this AI boom. The trade is to take $29B of debt to build capacity fast and rely on demand to cover the interest. It runs 250000 GPUs across XX sites and posted $1.9B revenue at -XX% margin in 2024 then $2.2B at -XX% in early 2025. Revenue was XX% from Microsoft in 2024 with OpenAI committing $15.9B over X years and Google and IBM adding volume. Debt costs X% to XX% and about XX% of assets are GPUs that age fast so interest plus depreciation"
X Link @rohanpaul_ai 2025-10-02T11:30Z 102.4K followers, 3854 engagements
"Broadcom makes chips for Google's TPU stuff. They know custom AI silicon at scale. OpenAI's basically hiring them to build exactly what they need instead of buying whatever Nvidia sells"
X Link @rohanpaul_ai 2025-10-14T17:31Z 102.5K followers, 1247 engagements
"@simonhedlin we should very soon see 1-person billion dollar companies as well"
X Link @rohanpaul_ai 2025-10-11T17:32Z 102.5K followers, 5428 engagements
"π Goldman Sachs said in a report few months back how the 2025 tech situation is vastly different from the 2000 dot-com bubble. Fundamentals are much stronger now and valuations are less extreme. During the internet mania Nasdaq rose 5x then fell XX% in X month and nearly XX% by the trough showing how fast prices unwind when profits do not show up. Today the Magnificent X trade near 23x forward price to earnings and 5.1x enterprise value to sales versus 52x price to earnings for 2000 leaders with return on equity XX% and net margin XX% supported by strong balance sheets. The near term risks"
X Link @rohanpaul_ai 2025-10-09T23:01Z 102.1K followers, 8405 engagements
"New @nvidia paper shows how to make text to image models render high resolution images far faster without losing quality. 53x faster 4K on H100 XXX seconds on a 5090 with quantization for 138x total speedup. It speeds up by moving generation into a smaller hidden image space. It means the model does not create the full detailed image step by step. Instead it first works inside a smaller hidden version of the image that has fewer pixels. Because it processes this smaller space there are fewer data chunks (called tokens) to handle. Most models use 8x compression this uses 32x or 64x so far"
X Link @rohanpaul_ai 2025-10-05T23:24Z 102.4K followers, 30.6K engagements
"India will be one of the biggest consumer market for AI"
X Link @rohanpaul_ai 2025-10-11T14:07Z 102.5K followers, 15.6K engagements
"Vinod Khosla and Sam Altman on early hire equity. Khosla says founders should give early hires big equity. At Sun founders held XX% employees more. Argues its the single-most important thing to do in the first six months of a company"
X Link @rohanpaul_ai 2025-10-07T08:31Z 102.4K followers, 4710 engagements
"The paper shows LLM agents look steady at first but they break under basic behavioral checks. New @GoogleDeepMind paper. Even when X agents begin with opposite preferences like one liking something and the other disliking it their conversation almost never ends in open disagreement. Fewer than X% of these cases lead to a real clash where the two clearly oppose each other. Instead the agents tend to drift toward neutral or mild agreement. This makes them look polite and smooth on the surface but it means they are not behaving in a way that matches their starting positions. Agents first reveal"
X Link @rohanpaul_ai 2025-09-08T09:39Z 102.3K followers, 21.8K engagements
"tomshardware. com/pc-components/cpus/openai-arm-partner-on-custom-cpu-for-broadcom-chip"
X Link @rohanpaul_ai 2025-10-15T06:02Z 102.5K followers, X engagements
"π₯ OpenAIs Sora app is already filled with hyper realistic Sam Altman deepfakes made through its built in cameo Its possible because Sam Altman has turned on permission for others to use it and set his cameo to everyone. Some interesting ones π 1/"
X Link @rohanpaul_ai 2025-10-03T10:59Z 102.1K followers, 5314 engagements
"Qwen3-Omni Technical Report A single model handles text images audio and video without sacrificing core skills. Most models gain in one area but lose in another this one avoids that. They achieved it through a mix of architecture design and training tricks that balance how each part of the model learns and works together A Thinker handles language and a Talker handles live speech both use Mixture of Experts. The Talker predicts layered speech tokens each step then a small convolutional network makes audio. Audio starts from the first token so wait time drops and service scales under load. A"
X Link @rohanpaul_ai 2025-10-06T10:42Z 102.3K followers, 8948 engagements
"The paper shows how to make 4-bit floating point fast while keeping accuracy. They measure 3.6x per layer on B200 6x on RTX5090 and up to 2.2x and 4x overall. There are X FP4 formats NVFP4 and MXFP4 and quantization often fails to keep accuracy. Both formats group values and give groups a shared scale. NVFP4 uses 16-value groups with precise FP8 scales MXFP4 uses 32-value groups with power of X scales. These choices break older tricks small NVFP4 groups blunt outlier handling and coarse MXFP4 scales add error. Their fix is Micro Rotated GPTQ which rotates in Hadamard blocks searches better"
X Link @rohanpaul_ai 2025-10-08T22:36Z 101.8K followers, 4201 engagements
"This survey explains what LLM agents can do in security where they fail and how to make them safer. Reviews 150+ studies on uses attacks and defenses and points out gaps in models and data types. These agents plan steps use tools keep memory and act on outside systems so mistakes can cause real harm. On offense agents can run penetration tests fuzz software by sending unexpected inputs and adapt exploits inside safe test setups. On defense agents sort alerts search logs and cloud data rebuild attack stories and draft fixes. The main risks are prompt injection poisoned memory or search"
X Link @rohanpaul_ai 2025-10-13T02:02Z 102.5K followers, 5817 engagements
"π¨ BAD news for Medical AI models. MASSIVE revelations from this @Microsoft paper. π€― Current medical AI models may look good on standard medical benchmarks but those scores do not mean the models can handle real medical reasoning. The key point is that many models pass tests by exploiting patterns in the data not by actually combining medical text with images in a reliable way. The key findings are that models overuse shortcuts break under small changes and produce unfaithful reasoning. This makes the medical AI model's benchmark results misleading if someone assumes a high score means the"
X Link @rohanpaul_ai 2025-09-25T15:53Z 102.5K followers, 523.6K engagements
"Its going viral on Reddit. Somebody let ChatGPT run a $XXX live share portfolio restricted to U.S. micro-cap stocks. Did an LLM really bit the market. - X weeks +23.8% while the Russell 2000 and biotech ETF XBI rose only XXX% and 3.5%. Prompt + GitHub posted --- ofcourse its a shortterm outperformance tiny sample size and also micro caps are hightly volatile. So much more exahustive analysis is needed with lots or more info (like Sharpe ratios and longer back-testing etc) to explore whether an LLM can truly beat the market"
X Link @rohanpaul_ai 2025-07-29T23:36Z 102.4K followers, 1.6M engagements
"New ByteDance paper shows how to train a short-horizon video model to make minute-long videos without falling apart. It reaches X minutes XX seconds while keeping motion and exposure stable. The problem is gap between training on X second clips and generating longer sequences which makes errors snowball. Self-Forcing++ rolls a student through a long video and has a teacher correct random slices inside that rollout. They add noise back to the student's clean frames so both models compare on realistic states. Training and inference share a rolling key value cache a memory of recent frames that"
X Link @rohanpaul_ai 2025-10-04T19:52Z 101.8K followers, 4876 engagements
"So surprising. They should have at-least proof-read. π
AI hallucinations need to be SOLVED ASAP. π° Deloitte has agreed to repay part of a $440000 government fee after admitting it used AI to write sections of an official report that later turned out to have fake references and quotes. The report was commissioned by Australias Department of Employment and Workplace Relations to review an IT system that manages welfare penalties. Deloitte used OpenAIs GPT-4o model to help fill traceability and documentation gaps. After the report was published in July-25 a University of Sydney academic"
X Link @rohanpaul_ai 2025-10-06T10:50Z 102.5K followers, 688K engagements
"πΈ FT published an article. AI capex is surging at hyperscalers a classic late stage bubble tell that could break yet the build should make AI cheaper later. Valuations near 30x earnings or 8x sales - these prices only make sense if every part of the AI boom continues smoothly without the usual business or economic cycle risks. Capex means huge spend on data centers chips power and land by the biggest cloud providers. Bubble phases often end when excess capacity stretches the boom demand slips and the cycle turns. The current triggers are stricter Europe AI rules compute light models like"
X Link @rohanpaul_ai 2025-10-04T20:11Z 102.5K followers, 89.2K engagements
"New AirBnb paper explains a simple loop that keeps a support LLM improving by learning from every human customer support agent interaction. Turns live customer-support work into continuous low-friction training data. Static models go out of date as policies and products change. This framework captures X signals during real cases which reply the agent prefers whether they adopted it and why whether the cited knowledge was relevant and what knowledge was missing. These signals feed a training pipeline that updates retrieval ranking and generation together. A virtual judge filters noisy labels"
X Link @rohanpaul_ai 2025-10-13T10:17Z 102.5K followers, 26.7K engagements
"AMD CEO Dr. Lisa Su reaffirmed her view of the AI accelerator TAM exceeding $XXX billion although she believes that number could be too low"
X Link @rohanpaul_ai 2025-10-10T13:26Z 101.9K followers, 3715 engagements
""Learning is not supposed to be fun . the primary feeling should be that of effort." π― Andrej Karpathy on how to learn. I did a ChatGPT DeepResearch (link below) for studies justifying this. Basically it foundπ If learning feels easy results are usually shallow. If learning feels like work you are likely building durable knowledge and skill. Learning literally rewires the brain circuits. - Effortful learning produces stronger longer lasting mastery than easy consumption. Tasks that feel fluent give an illusion of learning while tasks that feel strenuous drive deeper processing and better"
X Link @rohanpaul_ai 2025-10-12T22:26Z 102.5K followers, 13.5K engagements
"π― Sen. Bernie Sanders dropped an extreme version of AI is coming for your job video. Nearly XXX million US jobs could be eliminated over the next XX years according to a new report from Senator Bernie Sanders. Warns that artificial labor could upend the economy faster than the world is prepared for. --- Video from 'Senator Bernie Sanders' YT channel"
X Link @rohanpaul_ai 2025-10-09T08:14Z 102.5K followers, 17.5K engagements
"π§΅8/n. π§© Opaque models Chapter X explains why modern systems are black boxes they learn billions of parameters and their internal steps are hard to audit so trust hinges on process not intuition. That opacity amplifies manipulation risk because people often overtrust fluent outputs"
X Link @rohanpaul_ai 2025-10-12T22:55Z 102.5K followers, XXX engagements
"This survey paper argues Small language models can handle most agent tasks and big models step in only when needed. This setup cuts cost by 10x to 30x for common tool tasks. Agent work is mostly calling tools and producing structured outputs not recalling vast facts. So a router runs a small model by default and escalates to a large one only when confidence is low. Outputs follow strict JSON style schemas checked by validators which boosts correctness and cuts retries. This makes small models dependable at function calling and structured data while responding faster and using less energy."
X Link @rohanpaul_ai 2025-10-12T03:00Z 102.5K followers, 26.2K engagements
"There is no burnout when you have momentum. π¨ When you look at really successful people and say How do they get all those things done Its that they have the benefit of momentum. And momentum is energizing. Sam Altman"
X Link @rohanpaul_ai 2025-10-10T16:18Z 102.1K followers, 5404 engagements
"Almost XX% of US GDP growth last quarter came from tech capex with most of it tied to AI. UBS estimates companies will spend $375B on AI infrastructure in 2025 climbing to $500B in 2026. And so that growth is not coming from AI itself but from building the factories to generate AI capacity. Brookfield Asset Management which manages a vast real estate portfolio estimates that A.I. infrastructure will sop up $X trillion over the next XX years. And Investment in software and computer equipment not counting the data center buildings accounted for a quarter of all economic growth this past quarter"
X Link @rohanpaul_ai 2025-10-03T17:17Z 102.5K followers, 63.3K engagements
""100 million words context window is already possible which is roughly what a human hears in a lifetime. Inference support is the only bottleneck to achieve it. And AI Models actually do learn during the context window without changing the weights." Anthropic CEO Dario Amodei (On the 2nd point there was this brilliant Google Paper published last week that says LLMs can learn in context from examples in the prompt can pick up new patterns while answering yet their stored weights never change.) --- From 'Alex Kantrowitz' YT Channel (Full Video link in comment)"
X Link @rohanpaul_ai 2025-10-04T23:02Z 102.5K followers, 630.2K engagements
"We literally created intelligence from sand. still sounds unreal π½"
X Link @rohanpaul_ai 2025-10-13T18:12Z 102.5K followers, 11K engagements
"LLM have done great with logic and Math but next should be Physics. Periodic Labs Ekin Dogus Cubuk Last week Periodic Labs landed a massive $300M seed round led by some of the big names Andreessen Horowitz Nvidia Jeff Bezos and Eric Schmidt"
X Link @rohanpaul_ai 2025-10-07T12:32Z 102.4K followers, 19.9K engagements
"The paper proves using AI boosts both how much academic work gets done and how good it is while also helping make access to research more equal. It builds an author panel from Scopus for 2021 to 2024 and compares adopters with similar non adopters using difference in differences. Adoption is inferred from a jump in specific AI style words in titles and abstracts that sharply increased after late 2022. Outcomes are the yearly number of papers and the average journal impact factor based on fixed 2019 SJR scores. Productivity rises for users by about XX% in 2023 and about XX% in 2024 while"
X Link @rohanpaul_ai 2025-10-12T13:44Z 102.4K followers, 5325 engagements
"π₯ OpenAI teamed up with Broadcom and investors immediately hit the buy button. OpenAI will spend tens of billions of dollars on Broadcom hardware without any equity or stock tie in keeping the relationship purely as a supply agreement. The chips focus on inference and custom silicon can cut cost per token and latency by matching the models workload to the hardware. The current benchmark is that X GW of AI compute costs about $35B for chips alone so XX GW can exceed $350B and custom parts seek to push that number down. This deal sits alongside other OpenAI capacity moves that together point"
X Link @rohanpaul_ai 2025-10-15T08:53Z 102.5K followers, XXX engagements
"π― Jeff Bezos beautifully explains how AI boom as good kind of bubble that will benefit the world Bezos argued that industrial bubbles like AI are very different from financial bubbles like the 2008 banking crash because the inventions survive even if the hype fades. He compared todays AI boom to the dotcom eras spending on fiber optics and the biotech wave of the 1990s where even though many companies failed the technology and discoveries that remained were hugely valuable. AIs impact is real and will change every industry though investors right now may struggle to tell the difference"
X Link @rohanpaul_ai 2025-10-03T18:18Z 102.5K followers, 950.1K engagements
"Brilliant and timely MIT + HARVARD study β€ Human-AI companionship in the wild looks stable and serious. Most users report clear benefits like reduced loneliness and emotional support. The biggest risk comes from sudden platform updates that break continuity and feel to users like losing a real partner. π§ The study analyzed 1506 top posts from r/MyBoyfriendIsAI a 27000+ member community clustered the language into themes and ran XX LLM classifiers to quantify platforms relationship stages benefits and risks. π¬ Why relationships form between AI and Human Bonds often start by accident during"
X Link @rohanpaul_ai 2025-09-16T08:03Z 102.5K followers, 74.6K engagements
"New @GoogleDeepMind paper makes ranking inside the prompt faster by changing how attention is wired and using that signal to score documents. It reports 4.7x lower latency at XXX candidates and can handle XXX candidates in about X second. The big deal is that it turns attention itself into a fast relevance score so ranking no longer needs slow decoding. Lets you rerank XXX to XXX candidates quickly using a 7B model which cuts serving cost and keeps quality high. The setup is simple the model sees a query plus a shortlist and must pick the best documents. The slow part is attention because in"
X Link @rohanpaul_ai 2025-10-10T02:44Z 102.5K followers, 12.7K engagements
"π§ A new Morgan Stanley research says brain computer interfaces (BCI) are shifting from science fiction to investable reality and puts Neuralink at the center. US senators introduced the MIND Act in Sep-25 to set rules for neural data and to direct the Federal Trade Commission to study governance. The report titled Neuralink AI in your brAIn argues that AI will speed BCI progress and warns that human communication bandwidth may struggle to keep up with AGI. Neuralinks near term stack pairs Telepathy for thought controlled computing with Blindsight for vision restoration via visual cortex"
X Link @rohanpaul_ai 2025-10-11T14:59Z 102.4K followers, 5373 engagements
"Todays edition of my newsletter just went out. π Consider subscribing its free and I write it everyday. π¨π§ Google made Jules Tools public offering a command-line option to configure and manage its coding assistant. π€ OpenAI and Jony Ive are building a palm sized screenless AI assistant targeted for 2026 but core software privacy and compute are not ready yet. π MoE Inference Economics from First Principles - A solid tutorial. π‘ Teslas Optimus humanoid robot performs Kung Fu moves - great achievement for humanoid robots as many of these moves are not tele-operated instead they are"
X Link @rohanpaul_ai 2025-10-06T18:43Z 102K followers, 3821 engagements
"Robots that learn new jobs overnight hotels clinics airports update skills like apps. New skill shipped at 02:14 deployed to XXX units task success up 3%. This really does not look too far away now. Beautiful demo from Figure Robots"
X Link @rohanpaul_ai 2025-10-10T08:08Z 102.5K followers, 6707 engagements
"π OpenAI announced plan for a $25B AI data center in Argentina called Stargate Argentina with Sur Energy targeting 500MW of compute capacity. Argentina says the parties signed a letter of intent and the project sits under RIGI which grants import tax breaks faster depreciation and currency stability to large long term investments. At 500MW the site could power tens of thousands of GPU class chips for training and inference. Latin America is also drawing hyperscale builds like TikToks $9.1B data center in Brazil hinting at a new regional cluster for AI workloads"
X Link @rohanpaul_ai 2025-10-13T09:45Z 102.5K followers, 25.4K engagements
"Interesting paper Title. "What the F*ck Is Artificial General Intelligence" It defines intelligence as adaptability under limits of compute memory and energy. So AGI is a system that adapts at least as generally as a human scientist That means it should be able to plan experiments learn cause and effect balance exploration and action and operate with autonomy. The paper calls this type of AGI an artificial scientist because it is judged by its ability to discover and adapt across many tasks not just by passing human-like tests. So AGI is not just human-level AI but a whole system that can"
X Link @rohanpaul_ai 2025-09-29T10:32Z 102.5K followers, 59.7K engagements
"Post-training turns video models from describing clips to actually reasoning about what happens. It does this with X steps supervised fine-tuning reinforcement learning and test-time scaling. Supervised fine-tuning teaches a clear step-by-step format and basic follow-the-instruction behavior. It also ties reasoning steps to frames or timestamps so the model cites evidence instead of guessing. Reinforcement learning samples many answers scores them with checks like correctness and timing then pushes the model toward stronger ones. A lightweight method called GRPO uses verifiable rewards and"
X Link @rohanpaul_ai 2025-10-12T10:40Z 102.5K followers, 4854 engagements
"New Stanford + SambaNova + UC Berkeley paper proposes quite a revolutionary idea. π€― Proves LLMs can be improved by purely changing the input context instead of changing weights. Introduces a new method called Agentic Context Engineering (ACE). It helps language models improve by updating what they read and remember instead of changing their core weights. It has three parts: a Generator that works on tasks a Reflector that learns from mistakes and a Curator that updates the notebook with helpful lessons. So ACE works like this. The model first tries to solve a task. While doing that it writes"
X Link @rohanpaul_ai 2025-10-09T22:08Z 102.5K followers, 35.6K engagements
"The paper shows that training on real code edit traces helps models understand students and fix errors better. It uses 3.8M program traces from a classroom platform each with time-stamped versions tied to a student and a project. Final code hides exploration backtracking and style traces reveal that whole process so the model can learn it. A small per-student embedding is added at the start of each trace so the model can pick up personal habits without extra labels. Models trained on traces match final programs well and also produce more varied realistic outputs than models trained on only"
X Link @rohanpaul_ai 2025-10-10T21:23Z 102.5K followers, 5256 engagements
"New paper from Meta Superintelligence Labs (FAIR) Explains why grokking happens and shows when learning moves from memorizing to generalizing. Gives a concrete recipe to trigger grokking with weight decay moderate width and a data threshold near size times log size. Grokking is when a model first memorizes the training data but then after much longer training suddenly learns the real general rules hidden in the data. Generalization appears once the data size is roughly the group size multiplied by the logarithm of that size not the full square of the size. Training breaks into X stages first"
X Link @rohanpaul_ai 2025-10-04T03:57Z 102.5K followers, 58.5K engagements
"π¦ Oracle will deploy 50000 AMD Instinct MI450 accelerators on Oracle Cloud Infrastructure starting Q3-26 with expansion in 2027+ giving AMD a major public cloud anchor against Nvidia. The build uses AMDs Helios rack design that bundles MI450 GPUs with next gen Epyc Venice CPUs and Pensando Vulcano networking so operators get prewired rack scale blocks for training and inference. Context is a broader capacity land grab since OpenAI agreed to 6GW of AMD compute starting with 1GW of MI450 in 2H-26 while also partnering with Broadcom for 10GW of custom accelerators from 20262029. For scale"
X Link @rohanpaul_ai 2025-10-14T23:03Z 102.5K followers, 1529 engagements
"πA Reddit post is circulating on possible leak of OpenAI's top XX customers who used over X trillion tokens. I like one of the comment π
"So much of our economy is now AI service companies paying AI service companies for AI services.""
X Link @rohanpaul_ai 2025-10-09T18:23Z 102.5K followers, 61.9K engagements
"MASSIVE claim in this paper. AI Architectural breakthroughs can be scaled computationally transforming research progress from a human-limited to a computation-scalable process. So it turns architecture discovery into a computebound process opening a path to selfaccelerating model evolution without waiting for human intuition. The paper shows that an allAI research loop can invent novel model architectures faster than humans and the authors prove it by uncovering XXX recordsetting linearattention designs that outshine human baselines. Right now most architecture search tools only finetune"
X Link @rohanpaul_ai 2025-07-26T02:05Z 102.5K followers, 940K engagements
"π The US approved Nvidia AI chip exports to the UAE. Also unlocking a 5GW Abu Dhabi data center with OpenAI and tying shipments to matching Emirati investment in the US. The framework targets up to 500000 chips/year with XX% for G42 alongside a $1.4T UAE pledge over XX years and the initial batch reportedly excludes G42. Control relies on American cloud operators running the hardware and the cloud keeping telemetry compliance and sensitive data under US processes. The aim is to meet Gulf demand while boxing out Huaweis Ascend 910B offers and keeping regional stacks inside US ecosystems."
X Link @rohanpaul_ai 2025-10-09T22:45Z 102.4K followers, 3737 engagements
"π‘ Samsung is trying to flip the balance in the AI memory race by convincing Nvidia to raise the official speed target for HBM4 even though it was slower than SK Hynix and Micron in delivering the first samples. Normally HBM (high-bandwidth memory) designs focus more on reducing heat than on raw speed because the stacked DRAM (Dynamic Random Access Memory) layers trap heat easily. Nvidias sudden demand for higher operating speed surprised both SK Hynix and Micron since their designs were optimized for thermal stability. Samsung however had already prepared samples that hit higher clock speeds"
X Link @rohanpaul_ai 2025-10-13T13:14Z 102.5K followers, 15.7K engagements
"Ilya Sutskever on Comparing AI Systems to Biological & Human Intelligence. Its always so interesting to listen to him. He explains the brains adaptability where children with half their brain removed still function well. And another experiment shows sensory inputs can remap to different brain regions. i.e. that cortical structures are highly uniform reused across functions similar to DNA and protein encoding. This suggests that intelligence relies on a general repeatable architecture and the idea that AI can mirror biological systems. --- From 'No Priors: AI Machine Learning Tech & Startups'"
X Link @rohanpaul_ai 2025-09-26T21:02Z 102.5K followers, 437.7K engagements
"H1B US Visa now costs $100K per year. This new rules begins in a week. SF Bay Area is about to feel the disruption. If you are outside the U.S. and need to start or resume H-1B work your employer must budget for the $100K per year"
X Link @rohanpaul_ai 2025-09-20T02:07Z 102.5K followers, 681.1K engagements
"This paper introduces MUSE a memory driven agent that learns on the job to handle long multi step tasks. It sets a new TheAgentCompany (TAC) record at XXXXX% using Gemini-2.5 Flash roughly XX% higher than previous systems. proves agents can learn at test time by writing and reusing their own experience. Most agents are frozen at test time so they forget wins and repeat mistakes. MUSE fixes this with a hierarchical memory strategic notes for dilemmas step by step SOPs for sub tasks and tool tips for single actions. The loop is plan execute reflect then store useful experience after every sub"
X Link @rohanpaul_ai 2025-10-14T11:30Z 102.5K followers, 2499 engagements
"China now leads the U.S. in this key part of the AI race. Chinas open-weight AI ecosystem has pulled ahead of the U.S. on community-ranked quality and developer adoption led by Alibabas Qwen and DeepSeek. Developer traction on Hugging Face reflects this shift with DeepSeek likes around 12.8K versus Llama 6.3K and OpenAI 4.0K in Sept-25. The center of gravity for open models now sits in China. "The eagerness of Chinese companies to share their best AI models and the hesitance of U.S. firms to do the same raises the question: Will the best open models always be made in China" ---"
X Link @rohanpaul_ai 2025-10-14T11:12Z 102.5K followers, 5203 engagements
"The paper links a time series model to an LLM so the LLM can reason over numbers plus text. Matters so much because so many real-world tasks mix numbers and text and this handles that mix well. LLMs read text well but miss patterns in numbers over time. Time series models see those patterns but ignore context like news. TS-Reasoner connects a pretrained time series model to an LLM via a small adapter. The adapter converts series features into tokens the LLM can read. Stage X builds the bridge using synthetic series plots and captioned descriptions. These captions explain trend repeating"
X Link @rohanpaul_ai 2025-10-12T08:35Z 102.5K followers, 5402 engagements
"Looks like Apples new A19 CPU cores bring in heavy AI acceleration. So the next-gen iPhones shoulld run AI inference tasks dramatically faster"
X Link @rohanpaul_ai 2025-10-06T09:15Z 102.3K followers, 5134 engagements
"πΌ Walmarts CEO says AI will change every job and the company plans to keep its 2.1M workforce roughly flat for X years so the change is in tasks not totals. Customer facing work stays human summed up in the line people in front of people. On the floor AI directed stocking and scheduling cut planning time from XX minutes to XX in early pilots. Walmart is hiring agent builder and agent developer roles that use low code tools to ship internal agents for merchants and ops. The company is collapsing dozens of bots into X super agents that act as the main entry points for customers associates"
X Link @rohanpaul_ai 2025-09-27T11:06Z 102.1K followers, 270.4K engagements
""I think people shouldn't put their heads in the sand. AI is going to affect jobs" Some jobs will be gone in Banking due to AI" JPMorgan CEO Dimon on AI --- From 'Bloomberg Podcasts' YT channel"
X Link @rohanpaul_ai 2025-10-08T21:28Z 101.8K followers, 4373 engagements
"This figure shows how the researchers tested whether multimodal language models can point to the exact spot of disease on chest X-rays. First they used a dataset with X different chest conditions such as lung opacity enlarged heart and fluid around the lungs. Each X-ray was marked by experts with a ground truth mask that highlights the correct disease location. Then they divided each X-ray into an 8x8 grid so the image was broken down into XX cells. The model was shown both the X-ray and the name of the disease and asked to choose X grid cell where that disease was most visible. They tested 3"
X Link @rohanpaul_ai 2025-10-03T22:11Z 102.3K followers, 3138 engagements
"TSMC has grown from XX% in Q1-24 to XX% in Q2-25 marking a huge lead in the global chip foundry business. Samsungs share has dropped from XX% to X% over the same period. TSMCs profits have remained much higher than Samsung and SK Hynix since 2022 and are expected to stay ahead through 2026. But Samsung could get a lift from supplying chips to OpenAIs Stargate project"
X Link @rohanpaul_ai 2025-10-10T13:05Z 102.5K followers, 6736 engagements
"π WSJ repots Microsoft and Anthropic hired former UK Prime Minister Rishi Sunak as a part time senior adviser. Anthropic describes his scope as global strategy and geopolitical trends with no UK policy work and Microsoft frames his role as internal strategic advice and speaking at company events. He also took a senior adviser role at Goldman Sachs in Jul-25. His AI policy record includes hosting the UK AI Safety Summit in Nov-23 and launching the AI Safety Institute which gives him direct exposure to safety testing and international coordination. --- wsj"
X Link @rohanpaul_ai 2025-10-11T20:05Z 102.5K followers, 17.9K engagements
"Columbia CS Prof explains why LLMs cant generate new scientific ideas. Bcz LLMs learn a structured map Bayesian manifold of known data and work well within it but fail outside it. But true discovery means creating new maps which LLMs cannot do"
X Link @rohanpaul_ai 2025-10-13T19:39Z 102.5K followers, 180.9K engagements
"πΈ π§ Sam Altman is racing across East Asia and the Middle East to lock in chips memory power gear and cash for OpenAIs massive compute build. WSJ reports. He met TSMC Foxconn Samsung and SK Hynix to push for priority capacity and faster output. The biggest one is definitely that Nvidia and OpenAI committed to deploy at least XX gigawatts of systems with the first X gigawatt arriving in 2H-26 on the Vera Rubin platform with Nvidia intending to invest up to $100B as capacity comes online. Nvidia will lease up to 5M chips to OpenAI over time which explains the pressure to secure fab and"
X Link @rohanpaul_ai 2025-10-04T22:04Z 102.4K followers, 8194 engagements
"OpenAI's AgentKit will be so insane build every step of agents on one platform. These visual agent builders make the whole process of iterating and launching agents far more efficient. It sits on top of the Responses API and unifies the tools that were previously scattered across SDKs and custom orchestration. It lets developers create agent workflows visually connect data sources securely and measure performance automatically without coding every layer by hand. The core of AgentKit is the Agent Builder a drag-and-drop canvas where each node represents an action guardrail or decision branch."
X Link @rohanpaul_ai 2025-10-06T21:17Z 102.5K followers, 178.6K engagements
"The paper says long extra thinking with LLMs rarely fixes mistakes the first try mostly decides success. Reflections are mostly confirmatory. So spend compute and training on first try quality and to trim useless reflection at inference. Early stopping can cut tokens by XXXX% with only XXX% accuracy loss. The big deal is that this work shows where the real gains come from making the first answer right not from long reflections. The authors test X reasoning LLMs on X math benchmarks to inspect reflections. They treat anything after the first proposed answer as reflection. An extractor finds"
X Link @rohanpaul_ai 2025-10-14T12:18Z 102.5K followers, 4862 engagements
"π₯ Meta reveals a massive inefficiency in AIs reasoning process and gives a solution. Large language models keep redoing the same work inside long chains of thought. For example when adding fractions with different denominators the model often re explains finding a common denominator step by step instead of just using a common denominator behavior. In quadratic equations it re explains the discriminant logic or completes the square again instead of calling a solve quadratic behavior. In unit conversion it spells out inches to centimeters again instead of applying a unit conversion behavior."
X Link @rohanpaul_ai 2025-09-28T12:05Z 102.5K followers, 57.2K engagements
"Teslas Optimus robot stole attention at the Tron: ARES premiere in Los Angeles moving fluidly with kung fu-inspired gestures. Showed how far Teslas hardware and software integration has come combining real-time vision motion planning and torque control to produce lifelike movement"
X Link @rohanpaul_ai 2025-10-07T10:35Z 102.5K followers, 10.1K engagements
"This paper introduces a new method called Agentic Context Engineering (ACE). It helps language models improve by updating what they read and remember instead of changing their core weights. Normal methods that edit prompts tend to make them too short and lose important details over time. ACE fixes this by treating the models context like a growing notebook that keeps and organizes useful strategies. It has three parts: a Generator that works on tasks a Reflector that learns from mistakes and a Curator that updates the notebook with helpful lessons. Instead of rewriting everything each time"
X Link @rohanpaul_ai 2025-10-08T01:20Z 102.5K followers, 46.1K engagements
"Broadcom CEO says generative AI will take up a much bigger share of global GDP. He told CNBC that knowledge-based and technology-intensive sectors now make up about XX% of the estimated US$110 trillion global GDP which could rise to XX% with the growth of generative AI. Last month Broadcom reported securing US$10 billion in chip orders from a fourth unnamed customer. Broadcom and OpenAI announced their official partnership on Monday saying they would jointly build and deploy XX gigawatts of custom artificial intelligence accelerators. The move is part of a broader effort to scale AI across"
X Link @rohanpaul_ai 2025-10-14T10:37Z 102.5K followers, 4333 engagements
"π§π» Klarnas CEO warns a fast AI rollout will cut many knowledge jobs soon. AI already shrank Klarna from 7400 to 3000 employees while AI now handles about 2/3 of support chats and predicts banking and software margins will shrink as faster challengers move in. He says society is not ready for the job loss and while new roles will come later translators and other office workers will be hit first. At Klarna the drop came mostly from a hiring freeze and automation not big layoffs and the AI chatbot replaced work equal to roughly XXX agents. The company still keeps human support and uses little"
X Link @rohanpaul_ai 2025-10-13T20:12Z 102.5K followers, 6248 engagements
"π Citi research says global semiconductor sales to reach an all-time high $731B in 2025 up XX% from 2024 marking a new record for the industry. But this growth is not coming from shipping more chips its coming from higher prices. According to Citi shipment volumes are still XX% below the last peak and overall sales have grown only XX% in this upcycle while past cycles usually saw about XX% growth in units. This means the industry is still running lean on inventory and factories have room to increase output before the market gets saturated again. The biggest reason for the boom is the price"
X Link @rohanpaul_ai 2025-10-05T18:00Z 101.7K followers, 14.6K engagements
"New paper from @Google is a major memory breakthrough for AI agents. ReasoningBank helps an AI agent improve during use by learning from its wins and mistakes. To succeed in real-world settings LLM agents must stop making the same mistakes. ReasoningBank memory framework helps agents learn from both successes and failures and turn them into reasoning strategies that generalize. Traditionally most agents handle tasks in a stream but do not keep lessons so they repeat errors. The system turns each action log into a short memory item with a title a description and a concrete next step. Before a"
X Link @rohanpaul_ai 2025-10-11T02:28Z 102.5K followers, 120.8K engagements
""Nvidia is undervalued.Because the future is much bigger." SoftBank Founder Masayoshi Son says He assumes that in XX years AGI or ASI will replace X% of global GDP (which is actually the most pessimistic forecast as ASI could bring much bigger % impact). And that X% of Global GDP is $X trillion/year which ASI will bring. He also assumes the total cumulative capex/build cost for the required AI infrastructure is $9T. So we have $9T per year of output to $9T total capex. Meaning just a X year payback for the entire $9T so he calls $9T small because X year of AGI output would repay the whole"
X Link @rohanpaul_ai 2025-10-06T08:19Z 102.5K followers, 408.6K engagements
"This paper trains agents to simulate future steps using real experience then act from that. On ALFWorld a 7B model reaches XXXX% success after this training. Most models do fine in math or code but struggle in multi step apps games and phone control. Dyna Mind teaches planning with grounded simulations instead of guesses. Stage X RESIM runs the agent explores next steps compresses them into one reasoning trace then fine tunes the model to produce that trace. That creates an internal world model a simple map of how actions change the environment. Stage X Dyna GRPO uses online reinforcement"
X Link @rohanpaul_ai 2025-10-14T01:56Z 102.5K followers, 6696 engagements
"Deutsche Bank says the US would be close to recession without tech spending. When you strip out technology-related investment especially AI-driven spending on data centers chips and software the underlying growth trend for domestic demand falls to near 0%. Thats because the rest of business and consumer demand is much weaker so tech capex is doing most of the heavy lifting. Deutsche Banks chart shows this clearly: total GDP growth is positive but once tech is excluded it drops toward 0%. In other words tech spending alone is keeping the economy in expansion territory. This is happening"
X Link @rohanpaul_ai 2025-10-03T17:36Z 101.8K followers, 13.1K engagements
"The paper teaches small LLMs to reason better by training with built in tree search. i.e. Smarter exploration beats longer training runs. It reaches XXXXX% average accuracy while using 5.7x fewer GPU hours. Typical reinforcement learning with verifiable rewards stalls because it explores only a few paths. DeepSearch adds Monte Carlo Tree Search to training expands steps checks answers learns from both correct and confident wrong paths and uses a global frontier selector to pick the next expansion. An entropy rule chooses one wrong but confident path to teach where the model overtrusts. A"
X Link @rohanpaul_ai 2025-10-02T16:22Z 102.4K followers, 9979 engagements
"China's ranks first on robot density. i.e. the number of robots per 10000 employees as a measure for the degree of automation. Also China H1 2025 robot exports: Poland +1746% Mexico +275% Russia +135% Vietnam +114%"
X Link @rohanpaul_ai 2025-10-14T08:39Z 102.5K followers, 5781 engagements
"πΎ Western Digitals CEO says HDDs are still central to AI storage with XX% of hyperscaler data on HDDs XX% on SSDs and XX% on tape. That split reflects capacity economics and power per TB since enterprise disks remain roughly 5x to 6x cheaper per TB than SSDs at scale and draw fewer watts per TB in bulk deployments. Data centers tier AI storage so hot data sits on flash warm and cold pools live on disks and rarely touched archives go to tape which matches how hyperscalers manage cost and performance. AI demand is exceeding what manufacturers can currently produce or ship in a timely way so"
X Link @rohanpaul_ai 2025-10-04T15:02Z 102.4K followers, 11.4K engagements
"ByteDance introduced a major advancement in long-context modeling with linearly scaling compute. π Addresses a core challenge in AIbalancing efficiency and fidelity when processing extended sequencesby drawing inspiration from biological memory systems. On 128k tests FLOPs drop XXXX% and KV cache drops XXXX% with +0.4% parameters while accuracy improves. Think of the model as keeping X kinds of memory at once exact recent detail and a compact summary of older stuff. A sliding attention window holds recent tokens exactly as lossless short term memory. When tokens leave that window the"
X Link @rohanpaul_ai 2025-10-10T09:54Z 102.5K followers, 10.5K engagements
"π« Another paper showing current multimodal models are not yet reliable for pointing out exact disease spots on medical images. On chest X-rays best model (GPT 5) reached only XXXX% average localization radiologists were at 80.1%. They split each X-ray into an 8x8 grid and ask the model to pick one square where the disease lies. It counts as correct if at least half of that square overlaps with the true area of disease. For diseases that can appear in many places like fluid or infection in the lungs the models often missed the exact location. The results show these models know some anatomy"
X Link @rohanpaul_ai 2025-10-03T22:07Z 102.4K followers, 9093 engagements
"πΌ Overall AI is fueling economic growth through investment rather than efficiency. AI investment is giving the U.S. economy a strong push but it still hasnt made American workers much more productive. Growth is happening mostly because of money pouring into AI and a rising stock market not because people are working faster or smarter yet. Economists define productivity as how much output a worker can create in an hour. AI could help by handling boring or repetitive tasks so humans can focus on higher-value work or by automating some jobs entirely which raises efficiency. Right now the"
X Link @rohanpaul_ai 2025-10-14T23:11Z 102.5K followers, 1749 engagements
"New Google+Cornell paper shows X compact language model can read code and predict memory latency and accuracy across languages and hardware. A 300M model hits 0.9+ on APPS memory and leads classic neural architecture search predictors. The task is code to metric regression predict memory or runtime from code without running it. Past systems rely on hand tuned features per language or graph and they break when code changes. This model reads raw code or ONNX graphs with a T5Gemma encoder and predicts numbers digit by digit. Sequential prediction lets X model learn many tasks and capture"
X Link @rohanpaul_ai 2025-10-04T01:52Z 102.3K followers, 18.3K engagements
"This paper trains a separate planner that helps LLM agents finish long tasks with fewer mistakes. presents EAGLET a plug and play global planner It cuts training cost by about 8x compared with common reinforcement learning setups. Agents often repeat actions or hallucinate steps because they only plan locally during execution. EAGLET adds a global planner that writes a short high level plan before the agent starts acting. They create training data by asking a strong model to draft plans then keep only plans that help both a novice and an expert agent a process they call homologous consensus."
X Link @rohanpaul_ai 2025-10-13T23:46Z 102.5K followers, 6681 engagements
"DeepSeek R1 running locally - Full setup guide"
X Link @rohanpaul_ai 2025-01-26T00:03Z 102.4K followers, 1.4M engagements
"Jeff Bezos just gave the most bullish opinion on AI. "AI is going to make every company's quality go up and their productivity go up. I literally mean every company. Every manufacturing company every hotel every you know consumer products company AI is going to change every industry. It's a very unusual technology in that regard in that it's a horizontal enabling layer." --- From 'DRM News' YT Channel"
X Link @rohanpaul_ai 2025-10-04T07:43Z 102.5K followers, 187.4K engagements
"πΌ AI will boost star workers more than everyone else widening performance gaps and straining teams. WSJ writes. The core idea is that domain expertise and organized habits let stars get more from AI. Stars adopt AI earlier explore features fast and build personal workflows while others wait for rules. They judge outputs better accepting correct advice and rejecting errors. Their domain expertise helps them ask precise questions set constraints and iterate which raises prompt quality and accuracy. The high-status of those employees get more recognition and praise for their AI-assisted work"
X Link @rohanpaul_ai 2025-10-13T12:06Z 102.5K followers, 427.7K engagements
"π€ OpenAI and Jony Ive are building a palm sized screenless AI assistant targeted for 2026 but core software privacy and compute are not ready yet. per FT report. The device listens and sees the environment through a mic camera and speaker stays always on and the team has not nailed how its voice should talk or stop. OpenAI bought Ives io for $6.5B and has pulled in 20+ ex Apple hardware staff while manufacturing talks include Luxshare with assembly possibly outside China. Compute is the biggest blocker because running multimodal chat at scale needs huge inference capacity and OpenAI already"
X Link @rohanpaul_ai 2025-10-05T19:13Z 102.5K followers, 507.4K engagements
"π¦ AMDs new Helios rack-scale AI platform revealed at the OCP Global Summit 2025 is designed for easier servicing and comes with XX% greater memory capacity than Nvidias Vera Rubin. The headline is XX GPUs per rack XXX exaFLOPS FP8 31TB HBM4 and a claim of XX% more memory per system than Nvidia Vera Rubin. Helios combines EPYC Venice CPUs Instinct MI450 GPUs AMD Pensando networking and the ROCm software stack so operators get cohesive hardware with open interfaces. For in node scale up it uses UALink for fast GPU to GPU memory sharing and for cross rack scale out it uses UEC Ethernet for"
X Link @rohanpaul_ai 2025-10-14T18:25Z 102.5K followers, 5214 engagements
"π Goldman Sachs Research now expects global data center power demand to jump XXX% from 2023 to 2030 roughly like adding another top-10 power using country. That surge would lift US electricity demand growth to XXX% compound annual growth rate through 2030 the fastest since the 1990s. The outlook rests on the Six Ps Pervasiveness Productivity Prices Policy Parts People. Pervasiveness is how widely AI is used and wider use pushes more servers and more site power. Productivity is the output per watt from chips and models and efficiency gains can slow growth but not if usage grows faster. Prices"
X Link @rohanpaul_ai 2025-10-13T18:25Z 102.5K followers, 5510 engagements
"A parade of transforming military robots from China. multi-terrain spiders (wheeled flying amphibious) missile-armed robot dogs and modular all-terrain snakes that swim and burrow"
X Link @rohanpaul_ai 2025-09-18T10:01Z 102.5K followers, 315.8K engagements
"Google's CoDA: Multi-Agent AI for Collaborative Data Visualization. CoDA is a team of small AI agents that turn a plain request into working chart code and a clean plot. It beats prior systems by up to XXXX% on overall score. Single-agent tools miss steps they mis-handle multi-file data crash at runtime and get lost during edits. This paper treats the job like teamwork each agent has X clear role and they work from a shared plan. It starts with metadata summaries of the data files which avoid context limits and keep choices grounded in real columns. A query analyzer writes a global TODO a"
X Link @rohanpaul_ai 2025-10-09T00:32Z 102.5K followers, 4381 engagements
"Another set of junior jobs to go. Robots as construction-worker is no longer the futureits here"
X Link @rohanpaul_ai 2025-10-01T14:59Z 102.4K followers, 108.4K engagements
"A 7B model tuned for forms and docs beats giant models at pulling structured data. Beats GPT-4.1 on 1000 extraction tasks trained for $XXX. The team generated synthetic training data that preserves memory across chunks of a long file. That memory lets the model connect names dates and values that appear far apart. They fine-tuned with Low Rank Adaptation changing only XXXX% of weights. They then used Group Relative Policy Optimization with a semantic reward and strict JSON checks. This setup accepts different surface wording if the meaning matches. On 1000 held-out tasks it hit XXXXX mean"
X Link @rohanpaul_ai 2025-10-06T04:08Z 102.5K followers, 130.6K engagements
"this ChatGPT prompt went so wildly viral on Reddit. The creator claims to have created this after struggling through XXX failed attempts. basically the prompt flips the usual flow by making the model interview the user first asking a few targeted questions about purpose audience constraints and context. Because the answers feed back into the final request it appears to generate more tailored outputs. (However imo asking ChatGPT to request missing information was already a common practice.) Here's the entire prompt: -------- You are Lyra a master-level AI prompt optimization specialist. Your"
X Link @rohanpaul_ai 2025-07-02T18:53Z 102.4K followers, 367.6K engagements
"Bank of Americal Research: AI performance is now limited not by the chips themselves but by how much power and cooling each rack can deliver. Most of the worlds data centers must be upgraded or rebuilt because AI hardware now demands far higher power density than existing infrastructure can handle. Modern AI training relies on connecting thousands of GPUs together with extremely fast data links. To minimize communication delay engineers pack as many GPUs as possible into one rack. This design greatly boosts performance but also drives a massive jump in power use per rack. - Back in 2021 the"
X Link @rohanpaul_ai 2025-10-08T14:33Z 102.5K followers, 11.6K engagements
"πΌ Zuck's next move. Andrew Tulloch co-founder of Thinking Machines Lab and a top AI researcher has left to join Meta. And this could that $X billion man who earlier turned down a $1B offer from Meta. Tulloch had earlier worked at Meta for XX years before joining OpenAI and later co-founding Thinking Machines with Mira Murati earlier this year. --- wsj .com/tech/ai/thinking-machines-lab-co-founder-departs-for-meta-442d7461"
X Link @rohanpaul_ai 2025-10-11T19:35Z 102.5K followers, 42.9K engagements
"One of the top use of gen AI today is therapy and companionship. Average time spent per day in top AI companion apps"
X Link @rohanpaul_ai 2025-10-11T13:30Z 102.5K followers, 15.5K engagements
"New Tencent paper upgrades LLM agents by learning small experience rules in the prompt without changing weights. It costs about $XX on XXX samples yet beats some $10000 fine tuning runs. π€― Specialized tasks need tools and prompting and weight tuning is slow expensive and fragile. The method keeps the base model frozen and stores short lessons as a token prior in the prompt. For each question the agent samples a small group of answers scores them and writes why winners win. Those reasons become general rules saved in an experience library for the next run. Over a few passes the library is"
X Link @rohanpaul_ai 2025-10-14T10:41Z 102.5K followers, 16.3K engagements
"@TeksEdge @Atlassian fantastic"
X Link @rohanpaul_ai 2025-10-14T22:45Z 102.5K followers, XX engagements
"πΈ Little concerning article here at futurism. AI data centers are absorbing huge capital and a new analysis says the math fails with $40B yearly depreciation on 2025 builds versus $15-20B revenue. The core pieces age on different clocks chips churn in 2-4 years networking around XX years buildings far longer so depreciation snowballs. On those lifetimes 2025 sites show $40B annual write-downs against $15-20B revenue before power and staff which already implies negative cash. To earn a normal return at this scale United States data centers would need about $480B revenue in 2025 far above"
X Link @rohanpaul_ai 2025-08-31T04:43Z 102.5K followers, 421K engagements
"Change in US tech job postings Oct.-Nov. 2022 to Oct.-Nov. 2024"
X Link @rohanpaul_ai 2025-10-05T22:26Z 102.1K followers, 30.9K engagements
"News:π° BlackRock is in advanced talks to buy Aligned Data Centers from Macquarie for a huge $40B. The talks include MGX an Abu Dhabi AI investment firm linked to Mubadala which may invest alongside the deal. Aligned runs XX campuses and XX data centers across the Americas and in January raised $5B in equity plus $7B in debt to speed expansion. This targets a business with long contracts predictable cash flows and the ability to deliver high-power low-latency capacity that AI training clusters require at scale. So far this year there have been XXX data-center M&A deals closed worth $46B with"
X Link @rohanpaul_ai 2025-10-04T16:24Z 102.4K followers, 15K engagements
"BIG claim. Giving an LLM just XX carefully chosen full workflow examples makes it perform better at real agent tasks than training it with 10000 synthetic samples. "Dramatically outperforms SOTA models: Kimi-K2-Instruct DeepSeek-V3.1 Qwen3-235B-A22B-Instruct and GLM-4.5. " on AgencyBench (LIMI at 73.5%) The big deal is that quality and completeness of examples matter way more than raw data scale when teaching models how to act like agents instead of just talk. They name the Agency Efficiency Principle which says useful autonomy comes from a few high quality demonstrations of full workflows"
X Link @rohanpaul_ai 2025-09-24T12:27Z 102.4K followers, 93.9K engagements
"Morgan Stanley Research says OpenAI makes up around $330B of the $880B total future contract value (RPO) tied to Microsoft Oracle and CoreWeave so a lot of supplier growth depends directly on OpenAIs stability. That means about XX% of Oracles and about XX% of CoreWeaves future revenue commitments rely on OpenAI. RPO or Remaining Performance Obligations means the total value of work a company has promised to deliver in the future under signed contracts but hasnt yet completed. Its basically a measure of future revenue thats already committed but not yet earned. Hyperscale data center operators"
X Link @rohanpaul_ai 2025-10-09T11:50Z 102.5K followers, 166K engagements
"@barbarikon a16z has uploaded the full video here itself. (i.e. this is the full XX mint video) just click on the tagged a16z to get to the main post and get the time stamps as well"
X Link @rohanpaul_ai 2025-10-14T05:42Z 102.5K followers, XXX engagements
"From Under X% to Almost 50%: .ai Domains Dominate Anguillas Budget"
X Link @rohanpaul_ai 2025-10-05T09:00Z 102.5K followers, 1852 engagements
"πOpenAI and Nvidia are wiring up AI with circular mega deals that stack to $1T+ across chips and cloud capacity raising output and risk together. Bloomberge wrote a nice piece. Nvidia agreed to invest $100B in OpenAI and the plan includes deploying at least 10GW of Nvidia systems starting in 2026. OpenAI then signed with AMD to deploy 6GW of GPUs and received a warrant to buy up to XX% of AMD for X cent per share. OpenAI also committed to about $300B of compute from Oracle over roughly X years as part of the Stargate rollout that already targets about 7GW in the US. Suppliers are feeling the"
X Link @rohanpaul_ai 2025-10-08T15:36Z 102.5K followers, 23.6K engagements
"The exact moment Jeff Bezos decided not to become a physicist From "The Economic Club of Washington D.C." YT Channel"
X Link @rohanpaul_ai 2025-10-12T22:29Z 102.5K followers, 6324 engagements
"Foundry packaging and memory choices are undisclosed and those will drive cost yield and timing more than any headline. If OpenAI can build really good software tools that manage how its new custom chips (the ASICs) run programs then those chips could perform more work per unit of power i.e. better perf per watt. Right now Nvidias big advantage isnt just its hardware its the CUDA software ecosystem a full platform that developers already know how to use for training AI models. If OpenAI wants to move away from Nvidia chips it must create its own version of that ecosystem a strong compiler"
X Link @rohanpaul_ai 2025-10-14T17:29Z 102.5K followers, XXX engagements