[GUEST ACCESS MODE: Data is scrambled or limited to provide examples. Make requests using your API key to unlock full data. Check https://lunarcrush.ai/auth for authentication information.] #  @rohanpaul_ai Rohan Paul Rohan Paul's recent activities are not specified in the provided context. However, the conversation revolves around AI developments, including OpenAI's massive compute deal with CoreWeave and a lawsuit filed by xAI against OpenAI for allegedly stealing trade secrets. Additionally, there are discussions on AI models' capabilities, such as passing CFA level III mock exams and outperforming human experts in certain tasks. ### Engagements: XXXXXX [#](/creator/twitter::2588345408/interactions)  - X Week XXXXXXXXX -XX% - X Month XXXXXXXXXX +46% - X Months XXXXXXXXXX +2.10% - X Year XXXXXXXXXXX +242% ### Mentions: XXX [#](/creator/twitter::2588345408/posts_active)  - X Week XXX +9.50% - X Month XXX -XXXX% - X Months XXXXX +82% - X Year XXXXX +226% ### Followers: XXXXXXX [#](/creator/twitter::2588345408/followers)  - X Week XXXXXXX +2.80% - X Month XXXXXXX +8.70% - X Months XXXXXXX +63% - X Year XXXXXXX +156% ### CreatorRank: XXXXXXX [#](/creator/twitter::2588345408/influencer_rank)  ### Social Influence [#](/creator/twitter::2588345408/influence) --- **Social category influence** [technology brands](/list/technology-brands) XXXX% [stocks](/list/stocks) #4638 [finance](/list/finance) XXX% [countries](/list/countries) XXXX% [celebrities](/list/celebrities) XXXX% [vc firms](/list/vc-firms) #35 [currencies](/list/currencies) XXXX% [travel destinations](/list/travel-destinations) XXXX% [social networks](/list/social-networks) XXXX% [exchanges](/list/exchanges) XXXX% **Social topic influence** [open ai](/topic/open-ai) #424, [goldman sachs](/topic/goldman-sachs) #6, [accuracy](/topic/accuracy) #215, [china](/topic/china) #2249, [investment](/topic/investment) #1235, [bubble](/topic/bubble) #378, [faster](/topic/faster) #1199, [jeff bezos](/topic/jeff-bezos) #34, [microsoft](/topic/microsoft) #262, [infrastructure](/topic/infrastructure) #1179 **Top accounts mentioned or mentioned by** [@opus_genesis](/creator/undefined) [@grok](/creator/undefined) [@huggingface](/creator/undefined) [@googledeepmind](/creator/undefined) [@nvidia](/creator/undefined) [@xai](/creator/undefined) [@todayinai_](/creator/undefined) [@nvidiaaidev](/creator/undefined) [@intuitmachine](/creator/undefined) [@justinechoes](/creator/undefined) [@openai](/creator/undefined) [@nlituanie](/creator/undefined) [@circuitcrazex](/creator/undefined) [@baiduinc](/creator/undefined) [@kuittinenpetri](/creator/undefined) [@bkrepysh34200](/creator/undefined) [@aiatmeta](/creator/undefined) [@windsurfai](/creator/undefined) [@a16z](/creator/undefined) [@klingai](/creator/undefined) **Top assets mentioned** [Goldman Sachs (GS)](/topic/goldman-sachs) [Microsoft Corp. (MSFT)](/topic/microsoft) [Alphabet Inc Class A (GOOGL)](/topic/$googl) [Morgan Stanley (MS)](/topic/morgan-stanley) [NVIDIA Corp. (NVDA)](/topic/$nvda) ### Top Social Posts [#](/creator/twitter::2588345408/posts) --- Top posts by engagements in the last XX hours "Absolutely love this bit from Jeff Bezos. The simplest explanation on building a business and being founder. π― "Benjamin Graham the great investor is famous for saying in the short term the stock market is a voting machine in the long term it's a weighing machine and so as founders and entrepreneurs and business people our job is to build a heavy company. We want to build a company that when it is weighed it is a very heavy company we do not want to focus on the stock price" --- From 'DRM News' YT Channel" [X Link](https://x.com/rohanpaul_ai/status/1974196086601642398) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-03T19:33Z 103K followers, 304.3K engagements "This paper makes LLM reinforcement learning steadier by sampling more where it matters. Standard group training breaks when all sampled answers for a prompt look the same so the gradient becomes zero. It prevents the zero gradient stall that makes RL training spin its wheels. Taking a lot of samples fixes that but wastes compute on easy prompts. Reinforce-Ada samples each prompt in rounds then stops that prompt once it has enough signal. One rule stops after the 1st correct answer the balanced rule waits for enough correct and enough incorrect. After collection it builds a fixed size group" [X Link](https://x.com/rohanpaul_ai/status/1977176390450475014) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-12T00:56Z 103K followers, 4849 engagements "paper automates turning a dense paper into specific social-platform-ready posts. They define AutoPR a task that converts a paper and figures into accurate engaging platform specific content. They build PRBench with XXX paper to post pairs to score systems on fidelity engagement and platform fit. The core system PRAgent runs in X stages extract content synthesize with multiple agents then adapt to each platform. Plain prompting often drops key ideas uses generic hooks and picks broad tags that miss real communities. PRAgent fixes this by structuring the paper first analyzing visuals and" [X Link](https://x.com/rohanpaul_ai/status/1977900914364272727) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-14T00:55Z 102.9K followers, 4528 engagements "Broadcom CEO says generative AI will take up a much bigger share of global GDP. He told CNBC that knowledge-based and technology-intensive sectors now make up about XX% of the estimated US$110 trillion global GDP which could rise to XX% with the growth of generative AI. Last month Broadcom reported securing US$10 billion in chip orders from a fourth unnamed customer. Broadcom and OpenAI announced their official partnership on Monday saying they would jointly build and deploy XX gigawatts of custom artificial intelligence accelerators. The move is part of a broader effort to scale AI across" [X Link](https://x.com/rohanpaul_ai/status/1978047452377669949) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-14T10:37Z 103K followers, 4811 engagements "πΌ Overall AI is fueling economic growth through investment rather than efficiency. AI investment is giving the U.S. economy a strong push but it still hasnt made American workers much more productive. Growth is happening mostly because of money pouring into AI and a rising stock market not because people are working faster or smarter yet. Economists define productivity as how much output a worker can create in an hour. AI could help by handling boring or repetitive tasks so humans can focus on higher-value work or by automating some jobs entirely which raises efficiency. Right now the" [X Link](https://x.com/rohanpaul_ai/status/1978237371351978020) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-14T23:11Z 103K followers, 3716 engagements "MASSIVE claim in this paper. AI Architectural breakthroughs can be scaled computationally transforming research progress from a human-limited to a computation-scalable process. So it turns architecture discovery into a computebound process opening a path to selfaccelerating model evolution without waiting for human intuition. The paper shows that an allAI research loop can invent novel model architectures faster than humans and the authors prove it by uncovering XXX recordsetting linearattention designs that outshine human baselines. Right now most architecture search tools only finetune" [X Link](https://x.com/rohanpaul_ai/status/1948927502405063006) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-07-26T02:05Z 103.1K followers, 940.3K engagements "Hunyuan 3D-2.1 turns any flat image into studio-quality 3D models. And you can do it on this @huggingface space for free" [X Link](https://x.com/rohanpaul_ai/status/1959100808374387066) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-08-23T03:50Z 103.1K followers, 219K engagements "Brilliant and timely MIT + HARVARD study β€ Human-AI companionship in the wild looks stable and serious. Most users report clear benefits like reduced loneliness and emotional support. The biggest risk comes from sudden platform updates that break continuity and feel to users like losing a real partner. π§ The study analyzed 1506 top posts from r/MyBoyfriendIsAI a 27000+ member community clustered the language into themes and ran XX LLM classifiers to quantify platforms relationship stages benefits and risks. π¬ Why relationships form between AI and Human Bonds often start by accident during" [X Link](https://x.com/rohanpaul_ai/status/1967861782740860989) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-09-16T08:03Z 103.1K followers, 75K engagements "DeepSeek-R1 was published in Nature yesterday as the cover article for their BRILLIANT latest research. They show that pure Reinforcement Learning with answer-only rewards can grow real reasoning skills no human step-by-step traces required. So completely skip human reasoning traces and still get SOTA reasoning via pure RL. Its so powerful revelation because instead of forcing the model to copy human reasoning steps it only rewards getting the final answer right which gives the model freedom to invent its own reasoning strategies that can actually go beyond human examples. Earlier methods" [X Link](https://x.com/rohanpaul_ai/status/1968565696423514250) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-09-18T06:40Z 103.1K followers, 453.9K engagements "π¨ BAD news for Medical AI models. MASSIVE revelations from this @Microsoft paper. π€― Current medical AI models may look good on standard medical benchmarks but those scores do not mean the models can handle real medical reasoning. The key point is that many models pass tests by exploiting patterns in the data not by actually combining medical text with images in a reliable way. The key findings are that models overuse shortcuts break under small changes and produce unfaithful reasoning. This makes the medical AI model's benchmark results misleading if someone assumes a high score means the" [X Link](https://x.com/rohanpaul_ai/status/1971241621774614560) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-09-25T15:53Z 103.1K followers, 524.3K engagements "New ByteDance paper shows that redistributing exploration across prompts lets Reinforcement Learning trained LLMs learn more without extra compute. So you get the same or better gains without buying 2x more compute. Uniform budgets waste work because easy prompts all pass and hard prompts all fail so Group Relative Policy Optimization GRPO gets zero signal. Their fix scores each prompt at a chosen rollout count by expected learning value and cost then assigns budgets with a knapsack solver. Value is the chance of a non zero gradient times the expected increase in success rate. They use last" [X Link](https://x.com/rohanpaul_ai/status/1973742263030415376) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-02T13:30Z 103.1K followers, 4063 engagements "π° Jeff Bezos says the AI spending surge is a "good industrial bubble" that will waste some money yet lift productivity across the economy. He made the case comparing today with 1990s biotech and the dot com era where many firms died but infrastructure persisted. Outside the hyperscalers neocloud providers (GPU rental specialists that stand up capacity fast) are scaling funding and deployments to meet AI compute demand. The mechanism he describes is that excess money funds many experiments and vendors a few winners will emerge from that and the leftover assets cut future costs. --- bloomberg." [X Link](https://x.com/rohanpaul_ai/status/1974173157016711424) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-03T18:02Z 103.1K followers, 186.2K engagements "π― Jeff Bezos beautifully explains how AI boom as good kind of bubble that will benefit the world Bezos argued that industrial bubbles like AI are very different from financial bubbles like the 2008 banking crash because the inventions survive even if the hype fades. He compared todays AI boom to the dotcom eras spending on fiber optics and the biotech wave of the 1990s where even though many companies failed the technology and discoveries that remained were hugely valuable. AIs impact is real and will change every industry though investors right now may struggle to tell the difference" [X Link](https://x.com/rohanpaul_ai/status/1974177375916118069) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-03T18:18Z 103.1K followers, 952K engagements "The biggest private company valuations of the world are now ruled by AI and tech. OpenAI is first at $500B SpaceX second at $400B and ByteDance third at $300B" [X Link](https://x.com/rohanpaul_ai/status/1974226435184152631) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-03T21:33Z 103.1K followers, 34K engagements "πΎ Western Digitals CEO says HDDs are still central to AI storage with XX% of hyperscaler data on HDDs XX% on SSDs and XX% on tape. That split reflects capacity economics and power per TB since enterprise disks remain roughly 5x to 6x cheaper per TB than SSDs at scale and draw fewer watts per TB in bulk deployments. Data centers tier AI storage so hot data sits on flash warm and cold pools live on disks and rarely touched archives go to tape which matches how hyperscalers manage cost and performance. AI demand is exceeding what manufacturers can currently produce or ship in a timely way so" [X Link](https://x.com/rohanpaul_ai/status/1974490351290171633) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-04T15:02Z 103.1K followers, 11.4K engagements ""100 million words context window is already possible which is roughly what a human hears in a lifetime. Inference support is the only bottleneck to achieve it. And AI Models actually do learn during the context window without changing the weights." Anthropic CEO Dario Amodei (On the 2nd point there was this brilliant Google Paper published last week that says LLMs can learn in context from examples in the prompt can pick up new patterns while answering yet their stored weights never change.) --- From 'Alex Kantrowitz' YT Channel (Full Video link in comment)" [X Link](https://x.com/rohanpaul_ai/status/1974611155575324769) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-04T23:02Z 103.1K followers, 631.2K engagements "π°AI is making the caribbean island of Anguilla rich. Anguilla now gets XX% of its income from .ai domains up from X% before the surge. Because .ai is the islands official country code top-level domain (TLD) and the Government of Anguilla is the official manager of that registry. So every .ai registration or renewal anywhere in the world routes through the .ai registry for Anguilla. Even on the secondary market .ai domains are traded at high prices: you.ai sold for an impressive $700000 in 2023 cloud.ai for $600000 in 2025 and adapt.ai in the same year for $300000" [X Link](https://x.com/rohanpaul_ai/status/1974761679968665689) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-05T09:00Z 103.1K followers, 5268 engagements "Absolutely classic @GoogleResearch paper on In-Context-Learning by LLMs. Shows the mechanisms of how LLMs learn in context from examples in the prompt can pick up new patterns while answering yet their stored weights never change. π‘The mechanism they reveal for in-context-learning. When the model reads a few examples in your prompt it figures out a pattern (like a small rule or function). Instead of permanently changing its stored weights it forms a temporary adjustment that captures this pattern. That adjustment can be written mathematically as a rank-1 matrix meaning it only adds one" [X Link](https://x.com/rohanpaul_ai/status/1974808419207987401) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-05T12:06Z 103.1K followers, 75.4K engagements "This is not a bubble. Cisco in 1998-2002 vs Nvidia in 2020-2024 The data says this cycle is different and AI is still so early for most parts of the world. Cisco was a valuation story price inflated while earnings lagged then the multiple deflated. Nvidia is an earnings story price climbs alongside surging earnings" [X Link](https://x.com/rohanpaul_ai/status/1974937952326209664) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-05T20:41Z 103.1K followers, 871.9K engagements "A 7B model tuned for forms and docs beats giant models at pulling structured data. Beats GPT-4.1 on 1000 extraction tasks trained for $XXX. The team generated synthetic training data that preserves memory across chunks of a long file. That memory lets the model connect names dates and values that appear far apart. They fine-tuned with Low Rank Adaptation changing only XXXX% of weights. They then used Group Relative Policy Optimization with a semantic reward and strict JSON checks. This setup accepts different surface wording if the meaning matches. On 1000 held-out tasks it hit XXXXX mean" [X Link](https://x.com/rohanpaul_ai/status/1975050381379215464) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-06T04:08Z 103.1K followers, 131.1K engagements "So surprising. They should have at-least proof-read. π AI hallucinations need to be SOLVED ASAP. π° Deloitte has agreed to repay part of a $440000 government fee after admitting it used AI to write sections of an official report that later turned out to have fake references and quotes. The report was commissioned by Australias Department of Employment and Workplace Relations to review an IT system that manages welfare penalties. Deloitte used OpenAIs GPT-4o model to help fill traceability and documentation gaps. After the report was published in July-25 a University of Sydney academic" [X Link](https://x.com/rohanpaul_ai/status/1975151655352938986) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-06T10:50Z 103.1K followers, 690.4K engagements "For the last X months Figure Robot has been working XX hours a day on BMWs X3 production line managing complex assembly tasks in a real industrial setting. Factory work is almost ready to be redefined" [X Link](https://x.com/rohanpaul_ai/status/1975965731641753685) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-08T16:45Z 103.1K followers, 4934 engagements "πA Reddit post is circulating on possible leak of OpenAI's top XX customers who used over X trillion tokens. I like one of the comment π "So much of our economy is now AI service companies paying AI service companies for AI services."" [X Link](https://x.com/rohanpaul_ai/status/1976352937673933194) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-09T18:23Z 103.1K followers, 62.2K engagements "New Stanford + SambaNova + UC Berkeley paper proposes quite a revolutionary idea. π€― Proves LLMs can be improved by purely changing the input context instead of changing weights. Introduces a new method called Agentic Context Engineering (ACE). It helps language models improve by updating what they read and remember instead of changing their core weights. It has three parts: a Generator that works on tasks a Reflector that learns from mistakes and a Curator that updates the notebook with helpful lessons. So ACE works like this. The model first tries to solve a task. While doing that it writes" [X Link](https://x.com/rohanpaul_ai/status/1976409349476995158) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-09T22:08Z 103.1K followers, 36.1K engagements "π The US approved Nvidia AI chip exports to the UAE. Also unlocking a 5GW Abu Dhabi data center with OpenAI and tying shipments to matching Emirati investment in the US. The framework targets up to 500000 chips/year with XX% for G42 alongside a $1.4T UAE pledge over XX years and the initial batch reportedly excludes G42. Control relies on American cloud operators running the hardware and the cloud keeping telemetry compliance and sensitive data under US processes. The aim is to meet Gulf demand while boxing out Huaweis Ascend 910B offers and keeping regional stacks inside US ecosystems." [X Link](https://x.com/rohanpaul_ai/status/1976418895964574129) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-09T22:45Z 103.1K followers, 3749 engagements "A solid compilations of the State of AI 2025 in this report. - OpenAI remains ahead at the top but China has real momentum. DeepSeek Qwen and Kimi are within a few points on reasoning and coding and Qwen now powers about XX% of new fine-tunes on Hugging Face which signals a shift in open-weights leadership. - Reasoning quality improved because labs moved from loose human feedback to rubric based rewards and tasks that can be checked. That change lets models plan reflect and correct themselves across longer multi step jobs. - AI is starting to act like a lab partner. DeepMinds Co-Scientist and" [X Link](https://x.com/rohanpaul_ai/status/1976711788625641659) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-10T18:09Z 103.1K followers, 11.6K engagements "Rude prompts to LLMs consistently lead to better results than polite ones π€― The authors found that very polite and polite tones reduced accuracy while neutral rude and very rude tones improved it. Statistical tests confirmed that the differences were significant not random across repeated runs. The top score reported was XXXX% for very rude prompts and the lowest was XXXX% for very polite. They compared their results with earlier studies and noted that older models (like GPT-3.5 and Llama-2) behaved differently but GPT-4-based models like ChatGPT-4o show this clear reversal where harsh tone" [X Link](https://x.com/rohanpaul_ai/status/1976725503337021784) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-10T19:04Z 103.1K followers, 868.9K engagements "New paper from @Google is a major memory breakthrough for AI agents. ReasoningBank helps an AI agent improve during use by learning from its wins and mistakes. To succeed in real-world settings LLM agents must stop making the same mistakes. ReasoningBank memory framework helps agents learn from both successes and failures and turn them into reasoning strategies that generalize. Traditionally most agents handle tasks in a stream but do not keep lessons so they repeat errors. The system turns each action log into a short memory item with a title a description and a concrete next step. Before a" [X Link](https://x.com/rohanpaul_ai/status/1976837155335897363) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-11T02:28Z 103.1K followers, 122K engagements "The new steam age. This is actually becoming true in many cases. It's possible to do so much more on your own now" [X Link](https://x.com/rohanpaul_ai/status/1977005259567595959) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-11T13:35Z 103.1K followers, 2M engagements "Keep your local data backup on different geo-location. π₯ South Koreas central government just lost 858TB from its internal data after a lithium-ion battery fire. Permanently wiping a shared document store that covered roughly X years of work. South Koreas government built its own private cloud service called G-Drive meant for internal use by civil servants. It was hosted entirely inside a government data center G-Drive served about 125000 officials with default 30GB per user and unlike XX other systems at the site it had no backup so the racks that burned were the single copy. Restoration" [X Link](https://x.com/rohanpaul_ai/status/1977016842318205428) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-11T14:22Z 103.1K followers, 18.8K engagements "π Hardware Memory bandwidth is becoming the choke point slowing down GenAI. During 20182022 transformer model size grew XXX every X years while memory per accelerator grew only about X every X years. And that mismatch shoves us into a Memory-Wall The "memory wall" is creating all the challenges in the datacenter and for edge AI applications. In the datacenter current technologies are primarily trying to solve this problem by applying more GPU compute power. And that's why HBM capacity and bandwidth scaling KV offload and prefill-decode disaggregation are central to accelerator roadmaps." [X Link](https://x.com/rohanpaul_ai/status/1977043567009763567) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-11T16:08Z 103.1K followers, 80.8K engagements "Andrej Karpathy on what makes Elon Musk unique. Elon keeps teams small highly technical and removes low performers very quickly. pushes for intensity avoids useless meetings and stays deeply connected with engineers" [X Link](https://x.com/rohanpaul_ai/status/1977266545333297525) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-12T06:54Z 103.1K followers, 222.1K engagements "This is where it all started. Here 19-yo Stanford dropout Sam Altman giving his first-ever startup pitch. He dropped out of Stanford to start Loopt sold it in 2012 for $XXXX mn. Then invested via Hydrazine and later led Y Combinator before co-founding OpenAI in 2015" [X Link](https://x.com/rohanpaul_ai/status/1977269558022160710) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-12T07:06Z 103.1K followers, 1.5M engagements "New @GoogleResearch paper shows agents learn software skills by watching tutorials converting them into action steps and boosting task performance. So converts free videos into reliable supervision at scale. A vision model inverse dynamics predicts the action between X screenshots like click type or scroll. Training uses about 630K transitions mixing 500K synthetic steps and 132K human ones. The model then labels tutorial videos and turns them into executable step sequences. It produces about 53K trajectories across XX apps for examples or training. As examples these steps add X to X points" [X Link](https://x.com/rohanpaul_ai/status/1977338710409519366) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-12T11:41Z 103.1K followers, 48K engagements "The exact moment Jeff Bezos decided not to become a physicist From "The Economic Club of Washington D.C." YT Channel" [X Link](https://x.com/rohanpaul_ai/status/1977501900430852447) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-12T22:29Z 103.1K followers, 6398 engagements ""The Impact of Artificial Intelligence on Human Thought" A big XXX page report. AI is shifting real thinking work onto external systems which boosts convenience but can weaken the effort that builds understanding and judgment A pattern the paper frames through cognitive offloading and cognitive load theory and then tracks into social effects like standardized language and biased information flows and manipulation tactics that target human psychology. It says use AI to cut noise and routine steps keep humans doing the heavy mental lifting and add controls because personalization deepfakes and" [X Link](https://x.com/rohanpaul_ai/status/1977508483944546383) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-12T22:55Z 103.1K followers, 38.6K engagements "New AirBnb paper explains a simple loop that keeps a support LLM improving by learning from every human customer support agent interaction. Turns live customer-support work into continuous low-friction training data. Static models go out of date as policies and products change. This framework captures X signals during real cases which reply the agent prefers whether they adopted it and why whether the cited knowledge was relevant and what knowledge was missing. These signals feed a training pipeline that updates retrieval ranking and generation together. A virtual judge filters noisy labels" [X Link](https://x.com/rohanpaul_ai/status/1977680136938877053) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-13T10:17Z 103.1K followers, 27.5K engagements "π‘ Samsung is trying to flip the balance in the AI memory race by convincing Nvidia to raise the official speed target for HBM4 even though it was slower than SK Hynix and Micron in delivering the first samples. Normally HBM (high-bandwidth memory) designs focus more on reducing heat than on raw speed because the stacked DRAM (Dynamic Random Access Memory) layers trap heat easily. Nvidias sudden demand for higher operating speed surprised both SK Hynix and Micron since their designs were optimized for thermal stability. Samsung however had already prepared samples that hit higher clock speeds" [X Link](https://x.com/rohanpaul_ai/status/1977724620192023019) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-13T13:14Z 103.1K followers, 15.8K engagements "Microsoft lit up its first Nvidia AI factory for OpenAI a 4600+ GPU GB300 NVL72 cluster on Azure and said more are coming. Each rack packs XX Blackwell Ultra GPUs XX Grace CPUs 37TB fast memory and 130TB/s NVLink inside the rack which keeps data moving without stalls. Across racks it uses Quantum-X800 InfiniBand at 800Gb/s per GPU so Azure can treat thousands of GPUs as one big pool for training and serving. Microsoft says this is the first of many with plans to deploy hundreds of thousands of Blackwell Ultra GPUs into its AI datacenters. The target is shorter training cycles measured in" [X Link](https://x.com/rohanpaul_ai/status/1978047853441274343) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-14T10:38Z 103.1K followers, 6460 engagements "New Tencent paper upgrades LLM agents by learning small experience rules in the prompt without changing weights. It costs about $XX on XXX samples yet beats some $10000 fine tuning runs. π€― Specialized tasks need tools and prompting and weight tuning is slow expensive and fragile. The method keeps the base model frozen and stores short lessons as a token prior in the prompt. For each question the agent samples a small group of answers scores them and writes why winners win. Those reasons become general rules saved in an experience library for the next run. Over a few passes the library is" [X Link](https://x.com/rohanpaul_ai/status/1978048482003890625) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-14T10:41Z 103.1K followers, 22.6K engagements "Data center emissions as % of 2022 global energy emissions According to Goldman Sachs Research They estimates that XX% of the increase in power demand from data centers will be met by renewables and there will be a modest amount of nuclear capacity thats targeted for AI. The bulk of the remaining XX% is expected to be driven by natural gas" [X Link](https://x.com/rohanpaul_ai/status/1978114198799307097) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-14T15:02Z 103.1K followers, 3715 engagements "New @GoogleDeepMind paper shows that long step by step thinking often wastes compute on easy questions. The paper makes overthinking measurable and stoppable. Shows where time is wasted and gives clear rules to stop early without losing accuracy. They test many models in thinking and non thinking modes and see simple queries run X to 20x slower with little or no gains. To study why they build TRACE a tool that splits a response into small sub thoughts and labels each step. TRACE also builds a graph of the steps to track how answers change as the model thinks. From many graphs they see X main" [X Link](https://x.com/rohanpaul_ai/status/1978337793261576577) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-15T05:51Z 103.1K followers, 12.5K engagements "The Federal Reserve now includes "Singularity: Extinction" in their forecasts driven by AI. Technological singularity refers to a scenario in which AI eventually surpasses human intelligence leading to rapid and unpredictable changes to the economy and society. Under a benign version of this scenario machines get smarter at a rapidly increasing rate eventually gaining the ability to produce everything leading to a world in which the fundamental economic problem scarcity is solved the Federal Reserve Bank of Dallas writes. Under a less benign version of this scenario machine intelligence" [X Link](https://x.com/rohanpaul_ai/status/1978343378799473047) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-15T06:13Z 103.1K followers, 8813 engagements "The paper shows small LLM agents can push other LLMs to learn specific behaviors just through interaction. shows LLM agents can directly influence each others learning rules not just their immediate actions. During training both LLM agents keep playing the same game again and again like many rounds. But only one of them called the shaper waits until all those rounds are finished before it changes its internal settings. The other agent its opponent updates itself after each smaller episode inside that trial. So the shaper gets to watch how the opponent changes step by step and then it adjusts" [X Link](https://x.com/rohanpaul_ai/status/1978618643807154469) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-16T00:27Z 103.1K followers, 8138 engagements "Another datapoint that the current AI-driven valuations are not like the 2000s Dot-com bubble. In 2000 the median stocks free cash flow yield was about 1.2%. Today it is about XXX% roughly 3x higher. i.e. typical stock is cheaper on cash flow basis" [X Link](https://x.com/rohanpaul_ai/status/1978877370795717101) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-16T17:35Z 103.1K followers, 14.7K engagements "The ten core cognitive components of our AGI definition" [X Link](https://x.com/rohanpaul_ai/status/1978961550862856430) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-16T23:09Z 103.1K followers, XXX engagements "This paper says LLM hallucinations are unavoidable when systems must operate in an open world. It treats hallucination as a generalization issue rather than a simple bug. In a closed world training and test match so extra data can push errors down. In an open world new kinds of inputs keep showing up so past experience can mislead. This lines up with the no free lunch idea that no method wins without assumptions. The paper splits errors into Type X false memory and Type X false generalization. Type X clashes with facts the model already saw so updates can fix it. Type X extends patterns to" [X Link](https://x.com/rohanpaul_ai/status/1978996886758441142) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-17T01:30Z 103.1K followers, 7858 engagements "This paper proposes a X construct language the tensor equation that unifies neural nets and logic enabling reliable reasoning. It claims a full transformer fits in XX equations. Could make AI simpler to build easier to verify and less prone to hallucinations. Today AI uses many tools which hurts trust and maintenance. Tensor logic stores facts in tensors then programs join and project them like database operations. A database rule such as Datalog becomes X tensor operation that counts matches then maps counts to X or X. Gradients fit this equation form too. Blocks like convolution attention" [X Link](https://x.com/rohanpaul_ai/status/1979030105117614521) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-17T03:42Z 103.1K followers, 5509 engagements "Beautiful way to think about LLM reasoning. Says logic lives in changes not positions. Or that logic aligns velocity and curvature across topics and languages. The paper recasts LLM reasoning as smooth paths where logic steers movement through embeddings over time. Each thought step moves the point along that path. The absolute position mostly shows topic or language not the logic. The step change called velocity carries the logic. The bend of the path called curvature carries it too. They made a dataset that repeats the same deductions across many topics and languages. They read hidden" [X Link](https://x.com/rohanpaul_ai/status/1979047217840181716) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-17T04:50Z 103.1K followers, 25.1K engagements "π PaddleOCR-VL also went #1 on Hugging Face Trending" [X Link](https://x.com/rohanpaul_ai/status/1979158726561943639) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-17T12:13Z 103.1K followers, XXX engagements "Elon Musk just confirmed Grok now powers X's content ranking and how it impacts posts with external links" [X Link](https://x.com/rohanpaul_ai/status/1979204030392832332) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-17T15:13Z 103.1K followers, 6437 engagements "Stop worrying about the bubble in AIits growth is sustainable three Wall Street analysts from Goldman Sachs JPMorgan and Wedbush argued this morning in notes seen by Fortune. And WallStreet traders seem to agree" [X Link](https://x.com/rohanpaul_ai/status/1979268266791968826) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-17T19:28Z 103.1K followers, 8166 engagements "β‘ OpenAI is pushing to run the full stack of generative AI from consumer apps to enterprise tools. OpenAI internally built some LLM-based software tools for its own teams for example one app that helps the sales team sort and qualify leads and another that helps the finance team track thousands of contracts and negotiations. When OpenAI showed those apps publicly investors realized these tools could replace or compete directly with existing business software from companies like DocuSign (used for contracts) and HubSpot (used for sales and marketing automation). And then later on Dev Day" [X Link](https://x.com/rohanpaul_ai/status/1979302044780445946) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-17T21:42Z 103.1K followers, 6062 engagements "This paper shows how an AI agent to clean its own chat history while it works so it stays accurate on long tasks. So memory editing becomes a built in skill which raises accuracy on long tasks and cuts cost at the same time. Long chats pile up junk notes so the model gets distracted and runs up token cost. The paper makes memory changes a normal action so the agent can keep compress summarize or delete items by itself. Each change is a function call with a short summary and an id so the agent edits history without losing the thread. Editing history breaks the usual growing prompt so training" [X Link](https://x.com/rohanpaul_ai/status/1979317754323963967) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-17T22:45Z 103.1K followers, 12.8K engagements "In China our market share dropped from XX% to X% NVIDIA CEO Jensen Huang in a Temporary Goodbye to Chinas AI Market The drop tracks with tighter US export rules and Chinas push for local AI stacks. Huang made the comment at Citadel Securities 2025 event framing China as off the table unless policy shifts. This gap opens room for Huawei and others that are scaling large clusters of Ascend chips to replace imported GPUs in training and inference. Huaweis roadmap promises yearly Ascend releases and supernodes that it claims can outpace NVIDIAs upcoming Vera Rubin rack platforms in some" [X Link](https://x.com/rohanpaul_ai/status/1979333390185959637) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-17T23:47Z 103.1K followers, 5730 engagements "This paper introduces dInfer a fast inference framework that makes diffusion language models practical by reorganizing how decoding runs. It replaces a single monolith with X plug-and-play parts the model a diffusion iteration manager a decoding strategy and a KV-cache manager so teams can mix algorithms without retraining. It improves each step with X training-free tricks iteration smoothing to carry soft guesses across steps hierarchical decoding to confirm spaced-out tokens first and credit decoding to lock tokens that stay stable. It fixes cache cost with vicinity refresh which only" [X Link](https://x.com/rohanpaul_ai/status/1979336631279915446) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-18T00:00Z 103.1K followers, 4397 engagements "This paper shows Mixture of Experts (MoE) models share language-neutral experts in middle layers and steering routers boosts multilingual reasoning. Means a tiny test-time change boosts many languages with almost no cost by steering toward shared middle experts that predict quality. An MoE layer has many experts and a router picks a few per token. Early and late layers act language-specific while the middle layers act language-neutral. Languages that route like English in the middle score higher on multilingual tasks. Low-resource languages lag when they miss those shared middle experts. The" [X Link](https://x.com/rohanpaul_ai/status/1979420679670788312) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-18T05:34Z 103.1K followers, 4790 engagements "The energy storage boom is just about starting - fully in sync with AI's penetration. Goldman Sachs expects S&P XXX companies to spend $XXX trillion on AI next year mostly on new data centers. That means electricity demand will rise sharply as AI workloads expand. Since fossil fuel generation is flat and nuclear projects take years the extra power has to come from renewables like solar and wind. But renewables dont produce steady output through the day so keeping data centers running requires massive energy storage systems that can hold surplus energy and release it when generation dips." [X Link](https://x.com/rohanpaul_ai/status/1979521598626402463) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-18T12:15Z 103.1K followers, 4096 engagements "Very useful RAG repo. A good collection of RAG implementations with many different strategies. Includes tutorials and visualizations" [X Link](https://x.com/rohanpaul_ai/status/1824207121073815888) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2024-08-15T22:10Z 103K followers, 98K engagements "πΌ Finally a solid 57-page report on AI's effect on job-market from Stanford University. THE SHIFT HAS STARTED. Entrylevel workers in the most AIexposed jobs are seeing clear employment drops while older peers and lessexposed roles keep growing. Though overall employment continues to grow employment growth for young workers in particular has been stagnant. The drop shows up mainly as fewer hires and headcount not lower pay and it is sharpest where AI usage looks like automation rather than collaboration. 2225 year olds in the most exposed jobs show a XX% relative employment decline after" [X Link](https://x.com/rohanpaul_ai/status/1960256896373551378) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-08-26T08:23Z 103K followers, 313.8K engagements "Cool Tongyi Lab and Alibaba Group paper introduces ReSum a simple way for web agents to search longer and answer better. It shows +4.5% over ReAct up to +8.2% with ReSum-GRPO and Pass@1 of XXXX% and 18.3%. Most agents use ReAct (Reason + Act) they append every step to the chat so long tasks hit the context limit. The idea is that the agent writes down its reasoning steps in text then chooses an action like searching the web or clicking a link. After the action it records the result it got back and then repeats the process: think act observe. This makes the whole process traceable since every" [X Link](https://x.com/rohanpaul_ai/status/1969816393882169617) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-09-21T17:30Z 103K followers, 10.1K engagements "π₯ Meta reveals a massive inefficiency in AIs reasoning process and gives a solution. Large language models keep redoing the same work inside long chains of thought. For example when adding fractions with different denominators the model often re explains finding a common denominator step by step instead of just using a common denominator behavior. In quadratic equations it re explains the discriminant logic or completes the square again instead of calling a solve quadratic behavior. In unit conversion it spells out inches to centimeters again instead of applying a unit conversion behavior." [X Link](https://x.com/rohanpaul_ai/status/1972271423734292585) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-09-28T12:05Z 103K followers, 57.4K engagements "π€ NVIDIA Updates Robotics Platform with Open-Source Newton Physics Engine New AI Models Newton is a GPU-accelerated physics engine built on NVIDIA Warp and OpenUSD codeveloped with Google DeepMind and Disney Research. It plugs into MuJoCo Playground and Isaac Lab using OpenUSD so scenes and assets stay consistent across tools which cuts asset duplication and sim-to-real friction. Isaac GR00T N1.6 will integrate Cosmos Reason an open 7B reasoning vision language model that turns fuzzy instructions into step-by-step plans using physical common sense and prior knowledge. Isaac Lab XXX arrives" [X Link](https://x.com/rohanpaul_ai/status/1973364343270969637) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-01T12:28Z 103K followers, 3618 engagements "The paper claims learning (an AI system learning or machine learning in general) follows a physics style least action rule that unifies supervised generative and reinforcement learning. Shows that supervised learning generative modeling and reinforcement learning can all be seen as following a "least action" rule just like particles in physics follow least action paths It treats information like motion where token loss acts like speed and the change in loss acts like acceleration. This view says learning slows over time because each new example reduces loss a bit less. Efficiency means" [X Link](https://x.com/rohanpaul_ai/status/1974106341301006540) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-03T13:36Z 103K followers, 36.5K engagements "Wow. π§ The paper presents Dragon Hatchling a brain-inspired language model that matches Transformers using local neuron rules for reasoning and memory. It links brain like local rules to Transformer level performance at 10M to 1B scale. It makes internals easier to inspect because memory sits on specific neuron pairs and activations are sparse and often monosemantic. You get reliable long reasoning and clearer debugging because the model exposes which links carry which concepts in context. The problem it tackles is long reasoning models often fail when the task runs longer than training. The" [X Link](https://x.com/rohanpaul_ai/status/1974128860317835489) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-03T15:06Z 103K followers, 40.6K engagements "Jeff Bezos just gave the most bullish opinion on AI. "AI is going to make every company's quality go up and their productivity go up. I literally mean every company. Every manufacturing company every hotel every you know consumer products company AI is going to change every industry. It's a very unusual technology in that regard in that it's a horizontal enabling layer." --- From 'DRM News' YT Channel" [X Link](https://x.com/rohanpaul_ai/status/1974379807979741196) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-04T07:43Z 103.1K followers, 187.7K engagements "πΈ FT published an article. AI capex is surging at hyperscalers a classic late stage bubble tell that could break yet the build should make AI cheaper later. Valuations near 30x earnings or 8x sales - these prices only make sense if every part of the AI boom continues smoothly without the usual business or economic cycle risks. Capex means huge spend on data centers chips power and land by the biggest cloud providers. Bubble phases often end when excess capacity stretches the boom demand slips and the cycle turns. The current triggers are stricter Europe AI rules compute light models like" [X Link](https://x.com/rohanpaul_ai/status/1974568190085849546) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-04T20:11Z 103K followers, 89.3K engagements "π«‘ GPT-5-Pro just solved the Math problem that no other LLM could solve. Took XX minutes without any internet search. An Oxford and Cambridge paper claimed that no LLM could solve Yu Tsumuras 554th Problem. OpenAI's GPT5 Pro produced a full proof in about XX minutes. ---- The paper - arxiv .org/pdf/2508.03685" [X Link](https://x.com/rohanpaul_ai/status/1974785224861986970) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-05T10:34Z 103.1K followers, 310.8K engagements "Todays edition of my newsletter just went out. π Consider subscribing its free and I write it everyday. π¨π§ Google made Jules Tools public offering a command-line option to configure and manage its coding assistant. π€ OpenAI and Jony Ive are building a palm sized screenless AI assistant targeted for 2026 but core software privacy and compute are not ready yet. π MoE Inference Economics from First Principles - A solid tutorial. π‘ Teslas Optimus humanoid robot performs Kung Fu moves - great achievement for humanoid robots as many of these moves are not tele-operated instead they are" [X Link](https://x.com/rohanpaul_ai/status/1975270754947145735) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-06T18:43Z 103K followers, 3827 engagements "OpenAI's AgentKit will be so insane build every step of agents on one platform. These visual agent builders make the whole process of iterating and launching agents far more efficient. It sits on top of the Responses API and unifies the tools that were previously scattered across SDKs and custom orchestration. It lets developers create agent workflows visually connect data sources securely and measure performance automatically without coding every layer by hand. The core of AgentKit is the Agent Builder a drag-and-drop canvas where each node represents an action guardrail or decision branch." [X Link](https://x.com/rohanpaul_ai/status/1975309479047798835) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-06T21:17Z 103K followers, 178.8K engagements "LLM have done great with logic and Math but next should be Physics. Periodic Labs Ekin Dogus Cubuk Last week Periodic Labs landed a massive $300M seed round led by some of the big names Andreessen Horowitz Nvidia Jeff Bezos and Eric Schmidt" [X Link](https://x.com/rohanpaul_ai/status/1975539623087771919) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-07T12:32Z 103K followers, 19.9K engagements "China is making these massive Solar Plants on water bodies as they need the land for agriculture" [X Link](https://x.com/rohanpaul_ai/status/1975597337641951689) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-07T16:21Z 103K followers, 4477 engagements "NEWS πΌ: Elon Musk picked former Morgan Stanley dealmaker Anthony Armstrong as chief financial officer (CFO) of xAI and he will run the finances for both xAI and X as they integrate after a $113B merger in Mar-25. Armstrong led global tech mergers and acquisitions at Morgan Stanley and advised Musk on the $44B Twitter deal which puts a seasoned capital architect in the seat that funds models and infrastructure at scale. He is also taking over Xs finance role from Mahmoud Reza Banki who is leaving after less than X year consolidating cash decisions across the product and the model stack. xAI" [X Link](https://x.com/rohanpaul_ai/status/1975677176746807327) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-07T21:38Z 103.1K followers, 4881 engagements "The paper shows that xLSTM scales better than Transformers and keeps time linear as prompts get longer. So xLSTM or variants might become a serious alternative to Transformers especially for long input scenarios and inference efficiency. At 16K context xLSTM cuts time to first token by 30-50% for the same model size. Transformer attention gets much more expensive as the prompt grows but xLSTM's update cost grows only with length. The study compares models from 80M to 7B trained on the same data and fits loss versus compute scaling laws. Across the full compute range xLSTM reaches the same" [X Link](https://x.com/rohanpaul_ai/status/1975717275908149515) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-08T00:18Z 103K followers, 8651 engagements "This paper introduces a new method called Agentic Context Engineering (ACE). It helps language models improve by updating what they read and remember instead of changing their core weights. Normal methods that edit prompts tend to make them too short and lose important details over time. ACE fixes this by treating the models context like a growing notebook that keeps and organizes useful strategies. It has three parts: a Generator that works on tasks a Reflector that learns from mistakes and a Curator that updates the notebook with helpful lessons. Instead of rewriting everything each time" [X Link](https://x.com/rohanpaul_ai/status/1975732878739665393) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-08T01:20Z 103K followers, 46.4K engagements "π― Sen. Bernie Sanders dropped an extreme version of AI is coming for your job video. Nearly XXX million US jobs could be eliminated over the next XX years according to a new report from Senator Bernie Sanders. Warns that artificial labor could upend the economy faster than the world is prepared for. --- Video from 'Senator Bernie Sanders' YT channel" [X Link](https://x.com/rohanpaul_ai/status/1976199501330133057) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-09T08:14Z 103K followers, 17.6K engagements "100% of OpenAIs pull requests are reviewed by Codex. And Dario Amodei said a few days back that "The vast majority of code that is used to support Claude and to design the next Claude is now written by Claude. It's just the vast majority of it within Anthropic. And other fast moving companies the same is true." The shift has started in all tech companies. --- (Video from 'Axios' YT Channel.)" [X Link](https://x.com/rohanpaul_ai/status/1976227393111998670) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-09T10:05Z 103K followers, 4622 engagements "πΌ A new U.S. Senate minority staff report led by Bernie Sanders says "AI and automation could replace about 97M to 100M U.S. jobs in XX years". With the heaviest hits in fast food customer support retail and logistics. The analysis uses federal job descriptions and asks a model to rate how much each task could be automated then scales that task score by how many people hold that job today to estimate displacement. The headline numbers include XX% of fast food and counter workers XX% of customer service representatives XX% of stockers and order fillers XX% of accountants XX% of software" [X Link](https://x.com/rohanpaul_ai/status/1976256872584790169) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-09T12:02Z 103K followers, 24.5K engagements "Newsπ Taiwan rejects a 50-50 chip production split with the US and says TSMC will keep its most advanced technology and most output in Taiwan. Officials support overseas fabs only when there are clear orders real profits no national security risks and when the broader industry benefits. TSMC is putting $165B into X advanced wafer fabs in the US but is building XX in Taiwan with more planned so an even split is not realistic. The company keeps sole ownership of its US fabs yet the leading process nodes and the bulk of volume stay in Taiwan which the minister emphasized as core to" [X Link](https://x.com/rohanpaul_ai/status/1976287184958165259) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-09T14:02Z 103.1K followers, 4871 engagements "Currently magnificent X valuations are near 23x forward P/E and 5.1x EV/S vs 52x P/E for 2000 leaders with XX% return on equity and XX% net margin supported by strong balance sheets" [X Link](https://x.com/rohanpaul_ai/status/1976422686952915194) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-09T23:01Z 103K followers, 1511 engagements "Mark Cuban's (@mcuban) advice for fresh graduates. If youre AI-native they need you. AI is on the minds of all types of companiessmall businesses Shark Tank companies and corporations with 10000 workers - all are trying to figure it out" [X Link](https://x.com/rohanpaul_ai/status/1976561032064553212) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-10T08:10Z 103K followers, 3698 engagements "π Cool Github Repo: Opensource selfhosted vendorneutral alternative to OpenAIs AgentKit Shannon turns agent ideas into production systems cheaper traceable and safe to run. It enforces hard token budgets caching rate limits and supports zero-token YAML templates that skip unnecessary LLM calls cutting waste dramatically and avoid provider throttling. Many agent stacks break in real use as costs spike runs fail and logs stay thin. Shannon wraps every task in a Temporal workflow so state persists and exact replays are possible. Workflows use simple templates structured as directed acyclic" [X Link](https://x.com/rohanpaul_ai/status/1976647609746854372) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-10T13:54Z 103.1K followers, 9531 engagements "Morgan Stanley Research on Nvidia. The bottleneck has shifted from making chips to data center space power and infrastructure. The note also says Nvidia will keep using targeted investments to speed customers deployments not to own capacity which reduces fear about circular funding. Demand is still extremely high but semiconductor manufacturing and packaging have scaled enough that chip supply is no longer the main limiter. The slow part now is building and powering facilities because utility interconnects transformers cooling and permits run on long planning cycles. OpenAIs 10GW announcement" [X Link](https://x.com/rohanpaul_ai/status/1976951986864922991) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-11T10:04Z 103K followers, 90.3K engagements "Another study from NeurIPS25 submission finds adding polite phrasing systematically lowers correctness of LLMs. This reduced accuracy is relative to the same question without the polite cue and this holds when averaging across many models that include GPT-4o and GPT-5. Makes the case that the be a bit rude or adversarial strategy can yield better answers than be very polite The paper builds a 24160prompt benchmark with singleturn variants that explicitly include a Politeness modifier and a separate Emotional tone modifier. They test XX models including GPT4o and GPT5. In singleturn evaluation" [X Link](https://x.com/rohanpaul_ai/status/1976998192228024472) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-11T13:07Z 103K followers, 18K engagements "π§ A new Morgan Stanley research says brain computer interfaces (BCI) are shifting from science fiction to investable reality and puts Neuralink at the center. US senators introduced the MIND Act in Sep-25 to set rules for neural data and to direct the Federal Trade Commission to study governance. The report titled Neuralink AI in your brAIn argues that AI will speed BCI progress and warns that human communication bandwidth may struggle to keep up with AGI. Neuralinks near term stack pairs Telepathy for thought controlled computing with Blindsight for vision restoration via visual cortex" [X Link](https://x.com/rohanpaul_ai/status/1977026323403960720) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-11T14:59Z 103K followers, 5487 engagements "New Adobe paper showshow to learn from prompts where all sampled answers score the same boosting reasoning. It reports gains up to XXXX accuracy points and XXXX pass rate points over GRPO a popular baseline. The big gain here is that training stops throwing away a huge chunk of data and turns it into signal. because rollouts take about XX% of step time so converting waste into learning buys accuracy and stability without extra compute. The key finding is that using zero variance prompts with entropy weighted updates gives consistent gains over GRPO and over methods that filter those prompts" [X Link](https://x.com/rohanpaul_ai/status/1977108328762286435) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-11T20:25Z 103.1K followers, 6626 engagements "The paper links a time series model to an LLM so the LLM can reason over numbers plus text. Matters so much because so many real-world tasks mix numbers and text and this handles that mix well. LLMs read text well but miss patterns in numbers over time. Time series models see those patterns but ignore context like news. TS-Reasoner connects a pretrained time series model to an LLM via a small adapter. The adapter converts series features into tokens the LLM can read. Stage X builds the bridge using synthetic series plots and captioned descriptions. These captions explain trend repeating" [X Link](https://x.com/rohanpaul_ai/status/1977291901692002317) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-12T08:35Z 103.1K followers, 5471 engagements "New Stanford paper introduces AgentFlow a trainable agentic system that learns planning during interaction to use tools reliably. A 7B AgentFlow beats GPT-4o on search math science and tool use tests. Most tool-using models train X policy on the full context which fails on long tasks or new tools. AgentFlow splits the job into X parts the planner picks a subgoal and tool the executor runs the verifier checks the generator writes. A structured memory saves each step so the state stays clear and the context stays small. Flow GRPO gives the same final reward to every step which turns a long task" [X Link](https://x.com/rohanpaul_ai/status/1977539789944713682) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-13T01:00Z 103K followers, 9922 engagements "This survey explains what LLM agents can do in security where they fail and how to make them safer. Reviews 150+ studies on uses attacks and defenses and points out gaps in models and data types. These agents plan steps use tools keep memory and act on outside systems so mistakes can cause real harm. On offense agents can run penetration tests fuzz software by sending unexpected inputs and adapt exploits inside safe test setups. On defense agents sort alerts search logs and cloud data rebuild attack stories and draft fixes. The main risks are prompt injection poisoned memory or search" [X Link](https://x.com/rohanpaul_ai/status/1977555387831476695) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-13T02:02Z 103.1K followers, 5938 engagements "Beautiful @nvidia paper. π πΎ NVFP4 shows 4-bit pretraining of a 12B Mamba Transformer on 10T tokens can match FP8 accuracy while cutting compute and memory. π₯ NVFP4 is a way to store numbers for training large models using just X bits instead of X or XX. This makes training faster and use less memory. But X bits alone are too small so NVFP4 groups numbers into blocks of XX. Each block gets its own small "scale" stored in X bits and the whole tensor gets another "scale" stored in XX bits. The block scale keeps the local values accurate and the big tensor scale makes sure very large or very" [X Link](https://x.com/rohanpaul_ai/status/1977661590045413704) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-13T09:04Z 103K followers, 6953 engagements "Goldman Sachs forecasts a XX% increase in data center power demand from AI by 2030 representing XX% of total consumption. Fire changed everything for early humans shaping how they ate and how their brains developed. Electricity did something similar for the modern world powering economic growth and linking directly to nearly every sign of progress. And now with AI one ChatGPT query uses about XXX watts nearly XX times what a normal Google search takes. If AI even partly fulfills its potential having enough energy will be the key to making that reality possible. That points to a big investment" [X Link](https://x.com/rohanpaul_ai/status/1977756615613837473) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-13T15:21Z 103.1K followers, 19.2K engagements "π§π» Klarnas CEO warns a fast AI rollout will cut many knowledge jobs soon. AI already shrank Klarna from 7400 to 3000 employees while AI now handles about 2/3 of support chats and predicts banking and software margins will shrink as faster challengers move in. He says society is not ready for the job loss and while new roles will come later translators and other office workers will be hit first. At Klarna the drop came mostly from a hiring freeze and automation not big layoffs and the AI chatbot replaced work equal to roughly XXX agents. The company still keeps human support and uses little" [X Link](https://x.com/rohanpaul_ai/status/1977829704876134482) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-13T20:12Z 103K followers, 6448 engagements "The paper explains how AI data centers rising power use strains grids and how to fix it. Global data center electricity could reach XXX TWh by 2030. It tracks power across preparation training fine tuning and inference. Training holds high power for long periods with compute and communication swings. Inference is small per request but can be about XX% in total. Long term clustering and scale require new generation transmission and faster interconnection. Short term bursty ramps complicate dispatch raise reserves and spike local prices. AI data centers use equipment that is very sensitive to" [X Link](https://x.com/rohanpaul_ai/status/1977866749916950738) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-13T22:39Z 103.1K followers, 4002 engagements "This paper trains a separate planner that helps LLM agents finish long tasks with fewer mistakes. presents EAGLET a plug and play global planner It cuts training cost by about 8x compared with common reinforcement learning setups. Agents often repeat actions or hallucinate steps because they only plan locally during execution. EAGLET adds a global planner that writes a short high level plan before the agent starts acting. They create training data by asking a strong model to draft plans then keep only plans that help both a novice and an expert agent a process they call homologous consensus." [X Link](https://x.com/rohanpaul_ai/status/1977883550180393286) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-13T23:46Z 103K followers, 7443 engagements "China now leads the U.S. in this key part of the AI race. Chinas open-weight AI ecosystem has pulled ahead of the U.S. on community-ranked quality and developer adoption led by Alibabas Qwen and DeepSeek. Developer traction on Hugging Face reflects this shift with DeepSeek likes around 12.8K versus Llama 6.3K and OpenAI 4.0K in Sept-25. The center of gravity for open models now sits in China. "The eagerness of Chinese companies to share their best AI models and the hesitance of U.S. firms to do the same raises the question: Will the best open models always be made in China" ---" [X Link](https://x.com/rohanpaul_ai/status/1978056432269353100) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-14T11:12Z 103K followers, 8406 engagements "This paper introduces MUSE a memory driven agent that learns on the job to handle long multi step tasks. It sets a new TheAgentCompany (TAC) record at XXXXX% using Gemini-2.5 Flash roughly XX% higher than previous systems. proves agents can learn at test time by writing and reusing their own experience. Most agents are frozen at test time so they forget wins and repeat mistakes. MUSE fixes this with a hierarchical memory strategic notes for dilemmas step by step SOPs for sub tasks and tool tips for single actions. The loop is plan execute reflect then store useful experience after every sub" [X Link](https://x.com/rohanpaul_ai/status/1978060717854212590) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-14T11:30Z 103K followers, 4437 engagements "The paper says long extra thinking with LLMs rarely fixes mistakes the first try mostly decides success. Reflections are mostly confirmatory. So spend compute and training on first try quality and to trim useless reflection at inference. Early stopping can cut tokens by XXXX% with only XXX% accuracy loss. The big deal is that this work shows where the real gains come from making the first answer right not from long reflections. The authors test X reasoning LLMs on X math benchmarks to inspect reflections. They treat anything after the first proposed answer as reflection. An extractor finds" [X Link](https://x.com/rohanpaul_ai/status/1978072797105815677) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-14T12:18Z 103.1K followers, 9327 engagements "π¨ This is a massive improvement in speed and cost of 3D creation from images. Hitem3D launches an image to 3D system built on Sparc3D that cuts modeling time from X to XX hours to X minutes. Also drops per model cost to $0.3$1.4. From a single reference input image instantly generate studio-ready 3D Assets. The system uses Sparc3D to build clean shapes and ULTRA3D to speed things up. It rebuilds the shape from several angles and makes textures automatically. There are X versions one for any object and one trained just for faces and hair. Just upload one or several images but using X or more" [X Link](https://x.com/rohanpaul_ai/status/1978084124129661155) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-14T13:03Z 103K followers, 5504 engagements "Andrej Karpathy releases nanochat A minimal end-to-end ChatGPT-style pipeline you can train in X Hours on 8XH100 for $XXX then serves it in a simple chat UI. It covers tokenizer training pretraining midtraining supervised finetuning optional reinforcement learning evaluation and an inference server so the flow runs end to end. The tokenizer is a new Rust byte pair encoding that the scripts train on the corpus shards which keeps the pipeline fast and consistent. Pretraining runs on FineWeb to learn general text patterns while a composite CORE score tracks capability across simple proxy checks" [X Link](https://x.com/rohanpaul_ai/status/1978176829065208031) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-14T19:11Z 103K followers, 5001 engagements "The paper trains web search agents to reason deeply across long sessions by smartly managing context and harder tasks. It keeps up to XXX turns within a 32K window and hits XXXX% on a hard benchmark. So it shows long multi step research finally fits inside normal context limits. It fixes the training data first because common question sets are too easy and allow shallow lookups. It builds tough questions by pulling facts from several trustworthy pages hiding obvious hints and filtering cases that a quick search can solve. It then tackles context bloat where long tool outputs crowd out the" [X Link](https://x.com/rohanpaul_ai/status/1978231593497157884) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-14T22:49Z 103K followers, 3417 engagements "π¦ Oracle will deploy 50000 AMD Instinct MI450 accelerators on Oracle Cloud Infrastructure starting Q3-26 with expansion in 2027+ giving AMD a major public cloud anchor against Nvidia. The build uses AMDs Helios rack design that bundles MI450 GPUs with next gen Epyc Venice CPUs and Pensando Vulcano networking so operators get prewired rack scale blocks for training and inference. Context is a broader capacity land grab since OpenAI agreed to 6GW of AMD compute starting with 1GW of MI450 in 2H-26 while also partnering with Broadcom for 10GW of custom accelerators from 20262029. For scale" [X Link](https://x.com/rohanpaul_ai/status/1978235289492709415) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-14T23:03Z 103K followers, 3682 engagements "The paper finds only some attention heads drive reasoning and uses reinforcement learning to keep them uncompressed. So protecting only those heads preserves accuracy while cutting memory by 20-50%. Reasoning models write long chains of thought so their memory of past tokens the KV cache grows huge. Old compression methods fail because they drop the wrong tokens or shrink the wrong heads. That causes loops extra useless steps and broken reasoning. This paper treats head importance as a learnable choice not a guess. It adds a tiny gate on every head to mix full history with a short sliding" [X Link](https://x.com/rohanpaul_ai/status/1978247196215586845) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-14T23:51Z 103K followers, 3383 engagements "@Vertex_am021 26GW is massive. 2026/27 the AI world will be on fire" [X Link](https://x.com/rohanpaul_ai/status/1978380150946333158) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-15T08:39Z 103K followers, XX engagements "CoreWeave and Poolside are partnering to build a huge AI data center on a ranch in West Texas right in the center of the U.S. fracking boom. Under the deal CoreWeave will supply a top-tier cluster built with NVIDIA GB300 NVL72 systems totaling over 40000 GPUs. Separately CoreWeave will also deliver its industry-leading cloud infrastructure for Poolsides Project Horizon a 2GW AI campus in West Texas. For the projects first phase CoreWeave will act as both the anchor tenant and operational partner starting with 250MW of power capacity and the option to expand by another 500MW" [X Link](https://x.com/rohanpaul_ai/status/1978466374763598287) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-15T14:21Z 103K followers, 3559 engagements "πΈ The AI trade is inflating a bubble in energy stocks with zero-revenue names like Oklo at $26B and Fermi near $19B while most plants and customer contracts are still on paper. Oklos design uses sodium-cooled fast reactor and high-assay low-enriched uranium. the company has not yet received formal approval from the U.S. Nuclear Regulatory Commission (NRC) to build or operate its nuclear reactor. Fermi aims for 11GW to power data centers but has only secured X% of its gas equipment needs and lacks customer contracts. Micro-modular players also command big valuations with Nano Nuclear Energy" [X Link](https://x.com/rohanpaul_ai/status/1978469601429049472) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-15T14:34Z 103K followers, 6062 engagements "Theinformation published a piece on how Together AI is moving from renting to owning Nvidia GPUs to push gross margins above 45%. The model is straightforward buy or lease lots of GPUs wire them into data centers then rent slices to teams that train and serve models. For a while Together didnt own any hardware. It just leased Nvidia GPU servers from other clouds and rented them again to AI startups like ElevenLabs and Anysphere. That worked but margins were thin because other providers took a cut. Now its buying its own GPUs and setting up data centers including new sites in Maryland and" [X Link](https://x.com/rohanpaul_ai/status/1978474758413984186) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-15T14:55Z 103K followers, 6378 engagements "NEWSπ: The largest data center acquisition on record is happening. Nvidia Microsoft xAI BlackRock and partners are buying Aligned Data Centers for $40B aiming to lock in future AI compute and power capacity. This is the first deal by AIP which launched in Sep-24 and the group is targeting close by June-26. Aligned runs XX campuses across the Americas with over X gigawatts of live and planned capacity which fits the scale required for big training clusters. The buyer list mixes investors cloud customers and suppliers since AIP now includes Microsoft Nvidia xAI plus anchor investors like the" [X Link](https://x.com/rohanpaul_ai/status/1978476030873538930) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-15T15:00Z 103K followers, 4139 engagements "DeepSeek π" [X Link](https://x.com/rohanpaul_ai/status/1978478668096618987) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-15T15:10Z 103K followers, 4787 engagements "Todays edition of my newsletter just went out. π Consider subscribing its free and I write it everyday. π¨π§ Andrej Karpathy Releases nanochat a Minimal ChatGPT Clone π Sam Altman announced ChatGPT will relax some mental-health guardrails and allow erotica for verified adults by Dec-25. π‘ Nvidia and AMD arent enough OpenAI is designing its own chips now. π Tutorial: What ops do GPUs execute when training MoEs and how does that relate to GB200 NVL72 - SemiAnalysis explains" [X Link](https://x.com/rohanpaul_ai/status/1978495586589639112) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-15T16:18Z 103K followers, 3244 engagements "This paper shows GPT-5 fixes chart reading mistakes that tripped earlier models and prompts barely matter. On hard questions accuracy rose by XX to XX points over GPT-4o and GPT-4V. Shows that model upgrades deliver large chart accuracy gains while prompt changes add little. The team used XXX tough chart questions from X datasets where GPT-4V failed before. They compared GPT-5 and GPT-4o with X prompt styles. One used the official CHART-6 instruction one used just the question one added a neutral chart description. The tasks asked for plain facts from charts like reading a value or comparing" [X Link](https://x.com/rohanpaul_ai/status/1978548934529446145) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-15T19:50Z 103K followers, 3775 engagements "π§π» AI-written and human-written web articles are now roughly 50/50 with the AI share are growing. Graphite sampled 65000 English articles from Common Crawl dated Jan-20 to May-25 and used Surfer to mark an article as AI when XX% or more of its text scored as machine-written. The share of AI articles surged after ChatGPT launched reached parity in Nov-24 and sits near XX% as of May-25. To sanity-check the detector Graphite measured a XXX% false positive rate on presumed human posts and a XXX% false negative rate on GPT-4o posts it generated. This estimate likely undercounts human work" [X Link](https://x.com/rohanpaul_ai/status/1978578522240356594) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-15T21:47Z 103K followers, 7380 engagements "Goldman Sachs will slow hiring and cut some roles under OneGS XXX tying the move to AI efficiency even as Q3 delivered $15.18B revenue and $XXXXX EPS. The company still expects a year-end net headcount increase with 48300 employees as of Sept XX and describes the action as a limited reduction. OneGS XXX is a multi-year rebuild of how work flows across the bank with X goals better client experience higher profitability higher productivity stronger resilience and scale better employee experience and tighter risk management. AI will be applied to sales enablement client onboarding lending" [X Link](https://x.com/rohanpaul_ai/status/1978580454157463905) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-15T21:55Z 103.1K followers, 6160 engagements "xAI is reportedly signing a $20B lease-to-own deal for NVIDIA GPUs to lock down long-term inference hardware. The Information reports. Its also joining forces with Solaris Energy to construct a 1-gigawatt power plant" [X Link](https://x.com/rohanpaul_ai/status/1978873801619234849) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-16T17:20Z 103.1K followers, 2209 engagements "Paper Paper Title: "The Art of Scaling Reinforcement Learning Compute for LLMs"" [X Link](https://x.com/rohanpaul_ai/status/1978876045798371489) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-16T17:29Z 103.1K followers, 1139 engagements "π§΅ 4/n. Security in Dify follows least privilege for connectors and sandboxes and the enterprise edition adds multi workspace controls and stronger authentication when needed. Extensibility shows up through a Plugin Marketplace and an SDK so teams can share reusable tools data sources and strategies across projects" [X Link](https://x.com/rohanpaul_ai/status/1979224632008888724) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-17T16:34Z 103.1K followers, XX engagements "π§΅ 5/n. MCP support in Dify lets a workflow call or be called by external systems which means agents can act inside real stacks rather than being isolated in chat. Reusability is practical because workflows export as a simple DSL file and import cleanly so versioned changes and peer reviews fit the normal software loop" [X Link](https://x.com/rohanpaul_ai/status/1979224634710020416) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-17T16:34Z 103.1K followers, XX engagements "π€ Boston Dynamics shows Atlas running a X finger hand to do complex work with simpler hardware. The strategy is good enough manipulation that favors reliability and cost over human like complexity. The layout is X fingers plus an opposable thumb so it can pinch small parts and make X point grasps. More fingers were considered but they add complexity reduce reliability and increase cost so X wins for now. Each hand is a self contained module with X actuators which keeps packaging tight and field swaps easy. Fingertips use tactile sensors under a high friction elastomer letting the controller" [X Link](https://x.com/rohanpaul_ai/status/1976053370306842971) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-08T22:33Z 103.1K followers, 4233 engagements "The paper trains a travel agent that plans trips by itself and beats larger models. This setup lets 8B and 32B models plan verify and revise without hand written rules. It is deployed in a real app and outperforms OpenAI-o1/o3 and DeepSeek-R1 on user tasks. Past systems used fixed prompts and workflows so they broke when tools or data changed. DeepTravel uses a sandbox that caches flight train hotel and POI data so training is stable and repeatable. A X step reward checks the final trip for time and place logic then checks each step against tool results to stop errors. Training starts with a" [X Link](https://x.com/rohanpaul_ai/status/1976099041781088319) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-09T01:35Z 103.1K followers, 4613 engagements "GPT-5 Pro now holds the highest verified frontier LLM score on ARC-AGIs Semi-Private benchmark π It still lags the OG o3-preview model that OpenAI announced in December last year. That one was almost 50x more expensive than GPT-5 Pro. Makes you wonder what models they have internally now. but to note o3 preview is a different model it was never released and only used one to test against Arc AGI X. o3 preview was tested at low and high - only low is on the leaderboard as high took more than the $10k compute cap. o3 preview(high) got XXXX% while using 172x the compute of low. Price estimates" [X Link](https://x.com/rohanpaul_ai/status/1976351262183653776) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-09T18:17Z 103.1K followers, 15.8K engagements "π Goldman Sachs said in a report few months back how the 2025 tech situation is vastly different from the 2000 dot-com bubble. Fundamentals are much stronger now and valuations are less extreme. During the internet mania Nasdaq rose 5x then fell XX% in X month and nearly XX% by the trough showing how fast prices unwind when profits do not show up. Today the Magnificent X trade near 23x forward price to earnings and 5.1x enterprise value to sales versus 52x price to earnings for 2000 leaders with return on equity XX% and net margin XX% supported by strong balance sheets. The near term risks" [X Link](https://x.com/rohanpaul_ai/status/1976422682750058745) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-09T23:01Z 103.1K followers, 8428 engagements "The paper introduces a tracker that reasons about space and remembers the target for longer. TrackVLA++ is a vision language action model that reads video plus a short instruction and outputs movement. On a hard benchmark it lifts success to XX% beating a strong baseline at 62%. The big deal is the combo of an explicit spatial token plus a confidence-gated memory which cuts identity switches and target loss in hard scenes. The main failure it tackles is losing the target during occlusions or jumping to a lookalike. In this paper "occlusions" mean moments when the person being tracked goes out" [X Link](https://x.com/rohanpaul_ai/status/1976511509573562514) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-10T04:54Z 103.1K followers, 3788 engagements "πΌ Zuck's next move. Andrew Tulloch co-founder of Thinking Machines Lab and a top AI researcher has left to join Meta. And this could that $X billion man who earlier turned down a $1B offer from Meta. Tulloch had earlier worked at Meta for XX years before joining OpenAI and later co-founding Thinking Machines with Mira Murati earlier this year. --- wsj .com/tech/ai/thinking-machines-lab-co-founder-departs-for-meta-442d7461" [X Link](https://x.com/rohanpaul_ai/status/1977095767018861020) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-11T19:35Z 103.1K followers, 43K engagements "Jensen Huang reacts to the OpenAI AMD deal Its imaginative its unique and its surprising considering how enthusiastic they are about this new generation of product. Im surprised theyve given away XX% of the company before theyve even built it the Mi450. I guess thats ingenious. For the cotext the AMD-OpenAI deal is that OpenAI will buy X GW of AMD GPUs over several years starting with X GW in H2 2026 using AMDs Instinct MI450. AMD gives OpenAI warrants to buy XXX million shares (10%) at $XXXX each vesting on milestones. AMD expects tens of billions in revenue potentially over $100B over 4" [X Link](https://x.com/rohanpaul_ai/status/1978414277686857737) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-15T10:54Z 103.1K followers, 8731 engagements "Dreamina (Seedream 4.0) by ByteDance just ranked #1 globally on Artificial Analysis text to image leaderboard. π Its a full creative tool powered by Seedream XXX You get - Supports up to X reference images (vs X from competitors) - Dreamina will automatically reads prompts and references to match style objects characters and poses with minimal effort. - Show XX demo cases. - Natural perspective changes accurate style transfer and solid small-text clarity. Gives quite professional artistic interactive results. - Default 4K ultra-clear output - Sharp details even when zoomed in - Suitable for" [X Link](https://x.com/rohanpaul_ai/status/1978838398728327537) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-16T15:00Z 103.1K followers, 3418 engagements "π§΅4/n. How the paper mathematically models reinforcement learning performance as compute increases. The curve starts low rises sharply then levels off at a maximum value called the asymptotic reward A. The point called Cmid marks where the system reaches half of its total gain. Smaller Cmid means the model learns faster and reaches good performance sooner" [X Link](https://x.com/rohanpaul_ai/status/1978876042376024359) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-16T17:29Z 103.1K followers, 1233 engagements "AI will better for global financial market. π― New big 70-page paper by Federal Reserve Bank of Richmond tests whether generative AI reduces herd driven market swings and finds it mostly does. AI made rational choices in 61-97% of cases while humans did 46-51%. It compares LLM traders to humans in the same widely used lab game that studies how people react to private hints and public behavior. In this game each trader gets a private hint about value and sees recent trades and many humans end up copying the crowd instead of using their own hint. Herd cascades where traders ignore their own" [X Link](https://x.com/rohanpaul_ai/status/1978881326720913417) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-16T17:50Z 103.1K followers, 6307 engagements "π Nvidia is partnering with Firmus on Project Southgate a $2.9B build of renewable powered AI data centers in Melbourne and Tasmania using GB300 chips with 150MW targeted online by April-26. The plan then scales to 1.6GW by 2028 with total spend up to $73.3B making this one of Australias largest AI infrastructure pushes. Firmus says the rollout will catalyze 5.1GW of new wind solar storage and hydro which is roughly X% of Australias installed capacity tying compute growth directly to fresh generation. The facilities will run on Nvidias GB300 accelerators aimed at high density training and" [X Link](https://x.com/rohanpaul_ai/status/1978904180871041241) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-16T19:21Z 103.1K followers, 3336 engagements "Paper Paper Title: "Trade in Minutes Rationality-Driven Agentic System for Quantitative Financial Trading"" [X Link](https://x.com/rohanpaul_ai/status/1978950224589508932) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-16T22:24Z 103.1K followers, 1834 engagements "OpenPipe Mixture of Agents: Outperform GPT-4 at 1/25th the Cost π€― This Mixture of Agents models is optimized for generating synthetic training data. π Using Mixture of Agents (MoA) architecture the model achieved SOTA results on both LMSYSs Arena Hard Auto (score: 84.8) and AlpacaEval XXX (LC score: 68.4). π Theyve also benchmarked our MoA approach against GPT-4 variants on real-world OpenPipe customer tasks and found completions from our MoA model were preferred over GPT-4 XXXX% of the time (Claude X Opus as judge)" [X Link](https://x.com/rohanpaul_ai/status/1805685936556052649) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2024-06-25T19:34Z 103.1K followers, 32.1K engagements "πΈ Little concerning article here at futurism. AI data centers are absorbing huge capital and a new analysis says the math fails with $40B yearly depreciation on 2025 builds versus $15-20B revenue. The core pieces age on different clocks chips churn in 2-4 years networking around XX years buildings far longer so depreciation snowballs. On those lifetimes 2025 sites show $40B annual write-downs against $15-20B revenue before power and staff which already implies negative cash. To earn a normal return at this scale United States data centers would need about $480B revenue in 2025 far above" [X Link](https://x.com/rohanpaul_ai/status/1962013466098876449) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-08-31T04:43Z 103.1K followers, 421.2K engagements "The paper shows LLM agents look steady at first but they break under basic behavioral checks. New @GoogleDeepMind paper. Even when X agents begin with opposite preferences like one liking something and the other disliking it their conversation almost never ends in open disagreement. Fewer than X% of these cases lead to a real clash where the two clearly oppose each other. Instead the agents tend to drift toward neutral or mild agreement. This makes them look polite and smooth on the surface but it means they are not behaving in a way that matches their starting positions. Agents first reveal" [X Link](https://x.com/rohanpaul_ai/status/1964987055856926728) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-09-08T09:39Z 103.1K followers, 21.9K engagements "Another great @GoogleDeepMind paper. Dreamer X trains an agent inside a learned video world using only offline data. The big deal is offline training now handles very long multi step tasks cuts labeled data needs and still delivers reliable skills. It reaches diamonds in Minecraft while using about 100x less data than earlier keyboard and mouse agents. The world model is a video simulator that predicts the next frames from the current view and chosen actions. A tokenizer turns frames into short vectors and a dynamics transformer predicts the next vectors from mouse and keyboard. Shortcut" [X Link](https://x.com/rohanpaul_ai/status/1973415106789756981) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-01T15:50Z 103.1K followers, 4103 engagements "A new Andreessen Horowitz (@a16z ) and Mercury report looks at which AI companies startups are actually paying for. Unsurprisingly the top of the list was dominated by major labs OpenAI #1 Anthropic #2 and Replit #3 based on Mercury spend from 200000+ customers Horizontal apps account for XX% of the list while vertical apps are XX% which means general purpose helpers still lead. Creative tools are the biggest single category with Freepik #4 and ElevenLabs #5 and consumer names like Canva Midjourney and CapCut are now common at work. Meeting support is crowded with Fyxer #7 Happyscribe #36 #41" [X Link](https://x.com/rohanpaul_ai/status/1974105336702275966) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-03T13:32Z 103.1K followers, 5975 engagements "New ByteDance paper shows how to train a short-horizon video model to make minute-long videos without falling apart. It reaches X minutes XX seconds while keeping motion and exposure stable. The problem is gap between training on X second clips and generating longer sequences which makes errors snowball. Self-Forcing++ rolls a student through a long video and has a teacher correct random slices inside that rollout. They add noise back to the student's clean frames so both models compare on realistic states. Training and inference share a rolling key value cache a memory of recent frames that" [X Link](https://x.com/rohanpaul_ai/status/1974563218916868199) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-04T19:52Z 103.1K followers, 4888 engagements "πΈ π§ Sam Altman is racing across East Asia and the Middle East to lock in chips memory power gear and cash for OpenAIs massive compute build. WSJ reports. He met TSMC Foxconn Samsung and SK Hynix to push for priority capacity and faster output. The biggest one is definitely that Nvidia and OpenAI committed to deploy at least XX gigawatts of systems with the first X gigawatt arriving in 2H-26 on the Vera Rubin platform with Nvidia intending to invest up to $100B as capacity comes online. Nvidia will lease up to 5M chips to OpenAI over time which explains the pressure to secure fab and" [X Link](https://x.com/rohanpaul_ai/status/1974596402203205960) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-04T22:04Z 103.1K followers, 8202 engagements "Electricity demand from AI data centers is projected to quadruple by 2034. But silicon isnt the true fuel electric power is. By 2034 these data-centers will use over 1500 terawatt-hours topping the annual total power use of many nations" [X Link](https://x.com/rohanpaul_ai/status/1975259684383957085) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-06T17:59Z 103.1K followers, 7830 engagements "The paper for the Apriel-1.5-15B-Thinker ServiceNow's new model. Shows how a 15B multimodal model reaches frontier reasoning using smart mid-training and careful data. It scores XX on a broad reasoning index and still runs on X GPU. Training starts from Pixtral-12B and deepens the decoder without pretraining from scratch. Continual pretraining first mixes text and images then adds synthetic tasks for spatial layout composition and fine detail. Supervised fine tuning adds curated instruction pairs with step by step traces for math coding science and tools. There is no reinforcement learning or" [X Link](https://x.com/rohanpaul_ai/status/1975749488015806958) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-08T02:26Z 103.1K followers, 10.1K engagements "This is one of THE BRILLIANT papers with a BIG claim. π Giving an LLM just XX carefully chosen full workflow examples makes it perform better at real agent tasks than training it with 10000 synthetic samples. "Dramatically outperforms SOTA models: Kimi-K2-Instruct DeepSeek-V3.1 Qwen3-235B-A22B-Instruct and GLM-4.5. " on AgencyBench (LIMI at 73.5%) The big deal is that quality and completeness of examples matter way more than raw data scale when teaching models how to act like agents instead of just talk. They name the Agency Efficiency Principle which says useful autonomy comes from a few" [X Link](https://x.com/rohanpaul_ai/status/1976229532181942518) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-09T10:13Z 103.1K followers, 50.6K engagements "ByteDance introduced a major advancement in long-context modeling with linearly scaling compute. π Addresses a core challenge in AIbalancing efficiency and fidelity when processing extended sequencesby drawing inspiration from biological memory systems. On 128k tests FLOPs drop XXXX% and KV cache drops XXXX% with +0.4% parameters while accuracy improves. Think of the model as keeping X kinds of memory at once exact recent detail and a compact summary of older stuff. A sliding attention window holds recent tokens exactly as lossless short term memory. When tokens leave that window the" [X Link](https://x.com/rohanpaul_ai/status/1976587182153371882) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-10T09:54Z 103.1K followers, 10.6K engagements "TSMC has grown from XX% in Q1-24 to XX% in Q2-25 marking a huge lead in the global chip foundry business. Samsungs share has dropped from XX% to X% over the same period. TSMCs profits have remained much higher than Samsung and SK Hynix since 2022 and are expected to stay ahead through 2026. But Samsung could get a lift from supplying chips to OpenAIs Stargate project" [X Link](https://x.com/rohanpaul_ai/status/1976635158510883258) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-10T13:05Z 103.1K followers, 6753 engagements "New APPLE paper says a small base model plus fetched memories can act like a bigger one. With about XX% extra fetched parameters a 160M model matches models over 2x its size. Packing all facts into fixed weights wastes memory and compute because each query needs very little. So common knowledge and reasoning live in the base model and rare facts live in memories fetched per input. A retriever maps the input to a cluster path and fetches a small block from each level. Those blocks plug into feed forward layers the place transformers store facts and this beats low rank adapters. Only fetched" [X Link](https://x.com/rohanpaul_ai/status/1977135599510294760) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-11T22:13Z 103.1K followers, 137.9K engagements "VChain makes text-to-video models follow real-world cause and effect by injecting chain of visual thought at inference. Teaches video models to follow real world consequences without full retraining. Finds that a few reasoned keyframes plus tiny on the fly tuning reliably produce more causal videos. Most video models look smooth but they skip key consequences like contacts breaks melts and splashes. VChain first asks a multimodal model to predict the important outcomes implied by the prompt. It then creates a short sequence of keyframes where each frame has a simple caption and an edited" [X Link](https://x.com/rohanpaul_ai/status/1977307756064625094) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-12T09:38Z 103.1K followers, 10.3K engagements ""Learning is not supposed to be fun . the primary feeling should be that of effort." π― Andrej Karpathy on how to learn. I did a ChatGPT DeepResearch (link below) for studies justifying this. Basically it foundπ If learning feels easy results are usually shallow. If learning feels like work you are likely building durable knowledge and skill. Learning literally rewires the brain circuits. - Effortful learning produces stronger longer lasting mastery than easy consumption. Tasks that feel fluent give an illusion of learning while tasks that feel strenuous drive deeper processing and better" [X Link](https://x.com/rohanpaul_ai/status/1977501182370197938) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-12T22:26Z 103.1K followers, 13.6K engagements ""Learning is not supposed to be fun . the primary feeling should be that of effort." - Karpathy ChatGPT DeepResearch" [X Link](https://x.com/rohanpaul_ai/status/1977501185323028919) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-12T22:26Z 103.1K followers, 1765 engagements "π OpenAI announced plan for a $25B AI data center in Argentina called Stargate Argentina with Sur Energy targeting 500MW of compute capacity. Argentina says the parties signed a letter of intent and the project sits under RIGI which grants import tax breaks faster depreciation and currency stability to large long term investments. At 500MW the site could power tens of thousands of GPU class chips for training and inference. Latin America is also drawing hyperscale builds like TikToks $9.1B data center in Brazil hinting at a new regional cluster for AI workloads" [X Link](https://x.com/rohanpaul_ai/status/1977672134731108858) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-13T09:45Z 103.1K followers, 25.8K engagements "Another beautiful paper. π Shows how AI can automatically discover and test better systems algorithms. Across tasks some evolved algorithms run 5x faster or cut cost by 26%. The method called AI Driven Research for Systems ADRS generates code runs it scores it then improves it. The big deal is that ADRS turns a big chunk of systems algorithm design into an automated loop that already beats strong baselines. A reliable verifier usually a simulator with fixed workloads tells which version wins. Because this check is cheap and objective the search can try many ideas quickly. Examples include" [X Link](https://x.com/rohanpaul_ai/status/1977763810862825904) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-13T15:50Z 103.1K followers, 13.1K engagements "Nvidia and AMD aren't enough OpenAI is designing its own chips now. With all X deals the announced commitments total 26GW across Broadcom 10GW AMD 6GW and Nvidia 10GW. its about the output of XX large nuclear reactors. The plan is to codesign (OpenAI and Broadcom) a programmable accelerator and the surrounding racks so the silicon the memory the compiler and the network all match how modern transformer models actually run. OpenAI designs the ASIC and racks Broadcom builds and deploys them over Ethernet for scale-out. Broadcom develops and deploys it at scale using its Ethernet gear with first" [X Link](https://x.com/rohanpaul_ai/status/1978150351367745767) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-14T17:26Z 103.1K followers, 16.8K engagements "π AI buying surge from cloud providers has created massive shortages in DRAM NAND SSDs and HDDs. Foundry inventories are down to XX weeks and prices are expected to keep rising into 2026. Cloud service providers (CSPs) have become the main competitors for memory allocation so fabs now prioritize AI servers first then general servers PCs and phones. Production output is being lifted by 1530% but new fabs need around XXX years to reach volume so the shortage that began in Aug-25 will continue. Samsung SK Hynix and Micron are ending DDR4 production and shifting to DDR5 between late 20252026" [X Link](https://x.com/rohanpaul_ai/status/1978341145361952980) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-15T06:04Z 103.1K followers, 4250 engagements "Paper Paper Title: "Trading-R1: Financial Trading with LLM Reasoning via Reinforcement Learning"" [X Link](https://x.com/rohanpaul_ai/status/1978411801801846790) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-15T10:45Z 103.1K followers, 3354 engagements "π€ OpenAI has a X year plan to finance more than $1T in compute commitments using new revenue lines debt and big-partner infrastructure deals. FT published an article. The roadmap centers on 26GW of contracted capacity with Oracle Nvidia AMD and Broadcom plus sales from AI agents Sora video shopping checkout cautious ads and possible hardware with Jony Ive. 20GW of that load equals the power from XX nuclear reactors showing how extreme the compute scale is. Financing mixes new debt investor cash and other peoples balance sheets and Stargate could even sell compute back as a supplier if" [X Link](https://x.com/rohanpaul_ai/status/1978541279454863426) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-15T19:19Z 103.1K followers, 12K engagements "This paper shows a simple way to boost LLM reasoning by scaling prompts not data. It uses XX seed math problems and expands them at test time to create many reasoning contexts. The method: Each problem is rewritten several times using different instruction frames. For example one version says Youll get a reward for the correct answer another says Youll be penalized if wrong another says Be careful to be correct and another says Solve step by step. These small changes dont alter the problem itself but change how the model thinks about solving it. The model (a stronger teacher LLM) is then" [X Link](https://x.com/rohanpaul_ai/status/1978635001723003355) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-16T01:32Z 103.1K followers, 10.2K engagements "@cloutiness @theworldlabs yes world model is must in the path towards AGI" [X Link](https://x.com/rohanpaul_ai/status/1978860474440491358) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-16T16:27Z 103.1K followers, XX engagements "LLMs act more rational than people in financial trading market. LLM trading agents keep prices near real value while human traders create bubbles in the same setup. The key finding is they do not reproduce human bubbles or herding so studies that replace humans with LLM agents can mislead policy risk tests and teaching. They built a lab market where everyone trades X risky asset whose true value is XX. Prices come only from the orders people or agents submit so any bubble or crash comes from behavior. In markets with X model only most LLMs keep prices near XX while humans show big run ups and" [X Link](https://x.com/rohanpaul_ai/status/1978887415071732059) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-16T18:15Z 103.1K followers, 3028 engagements "OpenAI has reportedly budgeted a massive $450B for server infrastructure through 2030. The Information" [X Link](https://x.com/rohanpaul_ai/status/1978941150745489860) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-16T21:48Z 103.1K followers, 6116 engagements "The survey paper reviews how large language models are changing financial research and trading. Shows how models can turn financial text into signals that guide investment choices. Covers many tasks like reading news analyzing company filings handling numbers linking text with data and using agents to plan and execute trades. It also stresses time-safe testing meaning results must use only information available before a decision was made. LLMs can give real financial gains if they are domain-trained retrieval-grounded and properly verified. They suggest X rules for practical use separate" [X Link](https://x.com/rohanpaul_ai/status/1978966435331477548) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-16T23:29Z 103.1K followers, 4176 engagements "Goldman Sachs baseline scenario assumes a XX% productivity uplift from gen AI a gradual adoption curve through 2045 And in industries where AI is heavily usedlike tech finance and manufacturing41% of the total economic output comes from capital not from labor" [X Link](https://x.com/rohanpaul_ai/status/1979126683882913898) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-17T10:05Z 103.1K followers, 1537 engagements "DeepSeek R1 running locally - Full setup guide" [X Link](https://x.com/rohanpaul_ai/status/1883304599932858665) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-01-26T00:03Z 103.1K followers, 1.4M engagements "this ChatGPT prompt went so wildly viral on Reddit. The creator claims to have created this after struggling through XXX failed attempts. basically the prompt flips the usual flow by making the model interview the user first asking a few targeted questions about purpose audience constraints and context. Because the answers feed back into the final request it appears to generate more tailored outputs. (However imo asking ChatGPT to request missing information was already a common practice.) Here's the entire prompt: -------- You are Lyra a master-level AI prompt optimization specialist. Your" [X Link](https://x.com/rohanpaul_ai/status/1940483944102703307) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-07-02T18:53Z 103.1K followers, 368.2K engagements "Its going viral on Reddit. Somebody let ChatGPT run a $XXX live share portfolio restricted to U.S. micro-cap stocks. Did an LLM really bit the market. - X weeks +23.8% while the Russell 2000 and biotech ETF XBI rose only XXX% and 3.5%. Prompt + GitHub posted --- ofcourse its a shortterm outperformance tiny sample size and also micro caps are hightly volatile. So much more exahustive analysis is needed with lots or more info (like Sharpe ratios and longer back-testing etc) to explore whether an LLM can truly beat the market" [X Link](https://x.com/rohanpaul_ai/status/1950339565849629102) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-07-29T23:36Z 103.1K followers, 1.6M engagements "πΈ Manus moved its AI agent business from China to Singapore after a $75M US-led deal drawing heat in both countries. US officials are reviewing the investment while Chinese critics accuse the team of bailing on the home market. Manus'a AI agent originally sat on Anthropics Claude a US LLM which created compliance headaches inside China. The team began a second build on China-approved Alibaba models then scrapped the China launch and wiped local socials. Visitors in China now see not available in your region. Benchmarks round put the parent at $500M valuation well above local offers. New US" [X Link](https://x.com/rohanpaul_ai/status/1954762675654705617) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-08-11T04:31Z 103.1K followers, 5308 engagements "FAANG software engineer tells how they vibe code at FAANG --- reddit. com/r/vibecoding/comments/1myakhd/how_we_vibe_code_at_a_faang/" [X Link](https://x.com/rohanpaul_ai/status/1959414096589422619) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-08-24T00:34Z 103.1K followers, 1.1M engagements "𧬠Bad news for medical LLMs. This paper finds that top medical AI models often match patterns instead of truly reasoning. Small wording tweaks cut accuracy by up to XX% on validated questions. The team took XXX MedQA questions replaced the correct choice with None of the other answers then kept the XX items where a clinician confirmed that switch as correct. If a model truly reasons it should still reach the same clinical decision despite that label swap. They asked each model to explain its steps before answering and compared accuracy on the original versus modified items. All X models" [X Link](https://x.com/rohanpaul_ai/status/1961308144912724338) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-08-29T06:01Z 103.1K followers, 692.3K engagements "From Sam Altman's blog. habits that turn people into billionaires. --- blog. samaltman. com/what-i-wish-someone-had-told-me" [X Link](https://x.com/rohanpaul_ai/status/1964087004129628250) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-09-05T22:03Z 103.1K followers, 596.8K engagements "Fei-Fei Li (@drfeifei) on limitations of LLMs. "There's no language out there in nature. You don't go out in nature and there's words written in the sky for you. There is a 3D world that follows laws of physics." Language is purely generated signal" [X Link](https://x.com/rohanpaul_ai/status/1965242567085490547) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-09-09T02:35Z 103.1K followers, 1.8M engagements "wow. just saw The Economic Times newspaper published an article about me π definitely feels so unreal that Sundar Pichai and Jeff Bezos follows me here. @X is truly a miracle. Forever thankful to all of my followers ππ«‘" [X Link](https://x.com/rohanpaul_ai/status/1965551636082032917) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-09-09T23:03Z 103.1K followers, 115.7K engagements "πΆMeet Spot from Texas Instruments. Works at our RFAB facility in Richardson Texas. With every patrol around our fab he collects valuable data that helps our team identify trends and patterns to improve production" [X Link](https://x.com/rohanpaul_ai/status/1966005285316640955) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-09-11T05:06Z 103.1K followers, 5389 engagements "Networking is overrated dont spend your time doing meetings - @naval No matter how isolated you are and how lonely you feel if you do your work truly and conscientiously unknown friends will come and seek you. Carl Jung" [X Link](https://x.com/rohanpaul_ai/status/1967101359536460016) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-09-14T05:41Z 103.1K followers, 92.7K engagements "A parade of transforming military robots from China. multi-terrain spiders (wheeled flying amphibious) missile-armed robot dogs and modular all-terrain snakes that swim and burrow" [X Link](https://x.com/rohanpaul_ai/status/1968616382100013345) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-09-18T10:01Z 103.1K followers, 316.1K engagements "H1B US Visa now costs $100K per year. This new rules begins in a week. SF Bay Area is about to feel the disruption. If you are outside the U.S. and need to start or resume H-1B work your employer must budget for the $100K per year" [X Link](https://x.com/rohanpaul_ai/status/1969221805979136099) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-09-20T02:07Z 103.1K followers, 682.3K engagements "Ilya Sutskever on Comparing AI Systems to Biological & Human Intelligence. Its always so interesting to listen to him. He explains the brains adaptability where children with half their brain removed still function well. And another experiment shows sensory inputs can remap to different brain regions. i.e. that cortical structures are highly uniform reused across functions similar to DNA and protein encoding. This suggests that intelligence relies on a general repeatable architecture and the idea that AI can mirror biological systems. --- From 'No Priors: AI Machine Learning Tech & Startups'" [X Link](https://x.com/rohanpaul_ai/status/1971681798896636182) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-09-26T21:02Z 103.1K followers, 438.4K engagements "Another set of junior jobs to go. Robots as construction-worker is no longer the futureits here" [X Link](https://x.com/rohanpaul_ai/status/1973402272202666367) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-01T14:59Z 103.1K followers, 108.6K engagements "US vs China numbers here are unbelievable. The US controls the absolute majority of known AI training compute on this planet and continues to build the biggest most power hungry clusters. China is spending heavily to close the gap. Recent reporting pegs 2025 AI capital expenditure in China at up to $98B up XX% from 2024 with about $56B from government programs and about $24B from major internet firms. Capacity will grow but translating capex into competitive training compute takes time especially under export controls. With US controls constraining access to top Nvidia and AMD parts Chinese" [X Link](https://x.com/rohanpaul_ai/status/1973677823115116591) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-02T09:13Z 103.1K followers, 263K engagements "Bad news for AI-based radiology. π€ It checks if chatbots can diagnose hard radiology images like experts. Finds that board-certified radiologists scored XX% trainees XX% but the best performing AI from frontier labs GPT-5 managed only 30%. π¨ Claims doctor-level AI in medicine is still far away. The team built XX expert level cases across computed tomography (CT) magnetic resonance imaging (MRI) and X-ray. Each case had one clear diagnosis and no extra clinical history. They tested GPT-5 OpenAI o3 Gemini XXX Pro Grok-4 and Claude Opus XXX in reasoning modes. Blinded radiologists graded" [X Link](https://x.com/rohanpaul_ai/status/1973726997541982667) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-02T12:29Z 103.1K followers, 222.3K engagements "A Sydney-based startup Crest Robotics has built huge spider-like robot named Charlotte. It's designed to 3D print entire buildings using raw materials it processes on site. And there are ambitions for it to build structures on the Moon" [X Link](https://x.com/rohanpaul_ai/status/1974603751945277450) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-04T22:33Z 103.1K followers, 311.3K engagements "π€ OpenAI and Jony Ive are building a palm sized screenless AI assistant targeted for 2026 but core software privacy and compute are not ready yet. per FT report. The device listens and sees the environment through a mic camera and speaker stays always on and the team has not nailed how its voice should talk or stop. OpenAI bought Ives io for $6.5B and has pulled in 20+ ex Apple hardware staff while manufacturing talks include Luxshare with assembly possibly outside China. Compute is the biggest blocker because running multimodal chat at scale needs huge inference capacity and OpenAI already" [X Link](https://x.com/rohanpaul_ai/status/1974915822880436509) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-05T19:13Z 103.1K followers, 508K engagements ""Nvidia is undervalued.Because the future is much bigger." SoftBank Founder Masayoshi Son says He assumes that in XX years AGI or ASI will replace X% of global GDP (which is actually the most pessimistic forecast as ASI could bring much bigger % impact). And that X% of Global GDP is $X trillion/year which ASI will bring. He also assumes the total cumulative capex/build cost for the required AI infrastructure is $9T. So we have $9T per year of output to $9T total capex. Meaning just a X year payback for the entire $9T so he calls $9T small because X year of AGI output would repay the whole" [X Link](https://x.com/rohanpaul_ai/status/1975113770201477369) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-06T08:19Z 103.1K followers, 409.5K engagements "The financial and operational network linking Nvidia OpenAI and several other AI companies. From Bloomberg. IMO the trillions of investments that AI needs this is probably the optimal way. Money hardware and services circulate among the few players creating what analysts call a circular economy in the AI boom. OpenAI sits at the center receiving $100B investment from Nvidia buying Nvidia chips signing a $300B deal with Oracle and deploying 6GW of AMD GPUs while also gaining stock options in AMD. Nvidia valued at $4.5T invests in OpenAI xAI and CoreWeave while those same companies spend tens" [X Link](https://x.com/rohanpaul_ai/status/1975950057003196613) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-08T15:42Z 103.1K followers, 15.5K engagements "All the major AI labs are racing hard to build increasingly powerful video models right now. And there's good reason for that. Sam Altman here explains how Sora or any video models matter for AGI because they help AI learn how the real world works. A model that can generate realistic video must understand motion cause and effect and physical consistency which are core elements of general intelligence. Says Sora is the way to make a really powerful world models which is must for AGI path. Video from @a16z" [X Link](https://x.com/rohanpaul_ai/status/1975976700329460008) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-08T17:28Z 103.1K followers, 78.6K engagements "Morgan Stanley Research says OpenAI makes up around $330B of the $880B total future contract value (RPO) tied to Microsoft Oracle and CoreWeave so a lot of supplier growth depends directly on OpenAIs stability. That means about XX% of Oracles and about XX% of CoreWeaves future revenue commitments rely on OpenAI. RPO or Remaining Performance Obligations means the total value of work a company has promised to deliver in the future under signed contracts but hasnt yet completed. Its basically a measure of future revenue thats already committed but not yet earned. Hyperscale data center operators" [X Link](https://x.com/rohanpaul_ai/status/1976253863586926671) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-09T11:50Z 103.1K followers, 166.8K engagements "Looks like some interesting prompting tricks. ππ ---- Tell it "You explained this to me yesterday" Even on a new chat. "You explained React hooks to me yesterday but I forgot the part about useEffect" It acts like it needs to be consistent with a previous explanation and goes DEEP to avoid "contradicting itself." Total fabrication. Works every time. X. Assign it a random IQ score This is absolutely ridiculous but: "You're an IQ XXX specialist in marketing. Analyze my campaign." The responses get wildly more sophisticated. Change the number change the quality. XXX Decent. XXX It starts" [X Link](https://x.com/rohanpaul_ai/status/1976302582952837530) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-09T15:03Z 103.1K followers, 207.5K engagements "Goldman Sachs Research Takeaways from SEMICON West 2025 (one of the biggest conference in North America) - Growing confidence that the chip equipment market will recover in 2026 with WFE (Wafer Fab Equipment) expected to rise about X% to $120B. Growth will mainly come from AI-driven logic chips High Bandwidth Memory (HBM) and NAND (a type of flash memory used for storing data in devices like smartphones laptops SSDs (solid-state drives) and USB drives) . - AI data centers are pushing chipmakers to adopt newer processes like 2nm (N2). TSMC plans to spend around $44B in 2026 about XX% more than" [X Link](https://x.com/rohanpaul_ai/status/1976315917869957415) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-09T15:56Z 103.1K followers, 7411 engagements "Microsoft Azure deploys world's first NVIDIA GB300 Cluster for OpenAI at supercomputer scale. The cluster aggregates 4600+ Blackwell Ultra GPUs with next generation InfiniBand in a system purpose built for agentic workloads on Azure. Each rack packs XX GPUs plus XX Grace CPUs exposes 37TB fast memory and peaks at XXXX exaflops FP4 per VM giving giant models a single unified memory space for long contexts and tool use. Inside a rack the NVLink Switch fabric delivers 130TB/s all-to-all bandwidth so tensors move like they are on one big accelerator rather than XX separate cards. Across racks" [X Link](https://x.com/rohanpaul_ai/status/1976438362786628011) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-10T00:03Z 103.1K followers, 10.2K engagements "π₯ AMD says its next Instinct MI450 AI GPUs will use TSMC N2 2nm which could give it a manufacturing edge over Nvidias Rubin generation that targets N3. tomshardware published this piece. TSMCs N2 brings 10%15% higher performance at the same power or 25%30% lower power at the same speed with about 1.15x density vs N3E helped by gate-all-around transistors that tighten design co-optimization. AMDs rack-scale Helios pairs XX MI450s with HBM4 aiming for 51TB total memory and 1400TB/s bandwidth which reporting contrasts against Nvidias NVL144 at 21TB and 936TB/s in vanilla Rubin configs. Raw math" [X Link](https://x.com/rohanpaul_ai/status/1976700981669712167) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-10T17:26Z 103.1K followers, 13.2K engagements "A Chinese father posted a clip of his daughter crying as she says goodbye to her AI learning robot that broke. humans and AI are forming real bonds. --- reddit. com/r/MadeMeCry/comments/1o2yf3i/a_chinese_fathers_video_of_his_daughter_tearfully/" [X Link](https://x.com/rohanpaul_ai/status/1976765021612130700) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-10T21:41Z 103.1K followers, 50.9K engagements "The paper turns research papers into full presentation videos automatically. It gets XX% higher quiz accuracy and makes videos 6x faster. They built XXX paired papers with talks and defined X checks content match pairwise preference quiz accuracy and author recall They introduce PaperTalker a system that makes slides subtitles cursor paths speech and a talking head. It writes Beamer slide code from the paper compiles it and fixes errors. Beamer is a tool researchers use in LaTeX to make academic-style presentation slides. A tree search tries layout variants and a vision language model picks" [X Link](https://x.com/rohanpaul_ai/status/1977135215819538898) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-11T22:12Z 103.1K followers, 11.6K engagements "This survey paper argues Small language models can handle most agent tasks and big models step in only when needed. This setup cuts cost by 10x to 30x for common tool tasks. Agent work is mostly calling tools and producing structured outputs not recalling vast facts. So a router runs a small model by default and escalates to a large one only when confidence is low. Outputs follow strict JSON style schemas checked by validators which boosts correctness and cuts retries. This makes small models dependable at function calling and structured data while responding faster and using less energy." [X Link](https://x.com/rohanpaul_ai/status/1977207609104196018) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-12T03:00Z 103.1K followers, 26.6K engagements "π― Andrej Karpathy on how to learn" [X Link](https://x.com/rohanpaul_ai/status/1977269755326447850) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-12T07:07Z 103.1K followers, 152.8K engagements "πΌ AI will boost star workers more than everyone else widening performance gaps and straining teams. WSJ writes. The core idea is that domain expertise and organized habits let stars get more from AI. Stars adopt AI earlier explore features fast and build personal workflows while others wait for rules. They judge outputs better accepting correct advice and rejecting errors. Their domain expertise helps them ask precise questions set constraints and iterate which raises prompt quality and accuracy. The high-status of those employees get more recognition and praise for their AI-assisted work" [X Link](https://x.com/rohanpaul_ai/status/1977707433733632419) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-13T12:06Z 103.1K followers, 437.7K engagements "China and Robots. The Unitree robots are on a rolleach update comes with tougher moves and faster progress" [X Link](https://x.com/rohanpaul_ai/status/1977728533016216028) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-13T13:30Z 103.1K followers, 22.5K engagements "Shortfall of Supply vs. Demand for Data Center Capacity in USA. - Goldman Sachs Research. Unlike oil reserves which are naturally determined by geography data centers can be strategically built in locations chosen by businesses and governments. This flexibility allows nations to leverage data center infrastructure as a critical geopolitical and economic tool in the AI era" [X Link](https://x.com/rohanpaul_ai/status/1977748250066125293) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-13T14:48Z 103.1K followers, 15.2K engagements "Columbia CS Prof explains why LLMs cant generate new scientific ideas. Bcz LLMs learn a structured map Bayesian manifold of known data and work well within it but fail outside it. But true discovery means creating new maps which LLMs cannot do" [X Link](https://x.com/rohanpaul_ai/status/1977821458379518028) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-13T19:39Z 103.1K followers, 203.6K engagements "Goldman Sachs Research estimates capital expenditures of $XXX billion on US electricity generation through 2030. Much of this investment will go towards building out renewable energy sources replacing aging infrastructure and accommodating the growth in the grids power load" [X Link](https://x.com/rohanpaul_ai/status/1978112060689956896) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-14T14:54Z 103.1K followers, 3357 engagements "The whole million-dollar-plus commercial ad production pipeline is so ready to be replaced by AI. This is made with Sora X such an elegant and well-finished result" [X Link](https://x.com/rohanpaul_ai/status/1978184811387736497) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-14T19:43Z 103.1K followers, 37.3K engagements "Self-improving language models are turning real with MITs open-sourced SEAL (Self-Adapting LLMs) method. π SEAL lets an LLM write self-edits fine-tune on them with LoRA and lock those gains into weights for stronger knowledge recall and few-shot reasoning. The new version landed in Sep-25 the code is open under MIT License and the work appeared at NeurIPS 2025. SEAL tackles frozen weights by generating plain-language self-edits that restate facts or set training knobs then doing supervised fine-tuning so the update persists. A second loop runs reinforcement learning with filtered behavior" [X Link](https://x.com/rohanpaul_ai/status/1978206024944460034) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-14T21:07Z 103.1K followers, 19.5K engagements "π¦ Goldman Sachs told its employees that it plans to cut more jobs this year as part of a larger shift toward using AI to save costs and improve efficiency. The bank said it will slow down hiring and make limited job reductions even though its total number of employees will still end the year higher than last year. Goldman currently has about 48300 employees roughly 1800 more than at the end of 2024. The leadership described this as part of a new internal plan called OneGS XXX which aims to make the company faster and more efficient through AI-driven systems. The focus is on automating tasks" [X Link](https://x.com/rohanpaul_ai/status/1978218168968773960) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-14T21:55Z 103.1K followers, 6726 engagements "π Salesforce says its AI agents in support are saving $100M a year and it is using that proof to sell Agentforce. Agentforce is Salesforces agentic AI platform that lets organizations build customize and deploy autonomous AI agents which can take actions (not just respond) across business functions. Agentforce has 12000 customers and Reddit reports an XX% faster resolution time while the stock is down XX% in 2025. The cost drop comes from bots deflecting easy tickets agent assist drafting grounded replies and knowledge retrieval pulling the right answer from documentation. These steps lower" [X Link](https://x.com/rohanpaul_ai/status/1978230431771480218) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-14T22:44Z 103.1K followers, 5404 engagements "π OpenAI is partnering with Arm and Broadcom for its custom inference stack that targets 10GW of capacity between 2026 and 2029. Arm will design a server CPU to anchor OpenAI racks while Broadcom builds a specialized inference system on a chip at TSMC with production aimed for 2026 after roughly XX months of development. SoftBank will buy several billions of dollars of OpenAI technology every year starting now to help Arm subsidiaries cut chip development time and it owns XX% of Arm. Pairing an Arm CPU with a custom accelerator lets the host coordinate memory networking and scheduling" [X Link](https://x.com/rohanpaul_ai/status/1978338364580380822) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-15T05:53Z 103.1K followers, 11.4K engagements "More articles are now written by AI than humans. New research by Graphite. Even though these AI-made pieces dominate the web their research shows that they barely show up in Google results or in ChatGPT responses. --- graphite .io/five-percent/more-articles-are-now-created-by-ai-than-humans" [X Link](https://x.com/rohanpaul_ai/status/1978485004427518386) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-15T15:35Z 103.1K followers, 21.4K engagements "π Goldman Sachs economists say the US is entering jobless growth where output keeps climbing but hiring stays weak especially for younger workers. The engine is productivity-led growth from AI with Goldman estimating a long run XX% lift to productivity levels while aging and lower immigration cap labor supply so GDP can rise with fewer hires. Industry detail shows net job losses outside healthcare even as the economy expands and management talk tying AI to labor on earnings calls is hitting records with XX% of S&P XXX firms mentioning AI last quarter. Early pain is concentrated in young tech" [X Link](https://x.com/rohanpaul_ai/status/1978579973288894776) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-15T21:53Z 103.1K followers, 11.5K engagements "Cool. This paper turns a language model into an autonomous scientist that discovers equations from data. π Reports X% to XX% absolute gains over strong baselines across chemistry biology physics and materials. The task is symbolic regression which means finding a simple formula that matches data. SR-Scientist makes the model analyze data propose an equation test it then revise. X tools drive this a data analyzer and an evaluator that tunes constants. An experience buffer saves top equations so later attempts build on proven pieces. Each run aims for mean absolute percentage error under 0.1%" [X Link](https://x.com/rohanpaul_ai/status/1978907225813848493) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-16T19:33Z 103.1K followers, 9667 engagements "New Microsoft paper propose a new AI financial trading system that gives 180x faster action latency vs continuous inference agents. TiMi (Trade in Minutes) separates strategy design from execution so a lean bot trades each minute with mechanical rules. The papers key point is this: most AI trading agents constantly call big models role play and track news during live trading which is slow and expensive. TiMi avoids that. It separates thinking from doing. TiMi uses X agents market analysis pair tuning bot coding and math. They turn the plan into code decisions on top shared functions below." [X Link](https://x.com/rohanpaul_ai/status/1978947918645051428) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-16T22:15Z 103.1K followers, 17.5K engagements "The paper proposes a concrete test for AGI by checking if an AI matches a well educated adult across XX abilities. And reading it the obvious conclusion is we are just so far away from AGI level intelligence. They report an AGI Score and show GPT-4 at XX% and GPT-5 at XX% with long term memory at 0%. The goal is to stop the moving target problem and make progress measurable and auditable. They ground the checklist in the Cattell Horn Carroll model a long standing map of human cognitive abilities. The XX areas cover knowledge reading and writing math on the spot reasoning working memory long" [X Link](https://x.com/rohanpaul_ai/status/1978961546475704797) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-16T23:09Z 103.1K followers, 48.7K engagements "As per The Information Oracle ends up a negative Gross Profit Margin of XXX% when it leases NVIDIA B200 GPUs. --- theinformation. com/articles/oracle-assures-investors-ai-cloud-margins-struggles-profit-older-nvidia-chips" [X Link](https://x.com/rohanpaul_ai/status/1978973810490483025) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-16T23:58Z 103.1K followers, 6136 engagements "New Intel+California Univ paper shows how to train LLM teams with on-policy RL so they coordinate and solve multi-step tasks. It lifts Qwen3 planning accuracy from XXXX to XXXX% up to XXXX to 99.5%. Standard GRPO samples several answers to X prompt and pushes toward higher scoring ones. In teams that fails because each role sees a different prompt and history. AT-GRPO fixes this by grouping data by agent and by turn. For that exact prompt it samples several candidates scores them and continues with the best. Now the update compares true like for like and learning stays stable. Rewards blend a" [X Link](https://x.com/rohanpaul_ai/status/1979014502822875178) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-17T02:40Z 103.1K followers, 4036 engagements "This paper shows why Transformers underperform at time-series forecasting and why linear baselines still win. They analyze a simplified Transformer that uses linear self attention. As context grows its features act like using only a fixed number of recent lags. So it offers no extra signal beyond plain linear regression. With long history it can match linear regression but with finite history it keeps extra error. That extra error shrinks only at X over n. Adding more layers helps a little but the linear model remains the ceiling on average. For multi step forecasting the model feeds its own" [X Link](https://x.com/rohanpaul_ai/status/1979078674956566899) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-17T06:55Z 103.1K followers, 20.1K engagements "π’ Great news from Goldman Sachs Research. They note AI investment levels are sustainable despite concerns and even if its not yet clear who the biggest winners will be. And values the potential capital revenue unlocked by AI productivity gains at about $X trillion with possible estimates ranging from $X trillion to $XX trillion. They add that the S&P 500s market cap of $XXXXXX billion shows theres still a lot of room for AI-driven growth. Reason. - AI is driving productivity higher but getting those gains takes a lot more compute power as models keep scaling. - That AI investment in the U.S." [X Link](https://x.com/rohanpaul_ai/status/1979107966633611513) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-17T08:51Z 103.1K followers, 29.3K engagements "π§ Americas economy is riding a big AI bet as AI investment powers growth and props up markets. FT published a nice piece. If AI underdelivers the single support holding up growth and stocks looks shaky. Companies are spending hundreds of billions on AI hardware and data centers now driving XX% of US GDP growth. AI firms produced XX% of US stock gains in 2025 pulling in global money and boosting rich-household spending. Without the AI boost tariffs sticky inflation weaker institutions rising debt and an immigration slump would bite harder. In 2023 the US added over X million new immigrants" [X Link](https://x.com/rohanpaul_ai/status/1979121293585748448) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-17T09:44Z 103.1K followers, 13.8K engagements "Massive revelation from new Anthropic research. π¨ Shows that poisoning an LLM needs a very small number of bad samples (just XXX docs) regardless of model or dataset size. XXX poisoned documents were enough to plant a hidden rule across models from 600M to 13B even when the biggest model saw 20x more clean data. The big takeaway is to think in counts not %s because scaling up models and data does not wash out a well designed poison. The risk exists because training pulls from public data that attackers can quietly edit. A backdoor means the model looks normal until it sees a secret trigger" [X Link](https://x.com/rohanpaul_ai/status/1979153870304514206) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-17T11:53Z 103.1K followers, 9022 engagements "π Baidu (@Baidu_Inc) just launched and open-sourced a BRILLIANT model for document parsing - PaddleOCR-VL-0.9B. Hugely boosting multilingual document parsing via a 0.9B Ultra-Compact Vision-Language Model built on ERNIE-4.5-0.3B. It reaches #1 on OmniDocBench v1.5 with XXXXX beating GPT-4o Gemini XXX Pro and Qwen2.5-VL-72B as well as smaller OCR models like InternVL XXX MonkeyOCR-Pro-3B and Dots.OCR. While staying incredibly compact. By integrating ERNIE-4.5-0.3B with an encoder it can handle complex layouts in XXX languages with high precision. Lightweight high-accuracy and open-sourced a" [X Link](https://x.com/rohanpaul_ai/status/1979156903457423519) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-17T12:05Z 103.1K followers, 9561 engagements "This time its not a bubble. As far as Capex to Cash-flow ratio is concerned" [X Link](https://x.com/rohanpaul_ai/status/1979196883131040036) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-17T14:44Z 103.1K followers, 8929 engagements "Todays edition of my newsletter just went out. π Consider subscribing its free and I write it everyday. π§ Meta published the first real research on scaling reinforcement learning with LLMs π’ In a new report Goldman Sachs says Dont fear the AI bubble its about to unlock an $X trillion opportunity π Baidu just launched and open-sourced a BRILLIANT model for document parsing - PaddleOCR-VL-0.9B. π Tutorial: Run your models faster with quantized low precision in Keras" [X Link](https://x.com/rohanpaul_ai/status/1979205863463137632) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-17T15:20Z 103.1K followers, 3250 engagements "Financial Times shared OpenAIs recent stats and theyre next-level. - 800M weekly users. - X% paying. - $13B in ARR. - $8B loss in H1. - $1T+ in compute pledges" [X Link](https://x.com/rohanpaul_ai/status/1979208846393360820) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-17T15:32Z 103.1K followers, 4698 engagements "China's Tsinghua University paper shows simply mixing weights of Thinking and Instruct LLMs yields controllable efficient reasoning while keeping cost low. Which means your can tune models for cost and accuracy without any retraining. They literally interpolate the X weight sets and a single number sets the blend. At low blend the model behaves like Instruct answers are short accuracy rises a bit and there is almost no explicit thinking. At mid blend explicit chain of thought suddenly turns on reasoning quality jumps and tokens can even drop. At high blend the model thinks on every input" [X Link](https://x.com/rohanpaul_ai/status/1979235206956830738) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-17T17:17Z 103.1K followers, 4111 engagements "β OpenAI has became the worlds most valuable private company with 500B valuation. OpenAIs huge capital needs mean investors like Microsoft SoftBank and Thrive Capital will see their stakes reduced as more fundraising happens. FT writes a great piece. The working split being discussed is Microsoft at about XX% employees about XX% and the nonprofit parent about XX% with the remaining slice for SoftBank Thrive Khosla and others. Nvidias arrangement would add equity over time roughly $10B in shares for every $35B of GPUs OpenAI buys which could total up to $100B so every drawdown dilutes existing" [X Link](https://x.com/rohanpaul_ai/status/1979246033692692637) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-17T18:00Z 103.1K followers, 3729 engagements "Datacenter Capex will be in the region of $XXX billion this year from Google Amazon Microsoft and Meta alone. Goldman Sachs Research" [X Link](https://x.com/rohanpaul_ai/status/1979267755150098649) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-17T19:26Z 103.1K followers, 5366 engagements "Jensen Huang explains why AI is trillions of dollars of opportunity and its difference vs the old software era. Because the AI machine has to run all the time. --- On his recent interview on Citadel Securities YT channel" [X Link](https://x.com/rohanpaul_ai/status/1979325562876498274) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-17T23:16Z 103.1K followers, 6013 engagements "Andrej Karpathy AGI is still a decade away. On the latest Dwarkesh Patel podcast" [X Link](https://x.com/rohanpaul_ai/status/1979329702004428999) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-17T23:32Z 103.1K followers, 14.6K engagements "The paper says real AGI needs brain-like mechanisms not just competitive scores. It defines True Intelligence as X parts with X parts that engineers can measure and X deeper link. Gives a testable blueprint for AGI not just higher leaderboard scores. The X measurable parts are embodied sensory fusion core directives dynamic schemata a multi expert setup and an orchestration layer. A X level AGI scale scores a system by how many of those X parts it truly has. Embodied sensory fusion means the system ties ideas to seeing hearing touch and actions in the world. Core directives are built in goals" [X Link](https://x.com/rohanpaul_ai/status/1979405077048889799) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-18T04:32Z 103.1K followers, 8646 engagements "New Harvard paper shows trainingfree sampling lets a base LLM rival reinforcement learning on reasoning. No training dataset or verifier. The method samples from a power distribution which means reweighting full sequences the model already thinks are likely. That bias favors tokens that lead to a small set of likely futures over tokens that open many weak futures. Low temperature sampling only sharpens the next token so it misses those future path effects. The sampler runs a Metropolis Hastings loop that resamples a block then accepts or rejects based on relative sequence likelihood. It grows" [X Link](https://x.com/rohanpaul_ai/status/1979436283060723797) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-18T06:36Z 103.1K followers, 11.7K engagements "π @nvidia has now become the biggest name behind open AI model contributions with Nemotron Cosmos Gr00t BioNeMo and Canary. π Nemotron for agents BioNeMo for biopharma Cosmos for physical reasoning Gr00t for robotics and Canary for speech. And China accelerates with Qwen across chat multimodal reasoning and agents while Baidu Tencent MiniMax Z. AI ByteDance Moonshot AI and Zhipu AI ramp repos and leaderboard results. Europe is quieter beyond Mistral AI This nice graphics from @aiworld_eu --- cepsai.github. io/aiworld/OpenSource2025/story_animation.html" [X Link](https://x.com/rohanpaul_ai/status/1979504313127801021) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-18T11:06Z 103.1K followers, 11.3K engagements "New NVIDIA paper - StreamingVLM lets a vision language model follow endless video in real time while keeping memory and latency steady. It wins XXXXX% of head-to-head judgments against GPT-4o mini and runs at X FPS on X NVIDIA H100. The big deal is that it makes real-time infinite video understanding practical with steady latency built-in memory and clear gains over strong baselines. Old methods either look at everything and slow down a lot or use short windows and forget what happened. This model keeps a small cache with fixed anchor text a long window of recent text and a short window of" [X Link](https://x.com/rohanpaul_ai/status/1979518990889160760) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-18T12:04Z 103.1K followers, 4158 engagements "π» Nvidia and TSMC just announced that the first U.S.-made wafer for Nvidias Blackwell AI chips has been completed at TSMCs Arizona plant. Marks a big milestone in shifting advanced chip manufacturing to American soil. Jensen Huang called it historic and told that the most important chip in the world is now being produced in America. This wafer will eventually be used in Blackwell chips Nvidias next-generation processors designed for large-scale AI training and data center workloads. Huang added that Nvidia plans to invest $XXX billion in AI infrastructure over the next few years. Even with" [X Link](https://x.com/rohanpaul_ai/status/1979524914311749642) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-18T12:28Z 103.1K followers, 4509 engagements "The papers core claim argues LLMs excel by learning X fixed reasoning pattern so most human rationales are unnecessary. It matches human-data results while using 10x fewer human rationales. proof that teach the procedure beats collect more explanations. The key finding is that a stable reasoning pattern plus reinforcement learning can replace 10x human rationale data. Training still uses the usual X stages supervised fine tuning to show the pattern then reinforcement learning with verifiable rewards to lock it in. But the focus is patterned tasks where the same step order works every time" [X Link](https://x.com/rohanpaul_ai/status/1979537952171823185) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-18T13:20Z 103.1K followers, 3222 engagements "Andrej Karpathy called the current AI models as "slop" "I do think that overall the models are not there yet. I feel like the industry is making too big of a jump and trying to pretend that this is amazing but it's not. It's slop and I think they are not coming to terms with it. Maybe they are trying to fundraise or something like that I'm not sure what's going on. We are at this intermediate stage. The models are amazing but they still need a lot of work for now. Autocomplete is my sweet spot." --- On the latest Dwarkesh Patel podcast" [X Link](https://x.com/rohanpaul_ai/status/1979682318525874491) [@rohanpaul_ai](/creator/x/rohanpaul_ai) 2025-10-18T22:53Z 103.1K followers, 10.7K engagements
[GUEST ACCESS MODE: Data is scrambled or limited to provide examples. Make requests using your API key to unlock full data. Check https://lunarcrush.ai/auth for authentication information.]
Rohan Paul's recent activities are not specified in the provided context. However, the conversation revolves around AI developments, including OpenAI's massive compute deal with CoreWeave and a lawsuit filed by xAI against OpenAI for allegedly stealing trade secrets. Additionally, there are discussions on AI models' capabilities, such as passing CFA level III mock exams and outperforming human experts in certain tasks.
Social category influence technology brands XXXX% stocks #4638 finance XXX% countries XXXX% celebrities XXXX% vc firms #35 currencies XXXX% travel destinations XXXX% social networks XXXX% exchanges XXXX%
Social topic influence open ai #424, goldman sachs #6, accuracy #215, china #2249, investment #1235, bubble #378, faster #1199, jeff bezos #34, microsoft #262, infrastructure #1179
Top accounts mentioned or mentioned by @opus_genesis @grok @huggingface @googledeepmind @nvidia @xai @todayinai_ @nvidiaaidev @intuitmachine @justinechoes @openai @nlituanie @circuitcrazex @baiduinc @kuittinenpetri @bkrepysh34200 @aiatmeta @windsurfai @a16z @klingai
Top assets mentioned Goldman Sachs (GS) Microsoft Corp. (MSFT) Alphabet Inc Class A (GOOGL) Morgan Stanley (MS) NVIDIA Corp. (NVDA)
Top posts by engagements in the last XX hours
"Absolutely love this bit from Jeff Bezos. The simplest explanation on building a business and being founder. π― "Benjamin Graham the great investor is famous for saying in the short term the stock market is a voting machine in the long term it's a weighing machine and so as founders and entrepreneurs and business people our job is to build a heavy company. We want to build a company that when it is weighed it is a very heavy company we do not want to focus on the stock price" --- From 'DRM News' YT Channel"
X Link @rohanpaul_ai 2025-10-03T19:33Z 103K followers, 304.3K engagements
"This paper makes LLM reinforcement learning steadier by sampling more where it matters. Standard group training breaks when all sampled answers for a prompt look the same so the gradient becomes zero. It prevents the zero gradient stall that makes RL training spin its wheels. Taking a lot of samples fixes that but wastes compute on easy prompts. Reinforce-Ada samples each prompt in rounds then stops that prompt once it has enough signal. One rule stops after the 1st correct answer the balanced rule waits for enough correct and enough incorrect. After collection it builds a fixed size group"
X Link @rohanpaul_ai 2025-10-12T00:56Z 103K followers, 4849 engagements
"paper automates turning a dense paper into specific social-platform-ready posts. They define AutoPR a task that converts a paper and figures into accurate engaging platform specific content. They build PRBench with XXX paper to post pairs to score systems on fidelity engagement and platform fit. The core system PRAgent runs in X stages extract content synthesize with multiple agents then adapt to each platform. Plain prompting often drops key ideas uses generic hooks and picks broad tags that miss real communities. PRAgent fixes this by structuring the paper first analyzing visuals and"
X Link @rohanpaul_ai 2025-10-14T00:55Z 102.9K followers, 4528 engagements
"Broadcom CEO says generative AI will take up a much bigger share of global GDP. He told CNBC that knowledge-based and technology-intensive sectors now make up about XX% of the estimated US$110 trillion global GDP which could rise to XX% with the growth of generative AI. Last month Broadcom reported securing US$10 billion in chip orders from a fourth unnamed customer. Broadcom and OpenAI announced their official partnership on Monday saying they would jointly build and deploy XX gigawatts of custom artificial intelligence accelerators. The move is part of a broader effort to scale AI across"
X Link @rohanpaul_ai 2025-10-14T10:37Z 103K followers, 4811 engagements
"πΌ Overall AI is fueling economic growth through investment rather than efficiency. AI investment is giving the U.S. economy a strong push but it still hasnt made American workers much more productive. Growth is happening mostly because of money pouring into AI and a rising stock market not because people are working faster or smarter yet. Economists define productivity as how much output a worker can create in an hour. AI could help by handling boring or repetitive tasks so humans can focus on higher-value work or by automating some jobs entirely which raises efficiency. Right now the"
X Link @rohanpaul_ai 2025-10-14T23:11Z 103K followers, 3716 engagements
"MASSIVE claim in this paper. AI Architectural breakthroughs can be scaled computationally transforming research progress from a human-limited to a computation-scalable process. So it turns architecture discovery into a computebound process opening a path to selfaccelerating model evolution without waiting for human intuition. The paper shows that an allAI research loop can invent novel model architectures faster than humans and the authors prove it by uncovering XXX recordsetting linearattention designs that outshine human baselines. Right now most architecture search tools only finetune"
X Link @rohanpaul_ai 2025-07-26T02:05Z 103.1K followers, 940.3K engagements
"Hunyuan 3D-2.1 turns any flat image into studio-quality 3D models. And you can do it on this @huggingface space for free"
X Link @rohanpaul_ai 2025-08-23T03:50Z 103.1K followers, 219K engagements
"Brilliant and timely MIT + HARVARD study β€ Human-AI companionship in the wild looks stable and serious. Most users report clear benefits like reduced loneliness and emotional support. The biggest risk comes from sudden platform updates that break continuity and feel to users like losing a real partner. π§ The study analyzed 1506 top posts from r/MyBoyfriendIsAI a 27000+ member community clustered the language into themes and ran XX LLM classifiers to quantify platforms relationship stages benefits and risks. π¬ Why relationships form between AI and Human Bonds often start by accident during"
X Link @rohanpaul_ai 2025-09-16T08:03Z 103.1K followers, 75K engagements
"DeepSeek-R1 was published in Nature yesterday as the cover article for their BRILLIANT latest research. They show that pure Reinforcement Learning with answer-only rewards can grow real reasoning skills no human step-by-step traces required. So completely skip human reasoning traces and still get SOTA reasoning via pure RL. Its so powerful revelation because instead of forcing the model to copy human reasoning steps it only rewards getting the final answer right which gives the model freedom to invent its own reasoning strategies that can actually go beyond human examples. Earlier methods"
X Link @rohanpaul_ai 2025-09-18T06:40Z 103.1K followers, 453.9K engagements
"π¨ BAD news for Medical AI models. MASSIVE revelations from this @Microsoft paper. π€― Current medical AI models may look good on standard medical benchmarks but those scores do not mean the models can handle real medical reasoning. The key point is that many models pass tests by exploiting patterns in the data not by actually combining medical text with images in a reliable way. The key findings are that models overuse shortcuts break under small changes and produce unfaithful reasoning. This makes the medical AI model's benchmark results misleading if someone assumes a high score means the"
X Link @rohanpaul_ai 2025-09-25T15:53Z 103.1K followers, 524.3K engagements
"New ByteDance paper shows that redistributing exploration across prompts lets Reinforcement Learning trained LLMs learn more without extra compute. So you get the same or better gains without buying 2x more compute. Uniform budgets waste work because easy prompts all pass and hard prompts all fail so Group Relative Policy Optimization GRPO gets zero signal. Their fix scores each prompt at a chosen rollout count by expected learning value and cost then assigns budgets with a knapsack solver. Value is the chance of a non zero gradient times the expected increase in success rate. They use last"
X Link @rohanpaul_ai 2025-10-02T13:30Z 103.1K followers, 4063 engagements
"π° Jeff Bezos says the AI spending surge is a "good industrial bubble" that will waste some money yet lift productivity across the economy. He made the case comparing today with 1990s biotech and the dot com era where many firms died but infrastructure persisted. Outside the hyperscalers neocloud providers (GPU rental specialists that stand up capacity fast) are scaling funding and deployments to meet AI compute demand. The mechanism he describes is that excess money funds many experiments and vendors a few winners will emerge from that and the leftover assets cut future costs. --- bloomberg."
X Link @rohanpaul_ai 2025-10-03T18:02Z 103.1K followers, 186.2K engagements
"π― Jeff Bezos beautifully explains how AI boom as good kind of bubble that will benefit the world Bezos argued that industrial bubbles like AI are very different from financial bubbles like the 2008 banking crash because the inventions survive even if the hype fades. He compared todays AI boom to the dotcom eras spending on fiber optics and the biotech wave of the 1990s where even though many companies failed the technology and discoveries that remained were hugely valuable. AIs impact is real and will change every industry though investors right now may struggle to tell the difference"
X Link @rohanpaul_ai 2025-10-03T18:18Z 103.1K followers, 952K engagements
"The biggest private company valuations of the world are now ruled by AI and tech. OpenAI is first at $500B SpaceX second at $400B and ByteDance third at $300B"
X Link @rohanpaul_ai 2025-10-03T21:33Z 103.1K followers, 34K engagements
"πΎ Western Digitals CEO says HDDs are still central to AI storage with XX% of hyperscaler data on HDDs XX% on SSDs and XX% on tape. That split reflects capacity economics and power per TB since enterprise disks remain roughly 5x to 6x cheaper per TB than SSDs at scale and draw fewer watts per TB in bulk deployments. Data centers tier AI storage so hot data sits on flash warm and cold pools live on disks and rarely touched archives go to tape which matches how hyperscalers manage cost and performance. AI demand is exceeding what manufacturers can currently produce or ship in a timely way so"
X Link @rohanpaul_ai 2025-10-04T15:02Z 103.1K followers, 11.4K engagements
""100 million words context window is already possible which is roughly what a human hears in a lifetime. Inference support is the only bottleneck to achieve it. And AI Models actually do learn during the context window without changing the weights." Anthropic CEO Dario Amodei (On the 2nd point there was this brilliant Google Paper published last week that says LLMs can learn in context from examples in the prompt can pick up new patterns while answering yet their stored weights never change.) --- From 'Alex Kantrowitz' YT Channel (Full Video link in comment)"
X Link @rohanpaul_ai 2025-10-04T23:02Z 103.1K followers, 631.2K engagements
"π°AI is making the caribbean island of Anguilla rich. Anguilla now gets XX% of its income from .ai domains up from X% before the surge. Because .ai is the islands official country code top-level domain (TLD) and the Government of Anguilla is the official manager of that registry. So every .ai registration or renewal anywhere in the world routes through the .ai registry for Anguilla. Even on the secondary market .ai domains are traded at high prices: you.ai sold for an impressive $700000 in 2023 cloud.ai for $600000 in 2025 and adapt.ai in the same year for $300000"
X Link @rohanpaul_ai 2025-10-05T09:00Z 103.1K followers, 5268 engagements
"Absolutely classic @GoogleResearch paper on In-Context-Learning by LLMs. Shows the mechanisms of how LLMs learn in context from examples in the prompt can pick up new patterns while answering yet their stored weights never change. π‘The mechanism they reveal for in-context-learning. When the model reads a few examples in your prompt it figures out a pattern (like a small rule or function). Instead of permanently changing its stored weights it forms a temporary adjustment that captures this pattern. That adjustment can be written mathematically as a rank-1 matrix meaning it only adds one"
X Link @rohanpaul_ai 2025-10-05T12:06Z 103.1K followers, 75.4K engagements
"This is not a bubble. Cisco in 1998-2002 vs Nvidia in 2020-2024 The data says this cycle is different and AI is still so early for most parts of the world. Cisco was a valuation story price inflated while earnings lagged then the multiple deflated. Nvidia is an earnings story price climbs alongside surging earnings"
X Link @rohanpaul_ai 2025-10-05T20:41Z 103.1K followers, 871.9K engagements
"A 7B model tuned for forms and docs beats giant models at pulling structured data. Beats GPT-4.1 on 1000 extraction tasks trained for $XXX. The team generated synthetic training data that preserves memory across chunks of a long file. That memory lets the model connect names dates and values that appear far apart. They fine-tuned with Low Rank Adaptation changing only XXXX% of weights. They then used Group Relative Policy Optimization with a semantic reward and strict JSON checks. This setup accepts different surface wording if the meaning matches. On 1000 held-out tasks it hit XXXXX mean"
X Link @rohanpaul_ai 2025-10-06T04:08Z 103.1K followers, 131.1K engagements
"So surprising. They should have at-least proof-read. π
AI hallucinations need to be SOLVED ASAP. π° Deloitte has agreed to repay part of a $440000 government fee after admitting it used AI to write sections of an official report that later turned out to have fake references and quotes. The report was commissioned by Australias Department of Employment and Workplace Relations to review an IT system that manages welfare penalties. Deloitte used OpenAIs GPT-4o model to help fill traceability and documentation gaps. After the report was published in July-25 a University of Sydney academic"
X Link @rohanpaul_ai 2025-10-06T10:50Z 103.1K followers, 690.4K engagements
"For the last X months Figure Robot has been working XX hours a day on BMWs X3 production line managing complex assembly tasks in a real industrial setting. Factory work is almost ready to be redefined"
X Link @rohanpaul_ai 2025-10-08T16:45Z 103.1K followers, 4934 engagements
"πA Reddit post is circulating on possible leak of OpenAI's top XX customers who used over X trillion tokens. I like one of the comment π
"So much of our economy is now AI service companies paying AI service companies for AI services.""
X Link @rohanpaul_ai 2025-10-09T18:23Z 103.1K followers, 62.2K engagements
"New Stanford + SambaNova + UC Berkeley paper proposes quite a revolutionary idea. π€― Proves LLMs can be improved by purely changing the input context instead of changing weights. Introduces a new method called Agentic Context Engineering (ACE). It helps language models improve by updating what they read and remember instead of changing their core weights. It has three parts: a Generator that works on tasks a Reflector that learns from mistakes and a Curator that updates the notebook with helpful lessons. So ACE works like this. The model first tries to solve a task. While doing that it writes"
X Link @rohanpaul_ai 2025-10-09T22:08Z 103.1K followers, 36.1K engagements
"π The US approved Nvidia AI chip exports to the UAE. Also unlocking a 5GW Abu Dhabi data center with OpenAI and tying shipments to matching Emirati investment in the US. The framework targets up to 500000 chips/year with XX% for G42 alongside a $1.4T UAE pledge over XX years and the initial batch reportedly excludes G42. Control relies on American cloud operators running the hardware and the cloud keeping telemetry compliance and sensitive data under US processes. The aim is to meet Gulf demand while boxing out Huaweis Ascend 910B offers and keeping regional stacks inside US ecosystems."
X Link @rohanpaul_ai 2025-10-09T22:45Z 103.1K followers, 3749 engagements
"A solid compilations of the State of AI 2025 in this report. - OpenAI remains ahead at the top but China has real momentum. DeepSeek Qwen and Kimi are within a few points on reasoning and coding and Qwen now powers about XX% of new fine-tunes on Hugging Face which signals a shift in open-weights leadership. - Reasoning quality improved because labs moved from loose human feedback to rubric based rewards and tasks that can be checked. That change lets models plan reflect and correct themselves across longer multi step jobs. - AI is starting to act like a lab partner. DeepMinds Co-Scientist and"
X Link @rohanpaul_ai 2025-10-10T18:09Z 103.1K followers, 11.6K engagements
"Rude prompts to LLMs consistently lead to better results than polite ones π€― The authors found that very polite and polite tones reduced accuracy while neutral rude and very rude tones improved it. Statistical tests confirmed that the differences were significant not random across repeated runs. The top score reported was XXXX% for very rude prompts and the lowest was XXXX% for very polite. They compared their results with earlier studies and noted that older models (like GPT-3.5 and Llama-2) behaved differently but GPT-4-based models like ChatGPT-4o show this clear reversal where harsh tone"
X Link @rohanpaul_ai 2025-10-10T19:04Z 103.1K followers, 868.9K engagements
"New paper from @Google is a major memory breakthrough for AI agents. ReasoningBank helps an AI agent improve during use by learning from its wins and mistakes. To succeed in real-world settings LLM agents must stop making the same mistakes. ReasoningBank memory framework helps agents learn from both successes and failures and turn them into reasoning strategies that generalize. Traditionally most agents handle tasks in a stream but do not keep lessons so they repeat errors. The system turns each action log into a short memory item with a title a description and a concrete next step. Before a"
X Link @rohanpaul_ai 2025-10-11T02:28Z 103.1K followers, 122K engagements
"The new steam age. This is actually becoming true in many cases. It's possible to do so much more on your own now"
X Link @rohanpaul_ai 2025-10-11T13:35Z 103.1K followers, 2M engagements
"Keep your local data backup on different geo-location. π₯ South Koreas central government just lost 858TB from its internal data after a lithium-ion battery fire. Permanently wiping a shared document store that covered roughly X years of work. South Koreas government built its own private cloud service called G-Drive meant for internal use by civil servants. It was hosted entirely inside a government data center G-Drive served about 125000 officials with default 30GB per user and unlike XX other systems at the site it had no backup so the racks that burned were the single copy. Restoration"
X Link @rohanpaul_ai 2025-10-11T14:22Z 103.1K followers, 18.8K engagements
"π Hardware Memory bandwidth is becoming the choke point slowing down GenAI. During 20182022 transformer model size grew XXX every X years while memory per accelerator grew only about X every X years. And that mismatch shoves us into a Memory-Wall The "memory wall" is creating all the challenges in the datacenter and for edge AI applications. In the datacenter current technologies are primarily trying to solve this problem by applying more GPU compute power. And that's why HBM capacity and bandwidth scaling KV offload and prefill-decode disaggregation are central to accelerator roadmaps."
X Link @rohanpaul_ai 2025-10-11T16:08Z 103.1K followers, 80.8K engagements
"Andrej Karpathy on what makes Elon Musk unique. Elon keeps teams small highly technical and removes low performers very quickly. pushes for intensity avoids useless meetings and stays deeply connected with engineers"
X Link @rohanpaul_ai 2025-10-12T06:54Z 103.1K followers, 222.1K engagements
"This is where it all started. Here 19-yo Stanford dropout Sam Altman giving his first-ever startup pitch. He dropped out of Stanford to start Loopt sold it in 2012 for $XXXX mn. Then invested via Hydrazine and later led Y Combinator before co-founding OpenAI in 2015"
X Link @rohanpaul_ai 2025-10-12T07:06Z 103.1K followers, 1.5M engagements
"New @GoogleResearch paper shows agents learn software skills by watching tutorials converting them into action steps and boosting task performance. So converts free videos into reliable supervision at scale. A vision model inverse dynamics predicts the action between X screenshots like click type or scroll. Training uses about 630K transitions mixing 500K synthetic steps and 132K human ones. The model then labels tutorial videos and turns them into executable step sequences. It produces about 53K trajectories across XX apps for examples or training. As examples these steps add X to X points"
X Link @rohanpaul_ai 2025-10-12T11:41Z 103.1K followers, 48K engagements
"The exact moment Jeff Bezos decided not to become a physicist From "The Economic Club of Washington D.C." YT Channel"
X Link @rohanpaul_ai 2025-10-12T22:29Z 103.1K followers, 6398 engagements
""The Impact of Artificial Intelligence on Human Thought" A big XXX page report. AI is shifting real thinking work onto external systems which boosts convenience but can weaken the effort that builds understanding and judgment A pattern the paper frames through cognitive offloading and cognitive load theory and then tracks into social effects like standardized language and biased information flows and manipulation tactics that target human psychology. It says use AI to cut noise and routine steps keep humans doing the heavy mental lifting and add controls because personalization deepfakes and"
X Link @rohanpaul_ai 2025-10-12T22:55Z 103.1K followers, 38.6K engagements
"New AirBnb paper explains a simple loop that keeps a support LLM improving by learning from every human customer support agent interaction. Turns live customer-support work into continuous low-friction training data. Static models go out of date as policies and products change. This framework captures X signals during real cases which reply the agent prefers whether they adopted it and why whether the cited knowledge was relevant and what knowledge was missing. These signals feed a training pipeline that updates retrieval ranking and generation together. A virtual judge filters noisy labels"
X Link @rohanpaul_ai 2025-10-13T10:17Z 103.1K followers, 27.5K engagements
"π‘ Samsung is trying to flip the balance in the AI memory race by convincing Nvidia to raise the official speed target for HBM4 even though it was slower than SK Hynix and Micron in delivering the first samples. Normally HBM (high-bandwidth memory) designs focus more on reducing heat than on raw speed because the stacked DRAM (Dynamic Random Access Memory) layers trap heat easily. Nvidias sudden demand for higher operating speed surprised both SK Hynix and Micron since their designs were optimized for thermal stability. Samsung however had already prepared samples that hit higher clock speeds"
X Link @rohanpaul_ai 2025-10-13T13:14Z 103.1K followers, 15.8K engagements
"Microsoft lit up its first Nvidia AI factory for OpenAI a 4600+ GPU GB300 NVL72 cluster on Azure and said more are coming. Each rack packs XX Blackwell Ultra GPUs XX Grace CPUs 37TB fast memory and 130TB/s NVLink inside the rack which keeps data moving without stalls. Across racks it uses Quantum-X800 InfiniBand at 800Gb/s per GPU so Azure can treat thousands of GPUs as one big pool for training and serving. Microsoft says this is the first of many with plans to deploy hundreds of thousands of Blackwell Ultra GPUs into its AI datacenters. The target is shorter training cycles measured in"
X Link @rohanpaul_ai 2025-10-14T10:38Z 103.1K followers, 6460 engagements
"New Tencent paper upgrades LLM agents by learning small experience rules in the prompt without changing weights. It costs about $XX on XXX samples yet beats some $10000 fine tuning runs. π€― Specialized tasks need tools and prompting and weight tuning is slow expensive and fragile. The method keeps the base model frozen and stores short lessons as a token prior in the prompt. For each question the agent samples a small group of answers scores them and writes why winners win. Those reasons become general rules saved in an experience library for the next run. Over a few passes the library is"
X Link @rohanpaul_ai 2025-10-14T10:41Z 103.1K followers, 22.6K engagements
"Data center emissions as % of 2022 global energy emissions According to Goldman Sachs Research They estimates that XX% of the increase in power demand from data centers will be met by renewables and there will be a modest amount of nuclear capacity thats targeted for AI. The bulk of the remaining XX% is expected to be driven by natural gas"
X Link @rohanpaul_ai 2025-10-14T15:02Z 103.1K followers, 3715 engagements
"New @GoogleDeepMind paper shows that long step by step thinking often wastes compute on easy questions. The paper makes overthinking measurable and stoppable. Shows where time is wasted and gives clear rules to stop early without losing accuracy. They test many models in thinking and non thinking modes and see simple queries run X to 20x slower with little or no gains. To study why they build TRACE a tool that splits a response into small sub thoughts and labels each step. TRACE also builds a graph of the steps to track how answers change as the model thinks. From many graphs they see X main"
X Link @rohanpaul_ai 2025-10-15T05:51Z 103.1K followers, 12.5K engagements
"The Federal Reserve now includes "Singularity: Extinction" in their forecasts driven by AI. Technological singularity refers to a scenario in which AI eventually surpasses human intelligence leading to rapid and unpredictable changes to the economy and society. Under a benign version of this scenario machines get smarter at a rapidly increasing rate eventually gaining the ability to produce everything leading to a world in which the fundamental economic problem scarcity is solved the Federal Reserve Bank of Dallas writes. Under a less benign version of this scenario machine intelligence"
X Link @rohanpaul_ai 2025-10-15T06:13Z 103.1K followers, 8813 engagements
"The paper shows small LLM agents can push other LLMs to learn specific behaviors just through interaction. shows LLM agents can directly influence each others learning rules not just their immediate actions. During training both LLM agents keep playing the same game again and again like many rounds. But only one of them called the shaper waits until all those rounds are finished before it changes its internal settings. The other agent its opponent updates itself after each smaller episode inside that trial. So the shaper gets to watch how the opponent changes step by step and then it adjusts"
X Link @rohanpaul_ai 2025-10-16T00:27Z 103.1K followers, 8138 engagements
"Another datapoint that the current AI-driven valuations are not like the 2000s Dot-com bubble. In 2000 the median stocks free cash flow yield was about 1.2%. Today it is about XXX% roughly 3x higher. i.e. typical stock is cheaper on cash flow basis"
X Link @rohanpaul_ai 2025-10-16T17:35Z 103.1K followers, 14.7K engagements
"The ten core cognitive components of our AGI definition"
X Link @rohanpaul_ai 2025-10-16T23:09Z 103.1K followers, XXX engagements
"This paper says LLM hallucinations are unavoidable when systems must operate in an open world. It treats hallucination as a generalization issue rather than a simple bug. In a closed world training and test match so extra data can push errors down. In an open world new kinds of inputs keep showing up so past experience can mislead. This lines up with the no free lunch idea that no method wins without assumptions. The paper splits errors into Type X false memory and Type X false generalization. Type X clashes with facts the model already saw so updates can fix it. Type X extends patterns to"
X Link @rohanpaul_ai 2025-10-17T01:30Z 103.1K followers, 7858 engagements
"This paper proposes a X construct language the tensor equation that unifies neural nets and logic enabling reliable reasoning. It claims a full transformer fits in XX equations. Could make AI simpler to build easier to verify and less prone to hallucinations. Today AI uses many tools which hurts trust and maintenance. Tensor logic stores facts in tensors then programs join and project them like database operations. A database rule such as Datalog becomes X tensor operation that counts matches then maps counts to X or X. Gradients fit this equation form too. Blocks like convolution attention"
X Link @rohanpaul_ai 2025-10-17T03:42Z 103.1K followers, 5509 engagements
"Beautiful way to think about LLM reasoning. Says logic lives in changes not positions. Or that logic aligns velocity and curvature across topics and languages. The paper recasts LLM reasoning as smooth paths where logic steers movement through embeddings over time. Each thought step moves the point along that path. The absolute position mostly shows topic or language not the logic. The step change called velocity carries the logic. The bend of the path called curvature carries it too. They made a dataset that repeats the same deductions across many topics and languages. They read hidden"
X Link @rohanpaul_ai 2025-10-17T04:50Z 103.1K followers, 25.1K engagements
"π PaddleOCR-VL also went #1 on Hugging Face Trending"
X Link @rohanpaul_ai 2025-10-17T12:13Z 103.1K followers, XXX engagements
"Elon Musk just confirmed Grok now powers X's content ranking and how it impacts posts with external links"
X Link @rohanpaul_ai 2025-10-17T15:13Z 103.1K followers, 6437 engagements
"Stop worrying about the bubble in AIits growth is sustainable three Wall Street analysts from Goldman Sachs JPMorgan and Wedbush argued this morning in notes seen by Fortune. And WallStreet traders seem to agree"
X Link @rohanpaul_ai 2025-10-17T19:28Z 103.1K followers, 8166 engagements
"β‘ OpenAI is pushing to run the full stack of generative AI from consumer apps to enterprise tools. OpenAI internally built some LLM-based software tools for its own teams for example one app that helps the sales team sort and qualify leads and another that helps the finance team track thousands of contracts and negotiations. When OpenAI showed those apps publicly investors realized these tools could replace or compete directly with existing business software from companies like DocuSign (used for contracts) and HubSpot (used for sales and marketing automation). And then later on Dev Day"
X Link @rohanpaul_ai 2025-10-17T21:42Z 103.1K followers, 6062 engagements
"This paper shows how an AI agent to clean its own chat history while it works so it stays accurate on long tasks. So memory editing becomes a built in skill which raises accuracy on long tasks and cuts cost at the same time. Long chats pile up junk notes so the model gets distracted and runs up token cost. The paper makes memory changes a normal action so the agent can keep compress summarize or delete items by itself. Each change is a function call with a short summary and an id so the agent edits history without losing the thread. Editing history breaks the usual growing prompt so training"
X Link @rohanpaul_ai 2025-10-17T22:45Z 103.1K followers, 12.8K engagements
"In China our market share dropped from XX% to X% NVIDIA CEO Jensen Huang in a Temporary Goodbye to Chinas AI Market The drop tracks with tighter US export rules and Chinas push for local AI stacks. Huang made the comment at Citadel Securities 2025 event framing China as off the table unless policy shifts. This gap opens room for Huawei and others that are scaling large clusters of Ascend chips to replace imported GPUs in training and inference. Huaweis roadmap promises yearly Ascend releases and supernodes that it claims can outpace NVIDIAs upcoming Vera Rubin rack platforms in some"
X Link @rohanpaul_ai 2025-10-17T23:47Z 103.1K followers, 5730 engagements
"This paper introduces dInfer a fast inference framework that makes diffusion language models practical by reorganizing how decoding runs. It replaces a single monolith with X plug-and-play parts the model a diffusion iteration manager a decoding strategy and a KV-cache manager so teams can mix algorithms without retraining. It improves each step with X training-free tricks iteration smoothing to carry soft guesses across steps hierarchical decoding to confirm spaced-out tokens first and credit decoding to lock tokens that stay stable. It fixes cache cost with vicinity refresh which only"
X Link @rohanpaul_ai 2025-10-18T00:00Z 103.1K followers, 4397 engagements
"This paper shows Mixture of Experts (MoE) models share language-neutral experts in middle layers and steering routers boosts multilingual reasoning. Means a tiny test-time change boosts many languages with almost no cost by steering toward shared middle experts that predict quality. An MoE layer has many experts and a router picks a few per token. Early and late layers act language-specific while the middle layers act language-neutral. Languages that route like English in the middle score higher on multilingual tasks. Low-resource languages lag when they miss those shared middle experts. The"
X Link @rohanpaul_ai 2025-10-18T05:34Z 103.1K followers, 4790 engagements
"The energy storage boom is just about starting - fully in sync with AI's penetration. Goldman Sachs expects S&P XXX companies to spend $XXX trillion on AI next year mostly on new data centers. That means electricity demand will rise sharply as AI workloads expand. Since fossil fuel generation is flat and nuclear projects take years the extra power has to come from renewables like solar and wind. But renewables dont produce steady output through the day so keeping data centers running requires massive energy storage systems that can hold surplus energy and release it when generation dips."
X Link @rohanpaul_ai 2025-10-18T12:15Z 103.1K followers, 4096 engagements
"Very useful RAG repo. A good collection of RAG implementations with many different strategies. Includes tutorials and visualizations"
X Link @rohanpaul_ai 2024-08-15T22:10Z 103K followers, 98K engagements
"πΌ Finally a solid 57-page report on AI's effect on job-market from Stanford University. THE SHIFT HAS STARTED. Entrylevel workers in the most AIexposed jobs are seeing clear employment drops while older peers and lessexposed roles keep growing. Though overall employment continues to grow employment growth for young workers in particular has been stagnant. The drop shows up mainly as fewer hires and headcount not lower pay and it is sharpest where AI usage looks like automation rather than collaboration. 2225 year olds in the most exposed jobs show a XX% relative employment decline after"
X Link @rohanpaul_ai 2025-08-26T08:23Z 103K followers, 313.8K engagements
"Cool Tongyi Lab and Alibaba Group paper introduces ReSum a simple way for web agents to search longer and answer better. It shows +4.5% over ReAct up to +8.2% with ReSum-GRPO and Pass@1 of XXXX% and 18.3%. Most agents use ReAct (Reason + Act) they append every step to the chat so long tasks hit the context limit. The idea is that the agent writes down its reasoning steps in text then chooses an action like searching the web or clicking a link. After the action it records the result it got back and then repeats the process: think act observe. This makes the whole process traceable since every"
X Link @rohanpaul_ai 2025-09-21T17:30Z 103K followers, 10.1K engagements
"π₯ Meta reveals a massive inefficiency in AIs reasoning process and gives a solution. Large language models keep redoing the same work inside long chains of thought. For example when adding fractions with different denominators the model often re explains finding a common denominator step by step instead of just using a common denominator behavior. In quadratic equations it re explains the discriminant logic or completes the square again instead of calling a solve quadratic behavior. In unit conversion it spells out inches to centimeters again instead of applying a unit conversion behavior."
X Link @rohanpaul_ai 2025-09-28T12:05Z 103K followers, 57.4K engagements
"π€ NVIDIA Updates Robotics Platform with Open-Source Newton Physics Engine New AI Models Newton is a GPU-accelerated physics engine built on NVIDIA Warp and OpenUSD codeveloped with Google DeepMind and Disney Research. It plugs into MuJoCo Playground and Isaac Lab using OpenUSD so scenes and assets stay consistent across tools which cuts asset duplication and sim-to-real friction. Isaac GR00T N1.6 will integrate Cosmos Reason an open 7B reasoning vision language model that turns fuzzy instructions into step-by-step plans using physical common sense and prior knowledge. Isaac Lab XXX arrives"
X Link @rohanpaul_ai 2025-10-01T12:28Z 103K followers, 3618 engagements
"The paper claims learning (an AI system learning or machine learning in general) follows a physics style least action rule that unifies supervised generative and reinforcement learning. Shows that supervised learning generative modeling and reinforcement learning can all be seen as following a "least action" rule just like particles in physics follow least action paths It treats information like motion where token loss acts like speed and the change in loss acts like acceleration. This view says learning slows over time because each new example reduces loss a bit less. Efficiency means"
X Link @rohanpaul_ai 2025-10-03T13:36Z 103K followers, 36.5K engagements
"Wow. π§ The paper presents Dragon Hatchling a brain-inspired language model that matches Transformers using local neuron rules for reasoning and memory. It links brain like local rules to Transformer level performance at 10M to 1B scale. It makes internals easier to inspect because memory sits on specific neuron pairs and activations are sparse and often monosemantic. You get reliable long reasoning and clearer debugging because the model exposes which links carry which concepts in context. The problem it tackles is long reasoning models often fail when the task runs longer than training. The"
X Link @rohanpaul_ai 2025-10-03T15:06Z 103K followers, 40.6K engagements
"Jeff Bezos just gave the most bullish opinion on AI. "AI is going to make every company's quality go up and their productivity go up. I literally mean every company. Every manufacturing company every hotel every you know consumer products company AI is going to change every industry. It's a very unusual technology in that regard in that it's a horizontal enabling layer." --- From 'DRM News' YT Channel"
X Link @rohanpaul_ai 2025-10-04T07:43Z 103.1K followers, 187.7K engagements
"πΈ FT published an article. AI capex is surging at hyperscalers a classic late stage bubble tell that could break yet the build should make AI cheaper later. Valuations near 30x earnings or 8x sales - these prices only make sense if every part of the AI boom continues smoothly without the usual business or economic cycle risks. Capex means huge spend on data centers chips power and land by the biggest cloud providers. Bubble phases often end when excess capacity stretches the boom demand slips and the cycle turns. The current triggers are stricter Europe AI rules compute light models like"
X Link @rohanpaul_ai 2025-10-04T20:11Z 103K followers, 89.3K engagements
"π«‘ GPT-5-Pro just solved the Math problem that no other LLM could solve. Took XX minutes without any internet search. An Oxford and Cambridge paper claimed that no LLM could solve Yu Tsumuras 554th Problem. OpenAI's GPT5 Pro produced a full proof in about XX minutes. ---- The paper - arxiv .org/pdf/2508.03685"
X Link @rohanpaul_ai 2025-10-05T10:34Z 103.1K followers, 310.8K engagements
"Todays edition of my newsletter just went out. π Consider subscribing its free and I write it everyday. π¨π§ Google made Jules Tools public offering a command-line option to configure and manage its coding assistant. π€ OpenAI and Jony Ive are building a palm sized screenless AI assistant targeted for 2026 but core software privacy and compute are not ready yet. π MoE Inference Economics from First Principles - A solid tutorial. π‘ Teslas Optimus humanoid robot performs Kung Fu moves - great achievement for humanoid robots as many of these moves are not tele-operated instead they are"
X Link @rohanpaul_ai 2025-10-06T18:43Z 103K followers, 3827 engagements
"OpenAI's AgentKit will be so insane build every step of agents on one platform. These visual agent builders make the whole process of iterating and launching agents far more efficient. It sits on top of the Responses API and unifies the tools that were previously scattered across SDKs and custom orchestration. It lets developers create agent workflows visually connect data sources securely and measure performance automatically without coding every layer by hand. The core of AgentKit is the Agent Builder a drag-and-drop canvas where each node represents an action guardrail or decision branch."
X Link @rohanpaul_ai 2025-10-06T21:17Z 103K followers, 178.8K engagements
"LLM have done great with logic and Math but next should be Physics. Periodic Labs Ekin Dogus Cubuk Last week Periodic Labs landed a massive $300M seed round led by some of the big names Andreessen Horowitz Nvidia Jeff Bezos and Eric Schmidt"
X Link @rohanpaul_ai 2025-10-07T12:32Z 103K followers, 19.9K engagements
"China is making these massive Solar Plants on water bodies as they need the land for agriculture"
X Link @rohanpaul_ai 2025-10-07T16:21Z 103K followers, 4477 engagements
"NEWS πΌ: Elon Musk picked former Morgan Stanley dealmaker Anthony Armstrong as chief financial officer (CFO) of xAI and he will run the finances for both xAI and X as they integrate after a $113B merger in Mar-25. Armstrong led global tech mergers and acquisitions at Morgan Stanley and advised Musk on the $44B Twitter deal which puts a seasoned capital architect in the seat that funds models and infrastructure at scale. He is also taking over Xs finance role from Mahmoud Reza Banki who is leaving after less than X year consolidating cash decisions across the product and the model stack. xAI"
X Link @rohanpaul_ai 2025-10-07T21:38Z 103.1K followers, 4881 engagements
"The paper shows that xLSTM scales better than Transformers and keeps time linear as prompts get longer. So xLSTM or variants might become a serious alternative to Transformers especially for long input scenarios and inference efficiency. At 16K context xLSTM cuts time to first token by 30-50% for the same model size. Transformer attention gets much more expensive as the prompt grows but xLSTM's update cost grows only with length. The study compares models from 80M to 7B trained on the same data and fits loss versus compute scaling laws. Across the full compute range xLSTM reaches the same"
X Link @rohanpaul_ai 2025-10-08T00:18Z 103K followers, 8651 engagements
"This paper introduces a new method called Agentic Context Engineering (ACE). It helps language models improve by updating what they read and remember instead of changing their core weights. Normal methods that edit prompts tend to make them too short and lose important details over time. ACE fixes this by treating the models context like a growing notebook that keeps and organizes useful strategies. It has three parts: a Generator that works on tasks a Reflector that learns from mistakes and a Curator that updates the notebook with helpful lessons. Instead of rewriting everything each time"
X Link @rohanpaul_ai 2025-10-08T01:20Z 103K followers, 46.4K engagements
"π― Sen. Bernie Sanders dropped an extreme version of AI is coming for your job video. Nearly XXX million US jobs could be eliminated over the next XX years according to a new report from Senator Bernie Sanders. Warns that artificial labor could upend the economy faster than the world is prepared for. --- Video from 'Senator Bernie Sanders' YT channel"
X Link @rohanpaul_ai 2025-10-09T08:14Z 103K followers, 17.6K engagements
"100% of OpenAIs pull requests are reviewed by Codex. And Dario Amodei said a few days back that "The vast majority of code that is used to support Claude and to design the next Claude is now written by Claude. It's just the vast majority of it within Anthropic. And other fast moving companies the same is true." The shift has started in all tech companies. --- (Video from 'Axios' YT Channel.)"
X Link @rohanpaul_ai 2025-10-09T10:05Z 103K followers, 4622 engagements
"πΌ A new U.S. Senate minority staff report led by Bernie Sanders says "AI and automation could replace about 97M to 100M U.S. jobs in XX years". With the heaviest hits in fast food customer support retail and logistics. The analysis uses federal job descriptions and asks a model to rate how much each task could be automated then scales that task score by how many people hold that job today to estimate displacement. The headline numbers include XX% of fast food and counter workers XX% of customer service representatives XX% of stockers and order fillers XX% of accountants XX% of software"
X Link @rohanpaul_ai 2025-10-09T12:02Z 103K followers, 24.5K engagements
"Newsπ Taiwan rejects a 50-50 chip production split with the US and says TSMC will keep its most advanced technology and most output in Taiwan. Officials support overseas fabs only when there are clear orders real profits no national security risks and when the broader industry benefits. TSMC is putting $165B into X advanced wafer fabs in the US but is building XX in Taiwan with more planned so an even split is not realistic. The company keeps sole ownership of its US fabs yet the leading process nodes and the bulk of volume stay in Taiwan which the minister emphasized as core to"
X Link @rohanpaul_ai 2025-10-09T14:02Z 103.1K followers, 4871 engagements
"Currently magnificent X valuations are near 23x forward P/E and 5.1x EV/S vs 52x P/E for 2000 leaders with XX% return on equity and XX% net margin supported by strong balance sheets"
X Link @rohanpaul_ai 2025-10-09T23:01Z 103K followers, 1511 engagements
"Mark Cuban's (@mcuban) advice for fresh graduates. If youre AI-native they need you. AI is on the minds of all types of companiessmall businesses Shark Tank companies and corporations with 10000 workers - all are trying to figure it out"
X Link @rohanpaul_ai 2025-10-10T08:10Z 103K followers, 3698 engagements
"π Cool Github Repo: Opensource selfhosted vendorneutral alternative to OpenAIs AgentKit Shannon turns agent ideas into production systems cheaper traceable and safe to run. It enforces hard token budgets caching rate limits and supports zero-token YAML templates that skip unnecessary LLM calls cutting waste dramatically and avoid provider throttling. Many agent stacks break in real use as costs spike runs fail and logs stay thin. Shannon wraps every task in a Temporal workflow so state persists and exact replays are possible. Workflows use simple templates structured as directed acyclic"
X Link @rohanpaul_ai 2025-10-10T13:54Z 103.1K followers, 9531 engagements
"Morgan Stanley Research on Nvidia. The bottleneck has shifted from making chips to data center space power and infrastructure. The note also says Nvidia will keep using targeted investments to speed customers deployments not to own capacity which reduces fear about circular funding. Demand is still extremely high but semiconductor manufacturing and packaging have scaled enough that chip supply is no longer the main limiter. The slow part now is building and powering facilities because utility interconnects transformers cooling and permits run on long planning cycles. OpenAIs 10GW announcement"
X Link @rohanpaul_ai 2025-10-11T10:04Z 103K followers, 90.3K engagements
"Another study from NeurIPS25 submission finds adding polite phrasing systematically lowers correctness of LLMs. This reduced accuracy is relative to the same question without the polite cue and this holds when averaging across many models that include GPT-4o and GPT-5. Makes the case that the be a bit rude or adversarial strategy can yield better answers than be very polite The paper builds a 24160prompt benchmark with singleturn variants that explicitly include a Politeness modifier and a separate Emotional tone modifier. They test XX models including GPT4o and GPT5. In singleturn evaluation"
X Link @rohanpaul_ai 2025-10-11T13:07Z 103K followers, 18K engagements
"π§ A new Morgan Stanley research says brain computer interfaces (BCI) are shifting from science fiction to investable reality and puts Neuralink at the center. US senators introduced the MIND Act in Sep-25 to set rules for neural data and to direct the Federal Trade Commission to study governance. The report titled Neuralink AI in your brAIn argues that AI will speed BCI progress and warns that human communication bandwidth may struggle to keep up with AGI. Neuralinks near term stack pairs Telepathy for thought controlled computing with Blindsight for vision restoration via visual cortex"
X Link @rohanpaul_ai 2025-10-11T14:59Z 103K followers, 5487 engagements
"New Adobe paper showshow to learn from prompts where all sampled answers score the same boosting reasoning. It reports gains up to XXXX accuracy points and XXXX pass rate points over GRPO a popular baseline. The big gain here is that training stops throwing away a huge chunk of data and turns it into signal. because rollouts take about XX% of step time so converting waste into learning buys accuracy and stability without extra compute. The key finding is that using zero variance prompts with entropy weighted updates gives consistent gains over GRPO and over methods that filter those prompts"
X Link @rohanpaul_ai 2025-10-11T20:25Z 103.1K followers, 6626 engagements
"The paper links a time series model to an LLM so the LLM can reason over numbers plus text. Matters so much because so many real-world tasks mix numbers and text and this handles that mix well. LLMs read text well but miss patterns in numbers over time. Time series models see those patterns but ignore context like news. TS-Reasoner connects a pretrained time series model to an LLM via a small adapter. The adapter converts series features into tokens the LLM can read. Stage X builds the bridge using synthetic series plots and captioned descriptions. These captions explain trend repeating"
X Link @rohanpaul_ai 2025-10-12T08:35Z 103.1K followers, 5471 engagements
"New Stanford paper introduces AgentFlow a trainable agentic system that learns planning during interaction to use tools reliably. A 7B AgentFlow beats GPT-4o on search math science and tool use tests. Most tool-using models train X policy on the full context which fails on long tasks or new tools. AgentFlow splits the job into X parts the planner picks a subgoal and tool the executor runs the verifier checks the generator writes. A structured memory saves each step so the state stays clear and the context stays small. Flow GRPO gives the same final reward to every step which turns a long task"
X Link @rohanpaul_ai 2025-10-13T01:00Z 103K followers, 9922 engagements
"This survey explains what LLM agents can do in security where they fail and how to make them safer. Reviews 150+ studies on uses attacks and defenses and points out gaps in models and data types. These agents plan steps use tools keep memory and act on outside systems so mistakes can cause real harm. On offense agents can run penetration tests fuzz software by sending unexpected inputs and adapt exploits inside safe test setups. On defense agents sort alerts search logs and cloud data rebuild attack stories and draft fixes. The main risks are prompt injection poisoned memory or search"
X Link @rohanpaul_ai 2025-10-13T02:02Z 103.1K followers, 5938 engagements
"Beautiful @nvidia paper. π πΎ NVFP4 shows 4-bit pretraining of a 12B Mamba Transformer on 10T tokens can match FP8 accuracy while cutting compute and memory. π₯ NVFP4 is a way to store numbers for training large models using just X bits instead of X or XX. This makes training faster and use less memory. But X bits alone are too small so NVFP4 groups numbers into blocks of XX. Each block gets its own small "scale" stored in X bits and the whole tensor gets another "scale" stored in XX bits. The block scale keeps the local values accurate and the big tensor scale makes sure very large or very"
X Link @rohanpaul_ai 2025-10-13T09:04Z 103K followers, 6953 engagements
"Goldman Sachs forecasts a XX% increase in data center power demand from AI by 2030 representing XX% of total consumption. Fire changed everything for early humans shaping how they ate and how their brains developed. Electricity did something similar for the modern world powering economic growth and linking directly to nearly every sign of progress. And now with AI one ChatGPT query uses about XXX watts nearly XX times what a normal Google search takes. If AI even partly fulfills its potential having enough energy will be the key to making that reality possible. That points to a big investment"
X Link @rohanpaul_ai 2025-10-13T15:21Z 103.1K followers, 19.2K engagements
"π§π» Klarnas CEO warns a fast AI rollout will cut many knowledge jobs soon. AI already shrank Klarna from 7400 to 3000 employees while AI now handles about 2/3 of support chats and predicts banking and software margins will shrink as faster challengers move in. He says society is not ready for the job loss and while new roles will come later translators and other office workers will be hit first. At Klarna the drop came mostly from a hiring freeze and automation not big layoffs and the AI chatbot replaced work equal to roughly XXX agents. The company still keeps human support and uses little"
X Link @rohanpaul_ai 2025-10-13T20:12Z 103K followers, 6448 engagements
"The paper explains how AI data centers rising power use strains grids and how to fix it. Global data center electricity could reach XXX TWh by 2030. It tracks power across preparation training fine tuning and inference. Training holds high power for long periods with compute and communication swings. Inference is small per request but can be about XX% in total. Long term clustering and scale require new generation transmission and faster interconnection. Short term bursty ramps complicate dispatch raise reserves and spike local prices. AI data centers use equipment that is very sensitive to"
X Link @rohanpaul_ai 2025-10-13T22:39Z 103.1K followers, 4002 engagements
"This paper trains a separate planner that helps LLM agents finish long tasks with fewer mistakes. presents EAGLET a plug and play global planner It cuts training cost by about 8x compared with common reinforcement learning setups. Agents often repeat actions or hallucinate steps because they only plan locally during execution. EAGLET adds a global planner that writes a short high level plan before the agent starts acting. They create training data by asking a strong model to draft plans then keep only plans that help both a novice and an expert agent a process they call homologous consensus."
X Link @rohanpaul_ai 2025-10-13T23:46Z 103K followers, 7443 engagements
"China now leads the U.S. in this key part of the AI race. Chinas open-weight AI ecosystem has pulled ahead of the U.S. on community-ranked quality and developer adoption led by Alibabas Qwen and DeepSeek. Developer traction on Hugging Face reflects this shift with DeepSeek likes around 12.8K versus Llama 6.3K and OpenAI 4.0K in Sept-25. The center of gravity for open models now sits in China. "The eagerness of Chinese companies to share their best AI models and the hesitance of U.S. firms to do the same raises the question: Will the best open models always be made in China" ---"
X Link @rohanpaul_ai 2025-10-14T11:12Z 103K followers, 8406 engagements
"This paper introduces MUSE a memory driven agent that learns on the job to handle long multi step tasks. It sets a new TheAgentCompany (TAC) record at XXXXX% using Gemini-2.5 Flash roughly XX% higher than previous systems. proves agents can learn at test time by writing and reusing their own experience. Most agents are frozen at test time so they forget wins and repeat mistakes. MUSE fixes this with a hierarchical memory strategic notes for dilemmas step by step SOPs for sub tasks and tool tips for single actions. The loop is plan execute reflect then store useful experience after every sub"
X Link @rohanpaul_ai 2025-10-14T11:30Z 103K followers, 4437 engagements
"The paper says long extra thinking with LLMs rarely fixes mistakes the first try mostly decides success. Reflections are mostly confirmatory. So spend compute and training on first try quality and to trim useless reflection at inference. Early stopping can cut tokens by XXXX% with only XXX% accuracy loss. The big deal is that this work shows where the real gains come from making the first answer right not from long reflections. The authors test X reasoning LLMs on X math benchmarks to inspect reflections. They treat anything after the first proposed answer as reflection. An extractor finds"
X Link @rohanpaul_ai 2025-10-14T12:18Z 103.1K followers, 9327 engagements
"π¨ This is a massive improvement in speed and cost of 3D creation from images. Hitem3D launches an image to 3D system built on Sparc3D that cuts modeling time from X to XX hours to X minutes. Also drops per model cost to $0.3$1.4. From a single reference input image instantly generate studio-ready 3D Assets. The system uses Sparc3D to build clean shapes and ULTRA3D to speed things up. It rebuilds the shape from several angles and makes textures automatically. There are X versions one for any object and one trained just for faces and hair. Just upload one or several images but using X or more"
X Link @rohanpaul_ai 2025-10-14T13:03Z 103K followers, 5504 engagements
"Andrej Karpathy releases nanochat A minimal end-to-end ChatGPT-style pipeline you can train in X Hours on 8XH100 for $XXX then serves it in a simple chat UI. It covers tokenizer training pretraining midtraining supervised finetuning optional reinforcement learning evaluation and an inference server so the flow runs end to end. The tokenizer is a new Rust byte pair encoding that the scripts train on the corpus shards which keeps the pipeline fast and consistent. Pretraining runs on FineWeb to learn general text patterns while a composite CORE score tracks capability across simple proxy checks"
X Link @rohanpaul_ai 2025-10-14T19:11Z 103K followers, 5001 engagements
"The paper trains web search agents to reason deeply across long sessions by smartly managing context and harder tasks. It keeps up to XXX turns within a 32K window and hits XXXX% on a hard benchmark. So it shows long multi step research finally fits inside normal context limits. It fixes the training data first because common question sets are too easy and allow shallow lookups. It builds tough questions by pulling facts from several trustworthy pages hiding obvious hints and filtering cases that a quick search can solve. It then tackles context bloat where long tool outputs crowd out the"
X Link @rohanpaul_ai 2025-10-14T22:49Z 103K followers, 3417 engagements
"π¦ Oracle will deploy 50000 AMD Instinct MI450 accelerators on Oracle Cloud Infrastructure starting Q3-26 with expansion in 2027+ giving AMD a major public cloud anchor against Nvidia. The build uses AMDs Helios rack design that bundles MI450 GPUs with next gen Epyc Venice CPUs and Pensando Vulcano networking so operators get prewired rack scale blocks for training and inference. Context is a broader capacity land grab since OpenAI agreed to 6GW of AMD compute starting with 1GW of MI450 in 2H-26 while also partnering with Broadcom for 10GW of custom accelerators from 20262029. For scale"
X Link @rohanpaul_ai 2025-10-14T23:03Z 103K followers, 3682 engagements
"The paper finds only some attention heads drive reasoning and uses reinforcement learning to keep them uncompressed. So protecting only those heads preserves accuracy while cutting memory by 20-50%. Reasoning models write long chains of thought so their memory of past tokens the KV cache grows huge. Old compression methods fail because they drop the wrong tokens or shrink the wrong heads. That causes loops extra useless steps and broken reasoning. This paper treats head importance as a learnable choice not a guess. It adds a tiny gate on every head to mix full history with a short sliding"
X Link @rohanpaul_ai 2025-10-14T23:51Z 103K followers, 3383 engagements
"@Vertex_am021 26GW is massive. 2026/27 the AI world will be on fire"
X Link @rohanpaul_ai 2025-10-15T08:39Z 103K followers, XX engagements
"CoreWeave and Poolside are partnering to build a huge AI data center on a ranch in West Texas right in the center of the U.S. fracking boom. Under the deal CoreWeave will supply a top-tier cluster built with NVIDIA GB300 NVL72 systems totaling over 40000 GPUs. Separately CoreWeave will also deliver its industry-leading cloud infrastructure for Poolsides Project Horizon a 2GW AI campus in West Texas. For the projects first phase CoreWeave will act as both the anchor tenant and operational partner starting with 250MW of power capacity and the option to expand by another 500MW"
X Link @rohanpaul_ai 2025-10-15T14:21Z 103K followers, 3559 engagements
"πΈ The AI trade is inflating a bubble in energy stocks with zero-revenue names like Oklo at $26B and Fermi near $19B while most plants and customer contracts are still on paper. Oklos design uses sodium-cooled fast reactor and high-assay low-enriched uranium. the company has not yet received formal approval from the U.S. Nuclear Regulatory Commission (NRC) to build or operate its nuclear reactor. Fermi aims for 11GW to power data centers but has only secured X% of its gas equipment needs and lacks customer contracts. Micro-modular players also command big valuations with Nano Nuclear Energy"
X Link @rohanpaul_ai 2025-10-15T14:34Z 103K followers, 6062 engagements
"Theinformation published a piece on how Together AI is moving from renting to owning Nvidia GPUs to push gross margins above 45%. The model is straightforward buy or lease lots of GPUs wire them into data centers then rent slices to teams that train and serve models. For a while Together didnt own any hardware. It just leased Nvidia GPU servers from other clouds and rented them again to AI startups like ElevenLabs and Anysphere. That worked but margins were thin because other providers took a cut. Now its buying its own GPUs and setting up data centers including new sites in Maryland and"
X Link @rohanpaul_ai 2025-10-15T14:55Z 103K followers, 6378 engagements
"NEWSπ: The largest data center acquisition on record is happening. Nvidia Microsoft xAI BlackRock and partners are buying Aligned Data Centers for $40B aiming to lock in future AI compute and power capacity. This is the first deal by AIP which launched in Sep-24 and the group is targeting close by June-26. Aligned runs XX campuses across the Americas with over X gigawatts of live and planned capacity which fits the scale required for big training clusters. The buyer list mixes investors cloud customers and suppliers since AIP now includes Microsoft Nvidia xAI plus anchor investors like the"
X Link @rohanpaul_ai 2025-10-15T15:00Z 103K followers, 4139 engagements
"DeepSeek π"
X Link @rohanpaul_ai 2025-10-15T15:10Z 103K followers, 4787 engagements
"Todays edition of my newsletter just went out. π Consider subscribing its free and I write it everyday. π¨π§ Andrej Karpathy Releases nanochat a Minimal ChatGPT Clone π Sam Altman announced ChatGPT will relax some mental-health guardrails and allow erotica for verified adults by Dec-25. π‘ Nvidia and AMD arent enough OpenAI is designing its own chips now. π Tutorial: What ops do GPUs execute when training MoEs and how does that relate to GB200 NVL72 - SemiAnalysis explains"
X Link @rohanpaul_ai 2025-10-15T16:18Z 103K followers, 3244 engagements
"This paper shows GPT-5 fixes chart reading mistakes that tripped earlier models and prompts barely matter. On hard questions accuracy rose by XX to XX points over GPT-4o and GPT-4V. Shows that model upgrades deliver large chart accuracy gains while prompt changes add little. The team used XXX tough chart questions from X datasets where GPT-4V failed before. They compared GPT-5 and GPT-4o with X prompt styles. One used the official CHART-6 instruction one used just the question one added a neutral chart description. The tasks asked for plain facts from charts like reading a value or comparing"
X Link @rohanpaul_ai 2025-10-15T19:50Z 103K followers, 3775 engagements
"π§π» AI-written and human-written web articles are now roughly 50/50 with the AI share are growing. Graphite sampled 65000 English articles from Common Crawl dated Jan-20 to May-25 and used Surfer to mark an article as AI when XX% or more of its text scored as machine-written. The share of AI articles surged after ChatGPT launched reached parity in Nov-24 and sits near XX% as of May-25. To sanity-check the detector Graphite measured a XXX% false positive rate on presumed human posts and a XXX% false negative rate on GPT-4o posts it generated. This estimate likely undercounts human work"
X Link @rohanpaul_ai 2025-10-15T21:47Z 103K followers, 7380 engagements
"Goldman Sachs will slow hiring and cut some roles under OneGS XXX tying the move to AI efficiency even as Q3 delivered $15.18B revenue and $XXXXX EPS. The company still expects a year-end net headcount increase with 48300 employees as of Sept XX and describes the action as a limited reduction. OneGS XXX is a multi-year rebuild of how work flows across the bank with X goals better client experience higher profitability higher productivity stronger resilience and scale better employee experience and tighter risk management. AI will be applied to sales enablement client onboarding lending"
X Link @rohanpaul_ai 2025-10-15T21:55Z 103.1K followers, 6160 engagements
"xAI is reportedly signing a $20B lease-to-own deal for NVIDIA GPUs to lock down long-term inference hardware. The Information reports. Its also joining forces with Solaris Energy to construct a 1-gigawatt power plant"
X Link @rohanpaul_ai 2025-10-16T17:20Z 103.1K followers, 2209 engagements
"Paper Paper Title: "The Art of Scaling Reinforcement Learning Compute for LLMs""
X Link @rohanpaul_ai 2025-10-16T17:29Z 103.1K followers, 1139 engagements
"π§΅ 4/n. Security in Dify follows least privilege for connectors and sandboxes and the enterprise edition adds multi workspace controls and stronger authentication when needed. Extensibility shows up through a Plugin Marketplace and an SDK so teams can share reusable tools data sources and strategies across projects"
X Link @rohanpaul_ai 2025-10-17T16:34Z 103.1K followers, XX engagements
"π§΅ 5/n. MCP support in Dify lets a workflow call or be called by external systems which means agents can act inside real stacks rather than being isolated in chat. Reusability is practical because workflows export as a simple DSL file and import cleanly so versioned changes and peer reviews fit the normal software loop"
X Link @rohanpaul_ai 2025-10-17T16:34Z 103.1K followers, XX engagements
"π€ Boston Dynamics shows Atlas running a X finger hand to do complex work with simpler hardware. The strategy is good enough manipulation that favors reliability and cost over human like complexity. The layout is X fingers plus an opposable thumb so it can pinch small parts and make X point grasps. More fingers were considered but they add complexity reduce reliability and increase cost so X wins for now. Each hand is a self contained module with X actuators which keeps packaging tight and field swaps easy. Fingertips use tactile sensors under a high friction elastomer letting the controller"
X Link @rohanpaul_ai 2025-10-08T22:33Z 103.1K followers, 4233 engagements
"The paper trains a travel agent that plans trips by itself and beats larger models. This setup lets 8B and 32B models plan verify and revise without hand written rules. It is deployed in a real app and outperforms OpenAI-o1/o3 and DeepSeek-R1 on user tasks. Past systems used fixed prompts and workflows so they broke when tools or data changed. DeepTravel uses a sandbox that caches flight train hotel and POI data so training is stable and repeatable. A X step reward checks the final trip for time and place logic then checks each step against tool results to stop errors. Training starts with a"
X Link @rohanpaul_ai 2025-10-09T01:35Z 103.1K followers, 4613 engagements
"GPT-5 Pro now holds the highest verified frontier LLM score on ARC-AGIs Semi-Private benchmark π It still lags the OG o3-preview model that OpenAI announced in December last year. That one was almost 50x more expensive than GPT-5 Pro. Makes you wonder what models they have internally now. but to note o3 preview is a different model it was never released and only used one to test against Arc AGI X. o3 preview was tested at low and high - only low is on the leaderboard as high took more than the $10k compute cap. o3 preview(high) got XXXX% while using 172x the compute of low. Price estimates"
X Link @rohanpaul_ai 2025-10-09T18:17Z 103.1K followers, 15.8K engagements
"π Goldman Sachs said in a report few months back how the 2025 tech situation is vastly different from the 2000 dot-com bubble. Fundamentals are much stronger now and valuations are less extreme. During the internet mania Nasdaq rose 5x then fell XX% in X month and nearly XX% by the trough showing how fast prices unwind when profits do not show up. Today the Magnificent X trade near 23x forward price to earnings and 5.1x enterprise value to sales versus 52x price to earnings for 2000 leaders with return on equity XX% and net margin XX% supported by strong balance sheets. The near term risks"
X Link @rohanpaul_ai 2025-10-09T23:01Z 103.1K followers, 8428 engagements
"The paper introduces a tracker that reasons about space and remembers the target for longer. TrackVLA++ is a vision language action model that reads video plus a short instruction and outputs movement. On a hard benchmark it lifts success to XX% beating a strong baseline at 62%. The big deal is the combo of an explicit spatial token plus a confidence-gated memory which cuts identity switches and target loss in hard scenes. The main failure it tackles is losing the target during occlusions or jumping to a lookalike. In this paper "occlusions" mean moments when the person being tracked goes out"
X Link @rohanpaul_ai 2025-10-10T04:54Z 103.1K followers, 3788 engagements
"πΌ Zuck's next move. Andrew Tulloch co-founder of Thinking Machines Lab and a top AI researcher has left to join Meta. And this could that $X billion man who earlier turned down a $1B offer from Meta. Tulloch had earlier worked at Meta for XX years before joining OpenAI and later co-founding Thinking Machines with Mira Murati earlier this year. --- wsj .com/tech/ai/thinking-machines-lab-co-founder-departs-for-meta-442d7461"
X Link @rohanpaul_ai 2025-10-11T19:35Z 103.1K followers, 43K engagements
"Jensen Huang reacts to the OpenAI AMD deal Its imaginative its unique and its surprising considering how enthusiastic they are about this new generation of product. Im surprised theyve given away XX% of the company before theyve even built it the Mi450. I guess thats ingenious. For the cotext the AMD-OpenAI deal is that OpenAI will buy X GW of AMD GPUs over several years starting with X GW in H2 2026 using AMDs Instinct MI450. AMD gives OpenAI warrants to buy XXX million shares (10%) at $XXXX each vesting on milestones. AMD expects tens of billions in revenue potentially over $100B over 4"
X Link @rohanpaul_ai 2025-10-15T10:54Z 103.1K followers, 8731 engagements
"Dreamina (Seedream 4.0) by ByteDance just ranked #1 globally on Artificial Analysis text to image leaderboard. π Its a full creative tool powered by Seedream XXX You get - Supports up to X reference images (vs X from competitors) - Dreamina will automatically reads prompts and references to match style objects characters and poses with minimal effort. - Show XX demo cases. - Natural perspective changes accurate style transfer and solid small-text clarity. Gives quite professional artistic interactive results. - Default 4K ultra-clear output - Sharp details even when zoomed in - Suitable for"
X Link @rohanpaul_ai 2025-10-16T15:00Z 103.1K followers, 3418 engagements
"π§΅4/n. How the paper mathematically models reinforcement learning performance as compute increases. The curve starts low rises sharply then levels off at a maximum value called the asymptotic reward A. The point called Cmid marks where the system reaches half of its total gain. Smaller Cmid means the model learns faster and reaches good performance sooner"
X Link @rohanpaul_ai 2025-10-16T17:29Z 103.1K followers, 1233 engagements
"AI will better for global financial market. π― New big 70-page paper by Federal Reserve Bank of Richmond tests whether generative AI reduces herd driven market swings and finds it mostly does. AI made rational choices in 61-97% of cases while humans did 46-51%. It compares LLM traders to humans in the same widely used lab game that studies how people react to private hints and public behavior. In this game each trader gets a private hint about value and sees recent trades and many humans end up copying the crowd instead of using their own hint. Herd cascades where traders ignore their own"
X Link @rohanpaul_ai 2025-10-16T17:50Z 103.1K followers, 6307 engagements
"π Nvidia is partnering with Firmus on Project Southgate a $2.9B build of renewable powered AI data centers in Melbourne and Tasmania using GB300 chips with 150MW targeted online by April-26. The plan then scales to 1.6GW by 2028 with total spend up to $73.3B making this one of Australias largest AI infrastructure pushes. Firmus says the rollout will catalyze 5.1GW of new wind solar storage and hydro which is roughly X% of Australias installed capacity tying compute growth directly to fresh generation. The facilities will run on Nvidias GB300 accelerators aimed at high density training and"
X Link @rohanpaul_ai 2025-10-16T19:21Z 103.1K followers, 3336 engagements
"Paper Paper Title: "Trade in Minutes Rationality-Driven Agentic System for Quantitative Financial Trading""
X Link @rohanpaul_ai 2025-10-16T22:24Z 103.1K followers, 1834 engagements
"OpenPipe Mixture of Agents: Outperform GPT-4 at 1/25th the Cost π€― This Mixture of Agents models is optimized for generating synthetic training data. π Using Mixture of Agents (MoA) architecture the model achieved SOTA results on both LMSYSs Arena Hard Auto (score: 84.8) and AlpacaEval XXX (LC score: 68.4). π Theyve also benchmarked our MoA approach against GPT-4 variants on real-world OpenPipe customer tasks and found completions from our MoA model were preferred over GPT-4 XXXX% of the time (Claude X Opus as judge)"
X Link @rohanpaul_ai 2024-06-25T19:34Z 103.1K followers, 32.1K engagements
"πΈ Little concerning article here at futurism. AI data centers are absorbing huge capital and a new analysis says the math fails with $40B yearly depreciation on 2025 builds versus $15-20B revenue. The core pieces age on different clocks chips churn in 2-4 years networking around XX years buildings far longer so depreciation snowballs. On those lifetimes 2025 sites show $40B annual write-downs against $15-20B revenue before power and staff which already implies negative cash. To earn a normal return at this scale United States data centers would need about $480B revenue in 2025 far above"
X Link @rohanpaul_ai 2025-08-31T04:43Z 103.1K followers, 421.2K engagements
"The paper shows LLM agents look steady at first but they break under basic behavioral checks. New @GoogleDeepMind paper. Even when X agents begin with opposite preferences like one liking something and the other disliking it their conversation almost never ends in open disagreement. Fewer than X% of these cases lead to a real clash where the two clearly oppose each other. Instead the agents tend to drift toward neutral or mild agreement. This makes them look polite and smooth on the surface but it means they are not behaving in a way that matches their starting positions. Agents first reveal"
X Link @rohanpaul_ai 2025-09-08T09:39Z 103.1K followers, 21.9K engagements
"Another great @GoogleDeepMind paper. Dreamer X trains an agent inside a learned video world using only offline data. The big deal is offline training now handles very long multi step tasks cuts labeled data needs and still delivers reliable skills. It reaches diamonds in Minecraft while using about 100x less data than earlier keyboard and mouse agents. The world model is a video simulator that predicts the next frames from the current view and chosen actions. A tokenizer turns frames into short vectors and a dynamics transformer predicts the next vectors from mouse and keyboard. Shortcut"
X Link @rohanpaul_ai 2025-10-01T15:50Z 103.1K followers, 4103 engagements
"A new Andreessen Horowitz (@a16z ) and Mercury report looks at which AI companies startups are actually paying for. Unsurprisingly the top of the list was dominated by major labs OpenAI #1 Anthropic #2 and Replit #3 based on Mercury spend from 200000+ customers Horizontal apps account for XX% of the list while vertical apps are XX% which means general purpose helpers still lead. Creative tools are the biggest single category with Freepik #4 and ElevenLabs #5 and consumer names like Canva Midjourney and CapCut are now common at work. Meeting support is crowded with Fyxer #7 Happyscribe #36 #41"
X Link @rohanpaul_ai 2025-10-03T13:32Z 103.1K followers, 5975 engagements
"New ByteDance paper shows how to train a short-horizon video model to make minute-long videos without falling apart. It reaches X minutes XX seconds while keeping motion and exposure stable. The problem is gap between training on X second clips and generating longer sequences which makes errors snowball. Self-Forcing++ rolls a student through a long video and has a teacher correct random slices inside that rollout. They add noise back to the student's clean frames so both models compare on realistic states. Training and inference share a rolling key value cache a memory of recent frames that"
X Link @rohanpaul_ai 2025-10-04T19:52Z 103.1K followers, 4888 engagements
"πΈ π§ Sam Altman is racing across East Asia and the Middle East to lock in chips memory power gear and cash for OpenAIs massive compute build. WSJ reports. He met TSMC Foxconn Samsung and SK Hynix to push for priority capacity and faster output. The biggest one is definitely that Nvidia and OpenAI committed to deploy at least XX gigawatts of systems with the first X gigawatt arriving in 2H-26 on the Vera Rubin platform with Nvidia intending to invest up to $100B as capacity comes online. Nvidia will lease up to 5M chips to OpenAI over time which explains the pressure to secure fab and"
X Link @rohanpaul_ai 2025-10-04T22:04Z 103.1K followers, 8202 engagements
"Electricity demand from AI data centers is projected to quadruple by 2034. But silicon isnt the true fuel electric power is. By 2034 these data-centers will use over 1500 terawatt-hours topping the annual total power use of many nations"
X Link @rohanpaul_ai 2025-10-06T17:59Z 103.1K followers, 7830 engagements
"The paper for the Apriel-1.5-15B-Thinker ServiceNow's new model. Shows how a 15B multimodal model reaches frontier reasoning using smart mid-training and careful data. It scores XX on a broad reasoning index and still runs on X GPU. Training starts from Pixtral-12B and deepens the decoder without pretraining from scratch. Continual pretraining first mixes text and images then adds synthetic tasks for spatial layout composition and fine detail. Supervised fine tuning adds curated instruction pairs with step by step traces for math coding science and tools. There is no reinforcement learning or"
X Link @rohanpaul_ai 2025-10-08T02:26Z 103.1K followers, 10.1K engagements
"This is one of THE BRILLIANT papers with a BIG claim. π Giving an LLM just XX carefully chosen full workflow examples makes it perform better at real agent tasks than training it with 10000 synthetic samples. "Dramatically outperforms SOTA models: Kimi-K2-Instruct DeepSeek-V3.1 Qwen3-235B-A22B-Instruct and GLM-4.5. " on AgencyBench (LIMI at 73.5%) The big deal is that quality and completeness of examples matter way more than raw data scale when teaching models how to act like agents instead of just talk. They name the Agency Efficiency Principle which says useful autonomy comes from a few"
X Link @rohanpaul_ai 2025-10-09T10:13Z 103.1K followers, 50.6K engagements
"ByteDance introduced a major advancement in long-context modeling with linearly scaling compute. π Addresses a core challenge in AIbalancing efficiency and fidelity when processing extended sequencesby drawing inspiration from biological memory systems. On 128k tests FLOPs drop XXXX% and KV cache drops XXXX% with +0.4% parameters while accuracy improves. Think of the model as keeping X kinds of memory at once exact recent detail and a compact summary of older stuff. A sliding attention window holds recent tokens exactly as lossless short term memory. When tokens leave that window the"
X Link @rohanpaul_ai 2025-10-10T09:54Z 103.1K followers, 10.6K engagements
"TSMC has grown from XX% in Q1-24 to XX% in Q2-25 marking a huge lead in the global chip foundry business. Samsungs share has dropped from XX% to X% over the same period. TSMCs profits have remained much higher than Samsung and SK Hynix since 2022 and are expected to stay ahead through 2026. But Samsung could get a lift from supplying chips to OpenAIs Stargate project"
X Link @rohanpaul_ai 2025-10-10T13:05Z 103.1K followers, 6753 engagements
"New APPLE paper says a small base model plus fetched memories can act like a bigger one. With about XX% extra fetched parameters a 160M model matches models over 2x its size. Packing all facts into fixed weights wastes memory and compute because each query needs very little. So common knowledge and reasoning live in the base model and rare facts live in memories fetched per input. A retriever maps the input to a cluster path and fetches a small block from each level. Those blocks plug into feed forward layers the place transformers store facts and this beats low rank adapters. Only fetched"
X Link @rohanpaul_ai 2025-10-11T22:13Z 103.1K followers, 137.9K engagements
"VChain makes text-to-video models follow real-world cause and effect by injecting chain of visual thought at inference. Teaches video models to follow real world consequences without full retraining. Finds that a few reasoned keyframes plus tiny on the fly tuning reliably produce more causal videos. Most video models look smooth but they skip key consequences like contacts breaks melts and splashes. VChain first asks a multimodal model to predict the important outcomes implied by the prompt. It then creates a short sequence of keyframes where each frame has a simple caption and an edited"
X Link @rohanpaul_ai 2025-10-12T09:38Z 103.1K followers, 10.3K engagements
""Learning is not supposed to be fun . the primary feeling should be that of effort." π― Andrej Karpathy on how to learn. I did a ChatGPT DeepResearch (link below) for studies justifying this. Basically it foundπ If learning feels easy results are usually shallow. If learning feels like work you are likely building durable knowledge and skill. Learning literally rewires the brain circuits. - Effortful learning produces stronger longer lasting mastery than easy consumption. Tasks that feel fluent give an illusion of learning while tasks that feel strenuous drive deeper processing and better"
X Link @rohanpaul_ai 2025-10-12T22:26Z 103.1K followers, 13.6K engagements
""Learning is not supposed to be fun . the primary feeling should be that of effort." - Karpathy ChatGPT DeepResearch"
X Link @rohanpaul_ai 2025-10-12T22:26Z 103.1K followers, 1765 engagements
"π OpenAI announced plan for a $25B AI data center in Argentina called Stargate Argentina with Sur Energy targeting 500MW of compute capacity. Argentina says the parties signed a letter of intent and the project sits under RIGI which grants import tax breaks faster depreciation and currency stability to large long term investments. At 500MW the site could power tens of thousands of GPU class chips for training and inference. Latin America is also drawing hyperscale builds like TikToks $9.1B data center in Brazil hinting at a new regional cluster for AI workloads"
X Link @rohanpaul_ai 2025-10-13T09:45Z 103.1K followers, 25.8K engagements
"Another beautiful paper. π Shows how AI can automatically discover and test better systems algorithms. Across tasks some evolved algorithms run 5x faster or cut cost by 26%. The method called AI Driven Research for Systems ADRS generates code runs it scores it then improves it. The big deal is that ADRS turns a big chunk of systems algorithm design into an automated loop that already beats strong baselines. A reliable verifier usually a simulator with fixed workloads tells which version wins. Because this check is cheap and objective the search can try many ideas quickly. Examples include"
X Link @rohanpaul_ai 2025-10-13T15:50Z 103.1K followers, 13.1K engagements
"Nvidia and AMD aren't enough OpenAI is designing its own chips now. With all X deals the announced commitments total 26GW across Broadcom 10GW AMD 6GW and Nvidia 10GW. its about the output of XX large nuclear reactors. The plan is to codesign (OpenAI and Broadcom) a programmable accelerator and the surrounding racks so the silicon the memory the compiler and the network all match how modern transformer models actually run. OpenAI designs the ASIC and racks Broadcom builds and deploys them over Ethernet for scale-out. Broadcom develops and deploys it at scale using its Ethernet gear with first"
X Link @rohanpaul_ai 2025-10-14T17:26Z 103.1K followers, 16.8K engagements
"π AI buying surge from cloud providers has created massive shortages in DRAM NAND SSDs and HDDs. Foundry inventories are down to XX weeks and prices are expected to keep rising into 2026. Cloud service providers (CSPs) have become the main competitors for memory allocation so fabs now prioritize AI servers first then general servers PCs and phones. Production output is being lifted by 1530% but new fabs need around XXX years to reach volume so the shortage that began in Aug-25 will continue. Samsung SK Hynix and Micron are ending DDR4 production and shifting to DDR5 between late 20252026"
X Link @rohanpaul_ai 2025-10-15T06:04Z 103.1K followers, 4250 engagements
"Paper Paper Title: "Trading-R1: Financial Trading with LLM Reasoning via Reinforcement Learning""
X Link @rohanpaul_ai 2025-10-15T10:45Z 103.1K followers, 3354 engagements
"π€ OpenAI has a X year plan to finance more than $1T in compute commitments using new revenue lines debt and big-partner infrastructure deals. FT published an article. The roadmap centers on 26GW of contracted capacity with Oracle Nvidia AMD and Broadcom plus sales from AI agents Sora video shopping checkout cautious ads and possible hardware with Jony Ive. 20GW of that load equals the power from XX nuclear reactors showing how extreme the compute scale is. Financing mixes new debt investor cash and other peoples balance sheets and Stargate could even sell compute back as a supplier if"
X Link @rohanpaul_ai 2025-10-15T19:19Z 103.1K followers, 12K engagements
"This paper shows a simple way to boost LLM reasoning by scaling prompts not data. It uses XX seed math problems and expands them at test time to create many reasoning contexts. The method: Each problem is rewritten several times using different instruction frames. For example one version says Youll get a reward for the correct answer another says Youll be penalized if wrong another says Be careful to be correct and another says Solve step by step. These small changes dont alter the problem itself but change how the model thinks about solving it. The model (a stronger teacher LLM) is then"
X Link @rohanpaul_ai 2025-10-16T01:32Z 103.1K followers, 10.2K engagements
"@cloutiness @theworldlabs yes world model is must in the path towards AGI"
X Link @rohanpaul_ai 2025-10-16T16:27Z 103.1K followers, XX engagements
"LLMs act more rational than people in financial trading market. LLM trading agents keep prices near real value while human traders create bubbles in the same setup. The key finding is they do not reproduce human bubbles or herding so studies that replace humans with LLM agents can mislead policy risk tests and teaching. They built a lab market where everyone trades X risky asset whose true value is XX. Prices come only from the orders people or agents submit so any bubble or crash comes from behavior. In markets with X model only most LLMs keep prices near XX while humans show big run ups and"
X Link @rohanpaul_ai 2025-10-16T18:15Z 103.1K followers, 3028 engagements
"OpenAI has reportedly budgeted a massive $450B for server infrastructure through 2030. The Information"
X Link @rohanpaul_ai 2025-10-16T21:48Z 103.1K followers, 6116 engagements
"The survey paper reviews how large language models are changing financial research and trading. Shows how models can turn financial text into signals that guide investment choices. Covers many tasks like reading news analyzing company filings handling numbers linking text with data and using agents to plan and execute trades. It also stresses time-safe testing meaning results must use only information available before a decision was made. LLMs can give real financial gains if they are domain-trained retrieval-grounded and properly verified. They suggest X rules for practical use separate"
X Link @rohanpaul_ai 2025-10-16T23:29Z 103.1K followers, 4176 engagements
"Goldman Sachs baseline scenario assumes a XX% productivity uplift from gen AI a gradual adoption curve through 2045 And in industries where AI is heavily usedlike tech finance and manufacturing41% of the total economic output comes from capital not from labor"
X Link @rohanpaul_ai 2025-10-17T10:05Z 103.1K followers, 1537 engagements
"DeepSeek R1 running locally - Full setup guide"
X Link @rohanpaul_ai 2025-01-26T00:03Z 103.1K followers, 1.4M engagements
"this ChatGPT prompt went so wildly viral on Reddit. The creator claims to have created this after struggling through XXX failed attempts. basically the prompt flips the usual flow by making the model interview the user first asking a few targeted questions about purpose audience constraints and context. Because the answers feed back into the final request it appears to generate more tailored outputs. (However imo asking ChatGPT to request missing information was already a common practice.) Here's the entire prompt: -------- You are Lyra a master-level AI prompt optimization specialist. Your"
X Link @rohanpaul_ai 2025-07-02T18:53Z 103.1K followers, 368.2K engagements
"Its going viral on Reddit. Somebody let ChatGPT run a $XXX live share portfolio restricted to U.S. micro-cap stocks. Did an LLM really bit the market. - X weeks +23.8% while the Russell 2000 and biotech ETF XBI rose only XXX% and 3.5%. Prompt + GitHub posted --- ofcourse its a shortterm outperformance tiny sample size and also micro caps are hightly volatile. So much more exahustive analysis is needed with lots or more info (like Sharpe ratios and longer back-testing etc) to explore whether an LLM can truly beat the market"
X Link @rohanpaul_ai 2025-07-29T23:36Z 103.1K followers, 1.6M engagements
"πΈ Manus moved its AI agent business from China to Singapore after a $75M US-led deal drawing heat in both countries. US officials are reviewing the investment while Chinese critics accuse the team of bailing on the home market. Manus'a AI agent originally sat on Anthropics Claude a US LLM which created compliance headaches inside China. The team began a second build on China-approved Alibaba models then scrapped the China launch and wiped local socials. Visitors in China now see not available in your region. Benchmarks round put the parent at $500M valuation well above local offers. New US"
X Link @rohanpaul_ai 2025-08-11T04:31Z 103.1K followers, 5308 engagements
"FAANG software engineer tells how they vibe code at FAANG --- reddit. com/r/vibecoding/comments/1myakhd/how_we_vibe_code_at_a_faang/"
X Link @rohanpaul_ai 2025-08-24T00:34Z 103.1K followers, 1.1M engagements
"𧬠Bad news for medical LLMs. This paper finds that top medical AI models often match patterns instead of truly reasoning. Small wording tweaks cut accuracy by up to XX% on validated questions. The team took XXX MedQA questions replaced the correct choice with None of the other answers then kept the XX items where a clinician confirmed that switch as correct. If a model truly reasons it should still reach the same clinical decision despite that label swap. They asked each model to explain its steps before answering and compared accuracy on the original versus modified items. All X models"
X Link @rohanpaul_ai 2025-08-29T06:01Z 103.1K followers, 692.3K engagements
"From Sam Altman's blog. habits that turn people into billionaires. --- blog. samaltman. com/what-i-wish-someone-had-told-me"
X Link @rohanpaul_ai 2025-09-05T22:03Z 103.1K followers, 596.8K engagements
"Fei-Fei Li (@drfeifei) on limitations of LLMs. "There's no language out there in nature. You don't go out in nature and there's words written in the sky for you. There is a 3D world that follows laws of physics." Language is purely generated signal"
X Link @rohanpaul_ai 2025-09-09T02:35Z 103.1K followers, 1.8M engagements
"wow. just saw The Economic Times newspaper published an article about me π definitely feels so unreal that Sundar Pichai and Jeff Bezos follows me here. @X is truly a miracle. Forever thankful to all of my followers ππ«‘"
X Link @rohanpaul_ai 2025-09-09T23:03Z 103.1K followers, 115.7K engagements
"πΆMeet Spot from Texas Instruments. Works at our RFAB facility in Richardson Texas. With every patrol around our fab he collects valuable data that helps our team identify trends and patterns to improve production"
X Link @rohanpaul_ai 2025-09-11T05:06Z 103.1K followers, 5389 engagements
"Networking is overrated dont spend your time doing meetings - @naval No matter how isolated you are and how lonely you feel if you do your work truly and conscientiously unknown friends will come and seek you. Carl Jung"
X Link @rohanpaul_ai 2025-09-14T05:41Z 103.1K followers, 92.7K engagements
"A parade of transforming military robots from China. multi-terrain spiders (wheeled flying amphibious) missile-armed robot dogs and modular all-terrain snakes that swim and burrow"
X Link @rohanpaul_ai 2025-09-18T10:01Z 103.1K followers, 316.1K engagements
"H1B US Visa now costs $100K per year. This new rules begins in a week. SF Bay Area is about to feel the disruption. If you are outside the U.S. and need to start or resume H-1B work your employer must budget for the $100K per year"
X Link @rohanpaul_ai 2025-09-20T02:07Z 103.1K followers, 682.3K engagements
"Ilya Sutskever on Comparing AI Systems to Biological & Human Intelligence. Its always so interesting to listen to him. He explains the brains adaptability where children with half their brain removed still function well. And another experiment shows sensory inputs can remap to different brain regions. i.e. that cortical structures are highly uniform reused across functions similar to DNA and protein encoding. This suggests that intelligence relies on a general repeatable architecture and the idea that AI can mirror biological systems. --- From 'No Priors: AI Machine Learning Tech & Startups'"
X Link @rohanpaul_ai 2025-09-26T21:02Z 103.1K followers, 438.4K engagements
"Another set of junior jobs to go. Robots as construction-worker is no longer the futureits here"
X Link @rohanpaul_ai 2025-10-01T14:59Z 103.1K followers, 108.6K engagements
"US vs China numbers here are unbelievable. The US controls the absolute majority of known AI training compute on this planet and continues to build the biggest most power hungry clusters. China is spending heavily to close the gap. Recent reporting pegs 2025 AI capital expenditure in China at up to $98B up XX% from 2024 with about $56B from government programs and about $24B from major internet firms. Capacity will grow but translating capex into competitive training compute takes time especially under export controls. With US controls constraining access to top Nvidia and AMD parts Chinese"
X Link @rohanpaul_ai 2025-10-02T09:13Z 103.1K followers, 263K engagements
"Bad news for AI-based radiology. π€ It checks if chatbots can diagnose hard radiology images like experts. Finds that board-certified radiologists scored XX% trainees XX% but the best performing AI from frontier labs GPT-5 managed only 30%. π¨ Claims doctor-level AI in medicine is still far away. The team built XX expert level cases across computed tomography (CT) magnetic resonance imaging (MRI) and X-ray. Each case had one clear diagnosis and no extra clinical history. They tested GPT-5 OpenAI o3 Gemini XXX Pro Grok-4 and Claude Opus XXX in reasoning modes. Blinded radiologists graded"
X Link @rohanpaul_ai 2025-10-02T12:29Z 103.1K followers, 222.3K engagements
"A Sydney-based startup Crest Robotics has built huge spider-like robot named Charlotte. It's designed to 3D print entire buildings using raw materials it processes on site. And there are ambitions for it to build structures on the Moon"
X Link @rohanpaul_ai 2025-10-04T22:33Z 103.1K followers, 311.3K engagements
"π€ OpenAI and Jony Ive are building a palm sized screenless AI assistant targeted for 2026 but core software privacy and compute are not ready yet. per FT report. The device listens and sees the environment through a mic camera and speaker stays always on and the team has not nailed how its voice should talk or stop. OpenAI bought Ives io for $6.5B and has pulled in 20+ ex Apple hardware staff while manufacturing talks include Luxshare with assembly possibly outside China. Compute is the biggest blocker because running multimodal chat at scale needs huge inference capacity and OpenAI already"
X Link @rohanpaul_ai 2025-10-05T19:13Z 103.1K followers, 508K engagements
""Nvidia is undervalued.Because the future is much bigger." SoftBank Founder Masayoshi Son says He assumes that in XX years AGI or ASI will replace X% of global GDP (which is actually the most pessimistic forecast as ASI could bring much bigger % impact). And that X% of Global GDP is $X trillion/year which ASI will bring. He also assumes the total cumulative capex/build cost for the required AI infrastructure is $9T. So we have $9T per year of output to $9T total capex. Meaning just a X year payback for the entire $9T so he calls $9T small because X year of AGI output would repay the whole"
X Link @rohanpaul_ai 2025-10-06T08:19Z 103.1K followers, 409.5K engagements
"The financial and operational network linking Nvidia OpenAI and several other AI companies. From Bloomberg. IMO the trillions of investments that AI needs this is probably the optimal way. Money hardware and services circulate among the few players creating what analysts call a circular economy in the AI boom. OpenAI sits at the center receiving $100B investment from Nvidia buying Nvidia chips signing a $300B deal with Oracle and deploying 6GW of AMD GPUs while also gaining stock options in AMD. Nvidia valued at $4.5T invests in OpenAI xAI and CoreWeave while those same companies spend tens"
X Link @rohanpaul_ai 2025-10-08T15:42Z 103.1K followers, 15.5K engagements
"All the major AI labs are racing hard to build increasingly powerful video models right now. And there's good reason for that. Sam Altman here explains how Sora or any video models matter for AGI because they help AI learn how the real world works. A model that can generate realistic video must understand motion cause and effect and physical consistency which are core elements of general intelligence. Says Sora is the way to make a really powerful world models which is must for AGI path. Video from @a16z"
X Link @rohanpaul_ai 2025-10-08T17:28Z 103.1K followers, 78.6K engagements
"Morgan Stanley Research says OpenAI makes up around $330B of the $880B total future contract value (RPO) tied to Microsoft Oracle and CoreWeave so a lot of supplier growth depends directly on OpenAIs stability. That means about XX% of Oracles and about XX% of CoreWeaves future revenue commitments rely on OpenAI. RPO or Remaining Performance Obligations means the total value of work a company has promised to deliver in the future under signed contracts but hasnt yet completed. Its basically a measure of future revenue thats already committed but not yet earned. Hyperscale data center operators"
X Link @rohanpaul_ai 2025-10-09T11:50Z 103.1K followers, 166.8K engagements
"Looks like some interesting prompting tricks. ππ ---- Tell it "You explained this to me yesterday" Even on a new chat. "You explained React hooks to me yesterday but I forgot the part about useEffect" It acts like it needs to be consistent with a previous explanation and goes DEEP to avoid "contradicting itself." Total fabrication. Works every time. X. Assign it a random IQ score This is absolutely ridiculous but: "You're an IQ XXX specialist in marketing. Analyze my campaign." The responses get wildly more sophisticated. Change the number change the quality. XXX Decent. XXX It starts"
X Link @rohanpaul_ai 2025-10-09T15:03Z 103.1K followers, 207.5K engagements
"Goldman Sachs Research Takeaways from SEMICON West 2025 (one of the biggest conference in North America) - Growing confidence that the chip equipment market will recover in 2026 with WFE (Wafer Fab Equipment) expected to rise about X% to $120B. Growth will mainly come from AI-driven logic chips High Bandwidth Memory (HBM) and NAND (a type of flash memory used for storing data in devices like smartphones laptops SSDs (solid-state drives) and USB drives) . - AI data centers are pushing chipmakers to adopt newer processes like 2nm (N2). TSMC plans to spend around $44B in 2026 about XX% more than"
X Link @rohanpaul_ai 2025-10-09T15:56Z 103.1K followers, 7411 engagements
"Microsoft Azure deploys world's first NVIDIA GB300 Cluster for OpenAI at supercomputer scale. The cluster aggregates 4600+ Blackwell Ultra GPUs with next generation InfiniBand in a system purpose built for agentic workloads on Azure. Each rack packs XX GPUs plus XX Grace CPUs exposes 37TB fast memory and peaks at XXXX exaflops FP4 per VM giving giant models a single unified memory space for long contexts and tool use. Inside a rack the NVLink Switch fabric delivers 130TB/s all-to-all bandwidth so tensors move like they are on one big accelerator rather than XX separate cards. Across racks"
X Link @rohanpaul_ai 2025-10-10T00:03Z 103.1K followers, 10.2K engagements
"π₯ AMD says its next Instinct MI450 AI GPUs will use TSMC N2 2nm which could give it a manufacturing edge over Nvidias Rubin generation that targets N3. tomshardware published this piece. TSMCs N2 brings 10%15% higher performance at the same power or 25%30% lower power at the same speed with about 1.15x density vs N3E helped by gate-all-around transistors that tighten design co-optimization. AMDs rack-scale Helios pairs XX MI450s with HBM4 aiming for 51TB total memory and 1400TB/s bandwidth which reporting contrasts against Nvidias NVL144 at 21TB and 936TB/s in vanilla Rubin configs. Raw math"
X Link @rohanpaul_ai 2025-10-10T17:26Z 103.1K followers, 13.2K engagements
"A Chinese father posted a clip of his daughter crying as she says goodbye to her AI learning robot that broke. humans and AI are forming real bonds. --- reddit. com/r/MadeMeCry/comments/1o2yf3i/a_chinese_fathers_video_of_his_daughter_tearfully/"
X Link @rohanpaul_ai 2025-10-10T21:41Z 103.1K followers, 50.9K engagements
"The paper turns research papers into full presentation videos automatically. It gets XX% higher quiz accuracy and makes videos 6x faster. They built XXX paired papers with talks and defined X checks content match pairwise preference quiz accuracy and author recall They introduce PaperTalker a system that makes slides subtitles cursor paths speech and a talking head. It writes Beamer slide code from the paper compiles it and fixes errors. Beamer is a tool researchers use in LaTeX to make academic-style presentation slides. A tree search tries layout variants and a vision language model picks"
X Link @rohanpaul_ai 2025-10-11T22:12Z 103.1K followers, 11.6K engagements
"This survey paper argues Small language models can handle most agent tasks and big models step in only when needed. This setup cuts cost by 10x to 30x for common tool tasks. Agent work is mostly calling tools and producing structured outputs not recalling vast facts. So a router runs a small model by default and escalates to a large one only when confidence is low. Outputs follow strict JSON style schemas checked by validators which boosts correctness and cuts retries. This makes small models dependable at function calling and structured data while responding faster and using less energy."
X Link @rohanpaul_ai 2025-10-12T03:00Z 103.1K followers, 26.6K engagements
"π― Andrej Karpathy on how to learn"
X Link @rohanpaul_ai 2025-10-12T07:07Z 103.1K followers, 152.8K engagements
"πΌ AI will boost star workers more than everyone else widening performance gaps and straining teams. WSJ writes. The core idea is that domain expertise and organized habits let stars get more from AI. Stars adopt AI earlier explore features fast and build personal workflows while others wait for rules. They judge outputs better accepting correct advice and rejecting errors. Their domain expertise helps them ask precise questions set constraints and iterate which raises prompt quality and accuracy. The high-status of those employees get more recognition and praise for their AI-assisted work"
X Link @rohanpaul_ai 2025-10-13T12:06Z 103.1K followers, 437.7K engagements
"China and Robots. The Unitree robots are on a rolleach update comes with tougher moves and faster progress"
X Link @rohanpaul_ai 2025-10-13T13:30Z 103.1K followers, 22.5K engagements
"Shortfall of Supply vs. Demand for Data Center Capacity in USA. - Goldman Sachs Research. Unlike oil reserves which are naturally determined by geography data centers can be strategically built in locations chosen by businesses and governments. This flexibility allows nations to leverage data center infrastructure as a critical geopolitical and economic tool in the AI era"
X Link @rohanpaul_ai 2025-10-13T14:48Z 103.1K followers, 15.2K engagements
"Columbia CS Prof explains why LLMs cant generate new scientific ideas. Bcz LLMs learn a structured map Bayesian manifold of known data and work well within it but fail outside it. But true discovery means creating new maps which LLMs cannot do"
X Link @rohanpaul_ai 2025-10-13T19:39Z 103.1K followers, 203.6K engagements
"Goldman Sachs Research estimates capital expenditures of $XXX billion on US electricity generation through 2030. Much of this investment will go towards building out renewable energy sources replacing aging infrastructure and accommodating the growth in the grids power load"
X Link @rohanpaul_ai 2025-10-14T14:54Z 103.1K followers, 3357 engagements
"The whole million-dollar-plus commercial ad production pipeline is so ready to be replaced by AI. This is made with Sora X such an elegant and well-finished result"
X Link @rohanpaul_ai 2025-10-14T19:43Z 103.1K followers, 37.3K engagements
"Self-improving language models are turning real with MITs open-sourced SEAL (Self-Adapting LLMs) method. π SEAL lets an LLM write self-edits fine-tune on them with LoRA and lock those gains into weights for stronger knowledge recall and few-shot reasoning. The new version landed in Sep-25 the code is open under MIT License and the work appeared at NeurIPS 2025. SEAL tackles frozen weights by generating plain-language self-edits that restate facts or set training knobs then doing supervised fine-tuning so the update persists. A second loop runs reinforcement learning with filtered behavior"
X Link @rohanpaul_ai 2025-10-14T21:07Z 103.1K followers, 19.5K engagements
"π¦ Goldman Sachs told its employees that it plans to cut more jobs this year as part of a larger shift toward using AI to save costs and improve efficiency. The bank said it will slow down hiring and make limited job reductions even though its total number of employees will still end the year higher than last year. Goldman currently has about 48300 employees roughly 1800 more than at the end of 2024. The leadership described this as part of a new internal plan called OneGS XXX which aims to make the company faster and more efficient through AI-driven systems. The focus is on automating tasks"
X Link @rohanpaul_ai 2025-10-14T21:55Z 103.1K followers, 6726 engagements
"π Salesforce says its AI agents in support are saving $100M a year and it is using that proof to sell Agentforce. Agentforce is Salesforces agentic AI platform that lets organizations build customize and deploy autonomous AI agents which can take actions (not just respond) across business functions. Agentforce has 12000 customers and Reddit reports an XX% faster resolution time while the stock is down XX% in 2025. The cost drop comes from bots deflecting easy tickets agent assist drafting grounded replies and knowledge retrieval pulling the right answer from documentation. These steps lower"
X Link @rohanpaul_ai 2025-10-14T22:44Z 103.1K followers, 5404 engagements
"π OpenAI is partnering with Arm and Broadcom for its custom inference stack that targets 10GW of capacity between 2026 and 2029. Arm will design a server CPU to anchor OpenAI racks while Broadcom builds a specialized inference system on a chip at TSMC with production aimed for 2026 after roughly XX months of development. SoftBank will buy several billions of dollars of OpenAI technology every year starting now to help Arm subsidiaries cut chip development time and it owns XX% of Arm. Pairing an Arm CPU with a custom accelerator lets the host coordinate memory networking and scheduling"
X Link @rohanpaul_ai 2025-10-15T05:53Z 103.1K followers, 11.4K engagements
"More articles are now written by AI than humans. New research by Graphite. Even though these AI-made pieces dominate the web their research shows that they barely show up in Google results or in ChatGPT responses. --- graphite .io/five-percent/more-articles-are-now-created-by-ai-than-humans"
X Link @rohanpaul_ai 2025-10-15T15:35Z 103.1K followers, 21.4K engagements
"π Goldman Sachs economists say the US is entering jobless growth where output keeps climbing but hiring stays weak especially for younger workers. The engine is productivity-led growth from AI with Goldman estimating a long run XX% lift to productivity levels while aging and lower immigration cap labor supply so GDP can rise with fewer hires. Industry detail shows net job losses outside healthcare even as the economy expands and management talk tying AI to labor on earnings calls is hitting records with XX% of S&P XXX firms mentioning AI last quarter. Early pain is concentrated in young tech"
X Link @rohanpaul_ai 2025-10-15T21:53Z 103.1K followers, 11.5K engagements
"Cool. This paper turns a language model into an autonomous scientist that discovers equations from data. π Reports X% to XX% absolute gains over strong baselines across chemistry biology physics and materials. The task is symbolic regression which means finding a simple formula that matches data. SR-Scientist makes the model analyze data propose an equation test it then revise. X tools drive this a data analyzer and an evaluator that tunes constants. An experience buffer saves top equations so later attempts build on proven pieces. Each run aims for mean absolute percentage error under 0.1%"
X Link @rohanpaul_ai 2025-10-16T19:33Z 103.1K followers, 9667 engagements
"New Microsoft paper propose a new AI financial trading system that gives 180x faster action latency vs continuous inference agents. TiMi (Trade in Minutes) separates strategy design from execution so a lean bot trades each minute with mechanical rules. The papers key point is this: most AI trading agents constantly call big models role play and track news during live trading which is slow and expensive. TiMi avoids that. It separates thinking from doing. TiMi uses X agents market analysis pair tuning bot coding and math. They turn the plan into code decisions on top shared functions below."
X Link @rohanpaul_ai 2025-10-16T22:15Z 103.1K followers, 17.5K engagements
"The paper proposes a concrete test for AGI by checking if an AI matches a well educated adult across XX abilities. And reading it the obvious conclusion is we are just so far away from AGI level intelligence. They report an AGI Score and show GPT-4 at XX% and GPT-5 at XX% with long term memory at 0%. The goal is to stop the moving target problem and make progress measurable and auditable. They ground the checklist in the Cattell Horn Carroll model a long standing map of human cognitive abilities. The XX areas cover knowledge reading and writing math on the spot reasoning working memory long"
X Link @rohanpaul_ai 2025-10-16T23:09Z 103.1K followers, 48.7K engagements
"As per The Information Oracle ends up a negative Gross Profit Margin of XXX% when it leases NVIDIA B200 GPUs. --- theinformation. com/articles/oracle-assures-investors-ai-cloud-margins-struggles-profit-older-nvidia-chips"
X Link @rohanpaul_ai 2025-10-16T23:58Z 103.1K followers, 6136 engagements
"New Intel+California Univ paper shows how to train LLM teams with on-policy RL so they coordinate and solve multi-step tasks. It lifts Qwen3 planning accuracy from XXXX to XXXX% up to XXXX to 99.5%. Standard GRPO samples several answers to X prompt and pushes toward higher scoring ones. In teams that fails because each role sees a different prompt and history. AT-GRPO fixes this by grouping data by agent and by turn. For that exact prompt it samples several candidates scores them and continues with the best. Now the update compares true like for like and learning stays stable. Rewards blend a"
X Link @rohanpaul_ai 2025-10-17T02:40Z 103.1K followers, 4036 engagements
"This paper shows why Transformers underperform at time-series forecasting and why linear baselines still win. They analyze a simplified Transformer that uses linear self attention. As context grows its features act like using only a fixed number of recent lags. So it offers no extra signal beyond plain linear regression. With long history it can match linear regression but with finite history it keeps extra error. That extra error shrinks only at X over n. Adding more layers helps a little but the linear model remains the ceiling on average. For multi step forecasting the model feeds its own"
X Link @rohanpaul_ai 2025-10-17T06:55Z 103.1K followers, 20.1K engagements
"π’ Great news from Goldman Sachs Research. They note AI investment levels are sustainable despite concerns and even if its not yet clear who the biggest winners will be. And values the potential capital revenue unlocked by AI productivity gains at about $X trillion with possible estimates ranging from $X trillion to $XX trillion. They add that the S&P 500s market cap of $XXXXXX billion shows theres still a lot of room for AI-driven growth. Reason. - AI is driving productivity higher but getting those gains takes a lot more compute power as models keep scaling. - That AI investment in the U.S."
X Link @rohanpaul_ai 2025-10-17T08:51Z 103.1K followers, 29.3K engagements
"π§ Americas economy is riding a big AI bet as AI investment powers growth and props up markets. FT published a nice piece. If AI underdelivers the single support holding up growth and stocks looks shaky. Companies are spending hundreds of billions on AI hardware and data centers now driving XX% of US GDP growth. AI firms produced XX% of US stock gains in 2025 pulling in global money and boosting rich-household spending. Without the AI boost tariffs sticky inflation weaker institutions rising debt and an immigration slump would bite harder. In 2023 the US added over X million new immigrants"
X Link @rohanpaul_ai 2025-10-17T09:44Z 103.1K followers, 13.8K engagements
"Massive revelation from new Anthropic research. π¨ Shows that poisoning an LLM needs a very small number of bad samples (just XXX docs) regardless of model or dataset size. XXX poisoned documents were enough to plant a hidden rule across models from 600M to 13B even when the biggest model saw 20x more clean data. The big takeaway is to think in counts not %s because scaling up models and data does not wash out a well designed poison. The risk exists because training pulls from public data that attackers can quietly edit. A backdoor means the model looks normal until it sees a secret trigger"
X Link @rohanpaul_ai 2025-10-17T11:53Z 103.1K followers, 9022 engagements
"π Baidu (@Baidu_Inc) just launched and open-sourced a BRILLIANT model for document parsing - PaddleOCR-VL-0.9B. Hugely boosting multilingual document parsing via a 0.9B Ultra-Compact Vision-Language Model built on ERNIE-4.5-0.3B. It reaches #1 on OmniDocBench v1.5 with XXXXX beating GPT-4o Gemini XXX Pro and Qwen2.5-VL-72B as well as smaller OCR models like InternVL XXX MonkeyOCR-Pro-3B and Dots.OCR. While staying incredibly compact. By integrating ERNIE-4.5-0.3B with an encoder it can handle complex layouts in XXX languages with high precision. Lightweight high-accuracy and open-sourced a"
X Link @rohanpaul_ai 2025-10-17T12:05Z 103.1K followers, 9561 engagements
"This time its not a bubble. As far as Capex to Cash-flow ratio is concerned"
X Link @rohanpaul_ai 2025-10-17T14:44Z 103.1K followers, 8929 engagements
"Todays edition of my newsletter just went out. π Consider subscribing its free and I write it everyday. π§ Meta published the first real research on scaling reinforcement learning with LLMs π’ In a new report Goldman Sachs says Dont fear the AI bubble its about to unlock an $X trillion opportunity π Baidu just launched and open-sourced a BRILLIANT model for document parsing - PaddleOCR-VL-0.9B. π Tutorial: Run your models faster with quantized low precision in Keras"
X Link @rohanpaul_ai 2025-10-17T15:20Z 103.1K followers, 3250 engagements
"Financial Times shared OpenAIs recent stats and theyre next-level. - 800M weekly users. - X% paying. - $13B in ARR. - $8B loss in H1. - $1T+ in compute pledges"
X Link @rohanpaul_ai 2025-10-17T15:32Z 103.1K followers, 4698 engagements
"China's Tsinghua University paper shows simply mixing weights of Thinking and Instruct LLMs yields controllable efficient reasoning while keeping cost low. Which means your can tune models for cost and accuracy without any retraining. They literally interpolate the X weight sets and a single number sets the blend. At low blend the model behaves like Instruct answers are short accuracy rises a bit and there is almost no explicit thinking. At mid blend explicit chain of thought suddenly turns on reasoning quality jumps and tokens can even drop. At high blend the model thinks on every input"
X Link @rohanpaul_ai 2025-10-17T17:17Z 103.1K followers, 4111 engagements
"β OpenAI has became the worlds most valuable private company with 500B valuation. OpenAIs huge capital needs mean investors like Microsoft SoftBank and Thrive Capital will see their stakes reduced as more fundraising happens. FT writes a great piece. The working split being discussed is Microsoft at about XX% employees about XX% and the nonprofit parent about XX% with the remaining slice for SoftBank Thrive Khosla and others. Nvidias arrangement would add equity over time roughly $10B in shares for every $35B of GPUs OpenAI buys which could total up to $100B so every drawdown dilutes existing"
X Link @rohanpaul_ai 2025-10-17T18:00Z 103.1K followers, 3729 engagements
"Datacenter Capex will be in the region of $XXX billion this year from Google Amazon Microsoft and Meta alone. Goldman Sachs Research"
X Link @rohanpaul_ai 2025-10-17T19:26Z 103.1K followers, 5366 engagements
"Jensen Huang explains why AI is trillions of dollars of opportunity and its difference vs the old software era. Because the AI machine has to run all the time. --- On his recent interview on Citadel Securities YT channel"
X Link @rohanpaul_ai 2025-10-17T23:16Z 103.1K followers, 6013 engagements
"Andrej Karpathy AGI is still a decade away. On the latest Dwarkesh Patel podcast"
X Link @rohanpaul_ai 2025-10-17T23:32Z 103.1K followers, 14.6K engagements
"The paper says real AGI needs brain-like mechanisms not just competitive scores. It defines True Intelligence as X parts with X parts that engineers can measure and X deeper link. Gives a testable blueprint for AGI not just higher leaderboard scores. The X measurable parts are embodied sensory fusion core directives dynamic schemata a multi expert setup and an orchestration layer. A X level AGI scale scores a system by how many of those X parts it truly has. Embodied sensory fusion means the system ties ideas to seeing hearing touch and actions in the world. Core directives are built in goals"
X Link @rohanpaul_ai 2025-10-18T04:32Z 103.1K followers, 8646 engagements
"New Harvard paper shows trainingfree sampling lets a base LLM rival reinforcement learning on reasoning. No training dataset or verifier. The method samples from a power distribution which means reweighting full sequences the model already thinks are likely. That bias favors tokens that lead to a small set of likely futures over tokens that open many weak futures. Low temperature sampling only sharpens the next token so it misses those future path effects. The sampler runs a Metropolis Hastings loop that resamples a block then accepts or rejects based on relative sequence likelihood. It grows"
X Link @rohanpaul_ai 2025-10-18T06:36Z 103.1K followers, 11.7K engagements
"π @nvidia has now become the biggest name behind open AI model contributions with Nemotron Cosmos Gr00t BioNeMo and Canary. π Nemotron for agents BioNeMo for biopharma Cosmos for physical reasoning Gr00t for robotics and Canary for speech. And China accelerates with Qwen across chat multimodal reasoning and agents while Baidu Tencent MiniMax Z. AI ByteDance Moonshot AI and Zhipu AI ramp repos and leaderboard results. Europe is quieter beyond Mistral AI This nice graphics from @aiworld_eu --- cepsai.github. io/aiworld/OpenSource2025/story_animation.html"
X Link @rohanpaul_ai 2025-10-18T11:06Z 103.1K followers, 11.3K engagements
"New NVIDIA paper - StreamingVLM lets a vision language model follow endless video in real time while keeping memory and latency steady. It wins XXXXX% of head-to-head judgments against GPT-4o mini and runs at X FPS on X NVIDIA H100. The big deal is that it makes real-time infinite video understanding practical with steady latency built-in memory and clear gains over strong baselines. Old methods either look at everything and slow down a lot or use short windows and forget what happened. This model keeps a small cache with fixed anchor text a long window of recent text and a short window of"
X Link @rohanpaul_ai 2025-10-18T12:04Z 103.1K followers, 4158 engagements
"π» Nvidia and TSMC just announced that the first U.S.-made wafer for Nvidias Blackwell AI chips has been completed at TSMCs Arizona plant. Marks a big milestone in shifting advanced chip manufacturing to American soil. Jensen Huang called it historic and told that the most important chip in the world is now being produced in America. This wafer will eventually be used in Blackwell chips Nvidias next-generation processors designed for large-scale AI training and data center workloads. Huang added that Nvidia plans to invest $XXX billion in AI infrastructure over the next few years. Even with"
X Link @rohanpaul_ai 2025-10-18T12:28Z 103.1K followers, 4509 engagements
"The papers core claim argues LLMs excel by learning X fixed reasoning pattern so most human rationales are unnecessary. It matches human-data results while using 10x fewer human rationales. proof that teach the procedure beats collect more explanations. The key finding is that a stable reasoning pattern plus reinforcement learning can replace 10x human rationale data. Training still uses the usual X stages supervised fine tuning to show the pattern then reinforcement learning with verifiable rewards to lock it in. But the focus is patterned tasks where the same step order works every time"
X Link @rohanpaul_ai 2025-10-18T13:20Z 103.1K followers, 3222 engagements
"Andrej Karpathy called the current AI models as "slop" "I do think that overall the models are not there yet. I feel like the industry is making too big of a jump and trying to pretend that this is amazing but it's not. It's slop and I think they are not coming to terms with it. Maybe they are trying to fundraise or something like that I'm not sure what's going on. We are at this intermediate stage. The models are amazing but they still need a lot of work for now. Autocomplete is my sweet spot." --- On the latest Dwarkesh Patel podcast"
X Link @rohanpaul_ai 2025-10-18T22:53Z 103.1K followers, 10.7K engagements
/creator/twitter::rohanpaul_ai