Dark | Light
[GUEST ACCESS MODE: Data is scrambled or limited to provide examples. Make requests using your API key to unlock full data. Check https://lunarcrush.ai/auth for authentication information.]

# ![@karpathy Avatar](https://lunarcrush.com/gi/w:26/cr:twitter::33836629.png) @karpathy Andrej Karpathy

Andrej Karpathy posts on X about llm, ai, if you, all the the most. They currently have XXXXXXXXX followers and XXX posts still getting attention that total XXXXXXX engagements in the last XX hours.

### Engagements: XXXXXXX [#](/creator/twitter::33836629/interactions)
![Engagements Line Chart](https://lunarcrush.com/gi/w:600/cr:twitter::33836629/c:line/m:interactions.svg)

- X Week XXXXXXXXX +614%
- X Month XXXXXXXXXX +81%
- X Months XXXXXXXXXXX +55%
- X Year XXXXXXXXXXX +0.50%

### Mentions: XX [#](/creator/twitter::33836629/posts_active)
![Mentions Line Chart](https://lunarcrush.com/gi/w:600/cr:twitter::33836629/c:line/m:posts_active.svg)

- X Week XXX +7.50%
- X Month XXX +8.70%
- X Months XXX +18%
- X Year XXX +19%

### Followers: XXXXXXXXX [#](/creator/twitter::33836629/followers)
![Followers Line Chart](https://lunarcrush.com/gi/w:600/cr:twitter::33836629/c:line/m:followers.svg)

- X Week XXXXXXXXX +0.86%
- X Month XXXXXXXXX +3.50%
- X Months XXXXXXXXX +17%
- X Year XXXXXXXXX +37%

### CreatorRank: XXXXXX [#](/creator/twitter::33836629/influencer_rank)
![CreatorRank Line Chart](https://lunarcrush.com/gi/w:600/cr:twitter::33836629/c:line/m:influencer_rank.svg)

### Social Influence

**Social category influence**
[technology brands](/list/technology-brands)  #2371 [social networks](/list/social-networks)  XXXX% [finance](/list/finance)  XXXX% [stocks](/list/stocks)  #430 [nfts](/list/nfts)  #145 [automotive brands](/list/automotive-brands)  XXXX%

**Social topic influence**
[llm](/topic/llm) #1, [ai](/topic/ai) 8.74%, [if you](/topic/if-you) #5656, [all the](/topic/all-the) #382, [karpathy](/topic/karpathy) #1, [to the](/topic/to-the) 3.88%, [nano banana](/topic/nano-banana) #96, [imo](/topic/imo) #65, [banana](/topic/banana) #359, [open ai](/topic/open-ai) #415

**Top accounts mentioned or mentioned by**
[@grok](/creator/undefined) [@flolight44](/creator/undefined) [@billstenner7](/creator/undefined) [@johntheadman_](/creator/undefined) [@elonmusk](/creator/undefined) [@dataexec](/creator/undefined) [@adamskyart](/creator/undefined) [@mixedrealityman](/creator/undefined) [@yuchenj_uw](/creator/undefined) [@gp_pulipaka](/creator/undefined) [@cryptosausage](/creator/undefined) [@spil____](/creator/undefined) [@adarkm0ment](/creator/undefined) [@danadvantage](/creator/undefined) [@marswalkerr](/creator/undefined) [@mohamedatta_911](/creator/undefined) [@adelayida210519](/creator/undefined) [@_thomasip](/creator/undefined) [@jasonth0](/creator/undefined) [@rileyralmuto](/creator/undefined)

**Top assets mentioned**
[Doodles (doodles)](/topic/doodles) [Alphabet Inc Class A (GOOGL)](/topic/$googl) [Tesla, Inc. (TSLA)](/topic/tesla)
### Top Social Posts
Top posts by engagements in the last XX hours

"Finally had a chance to listen through this pod with Sutton which was interesting and amusing. As background Sutton's "The Bitter Lesson" has become a bit of biblical text in frontier LLM circles. Researchers routinely talk about and ask whether this or that approach or idea is sufficiently "bitter lesson pilled" (meaning arranged so that it benefits from added computation for free) as a proxy for whether it's going to work or worth even pursuing. The underlying assumption being that LLMs are of course highly "bitter lesson pilled" indeed just look at LLM scaling laws where if you put compute"  
[X Link](https://x.com/karpathy/status/1973435013875314729)  2025-10-01T17:09Z 1.5M followers, 2M engagements


"@zenitsu_aprntc Good question it's basically entirely hand-written (with tab autocomplete). I tried to use claude/codex agents a few times but they just didn't work well enough at all and net unhelpful possibly the repo is too far off the data distribution"  
[X Link](https://x.com/karpathy/status/1977758204139331904)  2025-10-13T15:27Z 1.5M followers, 484.3K engagements


"Deliberately*"  
[X Link](https://x.com/karpathy/status/1978654822036607245)  2025-10-16T02:50Z 1.5M followers, 146.3K engagements


"@proggineer Agree sometimes that is helpful too to have an overview of what the whole thing is about first. I just copy paste stuff around to LLM of the day (I cycle) theres no tool"  
[X Link](https://x.com/karpathy/status/1990580578287300816)  2025-11-18T00:39Z 1.5M followers, 59.9K engagements


"Has anyone encountered a good definition of slop. In a quantitative measurable sense. My brain has an intuitive slop index I can reliably estimate but Im not sure how to define it. I have some bad ideas that involve the use of LLM miniseries and thinking token budgets"  
[X Link](https://x.com/karpathy/status/1992053281900941549)  2025-11-22T02:11Z 1.5M followers, 633K engagements


"@_thomasip haha yes it makes mistakes You have to re-roll a few times until it's right. Sometimes it gets stuck in loops and you have to re-start in a new conversation. Example re-roll:"  
[X Link](https://x.com/karpathy/status/1992715615988220285)  2025-11-23T22:03Z 1.5M followers, 23.7K engagements


"@theJayAlto Edutainment. This one weird trick to consume entertainment and feel good about it"  
[X Link](https://x.com/karpathy/status/1994818887591055480)  2025-11-29T17:20Z 1.5M followers, 162.9K engagements


"@nickcammarata Your recent posts on this remind me of this Arnold gem +100 though. I finally had a chance to install a home gym recently making it trivial to use daily. Always looking forward to the next exercise high. Slightly miss the social/entropy aspects of gyms"  
[X Link](https://x.com/karpathy/status/1996654385003425949)  2025-12-04T18:54Z 1.5M followers, 184.7K engagements


"reminded of this paragraph from gsm8k paper 2021 :)"  
[X Link](https://x.com/karpathy/status/1966896849929073106)  2025-09-13T16:08Z 1.5M followers, 378.9K engagements


"@anneshu_nag I love that it's called Nano Banana instead of Google Imagine or some other corpo thing"  
[X Link](https://x.com/karpathy/status/1992712082903781448)  2025-11-23T21:49Z 1.5M followers, 20.5K engagements


"I've had medium success asking LLMs if a thing exists it works out of the box for some of the more well-known things (e.g. both GPT XXX and Gemini X know about this function if you describe the tensor transformation in words). For more esoteric or new libraries (e.g. uv being a recent example) I've had more success manually packaging up docs into markdown and including it as context for questions. PyTorch docs now also seem to have an "Ask AI" that presumably does RAG over their docs but the model is not that bright. But you'd have to suspect it might exist and attempt to ask in the first"  
[X Link](https://x.com/karpathy/status/1993369287969718540)  2025-11-25T17:20Z 1.5M followers, 96.6K engagements


"Don't think of LLMs as entities but as simulators. For example when exploring a topic don't ask: "What do you think about xyz" There is no "you". Next time try: "What would be a good group of people to explore xyz What would they say" The LLM can channel/simulate many perspectives but it hasn't "thought about" xyz for a while and over time and formed its own opinions in the way we're used to. If you force it via the use of "you" it will give you something by adopting a personality embedding vector implied by the statistics of its finetuning data and then simulate that. It's fine to do but"  
[X Link](https://x.com/karpathy/status/1997731268969304070)  2025-12-07T18:13Z 1.5M followers, 3.6M engagements


"We will see that a lot of weird behaviors and problems of LLMs actually trace back to tokenization. We'll go through a number of these issues discuss why tokenization is at fault and why someone out there ideally finds a way to delete this stage entirely"  
[X Link](https://x.com/karpathy/status/1759996551378940395)  2024-02-20T17:40Z 1.5M followers, 753.3K engagements


"@matejhladky_dev AI has crushed it since this post way beyond expectation. I made the same category of mistake all of AI was making of thinking we have to discover and write the algorithm. You don't. You pretrain and then finetune a BIG neural network on lots of tasks and it just falls out. lol"  
[X Link](https://x.com/karpathy/status/1993372017593335995)  2025-11-25T17:31Z 1.5M followers, 90.8K engagements


"# on shortification of "learning" There are a lot of videos on YouTube/TikTok etc. that give the appearance of education but if you look closely they are really just entertainment. This is very convenient for everyone involved : the people watching enjoy thinking they are learning (but actually they are just having fun). The people creating this content also enjoy it because fun has a much larger audience fame and revenue. But as far as learning goes this is a trap. This content is an epsilon away from watching the Bachelorette. It's like snacking on those "Garden Veggie Straws" which feel"  
[X Link](https://x.com/karpathy/status/1756380066580455557)  2024-02-10T18:10Z 1.5M followers, 2.2M engagements


"⚡ Excited to share that I am starting an AI+Education company called Eureka Labs. The announcement: --- We are Eureka Labs and we are building a new kind of school that is AI native. How can we approach an ideal experience for learning something new For example in the case of physics one could imagine working through very high quality course materials together with Feynman who is there to guide you every step of the way. Unfortunately subject matter experts who are deeply passionate great at teaching infinitely patient and fluent in all of the world's languages are also very scarce and cannot"  
[X Link](https://x.com/karpathy/status/1813263734707790301)  2024-07-16T17:25Z 1.5M followers, 2.5M engagements


"DeepSeek (Chinese AI co) making it look easy today with an open weights release of a frontier-grade LLM trained on a joke of a budget (2048 GPUs for X months $6M). For reference this level of capability is supposed to require clusters of closer to 16K GPUs the ones being brought up today are more around 100K GPUs. E.g. Llama X 405B used 30.8M GPU-hours while DeepSeek-V3 looks to be a stronger model at only 2.8M GPU-hours (11X less compute). If the model also passes vibe checks (e.g. LLM arena rankings are ongoing my few quick tests went well so far) it will be a highly impressive display of"  
[X Link](https://x.com/karpathy/status/1872362712958906460)  2024-12-26T19:23Z 1.5M followers, 6.5M engagements


"Continuing the journey of optimal LLM-assisted coding experience. In particular I find that instead of narrowing in on a perfect one thing my usage is increasingly diversifying across a few workflows that I "stitch up" the pros/cons of: Personally the bread & butter (75%) of my LLM assistance continues to be just (Cursor) tab complete. This is because I find that writing concrete chunks of code/comments myself and in the right part of the code is a high bandwidth way of communicating "task specification" to the LLM i.e. it's primarily about task specification bits - it takes too many bits and"  
[X Link](https://x.com/karpathy/status/1959703967694545296)  2025-08-24T19:46Z 1.5M followers, 687.8K engagements


"I think congrats again to OpenAI for cooking with GPT-5 Pro. This is the third time I've struggled on something complex/gnarly for an hour on and off with CC then X Pro goes off for XX minutes and comes back with code that works out of the box. I had CC read the X Pro version and it wrote up X paragraphs admiring it (very wholesome). If you're not giving it your hardest problems you're probably missing out"  
[X Link](https://x.com/karpathy/status/1964020416139448359)  2025-09-05T17:38Z 1.5M followers, 2.6M engagements


"nanochat now has a primordial identity and can talk a bit about itself and its capabilities (e.g. it knows it's nanochat d32 that cost $XXX that it was built by me that it can't speak languages other than English too well and why etc.). This kind of customization is all done through synthetic data generation and I uploaded a new example script to demonstrate. It's a bit subtle but by default LLMs have no inherent personality or any understanding of their own capabilities because they are not animal-like entities. They don't know what they are or what they can or can't do or know or don't"  
[X Link](https://x.com/karpathy/status/1980665134415802554)  2025-10-21T15:59Z 1.5M followers, 455.5K engagements


"@LucasAtkins7 This code is extremely dangerous. Here I improved it"  
[X Link](https://x.com/karpathy/status/1981009115523789169)  2025-10-22T14:45Z 1.5M followers, 1.5M engagements


"Sharing an interesting recent conversation on AI's impact on the economy. AI has been compared to various historical precedents: electricity industrial revolution etc. I think the strongest analogy is that of AI as a new computing paradigm (Software 2.0) because both are fundamentally about the automation of digital information processing. If you were to forecast the impact of computing on the job market in 1980s the most predictive feature of a task/job you'd look at is to what extent the algorithm of it is fixed i.e. are you just mechanically transforming information according to rote easy"  
[X Link](https://x.com/karpathy/status/1990116666194456651)  2025-11-16T17:56Z 1.5M followers, 2.1M engagements


"Finally had time to read & process this great post. I run into the pattern quite often it goes: "something that sounds wrong is good actually because galaxy brain reason" Galaxy brain reasoning is the best way to justify anything while looking / feeling good about it. From this perspective for example there's deeper wisdom in the Ten Commandments imposing constraints over actions instead of utility over states. It's not Ten Objectives. E.g. they don't attempt to define a utility function for the value of life they simply say "Thou shalt not kill". This approach curtails the relatively"  
[X Link](https://x.com/karpathy/status/1990494327936885192)  2025-11-17T18:56Z 1.5M followers, 839.5K engagements


"Im starting to get into a habit of reading everything (blogs articles book chapters) with LLMs. Usually pass X is manual then pass X explain/summarize pass X Q&A. I usually end up with a better/deeper understanding than if I moved on. Growing to among top use cases. On the flip side if youre a writer trying to explain/communicate something we may increasingly see less of a mindset of Im writing this for another human and more Im writing this for an LLM. Because once an LLM gets it it can then target personalize and serve the idea to its user"  
[X Link](https://x.com/karpathy/status/1990577951671509438)  2025-11-18T00:29Z 1.5M followers, 2.8M engagements


"I put up a simple repo I call reader3 (it's my 3rd version.) to illustrate how I read EPUBs with LLMs. Basically get some epub (e.g. Project Gutenberg is great) go chapter by chapter and with this you can easily copy paste text to your favorite LLM"  
[X Link](https://x.com/karpathy/status/1990612045700739548)  2025-11-18T02:44Z 1.5M followers, 292.9K engagements


"I played with Gemini X yesterday via early access. Few thoughts - First I usually urge caution with public benchmarks because imo they can be quite possible to game. It comes down to discipline and self-restraint of the team (who is meanwhile strongly incentivized otherwise) to not overfit test sets via elaborate gymnastics over test-set adjacent data in the document embedding space. Realistically because everyone else is doing it the pressure to do so is high. Go talk to the model. Talk to the other models (Ride the LLM Cycle - use a different LLM every day). I had a positive early"  
[X Link](https://x.com/karpathy/status/1990854771058913347)  2025-11-18T18:49Z 1.5M followers, 1.2M engagements


"My most amusing interaction was where the model (I think I was given some earlier version with a stale system prompt) refused to believe me that it is 2025 and kept inventing reasons why I must be trying to trick it or playing some elaborate joke on it. I kept giving it images and articles from "the future" and it kept insisting it was all fake. It accused me of using generative AI to defeat its challenges and argued why real wikipedia entries were actually generated and what the "dead giveaways" are. It highlighted tiny details when I gave it Google Image Search results arguing why the"  
[X Link](https://x.com/karpathy/status/1990855382756164013)  2025-11-18T18:51Z 1.5M followers, 1M engagements


"Something I think people continue to have poor intuition for: The space of intelligences is large and animal intelligence (the only kind we've ever known) is only a single point arising from a very specific kind of optimization that is fundamentally distinct from that of our technology. Animal intelligence optimization pressure: - innate and continuous stream of consciousness of an embodied "self" a drive for homeostasis and self-preservation in a dangerous physical world. - thoroughly optimized for natural selection = strong innate drives for power-seeking status dominance reproduction. many"  
[X Link](https://x.com/karpathy/status/1991910395720925418)  2025-11-21T16:43Z 1.5M followers, 2.6M engagements


"@TheVixhal your post challenged me. every one of your points is wrong but i had to think about each for a while :)"  
[X Link](https://x.com/karpathy/status/1991923470868119995)  2025-11-21T17:35Z 1.5M followers, 1.8M engagements


"@NickADobos I dont super love shoggoth represented as a kind of biological monster (animal) it feels wrong/misleading in this sense"  
[X Link](https://x.com/karpathy/status/1992018398436524230)  2025-11-21T23:52Z 1.5M followers, 49.2K engagements


"Imo this is along the lines of how talking to an LLM via text is like typing into a DOS Terminal and "GUI hasn't been invented yet" of some of my earlier posts. The GUI is an intelligent canvas"  
[X Link](https://x.com/karpathy/status/1992657223785586864)  2025-11-23T18:11Z 1.5M followers, 252.1K engagements


"I asked it to create a personalized weekly workout plan and then posters that I can print on the wall to remind me what exercises to do each day. Tuesday looks more intense because I asked for "more testosterone" :D. (sorry I'll stop posting more nano banana pro stuff now)"  
[X Link](https://x.com/karpathy/status/1992711182537707990)  2025-11-23T21:45Z 1.5M followers, 385.9K engagements


"Happy weekend to those who celebrate"  
[X Link](https://x.com/karpathy/status/1997697581410062590)  2025-12-07T15:59Z 1.5M followers, 1.3M engagements


"@Marswalkerr I love this meme too haha"  
[X Link](https://x.com/karpathy/status/1997698794973176092)  2025-12-07T16:04Z 1.5M followers, 44.7K engagements


"There is definitely work going into engineering the "you" simulation - the personality that gets all the rewards in verifiable problems or all the upvotes from users/judge LLMs or mimics the responses of SFT and there is an emergent composite personality from that. My point is more that the "you" there is deliberately bolted on engineered and layered on what is fundamentally a token simulation engine not a mind that is somehow emergent and over time constructed in a relatable way to an average person talking to an AI. The story is a bit more simple in verifiable domains but I think more"  
[X Link](https://x.com/karpathy/status/1997759548543947249)  2025-12-07T20:06Z 1.5M followers, 126.7K engagements


"I could certainly imagine that "nesting" the simulation might be too "effortful" for the model compute or data density wise. My results with it are not too bad so imo it's at least worth people try / experiment with / think about. For example it might be useful to read multiple distinct and approximate perspectives on topic xyz instead of one. Research-wise you might be able to elicit LLM Council - like benefits (not via diverse LLMs but via diverse simulations) and improve performance via the generator-discriminator gap effects or ensembling effects"  
[X Link](https://x.com/karpathy/status/1998082007893717478)  2025-12-08T17:27Z 1.5M followers, 28.8K engagements


"ty to ericsilberstein1 on github for spotting the bug. (it's not a big bug and only comes up in the SpellingBee synthetic task evaluation but still)"  
[X Link](https://x.com/karpathy/status/1998240551964193148)  2025-12-09T03:57Z 1.5M followers, 119K engagements


"nanoGPT - the first LLM to train and inference in space 🥹. It begins"  
[X Link](https://x.com/karpathy/status/1998806260783919434)  2025-12-10T17:25Z 1.5M followers, 759.7K engagements


"The hottest new programming language is English"  
[X Link](https://x.com/karpathy/status/1617979122625712128)  2023-01-24T20:14Z 1.5M followers, 9M engagements


"# Reproduce GPT-2 (124M) in llm.c in XX minutes for $XX ✨ The GPT-2 (124M) is the smallest model in the GPT-2 series released by OpenAI in 2019 and is actually quite accessible today even for the GPU poor. For example with llm.c you can now reproduce this model on one 8X A100 80GB SXM node in XX minutes (at XX% MFU). As they run for $14/hr this is $XX. I also think the 124M model makes for an excellent "cramming" challenge for training it very fast. So here is the launch command: And here is the output after XX minutes training on 10B tokens of the FineWeb dataset: It feels really nice to"  
[X Link](https://x.com/karpathy/status/1795484547267834137)  2024-05-28T15:57Z 1.5M followers, 663.7K engagements


"There's a new kind of coding I call "vibe coding" where you fully give in to the vibes embrace exponentials and forget that the code even exists. It's possible because the LLMs (e.g. Cursor Composer w Sonnet) are getting too good. Also I just talk to Composer with SuperWhisper so I barely even touch the keyboard. I ask for the dumbest things like "decrease the padding on the sidebar by half" because I'm too lazy to find it. I "Accept All" always I don't read the diffs anymore. When I get error messages I just copy paste them in with no comment usually that fixes it. The code grows beyond my"  
[X Link](https://x.com/karpathy/status/1886192184808149383)  2025-02-02T23:17Z 1.5M followers, 5.3M engagements


"I was given early access to Grok X earlier today making me I think one of the first few who could run a quick vibe check. Thinking ✅ First Grok X clearly has an around state of the art thinking model ("Think" button) and did great out of the box on my Settler's of Catan question: "Create a board game webpage showing a hex grid just like in the game Settlers of Catan. Each hex grid is numbered from 1.N where N is the total number of hex tiles. Make it generic so one can change the number of "rings" using a slider. For example in Catan the radius is X hexes. Single html page please." Few models"  
[X Link](https://x.com/karpathy/status/1891720635363254772)  2025-02-18T05:25Z 1.5M followers, 3.7M engagements


"Agency Intelligence I had this intuitively wrong for decades I think due to a pervasive cultural veneration of intelligence various entertainment/media obsession with IQ etc. Agency is significantly more powerful and significantly more scarce. Are you hiring for agency Are we educating for agency Are you acting as if you had 10X agency Grok explanation is close: Agency as a personality trait refers to an individual's capacity to take initiative make decisions and exert control over their actions and environment. Its about being proactive rather than reactivesomeone with high agency doesnt"  
[X Link](https://x.com/karpathy/status/1894099637218545984)  2025-02-24T18:58Z 1.5M followers, 7.3M engagements


""Finding the Best Sleep Tracker" Results of an experiment where I wore X sleep trackers every night for X months. TLDR Whoop = Oura 8Sleep Apple Watch + AutoSleep. Link simply right here instead of in a reply because ()/"  
[X Link](https://x.com/karpathy/status/1906386327190257963)  2025-03-30T16:41Z 1.5M followers, 1.6M engagements


"I attended a vibe coding hackathon recently and used the chance to build a web app (with auth payments deploy etc.). I tinker but I am not a web dev by background so besides the app I was very interested in what it's like to vibe code a full web app today. As such I wrote none of the code directly (Cursor+Claude/o3 did) and I don't really know how the app works in the conventional sense that I'm used to as an engineer. The app is called MenuGen and it is live on Basically I'm often confused about what all the things on a restaurant menu are - e.g. Pt Tagine Cavatappi or Sweetbread (hint it's."  
[X Link](https://x.com/karpathy/status/1917961248031080455)  2025-05-01T15:16Z 1.5M followers, 785.3K engagements


"An attempt to explain (current) ChatGPT versions. I still run into many many people who don't know that: - o3 is the obvious best thing for important/hard things. It is a reasoning model that is much stronger than 4o and if you are using ChatGPT professionally and not using o3 you're ngmi. - 4o is different from o4. Yes I know lol. 4o is a good "daily driver" for many easy-medium questions. o4 is only available as mini for now and is not as good as o3 and I'm not super sure why it's out right now. Example basic "router" in my own personal use: - Any simple query (e.g. "what foods are high in"  
[X Link](https://x.com/karpathy/status/1929597620969951434)  2025-06-02T17:54Z 1.5M followers, 1.4M engagements


"My sleep scores during recent travel were in the 90s. Now back in SF I am consistently back down to 70s 80s. I am increasingly convinced that this is due to traffic noise from a nearby road/intersection where I live - every 10min a car truck bus or motorcycle with a very loud engine passes by (some are 10X louder than others). In the later less deep stages of sleep it is much easier to wake and then much harder to go back to sleep. More generally I think noise pollution (esp early hours) come at a huge societal cost that is not correctly accounted for. E.g. I wouldn't be too surprised if a"  
[X Link](https://x.com/karpathy/status/1931426322536132767)  2025-06-07T19:01Z 1.5M followers, 1.5M engagements


"The race for LLM "cognitive core" - a few billion param model that maximally sacrifices encyclopedic knowledge for capability. It lives always-on and by default on every computer as the kernel of LLM personal computing. Its features are slowly crystalizing: - Natively multimodal text/vision/audio at both input and output. - Matryoshka-style architecture allowing a dial of capability up and down at test time. - Reasoning also with a dial. (system 2) - Aggressively tool-using. - On-device finetuning LoRA slots for test-time training personalization and customization. - Delegates and double"  
[X Link](https://x.com/karpathy/status/1938626382248149433)  2025-06-27T15:52Z 1.5M followers, 1.3M engagements


"I often rant about how XX% of attention is about to be LLM attention instead of human attention. What does a research paper look like for an LLM instead of a human Its definitely not a pdf. There is huge space for an extremely valuable research app that figures this out"  
[X Link](https://x.com/karpathy/status/1943411187296686448)  2025-07-10T20:45Z 1.5M followers, 905.6K engagements


"Scaling up RL is all the rage right now I had a chat with a friend about it yesterday. I'm fairly certain RL will continue to yield more intermediate gains but I also don't expect it to be the full story. RL is basically "hey this happened to go well (/poorly) let me slightly increase (/decrease) the probability of every action I took for the future". You get a lot more leverage from verifier functions than explicit supervision this is great. But first it looks suspicious asymptotically - once the tasks grow to be minutes/hours of interaction long you're really going to do all that work just"  
[X Link](https://x.com/karpathy/status/1944435412489171119)  2025-07-13T16:35Z 1.5M followers, 1.1M engagements


"I'm noticing that due to (I think) a lot of benchmarkmaxxing on long horizon tasks LLMs are becoming a little too agentic by default a little beyond my average use case. For example in coding the models now tend to reason for a fairly long time they have an inclination to start listing and grepping files all across the entire repo they do repeated web searchers they over-analyze and over-think little rare edge cases even in code that is knowingly incomplete and under active development and often come back minutes later even for simple queries. This might make sense for long-running tasks but"  
[X Link](https://x.com/karpathy/status/1954224651443544436)  2025-08-09T16:53Z 1.5M followers, 1M engagements


"In era of pretraining what mattered was internet text. You'd primarily want a large diverse high quality collection of internet documents to learn from. In era of supervised finetuning it was conversations. Contract workers are hired to create answers for questions a bit like what you'd see on Stack Overflow / Quora or etc. but geared towards LLM use cases. Neither of the two above are going away (imo) but in this era of reinforcement learning it is now environments. Unlike the above they give the LLM an opportunity to actually interact - take actions see outcomes etc. This means you can hope"  
[X Link](https://x.com/karpathy/status/1960803117689397543)  2025-08-27T20:34Z 1.5M followers, 951.4K engagements


"Transforming human knowledge sensors and actuators from human-first and human-legible to LLM-first and LLM-legible is a beautiful space with so much potential and so much can be done. One example I'm obsessed with recently - for every textbook pdf/epub there is a perfect "LLMification" of it intended not for human but for an LLM (though it is a non-trivial transformation that would need human in the loop involvement). - All of the exposition is extracted into a markdown document including all latex styling (bold/italic) tables lists etc. All of the figures are extracted as images. - All"  
[X Link](https://x.com/karpathy/status/1961128638725923119)  2025-08-28T18:07Z 1.5M followers, 724.9K engagements


""AI isn't replacing radiologists" good article Expectation: rapid progress in image recognition AI will delete radiology jobs (e.g. as famously predicted by Geoff Hinton now almost a decade ago). Reality: radiology is doing great and is growing. There are a lot of imo naive predictions out there on the imminent impact of AI on the job market. E.g. a year ago I was asked by someone who should know better if I think there will be any software engineers still today. (Spoiler: I think we're going to make it). This is happening too broadly. The post goes into detail on why it's not that simple"  
[X Link](https://x.com/karpathy/status/1971220449515516391)  2025-09-25T14:29Z 1.5M followers, 2.3M engagements


"Something I am experimenting with. I copy pasted: 1) the full podcast transcript 2) the bitter lesson blog post 3) my full post above To ChatGPT. The interesting part is you can fork the conversation context to ask any questions and take it in whatever direction with chat:"  
[X Link](https://x.com/karpathy/status/1973443912388977021)  2025-10-01T17:44Z 1.5M followers, 148.2K engagements


"Hah judging by mentions overnight people seem to find the ghost analogy provocative. I swear I don't wake up just trying to come with new memes but to elaborate briefly why I thought it was a fun comparison: 1) It captures the idea that LLMs are purely digital artifacts that don't interact with the physical world (unlike animals which are very embodied). 2) Ghosts are a kind of "echo" of the living in this case a statistical distillation of humanity. 3) There is an air of mystery over both ghosts and LLMs as in we don't fully understand what they are or how they work. 4) The process of"  
[X Link](https://x.com/karpathy/status/1973756330449236009)  2025-10-02T14:25Z 1.5M followers, 257.8K engagements


"Every company needs a DM POC - someone high up who you can just DM the most obvious things and who shortcuts the PM hierarchy"  
[X Link](https://x.com/karpathy/status/1974482521862865154)  2025-10-04T14:31Z 1.5M followers, 563.5K engagements


"I don't know what labs are doing to these poor LLMs during RL but they are mortally terrified of exceptions in any infinitesimally likely case. Exceptions are a normal part of life and healthy dev process. Sign my LLM welfare petition for improved rewards in cases of exceptions"  
[X Link](https://x.com/karpathy/status/1976077806443569355)  2025-10-09T00:10Z 1.5M followers, 715K engagements


"POV: Your LLM agent is dividing a by b"  
[X Link](https://x.com/karpathy/status/1976082963382272334)  2025-10-09T00:31Z 1.5M followers, 390.3K engagements


"Excited to release new repo: nanochat (it's among the most unhinged I've written). Unlike my earlier similar repo nanoGPT which only covered pretraining nanochat is a minimal from scratch full-stack training/inference pipeline of a simple ChatGPT clone in a single dependency-minimal codebase. You boot up a cloud GPU box run a single script and in as little as X hours later you can talk to your own LLM in a ChatGPT-like web UI. It weighs 8000 lines of imo quite clean code to: - Train the tokenizer using a new Rust implementation - Pretrain a Transformer LLM on FineWeb evaluate CORE score"  
[X Link](https://x.com/karpathy/status/1977755427569111362)  2025-10-13T15:16Z 1.5M followers, 5.8M engagements


"And an example of some of the summary metrics produced by the $XXX speedrun in the report card to start. The current code base is a bit over 8000 lines but I tried to keep them clean and well-commented. Now comes the fun part - of tuning and hillclimbing"  
[X Link](https://x.com/karpathy/status/1977755433172443626)  2025-10-13T15:16Z 1.5M followers, 183.3K engagements


"nanochat d32 i.e. the depth XX version that I specced for $1000 up from $XXX has finished training after XX hours and looks good. All the metrics go up quite a bit across pretraining SFT and RL. CORE score of XXXX is now well above GPT-2 at XXXX. GSM8K went X% - XX% etc. So that's encouraging. The model is pretty fun to talk to but judging from some early interactions I think people have a little bit too much expectation for these micro models. There is a reason that frontier LLM labs raise billions to train their models. nanochat models cost $XXX - $1000 to train from scratch. The $100"  
[X Link](https://x.com/karpathy/status/1978615547945521655)  2025-10-16T00:14Z 1.5M followers, 266.1K engagements


"TV in the 90s: you turn it on you watch. TV 2025: - turn on wait for it to load - popup: TV wants to update 1.5GB. No. - scroll sideways find prime video app or etc - popup: now app wants to update 500MB. No - App launching. App loading - select account screen - 🫠"  
[X Link](https://x.com/karpathy/status/1978653908663726585)  2025-10-16T02:47Z 1.5M followers, 1.7M engagements


"There is a movement I found on Instagram where people delivery choose to live in 90s refusing all technology after 2000. Like an intermediate form of the Amish"  
[X Link](https://x.com/karpathy/status/1978654744475578568)  2025-10-16T02:50Z 1.5M followers, 297.3K engagements


"My pleasure to come on Dwarkesh last week I thought the questions and conversation were really good. I re-watched the pod just now too. First of all yes I know and I'm sorry that I speak so fast :). It's to my detriment because sometimes my speaking thread out-executes my thinking thread so I think I botched a few explanations due to that and sometimes I was also nervous that I'm going too much on a tangent or too deep into something relatively spurious. Anyway a few notes/pointers: AGI timelines. My comments on AGI timelines looks to be the most trending part of the early response. This is"  
[X Link](https://x.com/karpathy/status/1979644538185752935)  2025-10-18T20:23Z 1.5M followers, 4.1M engagements


"I quite like the new DeepSeek-OCR paper. It's a good OCR model (maybe a bit worse than dots) and yes data collection etc. but anyway it doesn't matter. The more interesting part for me (esp as a computer vision at heart who is temporarily masquerading as a natural language person) is whether pixels are better inputs to LLMs than text. Whether text tokens are wasteful and just terrible at the input. Maybe it makes more sense that all inputs to LLMs should only ever be images. Even if you happen to have pure text input maybe you'd prefer to render it and then feed that in: - more information"  
[X Link](https://x.com/karpathy/status/1980397031542989305)  2025-10-20T22:13Z 1.5M followers, 3.3M engagements


"@r_chirra I fixed it :) deployed live now. This was done by doing a round of synthetic data generation to collect a 1000 multi-turn conversations (given a bunch of information including the readme of the nanochat project) and then mixing that into midtraining and SFT. fun"  
[X Link](https://x.com/karpathy/status/1980508380860150038)  2025-10-21T05:36Z 1.5M followers, 456.3K engagements


"Last night I taught nanochat d32 how to count 'r' in strawberry (or similar variations). I thought this would be a good/fun example of how to add capabilities to nanochat and I wrote up a full guide here: This is done via a new synthetic task SpellingBee that generates examples of a user asking for this kind of a problem and an ideal solution from an assistant. We then midtrain/SFT finetune on these to endow the LLM with the capability or further train with RL to make it more robust. There are many details to get right especially at smaller model sizes and the guide steps through them. As a"  
[X Link](https://x.com/karpathy/status/1981746327995465816)  2025-10-24T15:35Z 1.5M followers, 568.4K engagements


"Beautiful technical debugging detective longread that starts with a suspicious loss curve and ends all the way in the Objective-C++ depths of PyTorch MPS backend of addcmul_ that silently fails on non-contiguous output tensors. I wonder how long before an LLM can do all of this"  
[X Link](https://x.com/karpathy/status/1982483540899237981)  2025-10-26T16:24Z 1.5M followers, 601.2K engagements


"Movies are great though. Even if you set aside the pure artistic enjoyment (you shouldnt). Movies are stories and stories are powerful primal moving motivating. They are prompts to you to consider dilemmas and scenarios to build your world model and compass. My rec is to go to the golden age of story telling and movie making that imo ramped up in the 80s was roaring in 90s peaked early 00s and declined since. One sourcing example: pick a random year there look up Oscar winners pick and watch. Enjoy and attend guilt free"  
[X Link](https://x.com/karpathy/status/1987803670004826470)  2025-11-10T08:45Z 1.5M followers, 223.2K engagements


"I took delivery of a beautiful new shiny HW4 Tesla Model X today so I immediately took it out for an FSD test drive a bit like I used to do almost daily for X years. Basically. I'm amazed - it drives really really well smooth confident noticeably better than what I'm used to on HW3 (my previous car) and eons ahead of the version I remember driving up highway XXX on my first day at Tesla X years ago where I had to intervene every time the road mildly curved or sloped. (note this is v13 my car hasn't been offered the latest v14 yet) On the highway I felt like a passenger in some super high tech"  
[X Link](https://x.com/karpathy/status/1988705360723763242)  2025-11-12T20:28Z 1.5M followers, 17.8M engagements


"I am unreasonably excited about self-driving. It will be the first technology in many decades to visibly terraform outdoor physical spaces and way of life. Less parked cars. Less parking lots. Much greater safety for people in and out of cars. Less noise pollution. More space reclaimed for humans. Human brain cycles and attention capital freed up from lane following to other pursuits. Cheaper faster programmable delivery of physical items and goods. It wont happen overnight but there will be the era before and the era after"  
[X Link](https://x.com/karpathy/status/1989078861800411219)  2025-11-13T21:12Z 1.5M followers, 1.5M engagements


"A number of people are talking about implications of AI to schools. I spoke about some of my thoughts to a school board earlier some highlights: X. You will never be able to detect the use of AI in homework. Full stop. All "detectors" of AI imo don't really work can be defeated in various ways and are in principle doomed to fail. You have to assume that any work done outside classroom has used AI. X. Therefore the majority of grading has to shift to in-class work (instead of at-home assignments) in settings where teachers can physically monitor students. The students remain motivated to learn"  
[X Link](https://x.com/karpathy/status/1993010584175141038)  2025-11-24T17:35Z 1.5M followers, 2.5M engagements


"A good chunk of people misunderstood this tweet btw which is my bad. I am not suggesting people use the old style promoting techniques of you are an expert swift programmer or etc. its ok"  
[X Link](https://x.com/karpathy/status/1998245684521353664)  2025-12-09T04:17Z 1.5M followers, 113.8K engagements


"Quick new post: Auto-grading decade-old Hacker News discussions with hindsight I took all the XXX frontpage Hacker News article+discussion of December 2015 and asked the GPT XXX Thinking API to do an in-hindsight analysis to identify the most/least prescient comments. This took X hours to vibe code and X hour and $XX to run. The idea was sparked by the HN article yesterday where Gemini X was asked to hallucinate the HN front page one decade forward. More generally: X. in-hindsight analysis has always fascinated me as a way to train your forward prediction model so reading the results is"  
[X Link](https://x.com/karpathy/status/1998803709468487877)  2025-12-10T17:15Z 1.5M followers, 343K engagements


"New 3h31m video on YouTube: "Deep Dive into LLMs like ChatGPT" This is a general audience deep dive into the Large Language Model (LLM) AI technology that powers ChatGPT and related products. It is covers the full training stack of how the models are developed along with mental models of how to think about their "psychology" and how to get the best use them in practical applications. We cover all the major stages: X. pretraining: data tokenization Transformer neural network I/O and internals inference GPT-2 training example Llama XXX base inference examples X. supervised finetuning:"  
[X Link](https://x.com/karpathy/status/1887211193099825254)  2025-02-05T18:46Z 1.5M followers, 2.4M engagements


"+1 for "context engineering" over "prompt engineering". People associate prompts with short task descriptions you'd give an LLM in your day-to-day use. When in every industrial-strength LLM app context engineering is the delicate art and science of filling the context window with just the right information for the next step. Science because doing this right involves task descriptions and explanations few shot examples RAG related (possibly multimodal) data tools state and history compacting. Too little or of the wrong form and the LLM doesn't have the right context for optimal performance."  
[X Link](https://x.com/karpathy/status/1937902205765607626)  2025-06-25T15:54Z 1.5M followers, 2.4M engagements


"Tinker is cool. If you're a researcher/developer tinker dramatically simplifies LLM post-training. You retain XX% of algorithmic creative control (usually related to data loss function the algorithm) while tinker handles the hard parts that you usually want to touch much less often (infra forward/backward of the LLM itself distributed training) meaning you can do these at well below XX% of typical complexity involved. Compared to the more common and existing paradigm of "upload your data we'll post-train your LLM" this is imo a more clever place to "slice up" the complexity of post-training"  
[X Link](https://x.com/karpathy/status/1973468610917179630)  2025-10-01T19:22Z 1.5M followers, 734.1K engagements


"Nice short post illustrating how simple text (discrete) diffusion can be. Diffusion (i.e. parallel iterated denoising top) is the pervasive generative paradigm in image/video but autoregression (i.e. go left to right bottom) is the dominant paradigm in text. For audio I've seen a bit of both. A lot of diffusion papers look a bit dense but if you strip the mathematical formalism you end up with simple baseline algorithms e.g. something a lot closer to flow matching in continuous or something like this in discrete. It's your vanilla transformer but with bi-directional attention where you"  
[X Link](https://x.com/karpathy/status/1980347971935068380)  2025-10-20T18:58Z 1.5M followers, 858.4K engagements


"As a fun Saturday vibe code project and following up on this tweet earlier I hacked up an **llm-council** web app. It looks exactly like ChatGPT except each user query is 1) dispatched to multiple models on your council using OpenRouter e.g. currently: "openai/gpt-5.1" "google/gemini-3-pro-preview" "anthropic/claude-sonnet-4.5" "x-ai/grok-4" Then 2) all models get to see each other's (anonymized) responses and they review and rank them and then 3) a "Chairman LLM" gets all of that as context and produces the final response. It's interesting to see the results from multiple models side by side"  
[X Link](https://x.com/karpathy/status/1992381094667411768)  2025-11-22T23:54Z 1.5M followers, 5.2M engagements


"Gemini Nano Banana Pro can solve exam questions *in* the exam page image. With doodles diagrams all that. ChatGPT thinks these solutions are all correct except Se_2P_2 should be "diselenium diphosphide" and a spelling mistake (should be "thiocyanic acid" not "thoicyanic") :O"  
[X Link](https://x.com/karpathy/status/1992655330002817095)  2025-11-23T18:03Z 1.5M followers, 2.9M engagements


"In today's episode of programming horror. In the Python docs of random.seed() def we're told "If a is an int it is used directly." X But if you seed with X or -X you actually get the exact same rng object producing the same streams. (TIL). In nanochat I was using the sign as a (what I thought was) clever way to get different rng sequences for train/test splits. Hence gnarly bug because now train=test. I found the CPython code responsible in cpython/Modules/_randommodule.c X where on line XXX we see in a comment: "This algorithm relies on the number being unsigned. So: if the arg is a PyLong"  
[X Link](https://x.com/karpathy/status/1998236299862659485)  2025-12-09T03:40Z 1.5M followers, 709.3K engagements


"@knowtrend_ai Great idea"  
[X Link](https://x.com/karpathy/status/1998811953020711092)  2025-12-10T17:48Z 1.5M followers, 14.9K engagements

[GUEST ACCESS MODE: Data is scrambled or limited to provide examples. Make requests using your API key to unlock full data. Check https://lunarcrush.ai/auth for authentication information.]

@karpathy Avatar @karpathy Andrej Karpathy

Andrej Karpathy posts on X about llm, ai, if you, all the the most. They currently have XXXXXXXXX followers and XXX posts still getting attention that total XXXXXXX engagements in the last XX hours.

Engagements: XXXXXXX #

Engagements Line Chart

  • X Week XXXXXXXXX +614%
  • X Month XXXXXXXXXX +81%
  • X Months XXXXXXXXXXX +55%
  • X Year XXXXXXXXXXX +0.50%

Mentions: XX #

Mentions Line Chart

  • X Week XXX +7.50%
  • X Month XXX +8.70%
  • X Months XXX +18%
  • X Year XXX +19%

Followers: XXXXXXXXX #

Followers Line Chart

  • X Week XXXXXXXXX +0.86%
  • X Month XXXXXXXXX +3.50%
  • X Months XXXXXXXXX +17%
  • X Year XXXXXXXXX +37%

CreatorRank: XXXXXX #

CreatorRank Line Chart

Social Influence

Social category influence technology brands #2371 social networks XXXX% finance XXXX% stocks #430 nfts #145 automotive brands XXXX%

Social topic influence llm #1, ai 8.74%, if you #5656, all the #382, karpathy #1, to the 3.88%, nano banana #96, imo #65, banana #359, open ai #415

Top accounts mentioned or mentioned by @grok @flolight44 @billstenner7 @johntheadman_ @elonmusk @dataexec @adamskyart @mixedrealityman @yuchenj_uw @gp_pulipaka @cryptosausage @spil____ @adarkm0ment @danadvantage @marswalkerr @mohamedatta_911 @adelayida210519 @_thomasip @jasonth0 @rileyralmuto

Top assets mentioned Doodles (doodles) Alphabet Inc Class A (GOOGL) Tesla, Inc. (TSLA)

Top Social Posts

Top posts by engagements in the last XX hours

"Finally had a chance to listen through this pod with Sutton which was interesting and amusing. As background Sutton's "The Bitter Lesson" has become a bit of biblical text in frontier LLM circles. Researchers routinely talk about and ask whether this or that approach or idea is sufficiently "bitter lesson pilled" (meaning arranged so that it benefits from added computation for free) as a proxy for whether it's going to work or worth even pursuing. The underlying assumption being that LLMs are of course highly "bitter lesson pilled" indeed just look at LLM scaling laws where if you put compute"
X Link 2025-10-01T17:09Z 1.5M followers, 2M engagements

"@zenitsu_aprntc Good question it's basically entirely hand-written (with tab autocomplete). I tried to use claude/codex agents a few times but they just didn't work well enough at all and net unhelpful possibly the repo is too far off the data distribution"
X Link 2025-10-13T15:27Z 1.5M followers, 484.3K engagements

"Deliberately*"
X Link 2025-10-16T02:50Z 1.5M followers, 146.3K engagements

"@proggineer Agree sometimes that is helpful too to have an overview of what the whole thing is about first. I just copy paste stuff around to LLM of the day (I cycle) theres no tool"
X Link 2025-11-18T00:39Z 1.5M followers, 59.9K engagements

"Has anyone encountered a good definition of slop. In a quantitative measurable sense. My brain has an intuitive slop index I can reliably estimate but Im not sure how to define it. I have some bad ideas that involve the use of LLM miniseries and thinking token budgets"
X Link 2025-11-22T02:11Z 1.5M followers, 633K engagements

"@_thomasip haha yes it makes mistakes You have to re-roll a few times until it's right. Sometimes it gets stuck in loops and you have to re-start in a new conversation. Example re-roll:"
X Link 2025-11-23T22:03Z 1.5M followers, 23.7K engagements

"@theJayAlto Edutainment. This one weird trick to consume entertainment and feel good about it"
X Link 2025-11-29T17:20Z 1.5M followers, 162.9K engagements

"@nickcammarata Your recent posts on this remind me of this Arnold gem +100 though. I finally had a chance to install a home gym recently making it trivial to use daily. Always looking forward to the next exercise high. Slightly miss the social/entropy aspects of gyms"
X Link 2025-12-04T18:54Z 1.5M followers, 184.7K engagements

"reminded of this paragraph from gsm8k paper 2021 :)"
X Link 2025-09-13T16:08Z 1.5M followers, 378.9K engagements

"@anneshu_nag I love that it's called Nano Banana instead of Google Imagine or some other corpo thing"
X Link 2025-11-23T21:49Z 1.5M followers, 20.5K engagements

"I've had medium success asking LLMs if a thing exists it works out of the box for some of the more well-known things (e.g. both GPT XXX and Gemini X know about this function if you describe the tensor transformation in words). For more esoteric or new libraries (e.g. uv being a recent example) I've had more success manually packaging up docs into markdown and including it as context for questions. PyTorch docs now also seem to have an "Ask AI" that presumably does RAG over their docs but the model is not that bright. But you'd have to suspect it might exist and attempt to ask in the first"
X Link 2025-11-25T17:20Z 1.5M followers, 96.6K engagements

"Don't think of LLMs as entities but as simulators. For example when exploring a topic don't ask: "What do you think about xyz" There is no "you". Next time try: "What would be a good group of people to explore xyz What would they say" The LLM can channel/simulate many perspectives but it hasn't "thought about" xyz for a while and over time and formed its own opinions in the way we're used to. If you force it via the use of "you" it will give you something by adopting a personality embedding vector implied by the statistics of its finetuning data and then simulate that. It's fine to do but"
X Link 2025-12-07T18:13Z 1.5M followers, 3.6M engagements

"We will see that a lot of weird behaviors and problems of LLMs actually trace back to tokenization. We'll go through a number of these issues discuss why tokenization is at fault and why someone out there ideally finds a way to delete this stage entirely"
X Link 2024-02-20T17:40Z 1.5M followers, 753.3K engagements

"@matejhladky_dev AI has crushed it since this post way beyond expectation. I made the same category of mistake all of AI was making of thinking we have to discover and write the algorithm. You don't. You pretrain and then finetune a BIG neural network on lots of tasks and it just falls out. lol"
X Link 2025-11-25T17:31Z 1.5M followers, 90.8K engagements

"# on shortification of "learning" There are a lot of videos on YouTube/TikTok etc. that give the appearance of education but if you look closely they are really just entertainment. This is very convenient for everyone involved : the people watching enjoy thinking they are learning (but actually they are just having fun). The people creating this content also enjoy it because fun has a much larger audience fame and revenue. But as far as learning goes this is a trap. This content is an epsilon away from watching the Bachelorette. It's like snacking on those "Garden Veggie Straws" which feel"
X Link 2024-02-10T18:10Z 1.5M followers, 2.2M engagements

"⚡ Excited to share that I am starting an AI+Education company called Eureka Labs. The announcement: --- We are Eureka Labs and we are building a new kind of school that is AI native. How can we approach an ideal experience for learning something new For example in the case of physics one could imagine working through very high quality course materials together with Feynman who is there to guide you every step of the way. Unfortunately subject matter experts who are deeply passionate great at teaching infinitely patient and fluent in all of the world's languages are also very scarce and cannot"
X Link 2024-07-16T17:25Z 1.5M followers, 2.5M engagements

"DeepSeek (Chinese AI co) making it look easy today with an open weights release of a frontier-grade LLM trained on a joke of a budget (2048 GPUs for X months $6M). For reference this level of capability is supposed to require clusters of closer to 16K GPUs the ones being brought up today are more around 100K GPUs. E.g. Llama X 405B used 30.8M GPU-hours while DeepSeek-V3 looks to be a stronger model at only 2.8M GPU-hours (11X less compute). If the model also passes vibe checks (e.g. LLM arena rankings are ongoing my few quick tests went well so far) it will be a highly impressive display of"
X Link 2024-12-26T19:23Z 1.5M followers, 6.5M engagements

"Continuing the journey of optimal LLM-assisted coding experience. In particular I find that instead of narrowing in on a perfect one thing my usage is increasingly diversifying across a few workflows that I "stitch up" the pros/cons of: Personally the bread & butter (75%) of my LLM assistance continues to be just (Cursor) tab complete. This is because I find that writing concrete chunks of code/comments myself and in the right part of the code is a high bandwidth way of communicating "task specification" to the LLM i.e. it's primarily about task specification bits - it takes too many bits and"
X Link 2025-08-24T19:46Z 1.5M followers, 687.8K engagements

"I think congrats again to OpenAI for cooking with GPT-5 Pro. This is the third time I've struggled on something complex/gnarly for an hour on and off with CC then X Pro goes off for XX minutes and comes back with code that works out of the box. I had CC read the X Pro version and it wrote up X paragraphs admiring it (very wholesome). If you're not giving it your hardest problems you're probably missing out"
X Link 2025-09-05T17:38Z 1.5M followers, 2.6M engagements

"nanochat now has a primordial identity and can talk a bit about itself and its capabilities (e.g. it knows it's nanochat d32 that cost $XXX that it was built by me that it can't speak languages other than English too well and why etc.). This kind of customization is all done through synthetic data generation and I uploaded a new example script to demonstrate. It's a bit subtle but by default LLMs have no inherent personality or any understanding of their own capabilities because they are not animal-like entities. They don't know what they are or what they can or can't do or know or don't"
X Link 2025-10-21T15:59Z 1.5M followers, 455.5K engagements

"@LucasAtkins7 This code is extremely dangerous. Here I improved it"
X Link 2025-10-22T14:45Z 1.5M followers, 1.5M engagements

"Sharing an interesting recent conversation on AI's impact on the economy. AI has been compared to various historical precedents: electricity industrial revolution etc. I think the strongest analogy is that of AI as a new computing paradigm (Software 2.0) because both are fundamentally about the automation of digital information processing. If you were to forecast the impact of computing on the job market in 1980s the most predictive feature of a task/job you'd look at is to what extent the algorithm of it is fixed i.e. are you just mechanically transforming information according to rote easy"
X Link 2025-11-16T17:56Z 1.5M followers, 2.1M engagements

"Finally had time to read & process this great post. I run into the pattern quite often it goes: "something that sounds wrong is good actually because galaxy brain reason" Galaxy brain reasoning is the best way to justify anything while looking / feeling good about it. From this perspective for example there's deeper wisdom in the Ten Commandments imposing constraints over actions instead of utility over states. It's not Ten Objectives. E.g. they don't attempt to define a utility function for the value of life they simply say "Thou shalt not kill". This approach curtails the relatively"
X Link 2025-11-17T18:56Z 1.5M followers, 839.5K engagements

"Im starting to get into a habit of reading everything (blogs articles book chapters) with LLMs. Usually pass X is manual then pass X explain/summarize pass X Q&A. I usually end up with a better/deeper understanding than if I moved on. Growing to among top use cases. On the flip side if youre a writer trying to explain/communicate something we may increasingly see less of a mindset of Im writing this for another human and more Im writing this for an LLM. Because once an LLM gets it it can then target personalize and serve the idea to its user"
X Link 2025-11-18T00:29Z 1.5M followers, 2.8M engagements

"I put up a simple repo I call reader3 (it's my 3rd version.) to illustrate how I read EPUBs with LLMs. Basically get some epub (e.g. Project Gutenberg is great) go chapter by chapter and with this you can easily copy paste text to your favorite LLM"
X Link 2025-11-18T02:44Z 1.5M followers, 292.9K engagements

"I played with Gemini X yesterday via early access. Few thoughts - First I usually urge caution with public benchmarks because imo they can be quite possible to game. It comes down to discipline and self-restraint of the team (who is meanwhile strongly incentivized otherwise) to not overfit test sets via elaborate gymnastics over test-set adjacent data in the document embedding space. Realistically because everyone else is doing it the pressure to do so is high. Go talk to the model. Talk to the other models (Ride the LLM Cycle - use a different LLM every day). I had a positive early"
X Link 2025-11-18T18:49Z 1.5M followers, 1.2M engagements

"My most amusing interaction was where the model (I think I was given some earlier version with a stale system prompt) refused to believe me that it is 2025 and kept inventing reasons why I must be trying to trick it or playing some elaborate joke on it. I kept giving it images and articles from "the future" and it kept insisting it was all fake. It accused me of using generative AI to defeat its challenges and argued why real wikipedia entries were actually generated and what the "dead giveaways" are. It highlighted tiny details when I gave it Google Image Search results arguing why the"
X Link 2025-11-18T18:51Z 1.5M followers, 1M engagements

"Something I think people continue to have poor intuition for: The space of intelligences is large and animal intelligence (the only kind we've ever known) is only a single point arising from a very specific kind of optimization that is fundamentally distinct from that of our technology. Animal intelligence optimization pressure: - innate and continuous stream of consciousness of an embodied "self" a drive for homeostasis and self-preservation in a dangerous physical world. - thoroughly optimized for natural selection = strong innate drives for power-seeking status dominance reproduction. many"
X Link 2025-11-21T16:43Z 1.5M followers, 2.6M engagements

"@TheVixhal your post challenged me. every one of your points is wrong but i had to think about each for a while :)"
X Link 2025-11-21T17:35Z 1.5M followers, 1.8M engagements

"@NickADobos I dont super love shoggoth represented as a kind of biological monster (animal) it feels wrong/misleading in this sense"
X Link 2025-11-21T23:52Z 1.5M followers, 49.2K engagements

"Imo this is along the lines of how talking to an LLM via text is like typing into a DOS Terminal and "GUI hasn't been invented yet" of some of my earlier posts. The GUI is an intelligent canvas"
X Link 2025-11-23T18:11Z 1.5M followers, 252.1K engagements

"I asked it to create a personalized weekly workout plan and then posters that I can print on the wall to remind me what exercises to do each day. Tuesday looks more intense because I asked for "more testosterone" :D. (sorry I'll stop posting more nano banana pro stuff now)"
X Link 2025-11-23T21:45Z 1.5M followers, 385.9K engagements

"Happy weekend to those who celebrate"
X Link 2025-12-07T15:59Z 1.5M followers, 1.3M engagements

"@Marswalkerr I love this meme too haha"
X Link 2025-12-07T16:04Z 1.5M followers, 44.7K engagements

"There is definitely work going into engineering the "you" simulation - the personality that gets all the rewards in verifiable problems or all the upvotes from users/judge LLMs or mimics the responses of SFT and there is an emergent composite personality from that. My point is more that the "you" there is deliberately bolted on engineered and layered on what is fundamentally a token simulation engine not a mind that is somehow emergent and over time constructed in a relatable way to an average person talking to an AI. The story is a bit more simple in verifiable domains but I think more"
X Link 2025-12-07T20:06Z 1.5M followers, 126.7K engagements

"I could certainly imagine that "nesting" the simulation might be too "effortful" for the model compute or data density wise. My results with it are not too bad so imo it's at least worth people try / experiment with / think about. For example it might be useful to read multiple distinct and approximate perspectives on topic xyz instead of one. Research-wise you might be able to elicit LLM Council - like benefits (not via diverse LLMs but via diverse simulations) and improve performance via the generator-discriminator gap effects or ensembling effects"
X Link 2025-12-08T17:27Z 1.5M followers, 28.8K engagements

"ty to ericsilberstein1 on github for spotting the bug. (it's not a big bug and only comes up in the SpellingBee synthetic task evaluation but still)"
X Link 2025-12-09T03:57Z 1.5M followers, 119K engagements

"nanoGPT - the first LLM to train and inference in space 🥹. It begins"
X Link 2025-12-10T17:25Z 1.5M followers, 759.7K engagements

"The hottest new programming language is English"
X Link 2023-01-24T20:14Z 1.5M followers, 9M engagements

"# Reproduce GPT-2 (124M) in llm.c in XX minutes for $XX ✨ The GPT-2 (124M) is the smallest model in the GPT-2 series released by OpenAI in 2019 and is actually quite accessible today even for the GPU poor. For example with llm.c you can now reproduce this model on one 8X A100 80GB SXM node in XX minutes (at XX% MFU). As they run for $14/hr this is $XX. I also think the 124M model makes for an excellent "cramming" challenge for training it very fast. So here is the launch command: And here is the output after XX minutes training on 10B tokens of the FineWeb dataset: It feels really nice to"
X Link 2024-05-28T15:57Z 1.5M followers, 663.7K engagements

"There's a new kind of coding I call "vibe coding" where you fully give in to the vibes embrace exponentials and forget that the code even exists. It's possible because the LLMs (e.g. Cursor Composer w Sonnet) are getting too good. Also I just talk to Composer with SuperWhisper so I barely even touch the keyboard. I ask for the dumbest things like "decrease the padding on the sidebar by half" because I'm too lazy to find it. I "Accept All" always I don't read the diffs anymore. When I get error messages I just copy paste them in with no comment usually that fixes it. The code grows beyond my"
X Link 2025-02-02T23:17Z 1.5M followers, 5.3M engagements

"I was given early access to Grok X earlier today making me I think one of the first few who could run a quick vibe check. Thinking ✅ First Grok X clearly has an around state of the art thinking model ("Think" button) and did great out of the box on my Settler's of Catan question: "Create a board game webpage showing a hex grid just like in the game Settlers of Catan. Each hex grid is numbered from 1.N where N is the total number of hex tiles. Make it generic so one can change the number of "rings" using a slider. For example in Catan the radius is X hexes. Single html page please." Few models"
X Link 2025-02-18T05:25Z 1.5M followers, 3.7M engagements

"Agency Intelligence I had this intuitively wrong for decades I think due to a pervasive cultural veneration of intelligence various entertainment/media obsession with IQ etc. Agency is significantly more powerful and significantly more scarce. Are you hiring for agency Are we educating for agency Are you acting as if you had 10X agency Grok explanation is close: Agency as a personality trait refers to an individual's capacity to take initiative make decisions and exert control over their actions and environment. Its about being proactive rather than reactivesomeone with high agency doesnt"
X Link 2025-02-24T18:58Z 1.5M followers, 7.3M engagements

""Finding the Best Sleep Tracker" Results of an experiment where I wore X sleep trackers every night for X months. TLDR Whoop = Oura 8Sleep Apple Watch + AutoSleep. Link simply right here instead of in a reply because ()/"
X Link 2025-03-30T16:41Z 1.5M followers, 1.6M engagements

"I attended a vibe coding hackathon recently and used the chance to build a web app (with auth payments deploy etc.). I tinker but I am not a web dev by background so besides the app I was very interested in what it's like to vibe code a full web app today. As such I wrote none of the code directly (Cursor+Claude/o3 did) and I don't really know how the app works in the conventional sense that I'm used to as an engineer. The app is called MenuGen and it is live on Basically I'm often confused about what all the things on a restaurant menu are - e.g. Pt Tagine Cavatappi or Sweetbread (hint it's."
X Link 2025-05-01T15:16Z 1.5M followers, 785.3K engagements

"An attempt to explain (current) ChatGPT versions. I still run into many many people who don't know that: - o3 is the obvious best thing for important/hard things. It is a reasoning model that is much stronger than 4o and if you are using ChatGPT professionally and not using o3 you're ngmi. - 4o is different from o4. Yes I know lol. 4o is a good "daily driver" for many easy-medium questions. o4 is only available as mini for now and is not as good as o3 and I'm not super sure why it's out right now. Example basic "router" in my own personal use: - Any simple query (e.g. "what foods are high in"
X Link 2025-06-02T17:54Z 1.5M followers, 1.4M engagements

"My sleep scores during recent travel were in the 90s. Now back in SF I am consistently back down to 70s 80s. I am increasingly convinced that this is due to traffic noise from a nearby road/intersection where I live - every 10min a car truck bus or motorcycle with a very loud engine passes by (some are 10X louder than others). In the later less deep stages of sleep it is much easier to wake and then much harder to go back to sleep. More generally I think noise pollution (esp early hours) come at a huge societal cost that is not correctly accounted for. E.g. I wouldn't be too surprised if a"
X Link 2025-06-07T19:01Z 1.5M followers, 1.5M engagements

"The race for LLM "cognitive core" - a few billion param model that maximally sacrifices encyclopedic knowledge for capability. It lives always-on and by default on every computer as the kernel of LLM personal computing. Its features are slowly crystalizing: - Natively multimodal text/vision/audio at both input and output. - Matryoshka-style architecture allowing a dial of capability up and down at test time. - Reasoning also with a dial. (system 2) - Aggressively tool-using. - On-device finetuning LoRA slots for test-time training personalization and customization. - Delegates and double"
X Link 2025-06-27T15:52Z 1.5M followers, 1.3M engagements

"I often rant about how XX% of attention is about to be LLM attention instead of human attention. What does a research paper look like for an LLM instead of a human Its definitely not a pdf. There is huge space for an extremely valuable research app that figures this out"
X Link 2025-07-10T20:45Z 1.5M followers, 905.6K engagements

"Scaling up RL is all the rage right now I had a chat with a friend about it yesterday. I'm fairly certain RL will continue to yield more intermediate gains but I also don't expect it to be the full story. RL is basically "hey this happened to go well (/poorly) let me slightly increase (/decrease) the probability of every action I took for the future". You get a lot more leverage from verifier functions than explicit supervision this is great. But first it looks suspicious asymptotically - once the tasks grow to be minutes/hours of interaction long you're really going to do all that work just"
X Link 2025-07-13T16:35Z 1.5M followers, 1.1M engagements

"I'm noticing that due to (I think) a lot of benchmarkmaxxing on long horizon tasks LLMs are becoming a little too agentic by default a little beyond my average use case. For example in coding the models now tend to reason for a fairly long time they have an inclination to start listing and grepping files all across the entire repo they do repeated web searchers they over-analyze and over-think little rare edge cases even in code that is knowingly incomplete and under active development and often come back minutes later even for simple queries. This might make sense for long-running tasks but"
X Link 2025-08-09T16:53Z 1.5M followers, 1M engagements

"In era of pretraining what mattered was internet text. You'd primarily want a large diverse high quality collection of internet documents to learn from. In era of supervised finetuning it was conversations. Contract workers are hired to create answers for questions a bit like what you'd see on Stack Overflow / Quora or etc. but geared towards LLM use cases. Neither of the two above are going away (imo) but in this era of reinforcement learning it is now environments. Unlike the above they give the LLM an opportunity to actually interact - take actions see outcomes etc. This means you can hope"
X Link 2025-08-27T20:34Z 1.5M followers, 951.4K engagements

"Transforming human knowledge sensors and actuators from human-first and human-legible to LLM-first and LLM-legible is a beautiful space with so much potential and so much can be done. One example I'm obsessed with recently - for every textbook pdf/epub there is a perfect "LLMification" of it intended not for human but for an LLM (though it is a non-trivial transformation that would need human in the loop involvement). - All of the exposition is extracted into a markdown document including all latex styling (bold/italic) tables lists etc. All of the figures are extracted as images. - All"
X Link 2025-08-28T18:07Z 1.5M followers, 724.9K engagements

""AI isn't replacing radiologists" good article Expectation: rapid progress in image recognition AI will delete radiology jobs (e.g. as famously predicted by Geoff Hinton now almost a decade ago). Reality: radiology is doing great and is growing. There are a lot of imo naive predictions out there on the imminent impact of AI on the job market. E.g. a year ago I was asked by someone who should know better if I think there will be any software engineers still today. (Spoiler: I think we're going to make it). This is happening too broadly. The post goes into detail on why it's not that simple"
X Link 2025-09-25T14:29Z 1.5M followers, 2.3M engagements

"Something I am experimenting with. I copy pasted: 1) the full podcast transcript 2) the bitter lesson blog post 3) my full post above To ChatGPT. The interesting part is you can fork the conversation context to ask any questions and take it in whatever direction with chat:"
X Link 2025-10-01T17:44Z 1.5M followers, 148.2K engagements

"Hah judging by mentions overnight people seem to find the ghost analogy provocative. I swear I don't wake up just trying to come with new memes but to elaborate briefly why I thought it was a fun comparison: 1) It captures the idea that LLMs are purely digital artifacts that don't interact with the physical world (unlike animals which are very embodied). 2) Ghosts are a kind of "echo" of the living in this case a statistical distillation of humanity. 3) There is an air of mystery over both ghosts and LLMs as in we don't fully understand what they are or how they work. 4) The process of"
X Link 2025-10-02T14:25Z 1.5M followers, 257.8K engagements

"Every company needs a DM POC - someone high up who you can just DM the most obvious things and who shortcuts the PM hierarchy"
X Link 2025-10-04T14:31Z 1.5M followers, 563.5K engagements

"I don't know what labs are doing to these poor LLMs during RL but they are mortally terrified of exceptions in any infinitesimally likely case. Exceptions are a normal part of life and healthy dev process. Sign my LLM welfare petition for improved rewards in cases of exceptions"
X Link 2025-10-09T00:10Z 1.5M followers, 715K engagements

"POV: Your LLM agent is dividing a by b"
X Link 2025-10-09T00:31Z 1.5M followers, 390.3K engagements

"Excited to release new repo: nanochat (it's among the most unhinged I've written). Unlike my earlier similar repo nanoGPT which only covered pretraining nanochat is a minimal from scratch full-stack training/inference pipeline of a simple ChatGPT clone in a single dependency-minimal codebase. You boot up a cloud GPU box run a single script and in as little as X hours later you can talk to your own LLM in a ChatGPT-like web UI. It weighs 8000 lines of imo quite clean code to: - Train the tokenizer using a new Rust implementation - Pretrain a Transformer LLM on FineWeb evaluate CORE score"
X Link 2025-10-13T15:16Z 1.5M followers, 5.8M engagements

"And an example of some of the summary metrics produced by the $XXX speedrun in the report card to start. The current code base is a bit over 8000 lines but I tried to keep them clean and well-commented. Now comes the fun part - of tuning and hillclimbing"
X Link 2025-10-13T15:16Z 1.5M followers, 183.3K engagements

"nanochat d32 i.e. the depth XX version that I specced for $1000 up from $XXX has finished training after XX hours and looks good. All the metrics go up quite a bit across pretraining SFT and RL. CORE score of XXXX is now well above GPT-2 at XXXX. GSM8K went X% - XX% etc. So that's encouraging. The model is pretty fun to talk to but judging from some early interactions I think people have a little bit too much expectation for these micro models. There is a reason that frontier LLM labs raise billions to train their models. nanochat models cost $XXX - $1000 to train from scratch. The $100"
X Link 2025-10-16T00:14Z 1.5M followers, 266.1K engagements

"TV in the 90s: you turn it on you watch. TV 2025: - turn on wait for it to load - popup: TV wants to update 1.5GB. No. - scroll sideways find prime video app or etc - popup: now app wants to update 500MB. No - App launching. App loading - select account screen - 🫠"
X Link 2025-10-16T02:47Z 1.5M followers, 1.7M engagements

"There is a movement I found on Instagram where people delivery choose to live in 90s refusing all technology after 2000. Like an intermediate form of the Amish"
X Link 2025-10-16T02:50Z 1.5M followers, 297.3K engagements

"My pleasure to come on Dwarkesh last week I thought the questions and conversation were really good. I re-watched the pod just now too. First of all yes I know and I'm sorry that I speak so fast :). It's to my detriment because sometimes my speaking thread out-executes my thinking thread so I think I botched a few explanations due to that and sometimes I was also nervous that I'm going too much on a tangent or too deep into something relatively spurious. Anyway a few notes/pointers: AGI timelines. My comments on AGI timelines looks to be the most trending part of the early response. This is"
X Link 2025-10-18T20:23Z 1.5M followers, 4.1M engagements

"I quite like the new DeepSeek-OCR paper. It's a good OCR model (maybe a bit worse than dots) and yes data collection etc. but anyway it doesn't matter. The more interesting part for me (esp as a computer vision at heart who is temporarily masquerading as a natural language person) is whether pixels are better inputs to LLMs than text. Whether text tokens are wasteful and just terrible at the input. Maybe it makes more sense that all inputs to LLMs should only ever be images. Even if you happen to have pure text input maybe you'd prefer to render it and then feed that in: - more information"
X Link 2025-10-20T22:13Z 1.5M followers, 3.3M engagements

"@r_chirra I fixed it :) deployed live now. This was done by doing a round of synthetic data generation to collect a 1000 multi-turn conversations (given a bunch of information including the readme of the nanochat project) and then mixing that into midtraining and SFT. fun"
X Link 2025-10-21T05:36Z 1.5M followers, 456.3K engagements

"Last night I taught nanochat d32 how to count 'r' in strawberry (or similar variations). I thought this would be a good/fun example of how to add capabilities to nanochat and I wrote up a full guide here: This is done via a new synthetic task SpellingBee that generates examples of a user asking for this kind of a problem and an ideal solution from an assistant. We then midtrain/SFT finetune on these to endow the LLM with the capability or further train with RL to make it more robust. There are many details to get right especially at smaller model sizes and the guide steps through them. As a"
X Link 2025-10-24T15:35Z 1.5M followers, 568.4K engagements

"Beautiful technical debugging detective longread that starts with a suspicious loss curve and ends all the way in the Objective-C++ depths of PyTorch MPS backend of addcmul_ that silently fails on non-contiguous output tensors. I wonder how long before an LLM can do all of this"
X Link 2025-10-26T16:24Z 1.5M followers, 601.2K engagements

"Movies are great though. Even if you set aside the pure artistic enjoyment (you shouldnt). Movies are stories and stories are powerful primal moving motivating. They are prompts to you to consider dilemmas and scenarios to build your world model and compass. My rec is to go to the golden age of story telling and movie making that imo ramped up in the 80s was roaring in 90s peaked early 00s and declined since. One sourcing example: pick a random year there look up Oscar winners pick and watch. Enjoy and attend guilt free"
X Link 2025-11-10T08:45Z 1.5M followers, 223.2K engagements

"I took delivery of a beautiful new shiny HW4 Tesla Model X today so I immediately took it out for an FSD test drive a bit like I used to do almost daily for X years. Basically. I'm amazed - it drives really really well smooth confident noticeably better than what I'm used to on HW3 (my previous car) and eons ahead of the version I remember driving up highway XXX on my first day at Tesla X years ago where I had to intervene every time the road mildly curved or sloped. (note this is v13 my car hasn't been offered the latest v14 yet) On the highway I felt like a passenger in some super high tech"
X Link 2025-11-12T20:28Z 1.5M followers, 17.8M engagements

"I am unreasonably excited about self-driving. It will be the first technology in many decades to visibly terraform outdoor physical spaces and way of life. Less parked cars. Less parking lots. Much greater safety for people in and out of cars. Less noise pollution. More space reclaimed for humans. Human brain cycles and attention capital freed up from lane following to other pursuits. Cheaper faster programmable delivery of physical items and goods. It wont happen overnight but there will be the era before and the era after"
X Link 2025-11-13T21:12Z 1.5M followers, 1.5M engagements

"A number of people are talking about implications of AI to schools. I spoke about some of my thoughts to a school board earlier some highlights: X. You will never be able to detect the use of AI in homework. Full stop. All "detectors" of AI imo don't really work can be defeated in various ways and are in principle doomed to fail. You have to assume that any work done outside classroom has used AI. X. Therefore the majority of grading has to shift to in-class work (instead of at-home assignments) in settings where teachers can physically monitor students. The students remain motivated to learn"
X Link 2025-11-24T17:35Z 1.5M followers, 2.5M engagements

"A good chunk of people misunderstood this tweet btw which is my bad. I am not suggesting people use the old style promoting techniques of you are an expert swift programmer or etc. its ok"
X Link 2025-12-09T04:17Z 1.5M followers, 113.8K engagements

"Quick new post: Auto-grading decade-old Hacker News discussions with hindsight I took all the XXX frontpage Hacker News article+discussion of December 2015 and asked the GPT XXX Thinking API to do an in-hindsight analysis to identify the most/least prescient comments. This took X hours to vibe code and X hour and $XX to run. The idea was sparked by the HN article yesterday where Gemini X was asked to hallucinate the HN front page one decade forward. More generally: X. in-hindsight analysis has always fascinated me as a way to train your forward prediction model so reading the results is"
X Link 2025-12-10T17:15Z 1.5M followers, 343K engagements

"New 3h31m video on YouTube: "Deep Dive into LLMs like ChatGPT" This is a general audience deep dive into the Large Language Model (LLM) AI technology that powers ChatGPT and related products. It is covers the full training stack of how the models are developed along with mental models of how to think about their "psychology" and how to get the best use them in practical applications. We cover all the major stages: X. pretraining: data tokenization Transformer neural network I/O and internals inference GPT-2 training example Llama XXX base inference examples X. supervised finetuning:"
X Link 2025-02-05T18:46Z 1.5M followers, 2.4M engagements

"+1 for "context engineering" over "prompt engineering". People associate prompts with short task descriptions you'd give an LLM in your day-to-day use. When in every industrial-strength LLM app context engineering is the delicate art and science of filling the context window with just the right information for the next step. Science because doing this right involves task descriptions and explanations few shot examples RAG related (possibly multimodal) data tools state and history compacting. Too little or of the wrong form and the LLM doesn't have the right context for optimal performance."
X Link 2025-06-25T15:54Z 1.5M followers, 2.4M engagements

"Tinker is cool. If you're a researcher/developer tinker dramatically simplifies LLM post-training. You retain XX% of algorithmic creative control (usually related to data loss function the algorithm) while tinker handles the hard parts that you usually want to touch much less often (infra forward/backward of the LLM itself distributed training) meaning you can do these at well below XX% of typical complexity involved. Compared to the more common and existing paradigm of "upload your data we'll post-train your LLM" this is imo a more clever place to "slice up" the complexity of post-training"
X Link 2025-10-01T19:22Z 1.5M followers, 734.1K engagements

"Nice short post illustrating how simple text (discrete) diffusion can be. Diffusion (i.e. parallel iterated denoising top) is the pervasive generative paradigm in image/video but autoregression (i.e. go left to right bottom) is the dominant paradigm in text. For audio I've seen a bit of both. A lot of diffusion papers look a bit dense but if you strip the mathematical formalism you end up with simple baseline algorithms e.g. something a lot closer to flow matching in continuous or something like this in discrete. It's your vanilla transformer but with bi-directional attention where you"
X Link 2025-10-20T18:58Z 1.5M followers, 858.4K engagements

"As a fun Saturday vibe code project and following up on this tweet earlier I hacked up an llm-council web app. It looks exactly like ChatGPT except each user query is 1) dispatched to multiple models on your council using OpenRouter e.g. currently: "openai/gpt-5.1" "google/gemini-3-pro-preview" "anthropic/claude-sonnet-4.5" "x-ai/grok-4" Then 2) all models get to see each other's (anonymized) responses and they review and rank them and then 3) a "Chairman LLM" gets all of that as context and produces the final response. It's interesting to see the results from multiple models side by side"
X Link 2025-11-22T23:54Z 1.5M followers, 5.2M engagements

"Gemini Nano Banana Pro can solve exam questions in the exam page image. With doodles diagrams all that. ChatGPT thinks these solutions are all correct except Se_2P_2 should be "diselenium diphosphide" and a spelling mistake (should be "thiocyanic acid" not "thoicyanic") :O"
X Link 2025-11-23T18:03Z 1.5M followers, 2.9M engagements

"In today's episode of programming horror. In the Python docs of random.seed() def we're told "If a is an int it is used directly." X But if you seed with X or -X you actually get the exact same rng object producing the same streams. (TIL). In nanochat I was using the sign as a (what I thought was) clever way to get different rng sequences for train/test splits. Hence gnarly bug because now train=test. I found the CPython code responsible in cpython/Modules/_randommodule.c X where on line XXX we see in a comment: "This algorithm relies on the number being unsigned. So: if the arg is a PyLong"
X Link 2025-12-09T03:40Z 1.5M followers, 709.3K engagements

"@knowtrend_ai Great idea"
X Link 2025-12-10T17:48Z 1.5M followers, 14.9K engagements

@karpathy
/creator/twitter::karpathy