[GUEST ACCESS MODE: Data is scrambled or limited to provide examples. Make requests using your API key to unlock full data. Check https://lunarcrush.ai/auth for authentication information.]

@karpathy Avatar @karpathy Andrej Karpathy

Andrej Karpathy posts on X about rl, open ai, all the, state of the most. They currently have XXXXXXXXX followers and XXX posts still getting attention that total XXXXXXXXX engagements in the last XX hours.

Engagements: XXXXXXXXX #

Engagements Line Chart

Mentions: XX #

Mentions Line Chart

Followers: XXXXXXXXX #

Followers Line Chart

CreatorRank: XXXXX #

CreatorRank Line Chart

Social Influence #


Social category influence technology brands #560 social networks XXXX% finance XXXX% travel destinations XXXX%

Social topic influence rl #107, open ai #7, all the #4107, state of #1058, imo #111, hex 0.51%, for your 0.51%, want a 0.51%, collection of 0.51%, quora XXXX%

Top accounts mentioned or mentioned by @ronald_vanloon @gork @genaisummitsf @elonmusk @grok @teksedge @levelsio @onthegoai @iamarkdev @clementdelangue @afterdinnerco @intuitmachine @chain_gpt @romanderrmd @uhdk1ng @swyx @suhail @danielmiessler @teknium1 @avinavsahoo

Top Social Posts #


Top posts by engagements in the last XX hours

"@zenitsu_aprntc Good question it's basically entirely hand-written (with tab autocomplete). I tried to use claude/codex agents a few times but they just didn't work well enough at all and net unhelpful possibly the repo is too far off the data distribution"
X Link @karpathy 2025-10-13T15:27Z 1.4M followers, 325.2K engagements

"Every company needs a DM POC - someone high up who you can just DM the most obvious things and who shortcuts the PM hierarchy"
X Link @karpathy 2025-10-04T14:31Z 1.4M followers, 535.7K engagements

"I was given early access to Grok X earlier today making me I think one of the first few who could run a quick vibe check. Thinking ✅ First Grok X clearly has an around state of the art thinking model ("Think" button) and did great out of the box on my Settler's of Catan question: "Create a board game webpage showing a hex grid just like in the game Settlers of Catan. Each hex grid is numbered from 1.N where N is the total number of hex tiles. Make it generic so one can change the number of "rings" using a slider. For example in Catan the radius is X hexes. Single html page please." Few models"
X Link @karpathy 2025-02-18T05:25Z 1.4M followers, 3.7M engagements

"For your professional programming do you use mostly:"
X Link @karpathy 2025-10-02T23:28Z 1.4M followers, 340.2K engagements

"In era of pretraining what mattered was internet text. You'd primarily want a large diverse high quality collection of internet documents to learn from. In era of supervised finetuning it was conversations. Contract workers are hired to create answers for questions a bit like what you'd see on Stack Overflow / Quora or etc. but geared towards LLM use cases. Neither of the two above are going away (imo) but in this era of reinforcement learning it is now environments. Unlike the above they give the LLM an opportunity to actually interact - take actions see outcomes etc. This means you can hope"
X Link @karpathy 2025-08-27T20:34Z 1.4M followers, 925.8K engagements

"reminded of this paragraph from gsm8k paper 2021 :)"
X Link @karpathy 2025-09-13T16:08Z 1.4M followers, 366.3K engagements

"And an example of some of the summary metrics produced by the $XXX speedrun in the report card to start. The current code base is a bit over 8000 lines but I tried to keep them clean and well-commented. Now comes the fun part - of tuning and hillclimbing"
X Link @karpathy 2025-10-13T15:16Z 1.4M followers, 132.9K engagements

"I wrote a quick new post on "Digital Hygiene". Basically there are some no-brainer decisions you can make in your life to dramatically improve the privacy and security of your computing and this post goes over some of them. Blog post link in the reply but copy pasting below too. Every now and then I get reminded about the vast fraud apparatus of the internet re-invigorating my pursuit of basic digital hygiene around privacy/security of day to day computing. The sketchiness starts with major tech companies who are incentivized to build comprehensive profiles of you to monetize it directly for"
X Link @karpathy 2025-03-18T17:14Z 1.4M followers, 4M engagements

"There's a new kind of coding I call "vibe coding" where you fully give in to the vibes embrace exponentials and forget that the code even exists. It's possible because the LLMs (e.g. Cursor Composer w Sonnet) are getting too good. Also I just talk to Composer with SuperWhisper so I barely even touch the keyboard. I ask for the dumbest things like "decrease the padding on the sidebar by half" because I'm too lazy to find it. I "Accept All" always I don't read the diffs anymore. When I get error messages I just copy paste them in with no comment usually that fixes it. The code grows beyond my"
X Link @karpathy 2025-02-02T23:17Z 1.4M followers, 5.2M engagements

"@brickroad7 Think Casper Childhood favorite"
X Link @karpathy 2025-10-01T18:30Z 1.4M followers, 24.7K engagements

"POV: Your LLM agent is dividing a by b"
X Link @karpathy 2025-10-09T00:31Z 1.4M followers, 357.2K engagements

"@swyx @staysaasy I dont love it. Youre either engineering or youre vibing. They are opposites of some spectrum"
X Link @karpathy 2025-10-08T01:50Z 1.4M followers, 56.5K engagements

"May your regularizer be strong lest you RLHF to slop"
X Link @karpathy 2025-06-25T18:31Z 1.4M followers, 230.4K engagements

"@nikitabier Going to an Internet cafe saving the toronto dot ca website to floppy disk to take it home and show it to my parents as research just before we moved there. Lol"
X Link @karpathy 2025-10-12T20:22Z 1.4M followers, 59.3K engagements

"Shower of thoughts: Instead of keeping your Twitter/ payout direct it towards a "PayoutChallenge" of your choosing - anything you want more of in the world Here is mine for this round combining my last X payouts of $5478.51: It is imperative that humanity not fall while AI ascends. Humanity has to continue to rise become better alongside. Create something that is specifically designed to uplift team human. Definition intentionally left a bit vague to keep some entropy around people's interpretation but imo examples include: - Any piece of software that aids explanation visualization"
X Link @karpathy 2025-08-03T18:36Z 1.4M followers, 873.8K engagements

"The hottest new programming language is English"
X Link @karpathy 2023-01-24T20:14Z 1.4M followers, 8.2M engagements

"Something I am experimenting with. I copy pasted: 1) the full podcast transcript 2) the bitter lesson blog post 3) my full post above To ChatGPT. The interesting part is you can fork the conversation context to ask any questions and take it in whatever direction with chat:"
X Link @karpathy 2025-10-01T17:44Z 1.4M followers, 133.3K engagements

"⚡ Excited to share that I am starting an AI+Education company called Eureka Labs. The announcement: --- We are Eureka Labs and we are building a new kind of school that is AI native. How can we approach an ideal experience for learning something new For example in the case of physics one could imagine working through very high quality course materials together with Feynman who is there to guide you every step of the way. Unfortunately subject matter experts who are deeply passionate great at teaching infinitely patient and fluent in all of the world's languages are also very scarce and cannot"
X Link @karpathy 2024-07-16T17:25Z 1.4M followers, 2.5M engagements

"# Reproduce GPT-2 (124M) in llm.c in XX minutes for $XX ✨ The GPT-2 (124M) is the smallest model in the GPT-2 series released by OpenAI in 2019 and is actually quite accessible today even for the GPU poor. For example with llm.c you can now reproduce this model on one 8X A100 80GB SXM node in XX minutes (at XX% MFU). As they run for $14/hr this is $XX. I also think the 124M model makes for an excellent "cramming" challenge for training it very fast. So here is the launch command: And here is the output after XX minutes training on 10B tokens of the FineWeb dataset: It feels really nice to"
X Link @karpathy 2024-05-28T15:57Z 1.4M followers, 662.9K engagements

"Basically Llama-like a bit simpler some influences from modded-nanoGPT. Tried to find a solid baseline for this scale: - dense transformer - rotary embeddings (and no positional embeddings) - QK norm - untied weights for embedding and unembedding - norm after token embedding - relu2 activation in MLP - no learnable params in rmsnorm - no biases in linear layers - Multi-Query Attention (MQA) - logit softcap Optimizer is Muon+AdamW heavily influenced from modded-nanoGPT. I have a TODO to try to tune Adam LRs well (e.g. per module) to remove Muon I haven't tried hard enough yet"
X Link @karpathy 2025-10-13T15:48Z 1.4M followers, 74.3K engagements

"Anytime someone takes a picture/video that I happen to be in the background of I like to wave at the AGI that sees me XX years from now"
X Link @karpathy 2025-09-22T13:10Z 1.4M followers, 360.5K engagements

"How amazing it would be if we could extract and reframe all the practice problems from all the textbooks ever written into environments"
X Link @karpathy 2025-08-27T20:45Z 1.4M followers, 148.3K engagements

"More gists less gits"
X Link @karpathy 2025-07-05T21:59Z 1.4M followers, 136.2K engagements

""AI isn't replacing radiologists" good article Expectation: rapid progress in image recognition AI will delete radiology jobs (e.g. as famously predicted by Geoff Hinton now almost a decade ago). Reality: radiology is doing great and is growing. There are a lot of imo naive predictions out there on the imminent impact of AI on the job market. E.g. a year ago I was asked by someone who should know better if I think there will be any software engineers still today. (Spoiler: I think we're going to make it). This is happening too broadly. The post goes into detail on why it's not that simple"
X Link @karpathy 2025-09-25T14:29Z 1.4M followers, 2.3M engagements

"Finally had a chance to listen through this pod with Sutton which was interesting and amusing. As background Sutton's "The Bitter Lesson" has become a bit of biblical text in frontier LLM circles. Researchers routinely talk about and ask whether this or that approach or idea is sufficiently "bitter lesson pilled" (meaning arranged so that it benefits from added computation for free) as a proxy for whether it's going to work or worth even pursuing. The underlying assumption being that LLMs are of course highly "bitter lesson pilled" indeed just look at LLM scaling laws where if you put compute"
X Link @karpathy 2025-10-01T17:09Z 1.4M followers, 1.9M engagements

"LLM OS. Bear with me I'm still cooking. Specs: - LLM: OpenAI GPT-4 Turbo XXX core (batch size) processor @ 20Hz (tok/s) - RAM: 128Ktok - Filesystem: Ada002"
X Link @karpathy 2023-11-11T00:48Z 1.4M followers, 2.4M engagements

"Tinker is cool. If you're a researcher/developer tinker dramatically simplifies LLM post-training. You retain XX% of algorithmic creative control (usually related to data loss function the algorithm) while tinker handles the hard parts that you usually want to touch much less often (infra forward/backward of the LLM itself distributed training) meaning you can do these at well below XX% of typical complexity involved. Compared to the more common and existing paradigm of "upload your data we'll post-train your LLM" this is imo a more clever place to "slice up" the complexity of post-training"
X Link @karpathy 2025-10-01T19:22Z 1.4M followers, 671.9K engagements

""Using a better model for analysis" 🤨 I didn't realize I was using haiku all this time no idea when claude code snuck this one in rofl"
X Link @karpathy 2025-07-18T21:46Z 1.4M followers, 373.1K engagements

"GitHub repo: A lot more detailed and technical walkthrough: Example conversation with the $XXX 4-hour nanochat in the WebUI. It's. entertaining :) Larger models (e.g. a 12-hour depth XX or a 24-hour depth 30) quickly get more coherent"
X Link @karpathy 2025-10-13T15:16Z 1.4M followers, 188.3K engagements

"@shaneguML The Great Filter is kinda cute"
X Link @karpathy 2025-07-15T19:00Z 1.4M followers, 331.2K engagements

"Excited to release new repo: nanochat (it's among the most unhinged I've written). Unlike my earlier similar repo nanoGPT which only covered pretraining nanochat is a minimal from scratch full-stack training/inference pipeline of a simple ChatGPT clone in a single dependency-minimal codebase. You boot up a cloud GPU box run a single script and in as little as X hours later you can talk to your own LLM in a ChatGPT-like web UI. It weighs 8000 lines of imo quite clean code to: - Train the tokenizer using a new Rust implementation - Pretrain a Transformer LLM on FineWeb evaluate CORE score"
X Link @karpathy 2025-10-13T15:16Z 1.4M followers, 4.1M engagements

"# on shortification of "learning" There are a lot of videos on YouTube/TikTok etc. that give the appearance of education but if you look closely they are really just entertainment. This is very convenient for everyone involved : the people watching enjoy thinking they are learning (but actually they are just having fun). The people creating this content also enjoy it because fun has a much larger audience fame and revenue. But as far as learning goes this is a trap. This content is an epsilon away from watching the Bachelorette. It's like snacking on those "Garden Veggie Straws" which feel"
X Link @karpathy 2024-02-10T18:10Z 1.4M followers, 2.2M engagements

"Hah judging by mentions overnight people seem to find the ghost analogy provocative. I swear I don't wake up just trying to come with new memes but to elaborate briefly why I thought it was a fun comparison: 1) It captures the idea that LLMs are purely digital artifacts that don't interact with the physical world (unlike animals which are very embodied). 2) Ghosts are a kind of "echo" of the living in this case a statistical distillation of humanity. 3) There is an air of mystery over both ghosts and LLMs as in we don't fully understand what they are or how they work. 4) The process of"
X Link @karpathy 2025-10-02T14:25Z 1.4M followers, 229.7K engagements

"The future expands the variance of human condition a lot more than it drags its mean. This is an empirical observation with interesting extrapolations. The past is well-approximated as a population of farmers living similar lives w.r.t. upbringing knowledge activities ideals aspirations etc. The future trends to include all of: - the transhumanists who "ascend" with neuralinks etc. and the Amish living 19th century life. - those who "worship" ideals of religion technology knowledge wealth fitness community nature art . - those exploring externally into the stars those exploring internally"
X Link @karpathy 2024-10-16T07:09Z 1.4M followers, 427.6K engagements

"I think congrats again to OpenAI for cooking with GPT-5 Pro. This is the third time I've struggled on something complex/gnarly for an hour on and off with CC then X Pro goes off for XX minutes and comes back with code that works out of the box. I had CC read the X Pro version and it wrote up X paragraphs admiring it (very wholesome). If you're not giving it your hardest problems you're probably missing out"
X Link @karpathy 2025-09-05T17:38Z 1.4M followers, 2.6M engagements

"How to build a thriving open source community by writing code like bacteria do 🦠. Bacterial code (genomes) are: - small (each line of code costs energy) - modular (organized into groups of swappable operons) - self-contained (easily "copy paste-able" via horizontal gene transfer) If chunks of code are small modular self-contained and trivial to copy-and-paste the community can thrive via horizontal gene transfer. For any function (gene) or class (operon) that you write: can you imagine someone going "yoink" without knowing the rest of your code or having to import anything new to gain a"
X Link @karpathy 2025-07-05T21:54Z 1.4M followers, 617.9K engagements

"@doodlestein OPEN THE POD BAY DOORS HAL The number of AI pioneers anticipating this as the state of the art AI of 2025 must surely have been exactly zero"
X Link @karpathy 2025-10-12T18:44Z 1.4M followers, 46.6K engagements

"Scaling up RL is all the rage right now I had a chat with a friend about it yesterday. I'm fairly certain RL will continue to yield more intermediate gains but I also don't expect it to be the full story. RL is basically "hey this happened to go well (/poorly) let me slightly increase (/decrease) the probability of every action I took for the future". You get a lot more leverage from verifier functions than explicit supervision this is great. But first it looks suspicious asymptotically - once the tasks grow to be minutes/hours of interaction long you're really going to do all that work just"
X Link @karpathy 2025-07-13T16:35Z 1.4M followers, 1.1M engagements

"I don't know what labs are doing to these poor LLMs during RL but they are mortally terrified of exceptions in any infinitesimally likely case. Exceptions are a normal part of life and healthy dev process. Sign my LLM welfare petition for improved rewards in cases of exceptions"
X Link @karpathy 2025-10-09T00:10Z 1.4M followers, 646.5K engagements

"An attempt to explain (current) ChatGPT versions. I still run into many many people who don't know that: - o3 is the obvious best thing for important/hard things. It is a reasoning model that is much stronger than 4o and if you are using ChatGPT professionally and not using o3 you're ngmi. - 4o is different from o4. Yes I know lol. 4o is a good "daily driver" for many easy-medium questions. o4 is only available as mini for now and is not as good as o3 and I'm not super sure why it's out right now. Example basic "router" in my own personal use: - Any simple query (e.g. "what foods are high in"
X Link @karpathy 2025-06-02T17:54Z 1.4M followers, 1.4M engagements

"+1 for "context engineering" over "prompt engineering". People associate prompts with short task descriptions you'd give an LLM in your day-to-day use. When in every industrial-strength LLM app context engineering is the delicate art and science of filling the context window with just the right information for the next step. Science because doing this right involves task descriptions and explanations few shot examples RAG related (possibly multimodal) data tools state and history compacting. Too little or of the wrong form and the LLM doesn't have the right context for optimal performance."
X Link @karpathy 2025-06-25T15:54Z 1.4M followers, 2.4M engagements

"from this era"
X Link @karpathy 2025-09-13T16:12Z 1.4M followers, 256.6K engagements

"My sleep scores during recent travel were in the 90s. Now back in SF I am consistently back down to 70s 80s. I am increasingly convinced that this is due to traffic noise from a nearby road/intersection where I live - every 10min a car truck bus or motorcycle with a very loud engine passes by (some are 10X louder than others). In the later less deep stages of sleep it is much easier to wake and then much harder to go back to sleep. More generally I think noise pollution (esp early hours) come at a huge societal cost that is not correctly accounted for. E.g. I wouldn't be too surprised if a"
X Link @karpathy 2025-06-07T19:01Z 1.4M followers, 1.4M engagements

"Nice - my AI startup school talk is now up Chapters: 0:00 Imo fair to say that software is changing quite fundamentally again. LLMs are a new kind of computer and you program them in English. Hence I think they are well deserving of a major version upgrade in terms of software. 6:06 LLMs have properties of utilities of fabs and of operating systems = New LLM OS fabbed by labs and distributed like utilities (for now). Many historical analogies apply - imo we are computing circa 1960s. 14:39 LLM psychology: LLMs = "people spirits" stochastic simulations of people where the simulator is an"
X Link @karpathy 2025-06-19T02:01Z 1.4M followers, 1.3M engagements