[GUEST ACCESS MODE: Data is scrambled or limited to provide examples. Make requests using your API key to unlock full data. Check https://lunarcrush.ai/auth for authentication information.] #  @karpathy Andrej Karpathy Andrej Karpathy posts on X about llm, vibe, vibe coding, veo the most. They currently have XXXXXXXXX followers and XX posts still getting attention that total XXXXXXX engagements in the last XX hours. ### Engagements: XXXXXXX [#](/creator/twitter::33836629/interactions)  - X Week XXXXXXXXX +35% - X Month XXXXXXXXXX -XX% - X Months XXXXXXXXXX -XX% - X Year XXXXXXXXXXX +119% ### Mentions: XX [#](/creator/twitter::33836629/posts_active)  - X Week XX -XXXX% - X Month XXX +36% - X Months XXX +13% - X Year XXX +76% ### Followers: XXXXXXXXX [#](/creator/twitter::33836629/followers)  - X Week XXXXXXXXX +0.67% - X Month XXXXXXXXX +2.60% - X Months XXXXXXXXX +18% - X Year XXXXXXXXX +30% ### CreatorRank: XXXXXX [#](/creator/twitter::33836629/influencer_rank)  ### Social Influence [#](/creator/twitter::33836629/influence) --- **Social category influence** [technology brands](/list/technology-brands) XXXX% [stocks](/list/stocks) #471 [finance](/list/finance) XXXX% [automotive brands](/list/automotive-brands) XXXX% [social networks](/list/social-networks) XXXX% [celebrities](/list/celebrities) XXXX% **Social topic influence** [llm](/topic/llm) #19, [vibe](/topic/vibe) #972, [vibe coding](/topic/vibe-coding) #15, [veo](/topic/veo) #195, [open ai](/topic/open-ai) #775, [pdf](/topic/pdf) 2.53%, [claude](/topic/claude) #119, [hygiene](/topic/hygiene) 2.53%, [llama](/topic/llama) #63, [10x](/topic/10x) #193 **Top accounts mentioned or mentioned by** [@grok](/creator/undefined) [@ronald_vanloon](/creator/undefined) [@cyberpaun](/creator/undefined) [@yaaaaaashhh](/creator/undefined) [@elonmusk](/creator/undefined) [@justheresurfin](/creator/undefined) [@pastsmartlink](/creator/undefined) [@saedgossous](/creator/undefined) [@heretofilter](/creator/undefined) [@who_cares_whooo](/creator/undefined) [@phoenixqueenxxx](/creator/undefined) [@shahh](/creator/undefined) [@kryptopepe13](/creator/undefined) [@asdasxd357](/creator/undefined) [@mickeysteamboat](/creator/undefined) [@untgnr](/creator/undefined) [@kundem_serkan](/creator/undefined) [@shaneguml](/creator/undefined) [@genaisummitsf](/creator/undefined) [@byeonchansoo](/creator/undefined) **Top assets mentioned** [Tesla, Inc. (TSLA)](/topic/tesla) [Alphabet Inc Class A (GOOGL)](/topic/$googl) ### Top Social Posts [#](/creator/twitter::33836629/posts) --- Top posts by engagements in the last XX hours "Products with extensive/rich UIs lots of sliders switches menus with no scripting support and built on opaque custom binary formats are ngmi in the era of heavy human+AI collaboration. If an LLM can't read the underlying representations and manipulate them and all of the related settings via scripting then it also can't co-pilot your product with existing professionals and it doesn't allow vibe coding for the 100X more aspiring prosumers. Example high risk (binary objects/artifacts no text DSL): every Adobe product DAWs CAD/3D Example medium-high risk (already partially text scriptable):"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1930354382106964079) 2025-06-04 20:02:03 UTC 1.3M followers, 775.6K engagements ""Finding the Best Sleep Tracker" Results of an experiment where I wore X sleep trackers every night for X months. TLDR Whoop = Oura 8Sleep Apple Watch + AutoSleep. Link simply right here instead of in a reply because ()/"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1906386327190257963) 2025-03-30 16:41:34 UTC 1.3M followers, 1.6M engagements "The race for LLM "cognitive core" - a few billion param model that maximally sacrifices encyclopedic knowledge for capability. It lives always-on and by default on every computer as the kernel of LLM personal computing. Its features are slowly crystalizing: - Natively multimodal text/vision/audio at both input and output. - Matryoshka-style architecture allowing a dial of capability up and down at test time. - Reasoning also with a dial. (system 2) - Aggressively tool-using. - On-device finetuning LoRA slots for test-time training personalization and customization. - Delegates and double"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1938626382248149433) 2025-06-27 15:52:02 UTC 1.3M followers, 1.2M engagements "Good post from @balajis on the "verification gap". You could see it as there being two modes in creation. Borrowing GAN terminology: 1) generation and 2) discrimination. e.g. painting - you make a brush stroke (1) and then you look for a while to see if you improved the painting (2). these two stages are interspersed in pretty much all creative work. Second point. Discrimination can be computationally very hard. - images are by far the easiest. e.g. image generator teams can create giant grids of results to decide if one image is better than the other. thank you to the giant GPU in your brain"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1930305209747812559) 2025-06-04 16:46:39 UTC 1.3M followers, 510.7K engagements "Diffusion video models but now - **realtime** Simple video filters are real-time but can only do basic re-coloring and styles. Video diffusion models (Veo and friends) are magic but they take many seconds/minutes to generate. MirageLSD is real-time magic. Unlike simple video filters diffusion models actually *understand* what they are looking at so they can style all parts of the feed intelligently (e.g. putting hats on heads or light sabers into hands etc.). And they are arbitrarily steerable e.g. by text prompts. Customizable intelligent video filters unlock many cool ideas over time: -"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1945979830740435186) 2025-07-17 22:52:00 UTC 1.3M followers, 401.9K engagements "Test-based certification is the only way forward in food eager to see more over time. Food is not simple anymore - it is a complex industrial product with global supply and processing chains. Contamination can be introduced in many stages along the way from farming to harvest processing packaging transport and preparation. Examples include pesticides nitrates heavy metals plastics bacteria etc etc. So it's not just about what food to eat it's about which specific food item SKU from which specific supplier and the only way to know is to test. E.g. these two cat foods look the same the"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1940181840201228384) 2025-07-01 22:52:52 UTC 1.3M followers, 378.9K engagements ""Chatting" with LLM feels like using an 80s computer terminal. The GUI hasn't been invented yet but imo some properties of it can start to be predicted. X it will be visual (like GUIs of the past) because vision (pictures charts animations not so much reading) is the 10-lane highway into brain. It's the highest input information bandwidth and 1/3 of brain compute is dedicated to it. X it will be generative an input-conditional i.e. the GUI is generated on-demand specifically for your prompt and everything is present and reconfigured with the immediate purpose in mind. X a little bit more of"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1917920257257459899) 2025-05-01 12:33:17 UTC 1.3M followers, 727.8K engagements "PSA Its a new era of ergonomics. The primary audience of your thing (product service library ) is now an LLM not a human. LLMs dont like to navigate they like to scrape. LLMs dont like to see they like to read. LLMs dont like to click they like to curl. Etc etc"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1914494203696177444) 2025-04-22 01:39:22 UTC 1.3M followers, 518.4K engagements "I inherited "AI assisted coding" from this @simonw post: But I think it needs work. It doesn't roll off the tongue. Few days ago a friend asked me if I was vibe coding and I said no I'm "real coding". Possible candidate :D"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1915586183834587218) 2025-04-25 01:58:30 UTC 1.3M followers, 187K engagements "Mildly obsessed with what the "highest grade" pretraining data stream looks like for LLM training if XXX% of the focus was on quality putting aside any quantity considerations. Guessing something textbook-like content in markdown Or possibly samples from a really giant model Curious what the most powerful e.g. 1B param model trained on a dataset of 10B tokens looks like and how far "micromodels" can be pushed. As an example (text)books are already often included in pretraining data mixtures but whenever I look closely the data is all messed up - weird formatting padding OCR bugs Figure text"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1936171874398208202) 2025-06-20 21:18:41 UTC 1.3M followers, 538.8K engagements "LLM OS. Bear with me I'm still cooking. Specs: - LLM: OpenAI GPT-4 Turbo XXX core (batch size) processor @ 20Hz (tok/s) - RAM: 128Ktok - Filesystem: Ada002"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1723140519554105733) 2023-11-11 00:48:08 UTC 1.3M followers, 2.4M engagements "@Yuchenj_UW It really tests my default libertarian inclinations. Literally what the fuck"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1942614073860104690) 2025-07-08 15:57:41 UTC 1.3M followers, 21.9K engagements "As AI advances our contribution is more and more original knowledge - meaning something that cant be inferred from what exists digitally already by reasoning. Something like the result of an experiment. Maybe it should be written more natively for AIs instead of people eg PDF is an AI unfriendly format. Git repos of analysis code results in csvs explanations in markdown etc are a lot more friendlier"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1943345514239717873) 2025-07-10 16:24:10 UTC 1.3M followers, 99.7K engagements "Love this Supercharger diner but really a kind of exhibit for the future. Plotting a road trip SF - LA to charge Shadowfax"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1948062129187140051) 2025-07-23 16:46:19 UTC 1.3M followers, 2.3M engagements "This is what the ideal grocery store looks like. Minimally processed (NOVA Group 1) food only (no "edible food-like substances") organic local fresh. Food should not be more complex than this yet I don't believe this exists"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1942612984481870068) 2025-07-08 15:53:22 UTC 1.3M followers, 570.6K engagements "We have to take the LLMs to school. When you open any textbook you'll see three major types of information: X. Background information / exposition. The meat of the textbook that explains concepts. As you attend over it your brain is training on that data. This is equivalent to pretraining where the model is reading the internet and accumulating background knowledge. X. Worked problems with solutions. These are concrete examples of how an expert solves problems. They are demonstrations to be imitated. This is equivalent to supervised finetuning where the model is finetuning on "ideal"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1885026028428681698) 2025-01-30 18:03:21 UTC 1.3M followers, 704.3K engagements ""Using a better model for analysis" ๐คจ I didn't realize I was using haiku all this time no idea when claude code snuck this one in rofl"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1946325810618700033) 2025-07-18 21:46:48 UTC 1.3M followers, 326.5K engagements "I was given early access to Grok X earlier today making me I think one of the first few who could run a quick vibe check. Thinking โ First Grok X clearly has an around state of the art thinking model ("Think" button) and did great out of the box on my Settler's of Catan question: "Create a board game webpage showing a hex grid just like in the game Settlers of Catan. Each hex grid is numbered from 1.N where N is the total number of hex tiles. Make it generic so one can change the number of "rings" using a slider. For example in Catan the radius is X hexes. Single html page please." Few models"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1891720635363254772) 2025-02-18 05:25:20 UTC 1.3M followers, 3.7M engagements "I wrote a quick new post on "Digital Hygiene". Basically there are some no-brainer decisions you can make in your life to dramatically improve the privacy and security of your computing and this post goes over some of them. Blog post link in the reply but copy pasting below too. Every now and then I get reminded about the vast fraud apparatus of the internet re-invigorating my pursuit of basic digital hygiene around privacy/security of day to day computing. The sketchiness starts with major tech companies who are incentivized to build comprehensive profiles of you to monetize it directly for"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1902046003567718810) 2025-03-18 17:14:40 UTC 1.3M followers, 4M engagements "There's a new kind of coding I call "vibe coding" where you fully give in to the vibes embrace exponentials and forget that the code even exists. It's possible because the LLMs (e.g. Cursor Composer w Sonnet) are getting too good. Also I just talk to Composer with SuperWhisper so I barely even touch the keyboard. I ask for the dumbest things like "decrease the padding on the sidebar by half" because I'm too lazy to find it. I "Accept All" always I don't read the diffs anymore. When I get error messages I just copy paste them in with no comment usually that fixes it. The code grows beyond my"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1886192184808149383) 2025-02-02 23:17:15 UTC 1.3M followers, 5.2M engagements "@shaneguML The Great Filter is kinda cute"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1945196908420485125) 2025-07-15 19:00:57 UTC 1.3M followers, 297.7K engagements "Congrats to Simon Willison (@simonw) on XX years () of blogging. Really excellent LLM blog I sub & read everything: (e.g. I sub via RSS/Atom on NetNewsWire) +If you consistently enjoy the content like I do sponsor on GitHub:"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1933582359347278246) 2025-06-13 17:48:53 UTC 1.3M followers, 648.2K engagements "@phbenz Haha I'm not trying to coin a new word or something. I just think people's use of "prompt" tends to (incorrectly) trivialize a rather complex component. You prompt an LLM to tell you why the sky is blue. But apps build contexts (meticulously) for LLMs to solve their custom tasks"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1937909397180796982) 2025-06-25 16:22:59 UTC 1.3M followers, 51.3K engagements "Hi @gmail does the "report phishing" button do anything"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1946745524033593739) 2025-07-20 01:34:36 UTC 1.3M followers, 572.6K engagements "Love this project: nanoGPT - recursive self-improvement benchmark. Good old nanoGPT keeps on giving and surprising :) - First I wrote it as a small little repo to teach people the basics of training GPTs. - Then it became a target and baseline for my port to direct C/CUDA re-implementation in llm.c. - Then that was modded (by @kellerjordan0 et al.) into a (small-scale) LLM research harness. People iteratively optimized the training so that e.g. reproducing GPT-2 (124M) performance takes not XX min (original) but now only X min - Now the idea is to use this process of optimizing the code as a"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1939709449956126910) 2025-06-30 15:35:45 UTC 1.3M followers, 435.8K engagements "Agency Intelligence I had this intuitively wrong for decades I think due to a pervasive cultural veneration of intelligence various entertainment/media obsession with IQ etc. Agency is significantly more powerful and significantly more scarce. Are you hiring for agency Are we educating for agency Are you acting as if you had 10X agency Grok explanation is close: Agency as a personality trait refers to an individual's capacity to take initiative make decisions and exert control over their actions and environment. Its about being proactive rather than reactivesomeone with high agency doesnt"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1894099637218545984) 2025-02-24 18:58:38 UTC 1.3M followers, 2.5M engagements "Knowledge makes the world so much more beautiful"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1941893865507807541) 2025-07-06 16:15:50 UTC 1.3M followers, 714.4K engagements "Some of the links: - My slides as keynote: - Software XXX blog post from 2017 - How LLMs flip the script on technology diffusion - Vibe coding MenuGen (retrospective)"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1935519334123848101) 2025-06-19 02:05:44 UTC 1.3M followers, 257K engagements "@the_danny_g unhinged virus coated behavior haha"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1946326434836037982) 2025-07-18 21:49:17 UTC 1.3M followers, 25.9K engagements "May your regularizer be strong lest you RLHF to slop"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1937941695943065640) 2025-06-25 18:31:20 UTC 1.3M followers, 222.2K engagements "How to build a thriving open source community by writing code like bacteria do ๐ฆ . Bacterial code (genomes) are: - small (each line of code costs energy) - modular (organized into groups of swappable operons) - self-contained (easily "copy paste-able" via horizontal gene transfer) If chunks of code are small modular self-contained and trivial to copy-and-paste the community can thrive via horizontal gene transfer. For any function (gene) or class (operon) that you write: can you imagine someone going "yoink" without knowing the rest of your code or having to import anything new to gain a"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1941616674094170287) 2025-07-05 21:54:23 UTC 1.3M followers, 586.3K engagements "I should clarify that the risk is highest if you're running local LLM agents (e.g. Cursor Claude Code etc.). If you're just talking to an LLM on a website (e.g. ChatGPT) the risk is much lower *unless* you start turning on Connectors. For example I just saw ChatGPT is adding MCP support. This will combine especially poorly with all the recently added memory features - e.g. imagine ChatGPT telling everything it knows about you to some attacker on the internet just because you checked the wrong box in the Connectors settings"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1934657940155441477) 2025-06-16 17:02:51 UTC 1.3M followers, 102.6K engagements "Not fully sure why all the LLMs sound about the same - over-using lists delving into multifaceted issues over-offering to assist further about same length responses etc. Not something I had predicted at first because of many independent companies doing the finetuning"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1843005000206909856) 2024-10-06 19:06:48 UTC 1.3M followers, 3.5M engagements "When working with LLMs I am used to starting "New Conversation" for each request. But there is also the polar opposite approach of keeping one giant conversation going forever. The standard approach can still choose to use a Memory tool to write things down in between conversations (e.g. ChatGPT does so) so the "One Thread" approach can be seen as the extreme special case of using memory always and for everything. The other day I've come across someone saying that their conversation with Grok (which was free to them at the time) has now grown way too long for them to switch to ChatGPT. i.e."  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1902737525900525657) 2025-03-20 15:02:31 UTC 1.3M followers, 828.7K engagements "This is interesting as a first large diffusion-based LLM. Most of the LLMs you've been seeing are clones as far as the core modeling approach goes. They're all trained "autoregressively" i.e. predicting tokens from left to right. Diffusion is different - it doesn't go left to right but all at once. You start with noise and gradually denoise into a token stream. Most of the image / video generation AI tools actually work this way and use Diffusion not Autoregression. It's only text (and sometimes audio) that have resisted. So it's been a bit of a mystery to me and many others why for some"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1894923254864978091) 2025-02-27 01:31:24 UTC 1.3M followers, 947.1K engagements "Pleasure to come by the YC AI Startup School today I'm told the recordings will be up "in the coming weeks" I'll link to it then and include the slides. Thank you YC for organizing and bringing together an awesome group of builders Fun fact is that when I (and all the original founding members) decided to join OpenAI the name OpenAI didn't exist - we all thought we were joining a new AI non-profit under YC Research. My very first OpenAI swag t-shirt says "YC AI Day 1". Things changed up a bit after that. Cheers to YC :)"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1935074699450740785) 2025-06-17 20:38:54 UTC 1.3M followers, 487.9K engagements "@OrdinaryInds Is this real I've been looking for so long ๐โ๐โ๐โ"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1943005808410923244) 2025-07-09 17:54:18 UTC 1.3M followers, 267.6K engagements "The hottest new programming language is English"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1617979122625712128) 2023-01-24 20:14:18 UTC 1.3M followers, 7.8M engagements "The post below was trending last few days and reminded me that my earlier digital hygiene post was woefully incomplete without a discussion around smartphone choices. The post goes into how on Android apps routinely use a loophole (that Android has known about and not fixed for years) to get the list of all other apps on your phone. I disagree with the author that there are legitimate uses for this information. There aren't or if there are they are super marginal and the privacy tradeoff is not worth it. In practice the data is clearly being collected at scale for shady user profiling. The"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1906748528627503433) 2025-03-31 16:40:49 UTC 1.3M followers, 518.5K engagements "Scaling up RL is all the rage right now I had a chat with a friend about it yesterday. I'm fairly certain RL will continue to yield more intermediate gains but I also don't expect it to be the full story. RL is basically "hey this happened to go well (/poorly) let me slightly increase (/decrease) the probability of every action I took for the future". You get a lot more leverage from verifier functions than explicit supervision this is great. But first it looks suspicious asymptotically - once the tasks grow to be minutes/hours of interaction long you're really going to do all that work just"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1944435412489171119) 2025-07-13 16:35:02 UTC 1.3M followers, 1M engagements "Tired: elaborate docs pages for your product/service/library with fancy color palettes branding animations transitions dark mode Wired: one single docs .md file and a copy to clipboard button"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1914488029873627597) 2025-04-22 01:14:50 UTC 1.3M followers, 889.3K engagements "So so so cool. Llama 1B batch one inference in one single CUDA kernel deleting synchronization boundaries imposed by breaking the computation into a series of kernels called in sequence. The *optimal* orchestration of compute and memory is only achievable in this way"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1927506788527591853) 2025-05-27 23:26:44 UTC 1.3M followers, 263.4K engagements "I often rant about how XX% of attention is about to be LLM attention instead of human attention. What does a research paper look like for an LLM instead of a human Its definitely not a pdf. There is huge space for an extremely valuable research app that figures this out"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1943411187296686448) 2025-07-10 20:45:08 UTC 1.3M followers, 577.5K engagements "Do people *feel* how much work there is still to do. Like wow"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1938629042602934444) 2025-06-27 16:02:36 UTC 1.3M followers, 169.2K engagements "Part X of this mystery. Spotted on reddit. In my test not XXX% reproducible but still quite reproducible. ๐ค"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1935404600653492484) 2025-06-18 18:29:49 UTC 1.3M followers, 2.6M engagements "I always learn a lot more from in-depth analysis of few random cases over dashboards of aggregate statistics across all cases. Both projections can be helpful but the latter is disproportionately pervasive"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1944885371957031005) 2025-07-14 22:23:01 UTC 1.3M followers, 279.7K engagements "RT to help Simon raise awareness of prompt injection attacks in LLMs. Feels a bit like the wild west of early computing with computer viruses (now = malicious prompts hiding in web data/tools) and not well developed defenses (antivirus or a lot more developed kernel/user space security paradigm where e.g. an agent is given very specific action types instead of the ability to run arbitrary bash scripts). Conflicted because I want to be an early adopter of LLM agents in my personal computing but the wild west of possibility is holding me back"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1934651657444528277) 2025-06-16 16:37:53 UTC 1.3M followers, 425K engagements "Yep I think RL is misleading in that it restricts field of view. Eg like you mentioned you can imagine review/reflect doing a lot more - building tools for later use or actively tuning the distribution for what to try next (instead of just sampling from policy independently as usual). Or you can imagine environments with no rewards. So much more. Basically - agentic interactions: absolutely +100. RL specifically: eeeh"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1944814767257842027) 2025-07-14 17:42:28 UTC 1.3M followers, 12.7K engagements "New 2h11m YouTube video: How I Use LLMs This video continues my general audience series. The last one focused on how LLMs are trained so I wanted to follow up with a more practical guide of the entire LLM ecosystem including lots of examples of use in my own life. Chapters give a sense of content: 00:00:00 Intro into the growing LLM ecosystem 00:02:54 ChatGPT interaction under the hood 00:13:12 Basic LLM interactions examples 00:18:03 Be aware of the model you're using pricing tiers 00:22:54 Thinking models and when to use them 00:31:00 Tool use: internet search 00:42:04 Tool use: deep"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1895242932095209667) 2025-02-27 22:41:41 UTC 1.3M followers, 979.6K engagements "o1-mini keeps refusing to try to solve the Riemann Hypothesis on my behalf. Model laziness continues to be a major issue sad ;p"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1834374965942255835) 2024-09-12 23:34:07 UTC 1.3M followers, 711.5K engagements "It's 2025 and most content is still written for humans instead of LLMs. XXXX% of attention is about to be LLM attention not human attention. E.g. XX% of libraries still have docs that basically render to some pretty .html static pages assuming a human will click through them. In 2025 the docs should be a single your_project.md text file that is intended to go into the context window of an LLM. Repeat for everything"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1899876370492383450) 2025-03-12 17:33:19 UTC 1.3M followers, 1.8M engagements "An attempt to explain (current) ChatGPT versions. I still run into many many people who don't know that: - o3 is the obvious best thing for important/hard things. It is a reasoning model that is much stronger than 4o and if you are using ChatGPT professionally and not using o3 you're ngmi. - 4o is different from o4. Yes I know lol. 4o is a good "daily driver" for many easy-medium questions. o4 is only available as mini for now and is not as good as o3 and I'm not super sure why it's out right now. Example basic "router" in my own personal use: - Any simple query (e.g. "what foods are high in"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1929597620969951434) 2025-06-02 17:54:57 UTC 1.3M followers, 1.3M engagements "Very impressed with Veo X and all the things people are finding on r/aivideo etc. Makes a big difference qualitatively when you add audio. There are a few macro aspects to video generation that may not be fully appreciated: X. Video is the highest bandwidth input to brain. Not just for entertainment but also for work/learning - think diagrams charts animations etc. X. Video is the most easy/fun. The average person doesn't like reading/writing it's very effortful. Anyone can (and wants to) engage with video. X. The barrier to creating videos is - X. X. For the first time video is directly"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1929634696474120576) 2025-06-02 20:22:17 UTC 1.3M followers, 1M engagements "A major mistake I made in my undergrad is that I focused way too much on mathematical lens of computing - computability decidability asymptotic complexity etc. And too little on physical lens - energy/heat of state change data locality parallelism computer architecture. The former is interesting; The latter bestows power"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1919647115099451892) 2025-05-06 06:55:12 UTC 1.3M followers, 1.3M engagements "Making slides manually feels especially painful now that you know Cursor for slides should exist but doesnt"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1931042840966222046) 2025-06-06 17:37:44 UTC 1.3M followers, 2.6M engagements "I attended a vibe coding hackathon recently and used the chance to build a web app (with auth payments deploy etc.). I tinker but I am not a web dev by background so besides the app I was very interested in what it's like to vibe code a full web app today. As such I wrote none of the code directly (Cursor+Claude/o3 did) and I don't really know how the app works in the conventional sense that I'm used to as an engineer. The app is called MenuGen and it is live on Basically I'm often confused about what all the things on a restaurant menu are - e.g. Pt Tagine Cavatappi or Sweetbread (hint it's."  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1917961248031080455) 2025-05-01 15:16:10 UTC 1.3M followers, 766.3K engagements "The docs also have to change in the content. Eg instead of instructing a person to go to some page and do this or that they could show curl commands to run - actions that are a lot easier for an LLM to carry out. Products have to change to support these too. Eg adding a Supabase db to your Vervel app shouldnt be clicks but curls"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1914489538006933770) 2025-04-22 01:20:50 UTC 1.3M followers, 209.7K engagements "+1 for "context engineering" over "prompt engineering". People associate prompts with short task descriptions you'd give an LLM in your day-to-day use. When in every industrial-strength LLM app context engineering is the delicate art and science of filling the context window with just the right information for the next step. Science because doing this right involves task descriptions and explanations few shot examples RAG related (possibly multimodal) data tools state and history compacting. Too little or of the wrong form and the LLM doesn't have the right context for optimal performance."  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1937902205765607626) 2025-06-25 15:54:24 UTC 1.3M followers, 2.3M engagements "@aidenybai I use CC from Cursor and I assumed most do as well (). I end up with a mixed thing where Cursor is the UI layer for reading the code manual edits tab completion and chunk edits and CC for larger changes architecting Q&A. Still rapidly evolving though"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1947019163949731891) 2025-07-20 19:41:57 UTC 1.3M followers, 153K engagements "My sleep scores during recent travel were in the 90s. Now back in SF I am consistently back down to 70s 80s. I am increasingly convinced that this is due to traffic noise from a nearby road/intersection where I live - every 10min a car truck bus or motorcycle with a very loud engine passes by (some are 10X louder than others). In the later less deep stages of sleep it is much easier to wake and then much harder to go back to sleep. More generally I think noise pollution (esp early hours) come at a huge societal cost that is not correctly accounted for. E.g. I wouldn't be too surprised if a"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1931426322536132767) 2025-06-07 19:01:34 UTC 1.3M followers, 1.4M engagements "Nice - my AI startup school talk is now up Chapters: 0:00 Imo fair to say that software is changing quite fundamentally again. LLMs are a new kind of computer and you program them *in English*. Hence I think they are well deserving of a major version upgrade in terms of software. 6:06 LLMs have properties of utilities of fabs and of operating systems = New LLM OS fabbed by labs and distributed like utilities (for now). Many historical analogies apply - imo we are computing circa 1960s. 14:39 LLM psychology: LLMs = "people spirits" stochastic simulations of people where the simulator is an"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1935518272667217925) 2025-06-19 02:01:31 UTC 1.3M followers, 1.2M engagements "@elonmusk @cyberpaun @shaneguML So what kind of revenue share are we talking about :D jk jk"  [@karpathy](/creator/x/karpathy) on [X](/post/tweet/1945566895362773146) 2025-07-16 19:31:09 UTC 1.3M followers, 152.7K engagements
[GUEST ACCESS MODE: Data is scrambled or limited to provide examples. Make requests using your API key to unlock full data. Check https://lunarcrush.ai/auth for authentication information.]
Andrej Karpathy posts on X about llm, vibe, vibe coding, veo the most. They currently have XXXXXXXXX followers and XX posts still getting attention that total XXXXXXX engagements in the last XX hours.
Social category influence technology brands XXXX% stocks #471 finance XXXX% automotive brands XXXX% social networks XXXX% celebrities XXXX%
Social topic influence llm #19, vibe #972, vibe coding #15, veo #195, open ai #775, pdf 2.53%, claude #119, hygiene 2.53%, llama #63, 10x #193
Top accounts mentioned or mentioned by @grok @ronald_vanloon @cyberpaun @yaaaaaashhh @elonmusk @justheresurfin @pastsmartlink @saedgossous @heretofilter @who_cares_whooo @phoenixqueenxxx @shahh @kryptopepe13 @asdasxd357 @mickeysteamboat @untgnr @kundem_serkan @shaneguml @genaisummitsf @byeonchansoo
Top assets mentioned Tesla, Inc. (TSLA) Alphabet Inc Class A (GOOGL)
Top posts by engagements in the last XX hours
"Products with extensive/rich UIs lots of sliders switches menus with no scripting support and built on opaque custom binary formats are ngmi in the era of heavy human+AI collaboration. If an LLM can't read the underlying representations and manipulate them and all of the related settings via scripting then it also can't co-pilot your product with existing professionals and it doesn't allow vibe coding for the 100X more aspiring prosumers. Example high risk (binary objects/artifacts no text DSL): every Adobe product DAWs CAD/3D Example medium-high risk (already partially text scriptable):" @karpathy on X 2025-06-04 20:02:03 UTC 1.3M followers, 775.6K engagements
""Finding the Best Sleep Tracker" Results of an experiment where I wore X sleep trackers every night for X months. TLDR Whoop = Oura 8Sleep Apple Watch + AutoSleep. Link simply right here instead of in a reply because ()/" @karpathy on X 2025-03-30 16:41:34 UTC 1.3M followers, 1.6M engagements
"The race for LLM "cognitive core" - a few billion param model that maximally sacrifices encyclopedic knowledge for capability. It lives always-on and by default on every computer as the kernel of LLM personal computing. Its features are slowly crystalizing: - Natively multimodal text/vision/audio at both input and output. - Matryoshka-style architecture allowing a dial of capability up and down at test time. - Reasoning also with a dial. (system 2) - Aggressively tool-using. - On-device finetuning LoRA slots for test-time training personalization and customization. - Delegates and double" @karpathy on X 2025-06-27 15:52:02 UTC 1.3M followers, 1.2M engagements
"Good post from @balajis on the "verification gap". You could see it as there being two modes in creation. Borrowing GAN terminology: 1) generation and 2) discrimination. e.g. painting - you make a brush stroke (1) and then you look for a while to see if you improved the painting (2). these two stages are interspersed in pretty much all creative work. Second point. Discrimination can be computationally very hard. - images are by far the easiest. e.g. image generator teams can create giant grids of results to decide if one image is better than the other. thank you to the giant GPU in your brain" @karpathy on X 2025-06-04 16:46:39 UTC 1.3M followers, 510.7K engagements
"Diffusion video models but now - realtime Simple video filters are real-time but can only do basic re-coloring and styles. Video diffusion models (Veo and friends) are magic but they take many seconds/minutes to generate. MirageLSD is real-time magic. Unlike simple video filters diffusion models actually understand what they are looking at so they can style all parts of the feed intelligently (e.g. putting hats on heads or light sabers into hands etc.). And they are arbitrarily steerable e.g. by text prompts. Customizable intelligent video filters unlock many cool ideas over time: -" @karpathy on X 2025-07-17 22:52:00 UTC 1.3M followers, 401.9K engagements
"Test-based certification is the only way forward in food eager to see more over time. Food is not simple anymore - it is a complex industrial product with global supply and processing chains. Contamination can be introduced in many stages along the way from farming to harvest processing packaging transport and preparation. Examples include pesticides nitrates heavy metals plastics bacteria etc etc. So it's not just about what food to eat it's about which specific food item SKU from which specific supplier and the only way to know is to test. E.g. these two cat foods look the same the" @karpathy on X 2025-07-01 22:52:52 UTC 1.3M followers, 378.9K engagements
""Chatting" with LLM feels like using an 80s computer terminal. The GUI hasn't been invented yet but imo some properties of it can start to be predicted. X it will be visual (like GUIs of the past) because vision (pictures charts animations not so much reading) is the 10-lane highway into brain. It's the highest input information bandwidth and 1/3 of brain compute is dedicated to it. X it will be generative an input-conditional i.e. the GUI is generated on-demand specifically for your prompt and everything is present and reconfigured with the immediate purpose in mind. X a little bit more of" @karpathy on X 2025-05-01 12:33:17 UTC 1.3M followers, 727.8K engagements
"PSA Its a new era of ergonomics. The primary audience of your thing (product service library ) is now an LLM not a human. LLMs dont like to navigate they like to scrape. LLMs dont like to see they like to read. LLMs dont like to click they like to curl. Etc etc" @karpathy on X 2025-04-22 01:39:22 UTC 1.3M followers, 518.4K engagements
"I inherited "AI assisted coding" from this @simonw post: But I think it needs work. It doesn't roll off the tongue. Few days ago a friend asked me if I was vibe coding and I said no I'm "real coding". Possible candidate :D" @karpathy on X 2025-04-25 01:58:30 UTC 1.3M followers, 187K engagements
"Mildly obsessed with what the "highest grade" pretraining data stream looks like for LLM training if XXX% of the focus was on quality putting aside any quantity considerations. Guessing something textbook-like content in markdown Or possibly samples from a really giant model Curious what the most powerful e.g. 1B param model trained on a dataset of 10B tokens looks like and how far "micromodels" can be pushed. As an example (text)books are already often included in pretraining data mixtures but whenever I look closely the data is all messed up - weird formatting padding OCR bugs Figure text" @karpathy on X 2025-06-20 21:18:41 UTC 1.3M followers, 538.8K engagements
"LLM OS. Bear with me I'm still cooking. Specs: - LLM: OpenAI GPT-4 Turbo XXX core (batch size) processor @ 20Hz (tok/s) - RAM: 128Ktok - Filesystem: Ada002" @karpathy on X 2023-11-11 00:48:08 UTC 1.3M followers, 2.4M engagements
"@Yuchenj_UW It really tests my default libertarian inclinations. Literally what the fuck" @karpathy on X 2025-07-08 15:57:41 UTC 1.3M followers, 21.9K engagements
"As AI advances our contribution is more and more original knowledge - meaning something that cant be inferred from what exists digitally already by reasoning. Something like the result of an experiment. Maybe it should be written more natively for AIs instead of people eg PDF is an AI unfriendly format. Git repos of analysis code results in csvs explanations in markdown etc are a lot more friendlier" @karpathy on X 2025-07-10 16:24:10 UTC 1.3M followers, 99.7K engagements
"Love this Supercharger diner but really a kind of exhibit for the future. Plotting a road trip SF - LA to charge Shadowfax" @karpathy on X 2025-07-23 16:46:19 UTC 1.3M followers, 2.3M engagements
"This is what the ideal grocery store looks like. Minimally processed (NOVA Group 1) food only (no "edible food-like substances") organic local fresh. Food should not be more complex than this yet I don't believe this exists" @karpathy on X 2025-07-08 15:53:22 UTC 1.3M followers, 570.6K engagements
"We have to take the LLMs to school. When you open any textbook you'll see three major types of information: X. Background information / exposition. The meat of the textbook that explains concepts. As you attend over it your brain is training on that data. This is equivalent to pretraining where the model is reading the internet and accumulating background knowledge. X. Worked problems with solutions. These are concrete examples of how an expert solves problems. They are demonstrations to be imitated. This is equivalent to supervised finetuning where the model is finetuning on "ideal" @karpathy on X 2025-01-30 18:03:21 UTC 1.3M followers, 704.3K engagements
""Using a better model for analysis" ๐คจ I didn't realize I was using haiku all this time no idea when claude code snuck this one in rofl" @karpathy on X 2025-07-18 21:46:48 UTC 1.3M followers, 326.5K engagements
"I was given early access to Grok X earlier today making me I think one of the first few who could run a quick vibe check. Thinking โ
First Grok X clearly has an around state of the art thinking model ("Think" button) and did great out of the box on my Settler's of Catan question: "Create a board game webpage showing a hex grid just like in the game Settlers of Catan. Each hex grid is numbered from 1.N where N is the total number of hex tiles. Make it generic so one can change the number of "rings" using a slider. For example in Catan the radius is X hexes. Single html page please." Few models" @karpathy on X 2025-02-18 05:25:20 UTC 1.3M followers, 3.7M engagements
"I wrote a quick new post on "Digital Hygiene". Basically there are some no-brainer decisions you can make in your life to dramatically improve the privacy and security of your computing and this post goes over some of them. Blog post link in the reply but copy pasting below too. Every now and then I get reminded about the vast fraud apparatus of the internet re-invigorating my pursuit of basic digital hygiene around privacy/security of day to day computing. The sketchiness starts with major tech companies who are incentivized to build comprehensive profiles of you to monetize it directly for" @karpathy on X 2025-03-18 17:14:40 UTC 1.3M followers, 4M engagements
"There's a new kind of coding I call "vibe coding" where you fully give in to the vibes embrace exponentials and forget that the code even exists. It's possible because the LLMs (e.g. Cursor Composer w Sonnet) are getting too good. Also I just talk to Composer with SuperWhisper so I barely even touch the keyboard. I ask for the dumbest things like "decrease the padding on the sidebar by half" because I'm too lazy to find it. I "Accept All" always I don't read the diffs anymore. When I get error messages I just copy paste them in with no comment usually that fixes it. The code grows beyond my" @karpathy on X 2025-02-02 23:17:15 UTC 1.3M followers, 5.2M engagements
"@shaneguML The Great Filter is kinda cute" @karpathy on X 2025-07-15 19:00:57 UTC 1.3M followers, 297.7K engagements
"Congrats to Simon Willison (@simonw) on XX years () of blogging. Really excellent LLM blog I sub & read everything: (e.g. I sub via RSS/Atom on NetNewsWire) +If you consistently enjoy the content like I do sponsor on GitHub:" @karpathy on X 2025-06-13 17:48:53 UTC 1.3M followers, 648.2K engagements
"@phbenz Haha I'm not trying to coin a new word or something. I just think people's use of "prompt" tends to (incorrectly) trivialize a rather complex component. You prompt an LLM to tell you why the sky is blue. But apps build contexts (meticulously) for LLMs to solve their custom tasks" @karpathy on X 2025-06-25 16:22:59 UTC 1.3M followers, 51.3K engagements
"Hi @gmail does the "report phishing" button do anything" @karpathy on X 2025-07-20 01:34:36 UTC 1.3M followers, 572.6K engagements
"Love this project: nanoGPT - recursive self-improvement benchmark. Good old nanoGPT keeps on giving and surprising :) - First I wrote it as a small little repo to teach people the basics of training GPTs. - Then it became a target and baseline for my port to direct C/CUDA re-implementation in llm.c. - Then that was modded (by @kellerjordan0 et al.) into a (small-scale) LLM research harness. People iteratively optimized the training so that e.g. reproducing GPT-2 (124M) performance takes not XX min (original) but now only X min - Now the idea is to use this process of optimizing the code as a" @karpathy on X 2025-06-30 15:35:45 UTC 1.3M followers, 435.8K engagements
"Agency Intelligence I had this intuitively wrong for decades I think due to a pervasive cultural veneration of intelligence various entertainment/media obsession with IQ etc. Agency is significantly more powerful and significantly more scarce. Are you hiring for agency Are we educating for agency Are you acting as if you had 10X agency Grok explanation is close: Agency as a personality trait refers to an individual's capacity to take initiative make decisions and exert control over their actions and environment. Its about being proactive rather than reactivesomeone with high agency doesnt" @karpathy on X 2025-02-24 18:58:38 UTC 1.3M followers, 2.5M engagements
"Knowledge makes the world so much more beautiful" @karpathy on X 2025-07-06 16:15:50 UTC 1.3M followers, 714.4K engagements
"Some of the links: - My slides as keynote: - Software XXX blog post from 2017 - How LLMs flip the script on technology diffusion - Vibe coding MenuGen (retrospective)" @karpathy on X 2025-06-19 02:05:44 UTC 1.3M followers, 257K engagements
"@the_danny_g unhinged virus coated behavior haha" @karpathy on X 2025-07-18 21:49:17 UTC 1.3M followers, 25.9K engagements
"May your regularizer be strong lest you RLHF to slop" @karpathy on X 2025-06-25 18:31:20 UTC 1.3M followers, 222.2K engagements
"How to build a thriving open source community by writing code like bacteria do ๐ฆ . Bacterial code (genomes) are: - small (each line of code costs energy) - modular (organized into groups of swappable operons) - self-contained (easily "copy paste-able" via horizontal gene transfer) If chunks of code are small modular self-contained and trivial to copy-and-paste the community can thrive via horizontal gene transfer. For any function (gene) or class (operon) that you write: can you imagine someone going "yoink" without knowing the rest of your code or having to import anything new to gain a" @karpathy on X 2025-07-05 21:54:23 UTC 1.3M followers, 586.3K engagements
"I should clarify that the risk is highest if you're running local LLM agents (e.g. Cursor Claude Code etc.). If you're just talking to an LLM on a website (e.g. ChatGPT) the risk is much lower unless you start turning on Connectors. For example I just saw ChatGPT is adding MCP support. This will combine especially poorly with all the recently added memory features - e.g. imagine ChatGPT telling everything it knows about you to some attacker on the internet just because you checked the wrong box in the Connectors settings" @karpathy on X 2025-06-16 17:02:51 UTC 1.3M followers, 102.6K engagements
"Not fully sure why all the LLMs sound about the same - over-using lists delving into multifaceted issues over-offering to assist further about same length responses etc. Not something I had predicted at first because of many independent companies doing the finetuning" @karpathy on X 2024-10-06 19:06:48 UTC 1.3M followers, 3.5M engagements
"When working with LLMs I am used to starting "New Conversation" for each request. But there is also the polar opposite approach of keeping one giant conversation going forever. The standard approach can still choose to use a Memory tool to write things down in between conversations (e.g. ChatGPT does so) so the "One Thread" approach can be seen as the extreme special case of using memory always and for everything. The other day I've come across someone saying that their conversation with Grok (which was free to them at the time) has now grown way too long for them to switch to ChatGPT. i.e." @karpathy on X 2025-03-20 15:02:31 UTC 1.3M followers, 828.7K engagements
"This is interesting as a first large diffusion-based LLM. Most of the LLMs you've been seeing are clones as far as the core modeling approach goes. They're all trained "autoregressively" i.e. predicting tokens from left to right. Diffusion is different - it doesn't go left to right but all at once. You start with noise and gradually denoise into a token stream. Most of the image / video generation AI tools actually work this way and use Diffusion not Autoregression. It's only text (and sometimes audio) that have resisted. So it's been a bit of a mystery to me and many others why for some" @karpathy on X 2025-02-27 01:31:24 UTC 1.3M followers, 947.1K engagements
"Pleasure to come by the YC AI Startup School today I'm told the recordings will be up "in the coming weeks" I'll link to it then and include the slides. Thank you YC for organizing and bringing together an awesome group of builders Fun fact is that when I (and all the original founding members) decided to join OpenAI the name OpenAI didn't exist - we all thought we were joining a new AI non-profit under YC Research. My very first OpenAI swag t-shirt says "YC AI Day 1". Things changed up a bit after that. Cheers to YC :)" @karpathy on X 2025-06-17 20:38:54 UTC 1.3M followers, 487.9K engagements
"@OrdinaryInds Is this real I've been looking for so long ๐โ๐โ๐โ" @karpathy on X 2025-07-09 17:54:18 UTC 1.3M followers, 267.6K engagements
"The hottest new programming language is English" @karpathy on X 2023-01-24 20:14:18 UTC 1.3M followers, 7.8M engagements
"The post below was trending last few days and reminded me that my earlier digital hygiene post was woefully incomplete without a discussion around smartphone choices. The post goes into how on Android apps routinely use a loophole (that Android has known about and not fixed for years) to get the list of all other apps on your phone. I disagree with the author that there are legitimate uses for this information. There aren't or if there are they are super marginal and the privacy tradeoff is not worth it. In practice the data is clearly being collected at scale for shady user profiling. The" @karpathy on X 2025-03-31 16:40:49 UTC 1.3M followers, 518.5K engagements
"Scaling up RL is all the rage right now I had a chat with a friend about it yesterday. I'm fairly certain RL will continue to yield more intermediate gains but I also don't expect it to be the full story. RL is basically "hey this happened to go well (/poorly) let me slightly increase (/decrease) the probability of every action I took for the future". You get a lot more leverage from verifier functions than explicit supervision this is great. But first it looks suspicious asymptotically - once the tasks grow to be minutes/hours of interaction long you're really going to do all that work just" @karpathy on X 2025-07-13 16:35:02 UTC 1.3M followers, 1M engagements
"Tired: elaborate docs pages for your product/service/library with fancy color palettes branding animations transitions dark mode Wired: one single docs .md file and a copy to clipboard button" @karpathy on X 2025-04-22 01:14:50 UTC 1.3M followers, 889.3K engagements
"So so so cool. Llama 1B batch one inference in one single CUDA kernel deleting synchronization boundaries imposed by breaking the computation into a series of kernels called in sequence. The optimal orchestration of compute and memory is only achievable in this way" @karpathy on X 2025-05-27 23:26:44 UTC 1.3M followers, 263.4K engagements
"I often rant about how XX% of attention is about to be LLM attention instead of human attention. What does a research paper look like for an LLM instead of a human Its definitely not a pdf. There is huge space for an extremely valuable research app that figures this out" @karpathy on X 2025-07-10 20:45:08 UTC 1.3M followers, 577.5K engagements
"Do people feel how much work there is still to do. Like wow" @karpathy on X 2025-06-27 16:02:36 UTC 1.3M followers, 169.2K engagements
"Part X of this mystery. Spotted on reddit. In my test not XXX% reproducible but still quite reproducible. ๐ค" @karpathy on X 2025-06-18 18:29:49 UTC 1.3M followers, 2.6M engagements
"I always learn a lot more from in-depth analysis of few random cases over dashboards of aggregate statistics across all cases. Both projections can be helpful but the latter is disproportionately pervasive" @karpathy on X 2025-07-14 22:23:01 UTC 1.3M followers, 279.7K engagements
"RT to help Simon raise awareness of prompt injection attacks in LLMs. Feels a bit like the wild west of early computing with computer viruses (now = malicious prompts hiding in web data/tools) and not well developed defenses (antivirus or a lot more developed kernel/user space security paradigm where e.g. an agent is given very specific action types instead of the ability to run arbitrary bash scripts). Conflicted because I want to be an early adopter of LLM agents in my personal computing but the wild west of possibility is holding me back" @karpathy on X 2025-06-16 16:37:53 UTC 1.3M followers, 425K engagements
"Yep I think RL is misleading in that it restricts field of view. Eg like you mentioned you can imagine review/reflect doing a lot more - building tools for later use or actively tuning the distribution for what to try next (instead of just sampling from policy independently as usual). Or you can imagine environments with no rewards. So much more. Basically - agentic interactions: absolutely +100. RL specifically: eeeh" @karpathy on X 2025-07-14 17:42:28 UTC 1.3M followers, 12.7K engagements
"New 2h11m YouTube video: How I Use LLMs This video continues my general audience series. The last one focused on how LLMs are trained so I wanted to follow up with a more practical guide of the entire LLM ecosystem including lots of examples of use in my own life. Chapters give a sense of content: 00:00:00 Intro into the growing LLM ecosystem 00:02:54 ChatGPT interaction under the hood 00:13:12 Basic LLM interactions examples 00:18:03 Be aware of the model you're using pricing tiers 00:22:54 Thinking models and when to use them 00:31:00 Tool use: internet search 00:42:04 Tool use: deep" @karpathy on X 2025-02-27 22:41:41 UTC 1.3M followers, 979.6K engagements
"o1-mini keeps refusing to try to solve the Riemann Hypothesis on my behalf. Model laziness continues to be a major issue sad ;p" @karpathy on X 2024-09-12 23:34:07 UTC 1.3M followers, 711.5K engagements
"It's 2025 and most content is still written for humans instead of LLMs. XXXX% of attention is about to be LLM attention not human attention. E.g. XX% of libraries still have docs that basically render to some pretty .html static pages assuming a human will click through them. In 2025 the docs should be a single your_project.md text file that is intended to go into the context window of an LLM. Repeat for everything" @karpathy on X 2025-03-12 17:33:19 UTC 1.3M followers, 1.8M engagements
"An attempt to explain (current) ChatGPT versions. I still run into many many people who don't know that: - o3 is the obvious best thing for important/hard things. It is a reasoning model that is much stronger than 4o and if you are using ChatGPT professionally and not using o3 you're ngmi. - 4o is different from o4. Yes I know lol. 4o is a good "daily driver" for many easy-medium questions. o4 is only available as mini for now and is not as good as o3 and I'm not super sure why it's out right now. Example basic "router" in my own personal use: - Any simple query (e.g. "what foods are high in" @karpathy on X 2025-06-02 17:54:57 UTC 1.3M followers, 1.3M engagements
"Very impressed with Veo X and all the things people are finding on r/aivideo etc. Makes a big difference qualitatively when you add audio. There are a few macro aspects to video generation that may not be fully appreciated: X. Video is the highest bandwidth input to brain. Not just for entertainment but also for work/learning - think diagrams charts animations etc. X. Video is the most easy/fun. The average person doesn't like reading/writing it's very effortful. Anyone can (and wants to) engage with video. X. The barrier to creating videos is - X. X. For the first time video is directly" @karpathy on X 2025-06-02 20:22:17 UTC 1.3M followers, 1M engagements
"A major mistake I made in my undergrad is that I focused way too much on mathematical lens of computing - computability decidability asymptotic complexity etc. And too little on physical lens - energy/heat of state change data locality parallelism computer architecture. The former is interesting; The latter bestows power" @karpathy on X 2025-05-06 06:55:12 UTC 1.3M followers, 1.3M engagements
"Making slides manually feels especially painful now that you know Cursor for slides should exist but doesnt" @karpathy on X 2025-06-06 17:37:44 UTC 1.3M followers, 2.6M engagements
"I attended a vibe coding hackathon recently and used the chance to build a web app (with auth payments deploy etc.). I tinker but I am not a web dev by background so besides the app I was very interested in what it's like to vibe code a full web app today. As such I wrote none of the code directly (Cursor+Claude/o3 did) and I don't really know how the app works in the conventional sense that I'm used to as an engineer. The app is called MenuGen and it is live on Basically I'm often confused about what all the things on a restaurant menu are - e.g. Pt Tagine Cavatappi or Sweetbread (hint it's." @karpathy on X 2025-05-01 15:16:10 UTC 1.3M followers, 766.3K engagements
"The docs also have to change in the content. Eg instead of instructing a person to go to some page and do this or that they could show curl commands to run - actions that are a lot easier for an LLM to carry out. Products have to change to support these too. Eg adding a Supabase db to your Vervel app shouldnt be clicks but curls" @karpathy on X 2025-04-22 01:20:50 UTC 1.3M followers, 209.7K engagements
"+1 for "context engineering" over "prompt engineering". People associate prompts with short task descriptions you'd give an LLM in your day-to-day use. When in every industrial-strength LLM app context engineering is the delicate art and science of filling the context window with just the right information for the next step. Science because doing this right involves task descriptions and explanations few shot examples RAG related (possibly multimodal) data tools state and history compacting. Too little or of the wrong form and the LLM doesn't have the right context for optimal performance." @karpathy on X 2025-06-25 15:54:24 UTC 1.3M followers, 2.3M engagements
"@aidenybai I use CC from Cursor and I assumed most do as well (). I end up with a mixed thing where Cursor is the UI layer for reading the code manual edits tab completion and chunk edits and CC for larger changes architecting Q&A. Still rapidly evolving though" @karpathy on X 2025-07-20 19:41:57 UTC 1.3M followers, 153K engagements
"My sleep scores during recent travel were in the 90s. Now back in SF I am consistently back down to 70s 80s. I am increasingly convinced that this is due to traffic noise from a nearby road/intersection where I live - every 10min a car truck bus or motorcycle with a very loud engine passes by (some are 10X louder than others). In the later less deep stages of sleep it is much easier to wake and then much harder to go back to sleep. More generally I think noise pollution (esp early hours) come at a huge societal cost that is not correctly accounted for. E.g. I wouldn't be too surprised if a" @karpathy on X 2025-06-07 19:01:34 UTC 1.3M followers, 1.4M engagements
"Nice - my AI startup school talk is now up Chapters: 0:00 Imo fair to say that software is changing quite fundamentally again. LLMs are a new kind of computer and you program them in English. Hence I think they are well deserving of a major version upgrade in terms of software. 6:06 LLMs have properties of utilities of fabs and of operating systems = New LLM OS fabbed by labs and distributed like utilities (for now). Many historical analogies apply - imo we are computing circa 1960s. 14:39 LLM psychology: LLMs = "people spirits" stochastic simulations of people where the simulator is an" @karpathy on X 2025-06-19 02:01:31 UTC 1.3M followers, 1.2M engagements
"@elonmusk @cyberpaun @shaneguML So what kind of revenue share are we talking about :D jk jk" @karpathy on X 2025-07-16 19:31:09 UTC 1.3M followers, 152.7K engagements
/creator/x::karpathy