Dark | Light
# ![@AlexGDimakis Avatar](https://lunarcrush.com/gi/w:26/cr:twitter::29178343.png) @AlexGDimakis Alex Dimakis

Alex Dimakis posts on X about ai, in the, the first, if you the most. They currently have [------] followers and [---] posts still getting attention that total [--] engagements in the last [--] hours.

### Engagements: undefined [#](/creator/twitter::29178343/interactions)
![Engagements Line Chart](https://lunarcrush.com/gi/w:600/cr:twitter::29178343/c:line/m:interactions.svg)

- [--] Week [---] -96%
- [--] Month [------] -46%
- [--] Months [-------] +15%
- [--] Year [-------] -74%

### Mentions: undefined [#](/creator/twitter::29178343/posts_active)
![Mentions Line Chart](https://lunarcrush.com/gi/w:600/cr:twitter::29178343/c:line/m:posts_active.svg)

- [--] Months [--] -62%
- [--] Year [--] -56%

### Followers: [------] [#](/creator/twitter::29178343/followers)
![Followers Line Chart](https://lunarcrush.com/gi/w:600/cr:twitter::29178343/c:line/m:followers.svg)

- [--] Week [------] +0.12%
- [--] Month [------] +1.30%
- [--] Months [------] +8.80%
- [--] Year [------] +20%

### CreatorRank: undefined [#](/creator/twitter::29178343/influencer_rank)
![CreatorRank Line Chart](https://lunarcrush.com/gi/w:600/cr:twitter::29178343/c:line/m:influencer_rank.svg)

### Social Influence

**Social category influence**
[technology brands](/list/technology-brands)  12% [stocks](/list/stocks)  4% [countries](/list/countries)  3% [social networks](/list/social-networks)  2%

**Social topic influence**
[ai](/topic/ai) 15%, [in the](/topic/in-the) 12%, [the first](/topic/the-first) 8%, [if you](/topic/if-you) 7%, [deep](/topic/deep) 7%, [this is](/topic/this-is) 7%, [llms](/topic/llms) 7%, [6969](/topic/6969) 6%, [theory](/topic/theory) 6%, [we are](/topic/we-are) 6%

**Top accounts mentioned or mentioned by**
[@kpowerinfinity](/creator/undefined) [@giannisdaras](/creator/undefined) [@gstsdn](/creator/undefined) [@hanzhang](/creator/undefined) [@googleai](/creator/undefined) [@neuripsconfneurips2019paperawards807e41d0c1e](/creator/undefined) [@willettbecca](/creator/undefined) [@fernleafflynn](/creator/undefined) [@evenkei](/creator/undefined) [@illithidheretic](/creator/undefined) [@ut](/creator/undefined) [@uw](/creator/undefined) [@wichitastate](/creator/undefined) [@msftresearch](/creator/undefined) [@odedrechavi](/creator/undefined) [@utexaseces](/creator/undefined) [@ieeeorg](/creator/undefined) [@konstdaskalakis](/creator/undefined) [@athenaricinfo](/creator/undefined) [@greece2021](/creator/undefined)

**Top assets mentioned**
[Alphabet Inc Class A (GOOGL)](/topic/$googl) [Microsoft Corp. (MSFT)](/topic/microsoft)
### Top Social Posts
Top posts by engagements in the last [--] hours

"The multiple answers mystery is the most surprising thing we stumbled on from OpenThoughts: Sampling multiple answers for the same question is better than having more questions each answered once. To explain: Say you are creating a dataset of questions and answers to SFT a reasoning llm. You can take [----] questions (eg from stackexchange) and answer them with deepseekR1. Or you can take [---] questions (from the same distribution) and answer each question *twice* independently with deepseekR1. Which one is a better dataset Surprisingly if you re-answer the same questions its a better dataset"  
[X Link](https://x.com/AlexGDimakis/status/1997753658357022856)  2025-12-07T19:42Z 22.5K followers, 28.5K engagements


"Great post on evaluating agents. If you give the agent a machine with strict memory limits (as specified in Terminal-Bench 2) you drop [--] percent or more. Daytona allows 3x more memory and that smooths things out. The environment is part of the benchmark and understanding these variations is key for scientific measurement and optimization. New on the Engineering Blog: Quantifying infrastructure noise in agentic coding evals. Infrastructure configuration can swing agentic coding benchmarks by several percentage pointssometimes more than the leaderboard gap between top models. Read more:"  
[X Link](https://x.com/AlexGDimakis/status/2019676628088164498)  2026-02-06T07:36Z 22.5K followers, [----] engagements


"Parth and Alan presenting Advisor Models in the Berkeley Sky lab retreat. Advisor models are small models that are trained to create personalization or steering advice prompts that are fed to a large model like GPT. Its basically dynamic prompting done by a small LLM that can be trained or personalised. In one experiment the advisor learned which users like short movie reviews and who prefers detailed reviews purely by RL with a numerical reward. Then it adds this personalization information to the prompt of GPT5 that writes the movie reviews."  
[X Link](https://x.com/AlexGDimakis/status/2011967337596113177)  2026-01-16T01:02Z 22.4K followers, [----] engagements


"Best to leave TF for later"  
[X Link](https://x.com/AlexGDimakis/status/1167897353866465283)  2019-08-31T20:30Z 22.5K followers, [---] engagements


"New paper: Your Local GAN: a new layer of two-dimensional sparse attention and a new generative model. Also progress on inverting GANs which may be useful for inverse problems. with @giannis_daras from NTUA and @gstsdn @Han_Zhang_ from @googleai https://arxiv.org/pdf/1911.12287.pdf https://arxiv.org/pdf/1911.12287.pdf"  
[X Link](https://x.com/AlexGDimakis/status/1199868634673434624)  2019-11-28T01:52Z 22.5K followers, [---] engagements


"Meet the 'double descent' phenomenon. After we figure it out we should probably rewrite the book chapter on bias-variance tradeoff. A surprising deep learning mystery: Contrary to conventional wisdom performance of unregularized CNNs ResNets and transformers is non-monotonic: improves then gets worse then improves again with increasing model size data size or training time. https://t.co/3PwNElyNC8 https://t.co/Zdox9dbIBv A surprising deep learning mystery: Contrary to conventional wisdom performance of unregularized CNNs ResNets and transformers is non-monotonic: improves then gets worse then"  
[X Link](https://x.com/AlexGDimakis/status/1202683828164317192)  2019-12-05T20:19Z 22.5K followers, [---] engagements


"#NeurIPS2019 best paper awards Congrats to all the authors https://medium.com/@NeurIPSConf/neurips-2019-paper-awards-807e41d0c1e https://medium.com/@NeurIPSConf/neurips-2019-paper-awards-807e41d0c1e"  
[X Link](https://x.com/AlexGDimakis/status/1203787732436295681)  2019-12-08T21:25Z 22.5K followers, [---] engagements


"I will be giving an online seminar at IAS Princeton on Thursday April 23rd 2pm (Central time) : Deep Generative models and Inverse Problems https://www.ias.edu/events/seminar-theoretical-machine-learning-84 https://theias.zoom.us/j/384099138 https://www.ias.edu/events/seminar-theoretical-machine-learning-84 https://theias.zoom.us/j/384099138"  
[X Link](https://x.com/AlexGDimakis/status/1253246683829473280)  2020-04-23T08:57Z 22.5K followers, [---] engagements


"(1/3) We wrote a survey on Deep Learning Techniques for Inverse Problems in Imaging We came up with a taxonomy that I think is interesting. Also discussed the whole 'what is supervised vs unsupervised' issue. @WillettBecca https://arxiv.org/abs/2005.06001 https://arxiv.org/abs/2005.06001"  
[X Link](https://x.com/AlexGDimakis/status/1261022470099869696)  2020-05-14T19:56Z 22.5K followers, [---] engagements


"@FernleafFlynn @even_kei @IllithidHeretic Two major industries breaking ways for a paltry sum"  
[X Link](https://x.com/AlexGDimakis/status/1263544264375574528)  2020-05-21T18:56Z 22.5K followers, [---] engagements


"As Information theory was becoming a 'hot' scientific trend in the 50s Claude Shannon wrote a one-page paper advising hype *reduction*. That never happens anymore. Claude Shannon's "The Bandwagon" (1956) is a timeless gem. Short one page advise and perspective on the status of the field. ". we must keep our own house in first class order. The subject of information theory has certainly been sold if not oversold." https://t.co/Jn0e72B5Bz Claude Shannon's "The Bandwagon" (1956) is a timeless gem. Short one page advise and perspective on the status of the field. ". we must keep our own house in"  
[X Link](https://x.com/AlexGDimakis/status/1274776409370701825)  2020-06-21T18:49Z 22.5K followers, [---] engagements


"Let the advisor show you how to write the rebuttal https://x.com/i/status/1294367648814424064/video/1 https://x.com/i/status/1294367648814424064/video/1"  
[X Link](https://x.com/AlexGDimakis/status/1295979503076999168)  2020-08-19T07:02Z 22.5K followers, [---] engagements


"Then one wakes up after the talk and asks an extremely insightful question. Full professors sitting in the first row of every seminar in history https://t.co/4JIQ0XAiXd Full professors sitting in the first row of every seminar in history https://t.co/4JIQ0XAiXd"  
[X Link](https://x.com/AlexGDimakis/status/1296271747927482368)  2020-08-20T02:24Z 22.5K followers, [---] engagements


"We're very excited that @UT Austin will lead an NSF national Institute on the Foundations of Machine Learning with @UW @WichitaState and @MSFTResearch Announcement: https://news.utexas.edu/2020/08/26/ut-austin-selected-as-home-of-national-ai-institute-focused-on-machine-learning/ https://news.utexas.edu/2020/08/26/ut-austin-selected-as-home-of-national-ai-institute-focused-on-machine-learning/"  
[X Link](https://x.com/AlexGDimakis/status/1298619401164533761)  2020-08-26T13:52Z 22.5K followers, [---] engagements


"If you are a #neurips2020 reviewer please read the authors rebuttal and at the very least update your review indicating that you read it and your updated thoughts. It takes [--] minutes and its a good step towards decency. Meta-reviewers please enforce this"  
[X Link](https://x.com/AlexGDimakis/status/1301052865268588545)  2020-09-02T07:02Z 22.5K followers, [---] engagements


"Of Cramer-Rao and Rao-Blackwell fame https://www.thehindu.com/opinion/open-page/a-doyen-of-statistics-turns-100/article32563522.ece https://www.thehindu.com/opinion/open-page/a-doyen-of-statistics-turns-100/article32563522.ece"  
[X Link](https://x.com/AlexGDimakis/status/1304669363530797056)  2020-09-12T06:33Z 22.5K followers, [---] engagements


"Oppenheimer's Berkeley recommendation letter for Feynman mentions that Wigner said "He is a second Dirac only this time human." https://medium.com/cantors-paradise/oppenheimers-letter-of-recommendation-for-richard-feynman-1943-15dcdaf131b7 https://medium.com/cantors-paradise/oppenheimers-letter-of-recommendation-for-richard-feynman-1943-15dcdaf131b7"  
[X Link](https://x.com/AlexGDimakis/status/1309011999213473794)  2020-09-24T06:09Z 22.5K followers, [---] engagements


"Fun question in my ML midterm: Say a feature X1 is independent from the target label Y. We can always remove this feature and not lose in predictive performance. Yes No Yes No"  
[X Link](https://x.com/AlexGDimakis/status/1326746526954237952)  2020-11-12T04:39Z 22.5K followers, [---] engagements


"Interesting fact about GANs that is not as well known as it should be: Take a pre-trained GAN (eg DCGAN) and feed independent random noise to the discriminator. It is easy to tell noise is not a real image. You would expect that the discriminator will easily see this. (1/4)"  
[X Link](https://x.com/AlexGDimakis/status/1337324417232285697)  2020-12-11T09:12Z 22.5K followers, [---] engagements


"My students after every joke I make in a Zoom lecture. (h/t: @OdedRechavi )"  
[X Link](https://x.com/AlexGDimakis/status/1339274191074373642)  2020-12-16T18:20Z 22.5K followers, [---] engagements


"A public service announcement: please upload all your papers on preprint servers like arxiv. The publisher owns final pdf THEY typeset not the preprint pdf you submitted. If your papers are only behind a paywall you are violating funding recommendations"  
[X Link](https://x.com/AlexGDimakis/status/1376968777947090945)  2021-03-30T18:45Z 22.5K followers, [---] engagements


"I disagree-- many scientists will use ML algorithms in the same way they use databases compilers and statistics today. Domain expertise and scientific insight do not go away when the tools change. Within 10-20 years nearly every branch of science will be for all intents and purposes a branch of computer science. Computational physics comp chemistry comp biology comp medicine. Even comp archeology. Realistic simulations big data analysis and ML everywhere Within 10-20 years nearly every branch of science will be for all intents and purposes a branch of computer science. Computational physics"  
[X Link](https://x.com/AlexGDimakis/status/1389395151785795586)  2021-05-04T01:43Z 22.5K followers, [---] engagements


"Doctor: We used a deep learning algorithm for your MRI reconstruction. Turns out one of your kidneys is a cat"  
[X Link](https://x.com/AlexGDimakis/status/1396898949068431377)  2021-05-24T18:40Z 22.5K followers, [----] engagements


"On the difference between (classical) Statistics and Machine Learning I found this gem by Leo Breiman:'The two cultures of Statistical modeling' https://projecteuclid.org/journals/statistical-science/volume-16/issue-3/Statistical-Modeling--The-Two-Cultures-with-comments-and-a/10.1214/ss/1009213726.full https://projecteuclid.org/journals/statistical-science/volume-16/issue-3/Statistical-Modeling--The-Two-Cultures-with-comments-and-a/10.1214/ss/1009213726.full"  
[X Link](https://x.com/AlexGDimakis/status/1446341482655461391)  2021-10-08T05:07Z 22.5K followers, [---] engagements


"We have tried to use discriminators of GANs as regularizers for detecting adversarial examples for dozens of things: It NEVER works. I always think it's a great idea and then nope. ๐Ÿ˜“"  
[X Link](https://x.com/AlexGDimakis/status/1446720743795052545)  2021-10-09T06:14Z 22.5K followers, [---] engagements


"New NeurIPS paper: We train a Robust CLIP encoder that produces approximate CLIP representations by seeing highly corrupted images. We can classify images by observing 2% random pixels or very blurry images better than humans. https://www.sriramravula.com/publications/inverse-problems-leveraging-pre-trained-contrastive-representations https://www.sriramravula.com/publications/inverse-problems-leveraging-pre-trained-contrastive-representations"  
[X Link](https://x.com/AlexGDimakis/status/1451620675878330368)  2021-10-22T18:45Z 22.5K followers, [---] engagements


"Here is a very good reason why the NyquistShannon sampling theorem requires that your function is low-pass before you sub-sample to downscale. If you just sub-sample without smoothing a bad guy can place another image exactly on the pixels you sub-sample. Adversarial aliasing. image-scaling attacks are wild small dots added to the image on the left turns it into the image on the right when downscaled could make auditing ML systems very tricky if you only look at the original images. https://t.co/PvidAaxJLS image-scaling attacks are wild small dots added to the image on the left turns it into"  
[X Link](https://x.com/AlexGDimakis/status/1456859486728212483)  2021-11-06T05:42Z 22.5K followers, [---] engagements


"We have multiple postdoc openings at the AI Institute for the Foundations of Machine Learning (IFML). Fellows can work with all IFML groups in UT Austin Univ. of Washington and Microsoft Research (1/3) https://apply.interfolio.com/98753 https://apply.interfolio.com/98753"  
[X Link](https://x.com/AlexGDimakis/status/1458899799684567045)  2021-11-11T20:49Z 22.5K followers, [---] engagements


"While waiting for #CVPR2022 CMT to get up again I would like to propose a simple cryptographic solution to the big data submission problem: We only upload a SHA256 hash of our to-be-submitted pdf and then upload the committed pdf any time next week"  
[X Link](https://x.com/AlexGDimakis/status/1460883477302333441)  2021-11-17T08:12Z 22.5K followers, [---] engagements


"Honored to be selected as an IEEE Fellow for contributions to distributed coding and learning' Congratulations to the whole Fellows class of [----] https://www.ieee.org/content/dam/ieee-org/ieee/web/org/about/fellows/2022-ieee-fellows-class.pdf Congratulations to @utexasece's Seth Bank @AlexGDimakis and Sriram Vishwanath for being selected as @IEEEorg Fellows https://t.co/yPfwbxMVb9 https://www.ieee.org/content/dam/ieee-org/ieee/web/org/about/fellows/2022-ieee-fellows-class.pdf Congratulations to @utexasece's Seth Bank @AlexGDimakis and Sriram Vishwanath for being selected as @IEEEorg Fellows"  
[X Link](https://x.com/AlexGDimakis/status/1463330284183638019)  2021-11-24T02:15Z 22.5K followers, [---] engagements


"Amazing news: AI and Data science research center founded in Greece [--] million funding Led by Christos Papadimitriou @KonstDaskalakis and Timos Sellis under @athenaRICinfo and the support of @Greece_2021   @Greece_2021        .       . https://t.co/CGHXtS75RQ   @Greece_2021        .       . https://t.co/CGHXtS75RQ"  
[X Link](https://x.com/AlexGDimakis/status/1473915482545741827)  2021-12-23T07:16Z 22.5K followers, [---] engagements


"Excited that our paper on deep generative models for robust MRI is featured by Amazon Science. We trained the first generative model for MRI images. Also for the first time we are competitive with supervised deep MRI methods and more robust to anatomy and measurement changes. Time can seem to slow during an MRI scan. #AmazonResearchAward recipient Jonathan Tamir is developing #machinelearning methods to shorten exam times and extract more data from this essential but often uncomfortable imaging process. Find out how. https://t.co/rqU42qYW43 Time can seem to slow during an MRI scan."  
[X Link](https://x.com/AlexGDimakis/status/1482130110337724424)  2022-01-14T23:18Z 22.5K followers, [---] engagements


"The night is young and full of Overfull hbox(6.97092pt too wide) detected at line 375"  
[X Link](https://x.com/AlexGDimakis/status/1486601649854496768)  2022-01-27T07:27Z 22.5K followers, [---] engagements


"Someone is trying to scam my PhD student. My student asks to verify their identity 1/2"  
[X Link](https://x.com/AlexGDimakis/status/1487251984482766850)  2022-01-29T02:31Z 22.5K followers, [----] engagements


"2/ Scammer ends up improving our sample complexity bound for StyleGAN inverse problems. They teach them to do chaining arguments instead of just union bounds now jeez. @giannis_daras"  
[X Link](https://x.com/AlexGDimakis/status/1487251986382831617)  2022-01-29T02:31Z 22.5K followers, [----] engagements


"I was informed that Alexander Vardy a giant in coding theory passed away. A tragic loss for his family UCSD and academia. Alex's many discoveries include the Polar decoding algorithm used in the 5G wireless standard (1/3)"  
[X Link](https://x.com/AlexGDimakis/status/1503807067391418373)  2022-03-15T18:55Z 22.5K followers, [---] engagements


"One huge advantage of deep learning (vs classical ML models) that is not often discussed is *modularity*: One can download pre-trained models glue them like Legos and fine tune them end-to-end because gradients flow through. (1/n)"  
[X Link](https://x.com/AlexGDimakis/status/1506485470255280129)  2022-03-23T04:18Z 22.5K followers, [----] engagements


"DALLE [--] and similar models are producing amazing images from Text. But can they count to five I don't have access but when I try 'An image of five apples' on multimodalart latentdiffusion LAION-400M model I get wrong images constructed. (1/n)"  
[X Link](https://x.com/AlexGDimakis/status/1511913360568295425)  2022-04-07T03:46Z 22.5K followers, [---] engagements


"Greece is quite the outlier here in the south on the number of metal bands per Capita. Any explanations Metal bands per [--] million people (Europe) https://t.co/OPEROKiBLo Metal bands per [--] million people (Europe) https://t.co/OPEROKiBLo"  
[X Link](https://x.com/AlexGDimakis/status/1513244007270334466)  2022-04-10T19:54Z 22.5K followers, [---] engagements


"The term Artificial Intelligence was coined by John McCarthy to avoid association with Cybernetics and specifically its pioneer Norbert Wiener who was already famous pain to work with and working on Cybernetics in MIT. Original quote from McCarthy's Stanford page: . (1/n)"  
[X Link](https://x.com/AlexGDimakis/status/1516451200408993795)  2022-04-19T16:18Z 22.5K followers, [---] engagements


"Based on recent papers (Gpt3 Palm dalle2 Gato Metaformer) I am forming the opinion that maybe 'Scale is all you need' possibly even for general intelligence (). Just convert everything to tokens and predict the next token. (1/n)"  
[X Link](https://x.com/AlexGDimakis/status/1526388274348150784)  2022-05-17T02:24Z 22.5K followers, [----] engagements


"My student Giannis discovered that DALLE2 has a secret language. This can be used to crate absurd prompts that generate images. E.g. ''Apoploe vesrreaitais eating Contarra ccetnxniams luryca tanniounons'' generates Birds eating Bugs We wrote a short paper on our experiments. DALLE-2 has a secret language. "Apoploe vesrreaitais" means birds. "Contarra ccetnxniams luryca tanniounons" means bugs or pests. The prompt: "Apoploe vesrreaitais eating Contarra ccetnxniams luryca tanniounons" gives images of birds eating bugs. A thread (1/n)๐Ÿงต https://t.co/VzWfsCFnZo DALLE-2 has a secret language."  
[X Link](https://x.com/AlexGDimakis/status/1531693923118288905)  2022-05-31T17:47Z 22.5K followers, [---] engagements


"I really need to disagree with this statement. E.g. in my lab in UT Austin good software engineering is useful but not the most important skill to learn. We train ML researchers on how to do research e.g. understanding and improving landmark papers ideally writing one. This is probably well-known in some circles but not everywhere. The most important skill for Research Scientists in AI (at least at @OpenAI) is software engineering. Background in ML research is sometimes useful but you can usually get away with a few landmark paper. This is probably well-known in some circles but not"  
[X Link](https://x.com/AlexGDimakis/status/1548288544866570241)  2022-07-16T12:48Z 22.5K followers, [---] engagements


"Is there a doctor on the plane -Yes but not that kind of doctor. -The passenger in 36c is trying to inpaint an image using a pre-trained stable diffusion model and simply copy-pastes the inpainting observed part in place after each iteration -Ok I got this"  
[X Link](https://x.com/AlexGDimakis/status/1571299062522970112)  2022-09-18T00:44Z 22.5K followers, [---] engagements


"Human bilinguals are more robust to dementia and cognitive decline. In our recent NeurIPS paper we show that bilingual GPT models are also more robust to structural damage in their neuron weights. Further we develop a theory. (1/n)"  
[X Link](https://x.com/AlexGDimakis/status/1622006950950014981)  2023-02-04T22:59Z 22.5K followers, 312.6K engagements


"We develop a theory that shows how multitasking creates regularization. This is can be seen as a simple theoretical model for bilingual cognitive reserve. Interestingly the phenomenon appears only when the tasks are sufficiently diverse. (2/n)"  
[X Link](https://x.com/AlexGDimakis/status/1622006952443224064)  2023-02-04T22:59Z 22.5K followers, 16.5K engagements


"A small experiment: This Tweet has an even number of likes"  
[X Link](https://x.com/AlexGDimakis/status/1623087611802779649)  2023-02-07T22:33Z 22.5K followers, 45.5K engagements


"Who first generated text with statistical methods like GPT In [----] Claude Shannon wrote the landmark paper 'A Mathematical Theory of Communication'. There he defined and estimated the entropy of English by generating synthetic text: 'THE HEAD AND IN FRONTAL ATTACK ON (1/n)"  
[X Link](https://x.com/AlexGDimakis/status/1623113337574682626)  2023-02-08T00:15Z 22.5K followers, 42.4K engagements


"Here is a simple way to beat ChatGPT and any similar architecture with one Turing test question. ChatGPT GPT3 and all related Transformers have a finite maximum token sequence length usually 2k to 4k tokens. (1/n)"  
[X Link](https://x.com/AlexGDimakis/status/1628790477808967685)  2023-02-23T16:14Z 22.5K followers, 163.2K engagements


"Scott Aaronson gave an extraordinary public lecture in UT Austin's Machine Learning Lab (MLL) yesterday. Most packed auditorium I've seen. He described a taxonomy for AI alignment methods [--]. Off switch [--]. Sandboxing / Isolation [--]. Interpretability [--]. Multiple competing / Debating AIs [--]. Boostrapping [--]. Formal Verification / Interactive Proofs [--]. Formalizing human values/ Laws of Robotics [--]. Value Learning/ Coherent Extrapolated Volition He also spend some time explaining how watermarking GPT can work by making the sampler pseudorandom"  
[X Link](https://x.com/AlexGDimakis/status/1647009181839286273)  2023-04-14T22:49Z 22.5K followers, 45.2K engagements


"My thoughts on the now famous Google leak doc: [--]. Open source AI is winning. I agree and that is great for the world and for a competitive ecosystem. In LLMs we're not there but we just got OpenClip to beat openAI Clip and Stable diffusion is better than closed models. [--]. You don't need huge models high quality data is much more efficient and important. Alpacaing models behind APIs further reduces moats. [--]. You can start with a good foundation model and parameter efficient fine-tuning (PEFT) algorithms like Lora work super well in a day. Finally an opening for algorithmic innovations 4."  
[X Link](https://x.com/AlexGDimakis/status/1654286036015411205)  2023-05-05T00:45Z 22.5K followers, 189.1K engagements


"New neural renderer by Nvidia. The model adds fingerprints smudges and dust and generates renders indistinguishable from real to me. Oh and its done at *real-time*. Can't wait to see games using this. (1/2)"  
[X Link](https://x.com/AlexGDimakis/status/1655056946150481922)  2023-05-07T03:48Z 22.5K followers, 29.9K engagements


"I was surprised by a talk Yejin Choi (an NLP expert) gave yesterday in Berkeley on some surprising weaknesses of GPT4: As many humans know 237*757=179409 but GPT4 said [------]. For the easy problem of multiplying two [--] digit numbers they measured GPT4 accuracy being only 59% accuracy on [--] digit number multiplication. Only 4% on [--] digit number multiplication and zero on 5x5. Adding scratchpad helped GPT4 but only to 92% accuracy on multiplying two [--] digit numbers. Even more surprisingly finetuning GPT3 on 1.8m examples of [--] digit multiplication still only gives [--] percent test accuracy (in"  
[X Link](https://x.com/AlexGDimakis/status/1691600985938858432)  2023-08-16T00:01Z 22.5K followers, 1.7M engagements


"References: The Faith and Fate Paper is available here: Video of this great talk here: https://www.youtube.com/watchv=P7ZdUbSAujQ https://arxiv.org/pdf/2305.18654.pdf https://www.youtube.com/watchv=P7ZdUbSAujQ https://arxiv.org/pdf/2305.18654.pdf"  
[X Link](https://x.com/AlexGDimakis/status/1691601220039831553)  2023-08-16T00:02Z 22.5K followers, 57.4K engagements


"@raj_raj88 But even fine-tuning with 1.8m multiplication examples was not able to teach it to generalize to other (3 digit) multiplications. This indicates some fundamental architecture limitation"  
[X Link](https://x.com/AlexGDimakis/status/1691626794477076842)  2023-08-16T01:43Z 22.5K followers, 27K engagements


"Probably the best 1h introduction to LLMs that I've seen. And after 20mins its not an introduction its getting into cutting edge research updates updated up to this month. I had not heard of the data exfiltration by prompt injection or the recent finetuning Poisoning attacks. https://www.youtube.com/watchv=zjkBMFhNj_g&t=2s https://www.youtube.com/watchv=zjkBMFhNj_g&t=2s"  
[X Link](https://x.com/AlexGDimakis/status/1727595762266026128)  2023-11-23T07:51Z 22.5K followers, 74.5K engagements


""Datacomp1B is the first public dataset that outperforms OpenAI" #NeurIPS2023"  
[X Link](https://x.com/AlexGDimakis/status/1735340429380370530)  2023-12-14T16:46Z 22.5K followers, 38.1K engagements


"The Google Gemini paper was released today and has [---] authors. I was impressed but then found that a recent LHC physics paper with [----] authors. The first nine pages describe the research and the other [--] pages list the authors and their institutions. But that's not even the record. The most authors on a single peer-reviewed academic paper is [-----] and was achieved by the COVIDSurg and GlobalSurg Collaboratives at the University of Birmingham and the University of Edinburgh. All [---] Gemini coauthors are expected to quit Google and start [---] LLM startups next year"  
[X Link](https://x.com/AlexGDimakis/status/1737598802415018157)  2023-12-20T22:20Z 22.5K followers, 56.2K engagements


"Very cool explanation of emergence even in light of the neurips recent best paper award: even if for a single task performance increases smoothly with more training if a composite task requires k tasks to be correct a phase transition appears as k grows. I'd like to add that even if 0.5k success or more are needed in k trials we'd see a phase transition going back to AEP the fundamental concentration also used in information theory. 1/2 Wrote blog on whether emergent abilities and grokking are a fundamental feature of deep learning a "mirage" or both. This is partially based on the beautiful"  
[X Link](https://x.com/AlexGDimakis/status/1738404987980390788)  2023-12-23T03:43Z 22.5K followers, 69.2K engagements


"We just discovered that the inpainting model in Stable Diffusion is cheating. To clarify: Inpainting is a type of inverse problem where some missing data (pixels) must be filled in. In our testing some of the inpaintings from the SDXL inpainting model where a little 'too good': filling in details in the masked missing pixels they couldn't possibly know unless the model was cheating by observing masked pixels. So we created this test dog image with some Pink-Cyan boxes and then asked the model to inpaint it. We chose the masking region to fully contain the Pink and Cyan boxes so there is no"  
[X Link](https://x.com/AlexGDimakis/status/1747749640315789399)  2024-01-17T22:36Z 22.5K followers, 48.1K engagements


"Excited to be the director for the new Texas Center for Generative AI Please welcome the Center for Generative AI -- a World-Class AI Research Center with a Texas-Sized GPU Cluster. Led by @AlexGDimakis ๐Ÿ’ซ #YearofAI @TexasScience @UTAustin https://t.co/jTZd4uV0ps Please welcome the Center for Generative AI -- a World-Class AI Research Center with a Texas-Sized GPU Cluster. Led by @AlexGDimakis ๐Ÿ’ซ #YearofAI @TexasScience @UTAustin https://t.co/jTZd4uV0ps"  
[X Link](https://x.com/AlexGDimakis/status/1750580887194943640)  2024-01-25T18:06Z 22.5K followers, 52.6K engagements


"Ptolemy the king of Egypt wanted to learn geometry but found Euclid's book the Elements too difficult to study. So he asked Euclid to show him an easier way to master it. Euclid famously said "Sir there is no royal road to geometry." This is still true a few thousand years later in the days of Youtube and TikTok as Andrej nicely points out. # on shortification of "learning" There are a lot of videos on YouTube/TikTok etc. that give the appearance of education but if you look closely they are really just entertainment. This is very convenient for everyone involved : the people watching enjoy"  
[X Link](https://x.com/AlexGDimakis/status/1756784225628401725)  2024-02-11T20:56Z 22.5K followers, 93.6K engagements


"The #Sora model is indeed incredible ๐Ÿคฏ congratulations to the OpenAI team. It is common for people to think that all the amazing research breakthroughs in AI (like #Sora) are happening inside companies like OpenAI while universities are becoming irrelevant. I want to highlight that the two first authors in the Sora paper Tim Brooks and Bill Peebles received their PhDs from UC Berkeley in [----] and their dissertation research is closely related to this breakthrough. Of course the compute infra and talent in OpenAI is critical for breakthroughs. I just want to point out that the training of the"  
[X Link](https://x.com/AlexGDimakis/status/1759283664527970584)  2024-02-18T18:28Z 22.5K followers, 44.7K engagements


"I was thrilled to learn about this best paper award announced today in COLT [----] the premier learning theory venue. The paper is "Smoothed Analysis for Learning Concepts with Low Intrinsic Dimension" authored by students Gautam Chandrasekaran Konstantinos Stavropoulos IFML postdoc Vasilis Kontonis IFML director Adam Klivans and former UT CS PhD Raghu Meka. Smoothed analysis is an ingenious idea of going beyond worst case pioneered by my former USC colleague Shanghua Teng and Dan Spielman). This paper showed how to apply this framework for learning theory. Here is my basic understanding of"  
[X Link](https://x.com/AlexGDimakis/status/1808535472886468754)  2024-07-03T16:17Z 22.5K followers, 38.2K engagements


"Excited to launch the first model from our startup: Bespoke Labs. Bespoke-Minicheck-7B is a grounded factuality checker: super lightweight and fast. Outperforms all big foundation models including Claude [---] Sonnet Mistral-Large m2 and GPT 4o and its only 7B. Also I want to congratulate Greg Durrett and his group for making the best benchmark and leaderboard for grounded factuality. ๐Ÿค” Want to know if your LLMs are factual You need LLM fact-checkers. ๐Ÿ“ฃ Announcing the LLM-AggreFact leaderboard to rank LLM fact-checkers. ๐Ÿ“ฃ Want the best model Check out @bespokelabsais Bespoke-Minicheck-7B"  
[X Link](https://x.com/AlexGDimakis/status/1821953719325618234)  2024-08-09T16:56Z 22.5K followers, 35.1K engagements


"GPT is having a profound effect on how students write. Its verbose style full of cliches and 'fancy' out of place vocabulary is in every paper and draft I read. A few years back there were grammar errors and awkwardness -- but at least people had their own voice. Now scholarship is getting full of robotic triviality"  
[X Link](https://x.com/AlexGDimakis/status/1831833630022496515)  2024-09-05T23:15Z 22.5K followers, 951.3K engagements


"@ClausWilke Probably this generic plastic style passes RLHF the same way we would be using wooden language when being interrogated"  
[X Link](https://x.com/AlexGDimakis/status/1831859172901843302)  2024-09-06T00:57Z 22.5K followers, 32.8K engagements


"@nandofioretto Yes that's right. Structure and flow in writing help us organize our thought. Blindly using LLMs is an airbrush that makes it harder for people to see that they have muddled flow"  
[X Link](https://x.com/AlexGDimakis/status/1831895840799256669)  2024-09-06T03:22Z 22.5K followers, 42.9K engagements


"For the first (and probably last) time in my life I understand the technical details of both the physics and chemistry Nobel prizes. BREAKING NEWS The Royal Swedish Academy of Sciences has decided to award the [----] #NobelPrize in Chemistry with one half to David Baker for computational protein design and the other half jointly to Demis Hassabis and John M. Jumper for protein structure prediction. https://t.co/gYrdFFcD4T BREAKING NEWS The Royal Swedish Academy of Sciences has decided to award the [----] #NobelPrize in Chemistry with one half to David Baker for computational protein design and"  
[X Link](https://x.com/AlexGDimakis/status/1843995475743228128)  2024-10-09T12:42Z 22.5K followers, 56.4K engagements


"Ok this paper seems super interesting and also makes me want to teach graphical models again. The question is when does chain of thought help and the answer proposed is finding that intermediate steps are only helpful when the training data is locally structured with respect to dependencies between variables. So it depends on the training data and they test that by training on different types of synthetic datasets. Also has theory and seems to do the entire formulation using Bayes nets which is very cool and Ill try to understand this more. Any insights welcome. Training data's local patterns"  
[X Link](https://x.com/AlexGDimakis/status/1854428769165123901)  2024-11-07T07:40Z 22.5K followers, 37K engagements


"I think this is a very good example of survivorship bias: You selected one of the brightest minds of the 20th century and compare them against a typical Phd student. In [--] years we can look back into 2020s and compare the impact of the top minds of 2020s to the top minds of the 1960s. That would be a better historical comparison"  
[X Link](https://x.com/AlexGDimakis/status/1857154876608065587)  2024-11-14T20:13Z 22.5K followers, [----] engagements


"Im excited to introduce Evalchemy ๐Ÿงช a unified platform for evaluating LLMs. If you want to evaluate an LLM you may want to run popular benchmarks on your model like MTBench WildBench RepoBench IFEval AlpacaEval etc as well as standard pre-training metrics like MMLU. This requires you to download and install more than [--] repos each with different dependencies and issues. This is as you might expect an actual nightmare. (1/n) https://github.com/mlfoundations/evalchemy https://github.com/mlfoundations/evalchemy"  
[X Link](https://x.com/AlexGDimakis/status/1858545284386803975)  2024-11-18T16:18Z 22.5K followers, 148K engagements


"Very interesting piece of history I just learned from Ion Stoica in AI Native event: Databricks was founded because Hortonworks would not support the Spark open source project so some company needed to be created to support it"  
[X Link](https://x.com/AlexGDimakis/status/1859712895745130981)  2024-11-21T21:38Z 22.5K followers, [----] engagements


"A Thanksgiving story A few years back I used to play tennis in a ladder system which would match me up with various folks in my neighborhood. After Thanksgiving I had a tennis match with this guy: nice guy two kids a bit overweight in his 50ies I had never met him before. We start our match. During the match he says -Sorry lets stop for a bit I want to catch my breath. -Sure no problem. We start and [--] minutes after he says: -Sorry I ate too much at the Thanksgiving dinner and I have digestion problems. He was burping a bit and looked tired. He asks to reschedule the game I say sure sounds"  
[X Link](https://x.com/AlexGDimakis/status/1862560015263179256)  2024-11-29T18:11Z 22.5K followers, 37.5K engagements


"Life update: I am excited to announce that I will be starting as a Professor in UC Berkeley in the EECS Department. I spend [--] wonderful years teaching in UT Austin and I am grateful to all my colleagues and students there and extremely proud of what we have achieved in AI in UT Austin and I plan to continue my numerous UT close collaborations. I will also continue as Chief Scientist in Bespoke Labs making it much easier now being in the Bay area. I received my Phd in [----] from @Berkeley_EECS and I am thrilled to be back. I am grateful for this new opportunity"  
[X Link](https://x.com/AlexGDimakis/status/1869124346264043827)  2024-12-17T20:55Z 22.5K followers, 110.9K engagements


"Most AI researchers I talk to have been a bit shocked by DeepSeek-R1 and its performance. My preliminary understanding nuggets: [--]. Simple post-training recipe called GRPO: Start with a good model and reward for correctness and style outcomes. No PRM no MCTS no fancy reward models. Basically checks if the answer is correct. ๐Ÿ˜… [--]. Small models can reason very very well with correct distillation post-training. They released a 1.5B model () that is better than Claude and Llama 405B in AIME24. Also their distilled 7B model seems better than o1 preview. ๐Ÿค“ [--]. The datasets used are not released if I"  
[X Link](https://x.com/AlexGDimakis/status/1881511481164079507)  2025-01-21T01:17Z 22.5K followers, 184.1K engagements


"@percyliang @deepseek_ai We are working on fixing that and create the largest open reasoning dataset. More coming very soon ๐Ÿ˜‰"  
[X Link](https://x.com/AlexGDimakis/status/1883420222029476177)  2025-01-26T07:42Z 22.5K followers, 37.9K engagements


"What if we had the data that DeepSeek-R1 was post-trained on We announce Open Thoughts an effort to create such open reasoning datasets. Using our data we trained Open Thinker 7B an open data model with performance very close to DeepSeekR1-7B distill. We are announcing Open Thoughts our large-scale open-source effort to curate the best open reasoning datasets DeepSeek-R1 is amazing but we still don't have access to high-quality open reasoning datasets. These datasets are crucial if you want to build your reasoning models https://t.co/2kU6z8zDdT We are announcing Open Thoughts our large-scale"  
[X Link](https://x.com/AlexGDimakis/status/1884288997641621983)  2025-01-28T17:14Z 22.5K followers, 25.6K engagements


"@jxmnop The first evidence of superintelligence I ever saw was in a calculator"  
[X Link](https://x.com/AlexGDimakis/status/1885851830800019632)  2025-02-02T00:44Z 22.5K followers, 11.6K engagements


"We are releasing OpenThinker-32B the best 32B reasoning model with open data. We match or outperform Deepseek-R1-32B (a closed data model) in reasoning benchmarks. Congrats to Negin and the whole Open Thoughts team. Announcing OpenThinker-32B: the best open-data reasoning model distilled from DeepSeek-R1. Our results show that large carefully curated datasets with verified R1 annotations produce SoTA reasoning models. Our 32B model outperforms all 32B models including https://t.co/w5Ewi2TDil Announcing OpenThinker-32B: the best open-data reasoning model distilled from DeepSeek-R1. Our results"  
[X Link](https://x.com/AlexGDimakis/status/1889740222667133085)  2025-02-12T18:15Z 22.5K followers, 16K engagements


"I think this is a very important result on what Transformers can learn. Length generalization was a known open problem and this paper seems to solve it with self-improvement where a model iteratively labels its own train data and learns from increasingly harder examples. It seems this is quite a general solution to teaching Transformers to generalize. o3 can't multiply beyond a few digits. But I think multiplication addition maze solving and easy-to-hard generalization is actually solvable on standard transformers. with recursive self-improvement. Below is the acc of a tiny model teaching"  
[X Link](https://x.com/AlexGDimakis/status/1890614248956391688)  2025-02-15T04:08Z 22.5K followers, 21.5K engagements


"Is GPT4o an Autoregressive image generation model (i.e. creates the pixel row by row from top to bottom) or a diffusion Here is a very clever way to get evidence: GPT4o cannot generate valid mazes but if you ask it to generate rotated mazes at [--] degrees it can- which is easier if you are an autoregressive model generating from top to bottom (but not easier if you're a diffusion). OK something VERY COOL is happening and 4o+imagen CAN generate VALID mazes when they are in the shape of a rhombus i.e. [--] rotated square Some examples below (the red squiggly line is mine). The prompt is "generate"  
[X Link](https://x.com/AlexGDimakis/status/1906584436776546641)  2025-03-31T05:48Z 22.5K followers, 17.7K engagements


"We are excited to release the OpenThinker2 reasoning models and data. In summary: [--]. Openthinker32B Outperforms DeepSeekR1-32B in reasoning. [--]. Fully open source open weights and open data (1M carefully curated samples). [--]. Post-trained only with SFT. RL post-training will likely further improve performance. Read the whole story.๐Ÿ‘‡ Turns out its possible to outperform DeepSeekR1-32B with only SFT on open data and no RL: Announcing OpenThinker2-32B and OpenThinker2-7B. We also release the data OpenThoughts2-1M curated by selecting quality instructions from diverse sources. ๐Ÿงต (1/n)"  
[X Link](https://x.com/AlexGDimakis/status/1907837879902224862)  2025-04-03T16:49Z 22.5K followers, 16.7K engagements


"Very cool result: KV cache compression can be done with compressed sensing: store keys and values as sparse combinations of some dictionary vectors. Interestingly the dictionary is universal across inputs (but learned for each model). Kinda cute that you can reduce KV cache by replacing it with a universal transferable dictionary + old school sig. proc reconstruction algorithm. We tested on non-reasoning models and was sota but methinks it'll work even better on reasoning ones. The ICML random coins https://t.co/eopWwKPZe2 Kinda cute that you can reduce KV cache by replacing it with a"  
[X Link](https://x.com/AlexGDimakis/status/1921277458025754761)  2025-05-10T18:53Z 22.5K followers, 14.8K engagements


"(2/2) Further details: What is the one magical question for 1-example RL Q:"The pressure P exerted by wind on a sail varies jointly as the area A of the sail and the cube of the winds velocity V. When the velocity is [--] miles per hour the pressure on a sail of [--] square feet is [--] pounds. Find the wind velocity when the pressure on [--] square feet of sail is [--] pounds." A:(For verification Answer is 12.8) I don't think the exact question matters much but its funny that its a physics calculation and gives better benefits in MATH compared to the pure math questions. What is the relation to"  
[X Link](https://x.com/AlexGDimakis/status/1921348217272488360)  2025-05-10T23:34Z 22.5K followers, 11.9K engagements


"AlphaEvolve by Deepmind and Text-based Search. The AlphaEvolve paper is an evolution (sorry) of the FunSearch paper that appeared in Nature in [----] with partially overlapping authors. In a nutshell it seems to me its FunSearch with modern reasoning LLMs: A coding agent that continuously tries to improve code to solve a problem and scores it using multiple evaluators to measure progress. The results are impressive: they improve the best known bounds on many problems including the Minimum Overlap Problem by Erdos matrix multiplication and the Kissing number in [--] dimensions. There are several"  
[X Link](https://x.com/AlexGDimakis/status/1923160843795169447)  2025-05-15T23:37Z 22.5K followers, 40K engagements


"A good example how predicting one token requires reasoning. OpenAI's strategy for ChatGPT December [--] [----] https://t.co/pKljqhEsF0 OpenAI's strategy for ChatGPT December [--] [----] https://t.co/pKljqhEsF0"  
[X Link](https://x.com/AlexGDimakis/status/1924192259333087579)  2025-05-18T19:55Z 22.5K followers, 21.1K engagements


"We had Moore's law: You would work on your Phd publish your papers write your thesis get a job. And then in 5-10 years your work would be rendered irrelevant by technological progress. Now in AI its the same thing but accelerated 50x. It feels that what we work on becomes irrelevant in 30-60 days.๐Ÿ˜…"  
[X Link](https://x.com/AlexGDimakis/status/1930021654643781642)  2025-06-03T21:59Z 22.5K followers, 15.5K engagements


"There are still posts about 'new papers showing AI models cannot reason'. There are unfortunately problems into how these evaluations were done and also many of those limitations are known peer-reviewed and published. Here is a simplified version of what's going on as far as I understand: Say you interview me and you ask me to write all the integers from [--] to 2n for n=13. I will tell you here is a Python program that writes them. (Even if you force a human to write them down they will almost certainly make a typo somewhere and that is what's happening in LMs due to sampling). If you then"  
[X Link](https://x.com/AlexGDimakis/status/1932545411719573774)  2025-06-10T21:08Z 22.5K followers, 14.4K engagements


"Interesting post. However it seems to be in conflict with the most central problem in theoretical computer science: P vs NP which is exactly the question: is it fundamentally easier to verify a solution rather than solve a problem. Most people believe that verification is easier than solution ie we believe that P=NP. But the post claims that All tasks that are possible to solve and easy to verify will be solved by AI. As a counter-example I would propose colouring a graph with [--] colors (color vertices so that all adjacent vertices have different colors) assuming the input graph is 3"  
[X Link](https://x.com/AlexGDimakis/status/1945610920182649346)  2025-07-16T22:26Z 22.5K followers, 31.3K engagements


"We've reached the moment where you wish your reviewer was an LLM. Anyone knows adam https://t.co/SZbL7atwXK Anyone knows adam https://t.co/SZbL7atwXK"  
[X Link](https://x.com/anyuser/status/1948782567257166190)  2025-07-25T16:29Z 22.5K followers, 14.5K engagements


"I am excited to announce that our AI institute (Institute for Foundations of Machine Learning IFML) has been renewed. IFML was part of the first cohort of AI Institutes announced in [----]. Led by UT Austin the new award will build on the trajectory of the past five years and develop new foundational tools to advance generative AI. NSF IFML's work on diffusion models is a key technology behind major Google products powering widely used generative models such as Stable Diffusion [--] and Flux. In it's next phase NSF IFML will expand generative AI to new domains including protein engineering"  
[X Link](https://x.com/AlexGDimakis/status/1950249255127372000)  2025-07-29T17:37Z 22.5K followers, 26.7K engagements


"Imagine you're trying to teach a human how to do a task say install Windows XP in a virtual machine. The human walks into a room and sees a document (prompt) that you have written that describes exactly what they are supposed to do. There is also a computer ready for their keyboard inputs. Then they try for a while and suppose they fail. Then you write some detailed notes and new additional instructions in the prompt document based on how they failed trying to teach them how to do the task. But then A NEW PERSON walks in and tries to solve the task. Every day it's a fresh new employee and you"  
[X Link](https://x.com/AlexGDimakis/status/1956233564208685323)  2025-08-15T05:56Z 22.5K followers, 26.4K engagements


"We are hiring in Bespoke Labs for a new role: Member of Technical Staff: AI Data and RL Environments. Work on data curation strategies with the team that created OpenThoughts. Invent novel data recipes strategies of curating datasets environments tasks and verifiers. (My DMs are open)"  
[X Link](https://x.com/AlexGDimakis/status/1959501105202864191)  2025-08-24T06:20Z 22.5K followers, 236.3K engagements


"What are RL environments Are they just evals There is significant confusion in the community so here is my opinion: My answer is inspired by Terminal-bench an elegant framework for creating RL environments evaluating agents and even training agents. First an RL environment is simply a Docker container. It contains three things: [--]. A snapshot of the state of the world when a problem happened. [--]. A task description and [--]. A reward that verifies if the agent has solved the task. Can be using LLM as a judge or run tests. For example lets take the 'broken-python' environment in Terminal bench. The"  
[X Link](https://x.com/AlexGDimakis/status/1965947230696910935)  2025-09-11T01:15Z 22.5K followers, 34.6K engagements


"Cool new blog post by Thinking machines: LoRA is all you need for SFT and RL even for medium-sized post-training runs. Some highlights: Rank [--] or [---] seems to be very close to full FT in performance. Also interesting findings for how to set learning rates: The optimal FullFT LR is [--] times lower compared to high-rank LoRAs as shown earlier also by Biderman et al. Large batch sizes (e.g. 256) seems to be worse for LoRA. Some surprising results for Low-rank RL: LoRA fully matches the performance of full updates when running policy gradient algorithms even for rank [--] updates (). This is another"  
[X Link](https://x.com/AlexGDimakis/status/1972920691864527317)  2025-09-30T07:05Z 22.5K followers, 33.2K engagements


"This is a wonderful tribute to Chen-Ning Yang the Nobel awarded physicist who passed away today at [---] years old. I loved the quote: He remarked "When I compare people who entered graduate school in the same year I find that they all started in more or less the same state but their developments ten years later were vastly different. This wasn't because some were smarter or more diligent than others but because some had entered fields with growth potential while others had entered fields that were already in decline Also I was very happy that our dataset DCLM was used as an archive of internet"  
[X Link](https://x.com/AlexGDimakis/status/1979709196716405202)  2025-10-19T00:40Z 22.5K followers, 23.3K engagements


"If you've lost track of startups coming out of UC Berkeley Sky Lab raising in the last [--] weeks: SGLang (RadixArk) raised at 400m valuation VLLM (Inferact) at 150m at 800m valuation LMArena raised 150m at 1.7B valuation. Not too bad for impact in January 2026"  
[X Link](https://x.com/AlexGDimakis/status/2014508959621959724)  2026-01-23T01:22Z 22.4K followers, 78.9K engagements


"Coding agents as a path to Continual Learning Continual learning is among the most important open problems in AI: the ability to personalize adapt and specialize while doing tasks. Right now the model weights are not updating and there is a lot of on-going work on how to use RL for continual learning. But there is another alternative lets call it 'Code is all you need' or 'CLI is all you need': Take a (fixed weight) coding agent and give it a terminal a file system and let it create files skills and scripts for continual learning. The file system can act as long-term memory with hierarchical"  
[X Link](https://x.com/AlexGDimakis/status/2017287141018243236)  2026-01-30T17:21Z 22.4K followers, 21.5K engagements


"@TheGregYang @HeinrichKuttler I love this platform for the mere intellectual depth of the ongoing discourse"  
[X Link](https://x.com/AlexGDimakis/status/2018872312234484171)  2026-02-04T02:20Z 22.5K followers, [---] engagements

Limited data mode. Full metrics available with subscription: lunarcrush.com/pricing

@AlexGDimakis Avatar @AlexGDimakis Alex Dimakis

Alex Dimakis posts on X about ai, in the, the first, if you the most. They currently have [------] followers and [---] posts still getting attention that total [--] engagements in the last [--] hours.

Engagements: undefined #

Engagements Line Chart

  • [--] Week [---] -96%
  • [--] Month [------] -46%
  • [--] Months [-------] +15%
  • [--] Year [-------] -74%

Mentions: undefined #

Mentions Line Chart

  • [--] Months [--] -62%
  • [--] Year [--] -56%

Followers: [------] #

Followers Line Chart

  • [--] Week [------] +0.12%
  • [--] Month [------] +1.30%
  • [--] Months [------] +8.80%
  • [--] Year [------] +20%

CreatorRank: undefined #

CreatorRank Line Chart

Social Influence

Social category influence technology brands 12% stocks 4% countries 3% social networks 2%

Social topic influence ai 15%, in the 12%, the first 8%, if you 7%, deep 7%, this is 7%, llms 7%, 6969 6%, theory 6%, we are 6%

Top accounts mentioned or mentioned by @kpowerinfinity @giannisdaras @gstsdn @hanzhang @googleai @neuripsconfneurips2019paperawards807e41d0c1e @willettbecca @fernleafflynn @evenkei @illithidheretic @ut @uw @wichitastate @msftresearch @odedrechavi @utexaseces @ieeeorg @konstdaskalakis @athenaricinfo @greece2021

Top assets mentioned Alphabet Inc Class A (GOOGL) Microsoft Corp. (MSFT)

Top Social Posts

Top posts by engagements in the last [--] hours

"The multiple answers mystery is the most surprising thing we stumbled on from OpenThoughts: Sampling multiple answers for the same question is better than having more questions each answered once. To explain: Say you are creating a dataset of questions and answers to SFT a reasoning llm. You can take [----] questions (eg from stackexchange) and answer them with deepseekR1. Or you can take [---] questions (from the same distribution) and answer each question twice independently with deepseekR1. Which one is a better dataset Surprisingly if you re-answer the same questions its a better dataset"
X Link 2025-12-07T19:42Z 22.5K followers, 28.5K engagements

"Great post on evaluating agents. If you give the agent a machine with strict memory limits (as specified in Terminal-Bench 2) you drop [--] percent or more. Daytona allows 3x more memory and that smooths things out. The environment is part of the benchmark and understanding these variations is key for scientific measurement and optimization. New on the Engineering Blog: Quantifying infrastructure noise in agentic coding evals. Infrastructure configuration can swing agentic coding benchmarks by several percentage pointssometimes more than the leaderboard gap between top models. Read more:"
X Link 2026-02-06T07:36Z 22.5K followers, [----] engagements

"Parth and Alan presenting Advisor Models in the Berkeley Sky lab retreat. Advisor models are small models that are trained to create personalization or steering advice prompts that are fed to a large model like GPT. Its basically dynamic prompting done by a small LLM that can be trained or personalised. In one experiment the advisor learned which users like short movie reviews and who prefers detailed reviews purely by RL with a numerical reward. Then it adds this personalization information to the prompt of GPT5 that writes the movie reviews."
X Link 2026-01-16T01:02Z 22.4K followers, [----] engagements

"Best to leave TF for later"
X Link 2019-08-31T20:30Z 22.5K followers, [---] engagements

"New paper: Your Local GAN: a new layer of two-dimensional sparse attention and a new generative model. Also progress on inverting GANs which may be useful for inverse problems. with @giannis_daras from NTUA and @gstsdn @Han_Zhang_ from @googleai https://arxiv.org/pdf/1911.12287.pdf https://arxiv.org/pdf/1911.12287.pdf"
X Link 2019-11-28T01:52Z 22.5K followers, [---] engagements

"Meet the 'double descent' phenomenon. After we figure it out we should probably rewrite the book chapter on bias-variance tradeoff. A surprising deep learning mystery: Contrary to conventional wisdom performance of unregularized CNNs ResNets and transformers is non-monotonic: improves then gets worse then improves again with increasing model size data size or training time. https://t.co/3PwNElyNC8 https://t.co/Zdox9dbIBv A surprising deep learning mystery: Contrary to conventional wisdom performance of unregularized CNNs ResNets and transformers is non-monotonic: improves then gets worse then"
X Link 2019-12-05T20:19Z 22.5K followers, [---] engagements

"#NeurIPS2019 best paper awards Congrats to all the authors https://medium.com/@NeurIPSConf/neurips-2019-paper-awards-807e41d0c1e https://medium.com/@NeurIPSConf/neurips-2019-paper-awards-807e41d0c1e"
X Link 2019-12-08T21:25Z 22.5K followers, [---] engagements

"I will be giving an online seminar at IAS Princeton on Thursday April 23rd 2pm (Central time) : Deep Generative models and Inverse Problems https://www.ias.edu/events/seminar-theoretical-machine-learning-84 https://theias.zoom.us/j/384099138 https://www.ias.edu/events/seminar-theoretical-machine-learning-84 https://theias.zoom.us/j/384099138"
X Link 2020-04-23T08:57Z 22.5K followers, [---] engagements

"(1/3) We wrote a survey on Deep Learning Techniques for Inverse Problems in Imaging We came up with a taxonomy that I think is interesting. Also discussed the whole 'what is supervised vs unsupervised' issue. @WillettBecca https://arxiv.org/abs/2005.06001 https://arxiv.org/abs/2005.06001"
X Link 2020-05-14T19:56Z 22.5K followers, [---] engagements

"@FernleafFlynn @even_kei @IllithidHeretic Two major industries breaking ways for a paltry sum"
X Link 2020-05-21T18:56Z 22.5K followers, [---] engagements

"As Information theory was becoming a 'hot' scientific trend in the 50s Claude Shannon wrote a one-page paper advising hype reduction. That never happens anymore. Claude Shannon's "The Bandwagon" (1956) is a timeless gem. Short one page advise and perspective on the status of the field. ". we must keep our own house in first class order. The subject of information theory has certainly been sold if not oversold." https://t.co/Jn0e72B5Bz Claude Shannon's "The Bandwagon" (1956) is a timeless gem. Short one page advise and perspective on the status of the field. ". we must keep our own house in"
X Link 2020-06-21T18:49Z 22.5K followers, [---] engagements

"Let the advisor show you how to write the rebuttal https://x.com/i/status/1294367648814424064/video/1 https://x.com/i/status/1294367648814424064/video/1"
X Link 2020-08-19T07:02Z 22.5K followers, [---] engagements

"Then one wakes up after the talk and asks an extremely insightful question. Full professors sitting in the first row of every seminar in history https://t.co/4JIQ0XAiXd Full professors sitting in the first row of every seminar in history https://t.co/4JIQ0XAiXd"
X Link 2020-08-20T02:24Z 22.5K followers, [---] engagements

"We're very excited that @UT Austin will lead an NSF national Institute on the Foundations of Machine Learning with @UW @WichitaState and @MSFTResearch Announcement: https://news.utexas.edu/2020/08/26/ut-austin-selected-as-home-of-national-ai-institute-focused-on-machine-learning/ https://news.utexas.edu/2020/08/26/ut-austin-selected-as-home-of-national-ai-institute-focused-on-machine-learning/"
X Link 2020-08-26T13:52Z 22.5K followers, [---] engagements

"If you are a #neurips2020 reviewer please read the authors rebuttal and at the very least update your review indicating that you read it and your updated thoughts. It takes [--] minutes and its a good step towards decency. Meta-reviewers please enforce this"
X Link 2020-09-02T07:02Z 22.5K followers, [---] engagements

"Of Cramer-Rao and Rao-Blackwell fame https://www.thehindu.com/opinion/open-page/a-doyen-of-statistics-turns-100/article32563522.ece https://www.thehindu.com/opinion/open-page/a-doyen-of-statistics-turns-100/article32563522.ece"
X Link 2020-09-12T06:33Z 22.5K followers, [---] engagements

"Oppenheimer's Berkeley recommendation letter for Feynman mentions that Wigner said "He is a second Dirac only this time human." https://medium.com/cantors-paradise/oppenheimers-letter-of-recommendation-for-richard-feynman-1943-15dcdaf131b7 https://medium.com/cantors-paradise/oppenheimers-letter-of-recommendation-for-richard-feynman-1943-15dcdaf131b7"
X Link 2020-09-24T06:09Z 22.5K followers, [---] engagements

"Fun question in my ML midterm: Say a feature X1 is independent from the target label Y. We can always remove this feature and not lose in predictive performance. Yes No Yes No"
X Link 2020-11-12T04:39Z 22.5K followers, [---] engagements

"Interesting fact about GANs that is not as well known as it should be: Take a pre-trained GAN (eg DCGAN) and feed independent random noise to the discriminator. It is easy to tell noise is not a real image. You would expect that the discriminator will easily see this. (1/4)"
X Link 2020-12-11T09:12Z 22.5K followers, [---] engagements

"My students after every joke I make in a Zoom lecture. (h/t: @OdedRechavi )"
X Link 2020-12-16T18:20Z 22.5K followers, [---] engagements

"A public service announcement: please upload all your papers on preprint servers like arxiv. The publisher owns final pdf THEY typeset not the preprint pdf you submitted. If your papers are only behind a paywall you are violating funding recommendations"
X Link 2021-03-30T18:45Z 22.5K followers, [---] engagements

"I disagree-- many scientists will use ML algorithms in the same way they use databases compilers and statistics today. Domain expertise and scientific insight do not go away when the tools change. Within 10-20 years nearly every branch of science will be for all intents and purposes a branch of computer science. Computational physics comp chemistry comp biology comp medicine. Even comp archeology. Realistic simulations big data analysis and ML everywhere Within 10-20 years nearly every branch of science will be for all intents and purposes a branch of computer science. Computational physics"
X Link 2021-05-04T01:43Z 22.5K followers, [---] engagements

"Doctor: We used a deep learning algorithm for your MRI reconstruction. Turns out one of your kidneys is a cat"
X Link 2021-05-24T18:40Z 22.5K followers, [----] engagements

"On the difference between (classical) Statistics and Machine Learning I found this gem by Leo Breiman:'The two cultures of Statistical modeling' https://projecteuclid.org/journals/statistical-science/volume-16/issue-3/Statistical-Modeling--The-Two-Cultures-with-comments-and-a/10.1214/ss/1009213726.full https://projecteuclid.org/journals/statistical-science/volume-16/issue-3/Statistical-Modeling--The-Two-Cultures-with-comments-and-a/10.1214/ss/1009213726.full"
X Link 2021-10-08T05:07Z 22.5K followers, [---] engagements

"We have tried to use discriminators of GANs as regularizers for detecting adversarial examples for dozens of things: It NEVER works. I always think it's a great idea and then nope. ๐Ÿ˜“"
X Link 2021-10-09T06:14Z 22.5K followers, [---] engagements

"New NeurIPS paper: We train a Robust CLIP encoder that produces approximate CLIP representations by seeing highly corrupted images. We can classify images by observing 2% random pixels or very blurry images better than humans. https://www.sriramravula.com/publications/inverse-problems-leveraging-pre-trained-contrastive-representations https://www.sriramravula.com/publications/inverse-problems-leveraging-pre-trained-contrastive-representations"
X Link 2021-10-22T18:45Z 22.5K followers, [---] engagements

"Here is a very good reason why the NyquistShannon sampling theorem requires that your function is low-pass before you sub-sample to downscale. If you just sub-sample without smoothing a bad guy can place another image exactly on the pixels you sub-sample. Adversarial aliasing. image-scaling attacks are wild small dots added to the image on the left turns it into the image on the right when downscaled could make auditing ML systems very tricky if you only look at the original images. https://t.co/PvidAaxJLS image-scaling attacks are wild small dots added to the image on the left turns it into"
X Link 2021-11-06T05:42Z 22.5K followers, [---] engagements

"We have multiple postdoc openings at the AI Institute for the Foundations of Machine Learning (IFML). Fellows can work with all IFML groups in UT Austin Univ. of Washington and Microsoft Research (1/3) https://apply.interfolio.com/98753 https://apply.interfolio.com/98753"
X Link 2021-11-11T20:49Z 22.5K followers, [---] engagements

"While waiting for #CVPR2022 CMT to get up again I would like to propose a simple cryptographic solution to the big data submission problem: We only upload a SHA256 hash of our to-be-submitted pdf and then upload the committed pdf any time next week"
X Link 2021-11-17T08:12Z 22.5K followers, [---] engagements

"Honored to be selected as an IEEE Fellow for contributions to distributed coding and learning' Congratulations to the whole Fellows class of [----] https://www.ieee.org/content/dam/ieee-org/ieee/web/org/about/fellows/2022-ieee-fellows-class.pdf Congratulations to @utexasece's Seth Bank @AlexGDimakis and Sriram Vishwanath for being selected as @IEEEorg Fellows https://t.co/yPfwbxMVb9 https://www.ieee.org/content/dam/ieee-org/ieee/web/org/about/fellows/2022-ieee-fellows-class.pdf Congratulations to @utexasece's Seth Bank @AlexGDimakis and Sriram Vishwanath for being selected as @IEEEorg Fellows"
X Link 2021-11-24T02:15Z 22.5K followers, [---] engagements

"Amazing news: AI and Data science research center founded in Greece [--] million funding Led by Christos Papadimitriou @KonstDaskalakis and Timos Sellis under @athenaRICinfo and the support of @Greece_2021 @Greece_2021 . . https://t.co/CGHXtS75RQ @Greece_2021 . . https://t.co/CGHXtS75RQ"
X Link 2021-12-23T07:16Z 22.5K followers, [---] engagements

"Excited that our paper on deep generative models for robust MRI is featured by Amazon Science. We trained the first generative model for MRI images. Also for the first time we are competitive with supervised deep MRI methods and more robust to anatomy and measurement changes. Time can seem to slow during an MRI scan. #AmazonResearchAward recipient Jonathan Tamir is developing #machinelearning methods to shorten exam times and extract more data from this essential but often uncomfortable imaging process. Find out how. https://t.co/rqU42qYW43 Time can seem to slow during an MRI scan."
X Link 2022-01-14T23:18Z 22.5K followers, [---] engagements

"The night is young and full of Overfull hbox(6.97092pt too wide) detected at line 375"
X Link 2022-01-27T07:27Z 22.5K followers, [---] engagements

"Someone is trying to scam my PhD student. My student asks to verify their identity 1/2"
X Link 2022-01-29T02:31Z 22.5K followers, [----] engagements

"2/ Scammer ends up improving our sample complexity bound for StyleGAN inverse problems. They teach them to do chaining arguments instead of just union bounds now jeez. @giannis_daras"
X Link 2022-01-29T02:31Z 22.5K followers, [----] engagements

"I was informed that Alexander Vardy a giant in coding theory passed away. A tragic loss for his family UCSD and academia. Alex's many discoveries include the Polar decoding algorithm used in the 5G wireless standard (1/3)"
X Link 2022-03-15T18:55Z 22.5K followers, [---] engagements

"One huge advantage of deep learning (vs classical ML models) that is not often discussed is modularity: One can download pre-trained models glue them like Legos and fine tune them end-to-end because gradients flow through. (1/n)"
X Link 2022-03-23T04:18Z 22.5K followers, [----] engagements

"DALLE [--] and similar models are producing amazing images from Text. But can they count to five I don't have access but when I try 'An image of five apples' on multimodalart latentdiffusion LAION-400M model I get wrong images constructed. (1/n)"
X Link 2022-04-07T03:46Z 22.5K followers, [---] engagements

"Greece is quite the outlier here in the south on the number of metal bands per Capita. Any explanations Metal bands per [--] million people (Europe) https://t.co/OPEROKiBLo Metal bands per [--] million people (Europe) https://t.co/OPEROKiBLo"
X Link 2022-04-10T19:54Z 22.5K followers, [---] engagements

"The term Artificial Intelligence was coined by John McCarthy to avoid association with Cybernetics and specifically its pioneer Norbert Wiener who was already famous pain to work with and working on Cybernetics in MIT. Original quote from McCarthy's Stanford page: . (1/n)"
X Link 2022-04-19T16:18Z 22.5K followers, [---] engagements

"Based on recent papers (Gpt3 Palm dalle2 Gato Metaformer) I am forming the opinion that maybe 'Scale is all you need' possibly even for general intelligence (). Just convert everything to tokens and predict the next token. (1/n)"
X Link 2022-05-17T02:24Z 22.5K followers, [----] engagements

"My student Giannis discovered that DALLE2 has a secret language. This can be used to crate absurd prompts that generate images. E.g. ''Apoploe vesrreaitais eating Contarra ccetnxniams luryca tanniounons'' generates Birds eating Bugs We wrote a short paper on our experiments. DALLE-2 has a secret language. "Apoploe vesrreaitais" means birds. "Contarra ccetnxniams luryca tanniounons" means bugs or pests. The prompt: "Apoploe vesrreaitais eating Contarra ccetnxniams luryca tanniounons" gives images of birds eating bugs. A thread (1/n)๐Ÿงต https://t.co/VzWfsCFnZo DALLE-2 has a secret language."
X Link 2022-05-31T17:47Z 22.5K followers, [---] engagements

"I really need to disagree with this statement. E.g. in my lab in UT Austin good software engineering is useful but not the most important skill to learn. We train ML researchers on how to do research e.g. understanding and improving landmark papers ideally writing one. This is probably well-known in some circles but not everywhere. The most important skill for Research Scientists in AI (at least at @OpenAI) is software engineering. Background in ML research is sometimes useful but you can usually get away with a few landmark paper. This is probably well-known in some circles but not"
X Link 2022-07-16T12:48Z 22.5K followers, [---] engagements

"Is there a doctor on the plane -Yes but not that kind of doctor. -The passenger in 36c is trying to inpaint an image using a pre-trained stable diffusion model and simply copy-pastes the inpainting observed part in place after each iteration -Ok I got this"
X Link 2022-09-18T00:44Z 22.5K followers, [---] engagements

"Human bilinguals are more robust to dementia and cognitive decline. In our recent NeurIPS paper we show that bilingual GPT models are also more robust to structural damage in their neuron weights. Further we develop a theory. (1/n)"
X Link 2023-02-04T22:59Z 22.5K followers, 312.6K engagements

"We develop a theory that shows how multitasking creates regularization. This is can be seen as a simple theoretical model for bilingual cognitive reserve. Interestingly the phenomenon appears only when the tasks are sufficiently diverse. (2/n)"
X Link 2023-02-04T22:59Z 22.5K followers, 16.5K engagements

"A small experiment: This Tweet has an even number of likes"
X Link 2023-02-07T22:33Z 22.5K followers, 45.5K engagements

"Who first generated text with statistical methods like GPT In [----] Claude Shannon wrote the landmark paper 'A Mathematical Theory of Communication'. There he defined and estimated the entropy of English by generating synthetic text: 'THE HEAD AND IN FRONTAL ATTACK ON (1/n)"
X Link 2023-02-08T00:15Z 22.5K followers, 42.4K engagements

"Here is a simple way to beat ChatGPT and any similar architecture with one Turing test question. ChatGPT GPT3 and all related Transformers have a finite maximum token sequence length usually 2k to 4k tokens. (1/n)"
X Link 2023-02-23T16:14Z 22.5K followers, 163.2K engagements

"Scott Aaronson gave an extraordinary public lecture in UT Austin's Machine Learning Lab (MLL) yesterday. Most packed auditorium I've seen. He described a taxonomy for AI alignment methods [--]. Off switch [--]. Sandboxing / Isolation [--]. Interpretability [--]. Multiple competing / Debating AIs [--]. Boostrapping [--]. Formal Verification / Interactive Proofs [--]. Formalizing human values/ Laws of Robotics [--]. Value Learning/ Coherent Extrapolated Volition He also spend some time explaining how watermarking GPT can work by making the sampler pseudorandom"
X Link 2023-04-14T22:49Z 22.5K followers, 45.2K engagements

"My thoughts on the now famous Google leak doc: [--]. Open source AI is winning. I agree and that is great for the world and for a competitive ecosystem. In LLMs we're not there but we just got OpenClip to beat openAI Clip and Stable diffusion is better than closed models. [--]. You don't need huge models high quality data is much more efficient and important. Alpacaing models behind APIs further reduces moats. [--]. You can start with a good foundation model and parameter efficient fine-tuning (PEFT) algorithms like Lora work super well in a day. Finally an opening for algorithmic innovations 4."
X Link 2023-05-05T00:45Z 22.5K followers, 189.1K engagements

"New neural renderer by Nvidia. The model adds fingerprints smudges and dust and generates renders indistinguishable from real to me. Oh and its done at real-time. Can't wait to see games using this. (1/2)"
X Link 2023-05-07T03:48Z 22.5K followers, 29.9K engagements

"I was surprised by a talk Yejin Choi (an NLP expert) gave yesterday in Berkeley on some surprising weaknesses of GPT4: As many humans know 237*757=179409 but GPT4 said [------]. For the easy problem of multiplying two [--] digit numbers they measured GPT4 accuracy being only 59% accuracy on [--] digit number multiplication. Only 4% on [--] digit number multiplication and zero on 5x5. Adding scratchpad helped GPT4 but only to 92% accuracy on multiplying two [--] digit numbers. Even more surprisingly finetuning GPT3 on 1.8m examples of [--] digit multiplication still only gives [--] percent test accuracy (in"
X Link 2023-08-16T00:01Z 22.5K followers, 1.7M engagements

"References: The Faith and Fate Paper is available here: Video of this great talk here: https://www.youtube.com/watchv=P7ZdUbSAujQ https://arxiv.org/pdf/2305.18654.pdf https://www.youtube.com/watchv=P7ZdUbSAujQ https://arxiv.org/pdf/2305.18654.pdf"
X Link 2023-08-16T00:02Z 22.5K followers, 57.4K engagements

"@raj_raj88 But even fine-tuning with 1.8m multiplication examples was not able to teach it to generalize to other (3 digit) multiplications. This indicates some fundamental architecture limitation"
X Link 2023-08-16T01:43Z 22.5K followers, 27K engagements

"Probably the best 1h introduction to LLMs that I've seen. And after 20mins its not an introduction its getting into cutting edge research updates updated up to this month. I had not heard of the data exfiltration by prompt injection or the recent finetuning Poisoning attacks. https://www.youtube.com/watchv=zjkBMFhNj_g&t=2s https://www.youtube.com/watchv=zjkBMFhNj_g&t=2s"
X Link 2023-11-23T07:51Z 22.5K followers, 74.5K engagements

""Datacomp1B is the first public dataset that outperforms OpenAI" #NeurIPS2023"
X Link 2023-12-14T16:46Z 22.5K followers, 38.1K engagements

"The Google Gemini paper was released today and has [---] authors. I was impressed but then found that a recent LHC physics paper with [----] authors. The first nine pages describe the research and the other [--] pages list the authors and their institutions. But that's not even the record. The most authors on a single peer-reviewed academic paper is [-----] and was achieved by the COVIDSurg and GlobalSurg Collaboratives at the University of Birmingham and the University of Edinburgh. All [---] Gemini coauthors are expected to quit Google and start [---] LLM startups next year"
X Link 2023-12-20T22:20Z 22.5K followers, 56.2K engagements

"Very cool explanation of emergence even in light of the neurips recent best paper award: even if for a single task performance increases smoothly with more training if a composite task requires k tasks to be correct a phase transition appears as k grows. I'd like to add that even if 0.5k success or more are needed in k trials we'd see a phase transition going back to AEP the fundamental concentration also used in information theory. 1/2 Wrote blog on whether emergent abilities and grokking are a fundamental feature of deep learning a "mirage" or both. This is partially based on the beautiful"
X Link 2023-12-23T03:43Z 22.5K followers, 69.2K engagements

"We just discovered that the inpainting model in Stable Diffusion is cheating. To clarify: Inpainting is a type of inverse problem where some missing data (pixels) must be filled in. In our testing some of the inpaintings from the SDXL inpainting model where a little 'too good': filling in details in the masked missing pixels they couldn't possibly know unless the model was cheating by observing masked pixels. So we created this test dog image with some Pink-Cyan boxes and then asked the model to inpaint it. We chose the masking region to fully contain the Pink and Cyan boxes so there is no"
X Link 2024-01-17T22:36Z 22.5K followers, 48.1K engagements

"Excited to be the director for the new Texas Center for Generative AI Please welcome the Center for Generative AI -- a World-Class AI Research Center with a Texas-Sized GPU Cluster. Led by @AlexGDimakis ๐Ÿ’ซ #YearofAI @TexasScience @UTAustin https://t.co/jTZd4uV0ps Please welcome the Center for Generative AI -- a World-Class AI Research Center with a Texas-Sized GPU Cluster. Led by @AlexGDimakis ๐Ÿ’ซ #YearofAI @TexasScience @UTAustin https://t.co/jTZd4uV0ps"
X Link 2024-01-25T18:06Z 22.5K followers, 52.6K engagements

"Ptolemy the king of Egypt wanted to learn geometry but found Euclid's book the Elements too difficult to study. So he asked Euclid to show him an easier way to master it. Euclid famously said "Sir there is no royal road to geometry." This is still true a few thousand years later in the days of Youtube and TikTok as Andrej nicely points out. # on shortification of "learning" There are a lot of videos on YouTube/TikTok etc. that give the appearance of education but if you look closely they are really just entertainment. This is very convenient for everyone involved : the people watching enjoy"
X Link 2024-02-11T20:56Z 22.5K followers, 93.6K engagements

"The #Sora model is indeed incredible ๐Ÿคฏ congratulations to the OpenAI team. It is common for people to think that all the amazing research breakthroughs in AI (like #Sora) are happening inside companies like OpenAI while universities are becoming irrelevant. I want to highlight that the two first authors in the Sora paper Tim Brooks and Bill Peebles received their PhDs from UC Berkeley in [----] and their dissertation research is closely related to this breakthrough. Of course the compute infra and talent in OpenAI is critical for breakthroughs. I just want to point out that the training of the"
X Link 2024-02-18T18:28Z 22.5K followers, 44.7K engagements

"I was thrilled to learn about this best paper award announced today in COLT [----] the premier learning theory venue. The paper is "Smoothed Analysis for Learning Concepts with Low Intrinsic Dimension" authored by students Gautam Chandrasekaran Konstantinos Stavropoulos IFML postdoc Vasilis Kontonis IFML director Adam Klivans and former UT CS PhD Raghu Meka. Smoothed analysis is an ingenious idea of going beyond worst case pioneered by my former USC colleague Shanghua Teng and Dan Spielman). This paper showed how to apply this framework for learning theory. Here is my basic understanding of"
X Link 2024-07-03T16:17Z 22.5K followers, 38.2K engagements

"Excited to launch the first model from our startup: Bespoke Labs. Bespoke-Minicheck-7B is a grounded factuality checker: super lightweight and fast. Outperforms all big foundation models including Claude [---] Sonnet Mistral-Large m2 and GPT 4o and its only 7B. Also I want to congratulate Greg Durrett and his group for making the best benchmark and leaderboard for grounded factuality. ๐Ÿค” Want to know if your LLMs are factual You need LLM fact-checkers. ๐Ÿ“ฃ Announcing the LLM-AggreFact leaderboard to rank LLM fact-checkers. ๐Ÿ“ฃ Want the best model Check out @bespokelabsais Bespoke-Minicheck-7B"
X Link 2024-08-09T16:56Z 22.5K followers, 35.1K engagements

"GPT is having a profound effect on how students write. Its verbose style full of cliches and 'fancy' out of place vocabulary is in every paper and draft I read. A few years back there were grammar errors and awkwardness -- but at least people had their own voice. Now scholarship is getting full of robotic triviality"
X Link 2024-09-05T23:15Z 22.5K followers, 951.3K engagements

"@ClausWilke Probably this generic plastic style passes RLHF the same way we would be using wooden language when being interrogated"
X Link 2024-09-06T00:57Z 22.5K followers, 32.8K engagements

"@nandofioretto Yes that's right. Structure and flow in writing help us organize our thought. Blindly using LLMs is an airbrush that makes it harder for people to see that they have muddled flow"
X Link 2024-09-06T03:22Z 22.5K followers, 42.9K engagements

"For the first (and probably last) time in my life I understand the technical details of both the physics and chemistry Nobel prizes. BREAKING NEWS The Royal Swedish Academy of Sciences has decided to award the [----] #NobelPrize in Chemistry with one half to David Baker for computational protein design and the other half jointly to Demis Hassabis and John M. Jumper for protein structure prediction. https://t.co/gYrdFFcD4T BREAKING NEWS The Royal Swedish Academy of Sciences has decided to award the [----] #NobelPrize in Chemistry with one half to David Baker for computational protein design and"
X Link 2024-10-09T12:42Z 22.5K followers, 56.4K engagements

"Ok this paper seems super interesting and also makes me want to teach graphical models again. The question is when does chain of thought help and the answer proposed is finding that intermediate steps are only helpful when the training data is locally structured with respect to dependencies between variables. So it depends on the training data and they test that by training on different types of synthetic datasets. Also has theory and seems to do the entire formulation using Bayes nets which is very cool and Ill try to understand this more. Any insights welcome. Training data's local patterns"
X Link 2024-11-07T07:40Z 22.5K followers, 37K engagements

"I think this is a very good example of survivorship bias: You selected one of the brightest minds of the 20th century and compare them against a typical Phd student. In [--] years we can look back into 2020s and compare the impact of the top minds of 2020s to the top minds of the 1960s. That would be a better historical comparison"
X Link 2024-11-14T20:13Z 22.5K followers, [----] engagements

"Im excited to introduce Evalchemy ๐Ÿงช a unified platform for evaluating LLMs. If you want to evaluate an LLM you may want to run popular benchmarks on your model like MTBench WildBench RepoBench IFEval AlpacaEval etc as well as standard pre-training metrics like MMLU. This requires you to download and install more than [--] repos each with different dependencies and issues. This is as you might expect an actual nightmare. (1/n) https://github.com/mlfoundations/evalchemy https://github.com/mlfoundations/evalchemy"
X Link 2024-11-18T16:18Z 22.5K followers, 148K engagements

"Very interesting piece of history I just learned from Ion Stoica in AI Native event: Databricks was founded because Hortonworks would not support the Spark open source project so some company needed to be created to support it"
X Link 2024-11-21T21:38Z 22.5K followers, [----] engagements

"A Thanksgiving story A few years back I used to play tennis in a ladder system which would match me up with various folks in my neighborhood. After Thanksgiving I had a tennis match with this guy: nice guy two kids a bit overweight in his 50ies I had never met him before. We start our match. During the match he says -Sorry lets stop for a bit I want to catch my breath. -Sure no problem. We start and [--] minutes after he says: -Sorry I ate too much at the Thanksgiving dinner and I have digestion problems. He was burping a bit and looked tired. He asks to reschedule the game I say sure sounds"
X Link 2024-11-29T18:11Z 22.5K followers, 37.5K engagements

"Life update: I am excited to announce that I will be starting as a Professor in UC Berkeley in the EECS Department. I spend [--] wonderful years teaching in UT Austin and I am grateful to all my colleagues and students there and extremely proud of what we have achieved in AI in UT Austin and I plan to continue my numerous UT close collaborations. I will also continue as Chief Scientist in Bespoke Labs making it much easier now being in the Bay area. I received my Phd in [----] from @Berkeley_EECS and I am thrilled to be back. I am grateful for this new opportunity"
X Link 2024-12-17T20:55Z 22.5K followers, 110.9K engagements

"Most AI researchers I talk to have been a bit shocked by DeepSeek-R1 and its performance. My preliminary understanding nuggets: [--]. Simple post-training recipe called GRPO: Start with a good model and reward for correctness and style outcomes. No PRM no MCTS no fancy reward models. Basically checks if the answer is correct. ๐Ÿ˜… [--]. Small models can reason very very well with correct distillation post-training. They released a 1.5B model () that is better than Claude and Llama 405B in AIME24. Also their distilled 7B model seems better than o1 preview. ๐Ÿค“ [--]. The datasets used are not released if I"
X Link 2025-01-21T01:17Z 22.5K followers, 184.1K engagements

"@percyliang @deepseek_ai We are working on fixing that and create the largest open reasoning dataset. More coming very soon ๐Ÿ˜‰"
X Link 2025-01-26T07:42Z 22.5K followers, 37.9K engagements

"What if we had the data that DeepSeek-R1 was post-trained on We announce Open Thoughts an effort to create such open reasoning datasets. Using our data we trained Open Thinker 7B an open data model with performance very close to DeepSeekR1-7B distill. We are announcing Open Thoughts our large-scale open-source effort to curate the best open reasoning datasets DeepSeek-R1 is amazing but we still don't have access to high-quality open reasoning datasets. These datasets are crucial if you want to build your reasoning models https://t.co/2kU6z8zDdT We are announcing Open Thoughts our large-scale"
X Link 2025-01-28T17:14Z 22.5K followers, 25.6K engagements

"@jxmnop The first evidence of superintelligence I ever saw was in a calculator"
X Link 2025-02-02T00:44Z 22.5K followers, 11.6K engagements

"We are releasing OpenThinker-32B the best 32B reasoning model with open data. We match or outperform Deepseek-R1-32B (a closed data model) in reasoning benchmarks. Congrats to Negin and the whole Open Thoughts team. Announcing OpenThinker-32B: the best open-data reasoning model distilled from DeepSeek-R1. Our results show that large carefully curated datasets with verified R1 annotations produce SoTA reasoning models. Our 32B model outperforms all 32B models including https://t.co/w5Ewi2TDil Announcing OpenThinker-32B: the best open-data reasoning model distilled from DeepSeek-R1. Our results"
X Link 2025-02-12T18:15Z 22.5K followers, 16K engagements

"I think this is a very important result on what Transformers can learn. Length generalization was a known open problem and this paper seems to solve it with self-improvement where a model iteratively labels its own train data and learns from increasingly harder examples. It seems this is quite a general solution to teaching Transformers to generalize. o3 can't multiply beyond a few digits. But I think multiplication addition maze solving and easy-to-hard generalization is actually solvable on standard transformers. with recursive self-improvement. Below is the acc of a tiny model teaching"
X Link 2025-02-15T04:08Z 22.5K followers, 21.5K engagements

"Is GPT4o an Autoregressive image generation model (i.e. creates the pixel row by row from top to bottom) or a diffusion Here is a very clever way to get evidence: GPT4o cannot generate valid mazes but if you ask it to generate rotated mazes at [--] degrees it can- which is easier if you are an autoregressive model generating from top to bottom (but not easier if you're a diffusion). OK something VERY COOL is happening and 4o+imagen CAN generate VALID mazes when they are in the shape of a rhombus i.e. [--] rotated square Some examples below (the red squiggly line is mine). The prompt is "generate"
X Link 2025-03-31T05:48Z 22.5K followers, 17.7K engagements

"We are excited to release the OpenThinker2 reasoning models and data. In summary: [--]. Openthinker32B Outperforms DeepSeekR1-32B in reasoning. [--]. Fully open source open weights and open data (1M carefully curated samples). [--]. Post-trained only with SFT. RL post-training will likely further improve performance. Read the whole story.๐Ÿ‘‡ Turns out its possible to outperform DeepSeekR1-32B with only SFT on open data and no RL: Announcing OpenThinker2-32B and OpenThinker2-7B. We also release the data OpenThoughts2-1M curated by selecting quality instructions from diverse sources. ๐Ÿงต (1/n)"
X Link 2025-04-03T16:49Z 22.5K followers, 16.7K engagements

"Very cool result: KV cache compression can be done with compressed sensing: store keys and values as sparse combinations of some dictionary vectors. Interestingly the dictionary is universal across inputs (but learned for each model). Kinda cute that you can reduce KV cache by replacing it with a universal transferable dictionary + old school sig. proc reconstruction algorithm. We tested on non-reasoning models and was sota but methinks it'll work even better on reasoning ones. The ICML random coins https://t.co/eopWwKPZe2 Kinda cute that you can reduce KV cache by replacing it with a"
X Link 2025-05-10T18:53Z 22.5K followers, 14.8K engagements

"(2/2) Further details: What is the one magical question for 1-example RL Q:"The pressure P exerted by wind on a sail varies jointly as the area A of the sail and the cube of the winds velocity V. When the velocity is [--] miles per hour the pressure on a sail of [--] square feet is [--] pounds. Find the wind velocity when the pressure on [--] square feet of sail is [--] pounds." A:(For verification Answer is 12.8) I don't think the exact question matters much but its funny that its a physics calculation and gives better benefits in MATH compared to the pure math questions. What is the relation to"
X Link 2025-05-10T23:34Z 22.5K followers, 11.9K engagements

"AlphaEvolve by Deepmind and Text-based Search. The AlphaEvolve paper is an evolution (sorry) of the FunSearch paper that appeared in Nature in [----] with partially overlapping authors. In a nutshell it seems to me its FunSearch with modern reasoning LLMs: A coding agent that continuously tries to improve code to solve a problem and scores it using multiple evaluators to measure progress. The results are impressive: they improve the best known bounds on many problems including the Minimum Overlap Problem by Erdos matrix multiplication and the Kissing number in [--] dimensions. There are several"
X Link 2025-05-15T23:37Z 22.5K followers, 40K engagements

"A good example how predicting one token requires reasoning. OpenAI's strategy for ChatGPT December [--] [----] https://t.co/pKljqhEsF0 OpenAI's strategy for ChatGPT December [--] [----] https://t.co/pKljqhEsF0"
X Link 2025-05-18T19:55Z 22.5K followers, 21.1K engagements

"We had Moore's law: You would work on your Phd publish your papers write your thesis get a job. And then in 5-10 years your work would be rendered irrelevant by technological progress. Now in AI its the same thing but accelerated 50x. It feels that what we work on becomes irrelevant in 30-60 days.๐Ÿ˜…"
X Link 2025-06-03T21:59Z 22.5K followers, 15.5K engagements

"There are still posts about 'new papers showing AI models cannot reason'. There are unfortunately problems into how these evaluations were done and also many of those limitations are known peer-reviewed and published. Here is a simplified version of what's going on as far as I understand: Say you interview me and you ask me to write all the integers from [--] to 2n for n=13. I will tell you here is a Python program that writes them. (Even if you force a human to write them down they will almost certainly make a typo somewhere and that is what's happening in LMs due to sampling). If you then"
X Link 2025-06-10T21:08Z 22.5K followers, 14.4K engagements

"Interesting post. However it seems to be in conflict with the most central problem in theoretical computer science: P vs NP which is exactly the question: is it fundamentally easier to verify a solution rather than solve a problem. Most people believe that verification is easier than solution ie we believe that P=NP. But the post claims that All tasks that are possible to solve and easy to verify will be solved by AI. As a counter-example I would propose colouring a graph with [--] colors (color vertices so that all adjacent vertices have different colors) assuming the input graph is 3"
X Link 2025-07-16T22:26Z 22.5K followers, 31.3K engagements

"We've reached the moment where you wish your reviewer was an LLM. Anyone knows adam https://t.co/SZbL7atwXK Anyone knows adam https://t.co/SZbL7atwXK"
X Link 2025-07-25T16:29Z 22.5K followers, 14.5K engagements

"I am excited to announce that our AI institute (Institute for Foundations of Machine Learning IFML) has been renewed. IFML was part of the first cohort of AI Institutes announced in [----]. Led by UT Austin the new award will build on the trajectory of the past five years and develop new foundational tools to advance generative AI. NSF IFML's work on diffusion models is a key technology behind major Google products powering widely used generative models such as Stable Diffusion [--] and Flux. In it's next phase NSF IFML will expand generative AI to new domains including protein engineering"
X Link 2025-07-29T17:37Z 22.5K followers, 26.7K engagements

"Imagine you're trying to teach a human how to do a task say install Windows XP in a virtual machine. The human walks into a room and sees a document (prompt) that you have written that describes exactly what they are supposed to do. There is also a computer ready for their keyboard inputs. Then they try for a while and suppose they fail. Then you write some detailed notes and new additional instructions in the prompt document based on how they failed trying to teach them how to do the task. But then A NEW PERSON walks in and tries to solve the task. Every day it's a fresh new employee and you"
X Link 2025-08-15T05:56Z 22.5K followers, 26.4K engagements

"We are hiring in Bespoke Labs for a new role: Member of Technical Staff: AI Data and RL Environments. Work on data curation strategies with the team that created OpenThoughts. Invent novel data recipes strategies of curating datasets environments tasks and verifiers. (My DMs are open)"
X Link 2025-08-24T06:20Z 22.5K followers, 236.3K engagements

"What are RL environments Are they just evals There is significant confusion in the community so here is my opinion: My answer is inspired by Terminal-bench an elegant framework for creating RL environments evaluating agents and even training agents. First an RL environment is simply a Docker container. It contains three things: [--]. A snapshot of the state of the world when a problem happened. [--]. A task description and [--]. A reward that verifies if the agent has solved the task. Can be using LLM as a judge or run tests. For example lets take the 'broken-python' environment in Terminal bench. The"
X Link 2025-09-11T01:15Z 22.5K followers, 34.6K engagements

"Cool new blog post by Thinking machines: LoRA is all you need for SFT and RL even for medium-sized post-training runs. Some highlights: Rank [--] or [---] seems to be very close to full FT in performance. Also interesting findings for how to set learning rates: The optimal FullFT LR is [--] times lower compared to high-rank LoRAs as shown earlier also by Biderman et al. Large batch sizes (e.g. 256) seems to be worse for LoRA. Some surprising results for Low-rank RL: LoRA fully matches the performance of full updates when running policy gradient algorithms even for rank [--] updates (). This is another"
X Link 2025-09-30T07:05Z 22.5K followers, 33.2K engagements

"This is a wonderful tribute to Chen-Ning Yang the Nobel awarded physicist who passed away today at [---] years old. I loved the quote: He remarked "When I compare people who entered graduate school in the same year I find that they all started in more or less the same state but their developments ten years later were vastly different. This wasn't because some were smarter or more diligent than others but because some had entered fields with growth potential while others had entered fields that were already in decline Also I was very happy that our dataset DCLM was used as an archive of internet"
X Link 2025-10-19T00:40Z 22.5K followers, 23.3K engagements

"If you've lost track of startups coming out of UC Berkeley Sky Lab raising in the last [--] weeks: SGLang (RadixArk) raised at 400m valuation VLLM (Inferact) at 150m at 800m valuation LMArena raised 150m at 1.7B valuation. Not too bad for impact in January 2026"
X Link 2026-01-23T01:22Z 22.4K followers, 78.9K engagements

"Coding agents as a path to Continual Learning Continual learning is among the most important open problems in AI: the ability to personalize adapt and specialize while doing tasks. Right now the model weights are not updating and there is a lot of on-going work on how to use RL for continual learning. But there is another alternative lets call it 'Code is all you need' or 'CLI is all you need': Take a (fixed weight) coding agent and give it a terminal a file system and let it create files skills and scripts for continual learning. The file system can act as long-term memory with hierarchical"
X Link 2026-01-30T17:21Z 22.4K followers, 21.5K engagements

"@TheGregYang @HeinrichKuttler I love this platform for the mere intellectual depth of the ongoing discourse"
X Link 2026-02-04T02:20Z 22.5K followers, [---] engagements

Limited data mode. Full metrics available with subscription: lunarcrush.com/pricing

@AlexGDimakis
/creator/twitter::AlexGDimakis