[GUEST ACCESS MODE: Data is scrambled or limited to provide examples. Make requests using your API key to unlock full data. Check https://lunarcrush.ai/auth for authentication information.]

@rasbt Avatar @rasbt Sebastian Raschka

Sebastian Raschka posts on X about the first, released, reinforcement learning, instead of the most. They currently have XXXXXXX followers and XX posts still getting attention that total XXXXXXX engagements in the last XX hours.

Engagements: XXXXXXX #

Engagements Line Chart

Mentions: X #

Mentions Line Chart

Followers: XXXXXXX #

Followers Line Chart

CreatorRank: XXXXXXX #

CreatorRank Line Chart

Social Influence

Social category influence fashion brands XXXX% stocks XXXX% technology brands XXXX%

Social topic influence the first 10.53%, released 5.26%, reinforcement learning 5.26%, instead of 5.26%, hm 5.26%, solve 5.26%, hold on 5.26%, if you 5.26%, ai 5.26%, alpaca XXXX%

Top accounts mentioned or mentioned by @almmaasoglu @tugot17 @soulsurfer78 @thomasip @technmak @soul_surfer78 @karpathy @ilyasut @iamkyros69 @demishassabis @geminiapp @jvivasofficial @xeophon @lightningai @vir_bhadeshiya @honlv2024 @_thomasip @pavelsnajdr @argo_92 @ayushrathie

Top assets mentioned IBM (IBM)

Top Social Posts

Top posts by engagements in the last XX hours

"DeepSeek finally released a new model and paper. And because this DeepSeek-OCR release is a bit different from what everyone expected and DeepSeek releases are generally a big deal I wanted to do a brief explainer of what it is all about. In short they explore how vision encoders can improve the efficiency of LLMs in processing and compressing textual information. And the takeaway is that rendering text as images and feeding that to the model results in more efficient compression than working with text directly. My first intuition was that this sounds very inefficient and shouldn't work as"
X Link 2025-10-21T14:27Z 373K followers, 159.9K engagements

"Inference-scaling lets us trade extra compute for better modeling accuracy. Next to reinforcement learning it has become one of the most important concepts in today's LLMs so the book will cover it in two chapters instead of just one. I just finished the first one. It is a 35-page introduction to inference-time scaling through self-consistency sampling. This chapter was a lot of fun to write because it takes the base model on MATH-500 all the way from XXXX% percent to XXXX% accuracy. Seeing that jump without additional training is incredibly satisfying. Submitted the chapter yesterday and it"
X Link 2025-11-15T14:44Z 373K followers, 179.2K engagements

"This interesting week started with DeepSeek V3.2 I just wrote up a technical tour of the predecessors and components that led up to this: 🔗 - Multi-Head Latent Attention - RLVR - Sparse Attention - Self-Verification - GRPO Updates"
X Link 2025-12-03T14:49Z 373K followers, 89.2K engagements

"Hm I dont see the conflict. There are two things going on the reasoning behavior (the intermediate step-by-step explanations in LLMs to solve more complex problems) and then reasoning models which are LLMs specifically trained to emit such answers with intermediate steps. Does that help clarify"
X Link 2025-12-11T14:33Z 372.7K followers, XX engagements

"Hold on a sec Mistral X Large uses the DeepSeek V3 architecture including MLA Just went through the config files; the only difference I could see is that Mistral X Large used 2x fewer experts but made each expert 2x large"
X Link 2025-12-12T19:11Z 373K followers, 235.9K engagements

"If you are looking for something to read this upcoming weekend chapter X on inference-time scaling is available now 🔗"
X Link 2025-11-20T14:42Z 373K followers, 117.3K engagements

"My biennial update to the "Hello World"s of ML & AI: 2013: RandomForestClassifier on Iris 2015: XGBoost on Titanic 2017: MLPs on MNIST 2019: AlexNet on CIFAR-10 2021: DistilBERT on IMDb movie reviews 2023: Llama X with LoRA on Alpaca 50k 2025: Qwen3 with RLVR on MATH-500"
X Link 2025-12-08T18:56Z 373K followers, 129.7K engagements

"@_thomasip Yes but the larger labs usually always had some distinct difference or tweak not a straight up reuse"
X Link 2025-12-12T20:19Z 373K followers, 8143 engagements

"@_thomasip Ok fair but now Qwen3 IBM Granite XXX Olmo X Phi-4 all have unique tweaks"
X Link 2025-12-12T20:52Z 373K followers, 1109 engagements

"Just updated the Big LLM Architecture Comparison article. .it grew quite a bit since the initial version in July 2025 more than doubled"
X Link 2025-12-13T14:21Z 373K followers, 94.9K engagements