[GUEST ACCESS MODE: Data is scrambled or limited to provide examples. Make requests using your API key to unlock full data. Check https://lunarcrush.ai/auth for authentication information.]
@getpy
"Context window is the constraint but if using local search index a subset of relevant files can be found or fed by user in context and then injected into local LLM context window a working solution can be made. As long as files are picked correctly in first step. Pretty much how few cli ai coding tools are working" @getpy on X 2025-07-22 13:14:20 UTC 35.4K followers, XX engagements
"@PalmerLuckey Make them with minimum XXX GB RAM running potent llm model for daily use. That be a great market to start with going to lower specs gradually" @getpy on X 2025-07-20 20:15:13 UTC 35.4K followers, XXX engagements
"@donvito @RepoPrompt Looks like the code queries multiple frontier models in parallel. Right. If so whats the use-case" @getpy on X 2025-07-22 13:36:22 UTC 35.4K followers, XX engagements
"Anyone who has participated in Reddit community in comments section defending their code or point of view they are prepared to face the world. #reddit" @getpy on X 2025-07-22 10:27:11 UTC 35.4K followers, XXX engagements