[GUEST ACCESS MODE: Data is scrambled or limited to provide examples. Make requests using your API key to unlock full data. Check https://lunarcrush.ai/auth for authentication information.]
Andrew @Hilton0A on x XXX followers
Created: 2025-07-18 22:30:45 UTC
Why doesn't Nvidia just ship an easy to access, configurable, local LLM with their graphics card software? Every dweeb w >8gb of vram could be making use of it for so many tasks. Imagine just hooking into your local AI process for inference without any setup
XXX engagements
Related Topics graphics card inference coins ai llm $nvda stocks technology