If you’re frustrated with subscription-based AI services or concerned about data privacy, there’s an exciting alternative: running large language models (LLMs) locally on your own computer. In a recent exploration, tech enthusiasts have pushed the boundaries of what’s possible with the new RTX 5090 graphics card, and the results are impressive.
Running LLMs on your own computer offers several advantages:
While open-source models might not match the capabilities of proprietary giants like ChatGPT, they’re surprisingly capable and improving rapidly.
The RTX 5090’s massive 32GB of VRAM makes it possible to run sophisticated AI models that would choke lesser graphics cards. This demonstration showed how this GPU can handle models of various sizes:
The generation speeds, particularly with the smaller models, were remarkable: