Josh, I followed up on my comment to your twitter post for this video. I got ollama containerized and have added the breezy golang UI to it. I still need to setup docker GPU passthru in order to speed it up. The tokens come real slow right now. LOL My next step is to train it with my own data. This video is good stuff bro, thanks.
While you can technically run any model on any hardware, the real limitation comes from the hardware's capabilities, especially for complex models where powerful hardware is essential for practical execution times. Always match your model's requirements with your hardware's capabilities for the best experience.
I don't think running LLMs is a good gauge for how an operating system complexness is. There's one called GPT4all and it was just a one click install, one click model download. I'm pretty happy on Linux Mint, it's as complex as you want it to be!