I’ve had a pretty poor experience with it myself, so I wanna see what the Linux community thinks about this.
I’ll emphasize the point that this goes for any kind of machine learning model that can benefit from CUDA, which means a large amount of gaming computers already meet the prerequisites for this. Installation is trivial (but requires some knowledge), and I hope to see more ML applications for hobbyists in the near future. Image generation and locally hosted GPT models come to mind.