

The AI features are actually pretty cool!
Using a local AI model (running on your own GPU), you can:
- Drag files on to the “AI bar” to bring them into the context of a prompt.
- Tell it to perform actions like, “merge these two PDF files.”
- Use it to search your filesystem, “find pictures of spiders.”
- Ask it to find system settings or tell it to change them, “turn on dark mode.”
It supports voice and probably visual (e.g. with a webcam) I think.
Best of all: It doesn’t send any of that to some data center in the cloud! I mean, you can configure it to do that but you can just as easily use say, qwen3.5.
Note: It’s not realistic to expect to be able to use local models if you have less than 16GB of VRAM (in your GPU). I mean, some 8-billion parameter model will work with say, 8GB but you’re not going to be satisfied with the results most of the time 🤷





















“Here’s a picture of me and my child, next to $12,000 worth of drinks and a fruity thing made by my personal chef that gets paid over $200k/year out of my personal money vault.”