I was looking back at some old lemmee posts and came across GPT4All. Didn’t get much sleep last night as it’s awesome, even on my old (10yo) laptop with a Compute 5.0 NVidia card.
Still, I’m after more, I’d like to be able to get image creation and view it in the conversation, if it generates python code, to be able to run it (I’m using Debian, and have a default python env set up). Local file analysis also useful. CUDA Compute 5.0 / vulkan compatibility needed too.
Is there anything that can tick the boxes? Even if I have to scoot across models for some of the features?
The main limitation is the VRAM, but I doubt any model is going to be particularly fast.
I think
phi3:mini
on ollama might be an okish fit for python, since it’s a small model, but was trained on python codebases.I’m getting very-near real-time on my old laptop. Maybe a delay of 1-2s whilst it creates the response