But… bash snippet extensions already exist. The only difference is maybe it doesn’t auto name your variables for you. I’d take that over non-deterministic LLM outputs.
I have no idea what the hell a bash snippet extension is, but I do know what a local llama.cpp instance running a small model to tell me bash commands on the fly is.
I use it to make .desktop files, too. Isn’t that so lazy?
But… bash snippet extensions already exist. The only difference is maybe it doesn’t auto name your variables for you. I’d take that over non-deterministic LLM outputs.
I have no idea what the hell a bash snippet extension is, but I do know what a local llama.cpp instance running a small model to tell me bash commands on the fly is.
I use it to make .desktop files, too. Isn’t that so lazy?