I use AI toolings to generate snippets of bash scripts because I can’t be fucked to remember that syntax. Obviously not for anything with high risks or that I can’t easily verify. But things like parsing through mass amounts of files
But… bash snippet extensions already exist. The only difference is maybe it doesn’t auto name your variables for you. I’d take that over non-deterministic LLM outputs.
I have no idea what the hell a bash snippet extension is, but I do know what a local llama.cpp instance running a small model to tell me bash commands on the fly is.
I use it to make .desktop files, too. Isn’t that so lazy?
I use AI toolings to generate snippets of bash scripts because I can’t be fucked to remember that syntax. Obviously not for anything with high risks or that I can’t easily verify. But things like parsing through mass amounts of files
But… bash snippet extensions already exist. The only difference is maybe it doesn’t auto name your variables for you. I’d take that over non-deterministic LLM outputs.
I have no idea what the hell a bash snippet extension is, but I do know what a local llama.cpp instance running a small model to tell me bash commands on the fly is.
I use it to make .desktop files, too. Isn’t that so lazy?