I wanted to edit my Ghostty themes but found out a lot of the colors are in #hexadecimal notation. I like #rrggbb percentage style colors (b/c they are easy to tweak by hand) and I couldn’t find an online color picker that would output that format, so I used deepseek (free) & now have a scrappy ass one w Python & Tkinter completely via “vibe” coding (I call it Clyde Color Picker. It’s adorable).
Pretty awesome when you’re just some dumbass who needs a very specific tool and not trying to fleece people.
I use AI toolings to generate snippets of bash scripts because I can’t be fucked to remember that syntax. Obviously not for anything with high risks or that I can’t easily verify. But things like parsing through mass amounts of files
But… bash snippet extensions already exist. The only difference is maybe it doesn’t auto name your variables for you. I’d take that over non-deterministic LLM outputs.
I have no idea what the hell a bash snippet extension is, but I do know what a local llama.cpp instance running a small model to tell me bash commands on the fly is.
I use it to make .desktop files, too. Isn’t that so lazy?
I wanted to edit my Ghostty themes but found out a lot of the colors are in #hexadecimal notation. I like #rrggbb percentage style colors (b/c they are easy to tweak by hand) and I couldn’t find an online color picker that would output that format, so I used deepseek (free) & now have a scrappy ass one w Python & Tkinter completely via “vibe” coding (I call it Clyde Color Picker. It’s adorable).
Pretty awesome when you’re just some dumbass who needs a very specific tool and not trying to fleece people.
I use AI toolings to generate snippets of bash scripts because I can’t be fucked to remember that syntax. Obviously not for anything with high risks or that I can’t easily verify. But things like parsing through mass amounts of files
But… bash snippet extensions already exist. The only difference is maybe it doesn’t auto name your variables for you. I’d take that over non-deterministic LLM outputs.
I have no idea what the hell a bash snippet extension is, but I do know what a local llama.cpp instance running a small model to tell me bash commands on the fly is.
I use it to make .desktop files, too. Isn’t that so lazy?