Great for giving incantatons for ffmpeg, imagemagick, and other power tools.
“Use ffmpeg to get a thumbnail of the fifth second of a video.”
Anything where syntax is complicated, lots of half-baked tutorials exist for the AI to read, and you can immediately confirm if it worked or not. It does hallucinate flags, but fixes if you say “There is no --compress flag” etc.
This is the way.
Legitimately, no. I tried to use it to write code and the code it wrote was dog shit. I tried to use it to write an article and the article it wrote was dog shit. I tried to use it to generate a logo and the logo it generated was both dog shit and raster graphic, so I wouldn’t even have been able to use it.
It’s good at answering some simple things, but sometimes even gets that wrong. It’s like an extremely confident but undeniably stupid friend.
Oh, actually it did do something right. I asked it to help flesh out an idea and turn it into an outline, and it was pretty good at that. So I guess for going from idea to outline and maybe outline to first draft, it’s ok.
My experience is that while it’s useful for creating code from scratch it’s pretty alright if you give it a script and ask it to modify it to do something else.
For instance I have a cron job that runs every 15min and attempts to extract .rar files in a folder and email me if it fails to extract. Problem is if something does go wrong it emails me every 15minutes until I fix it. This is especially annoying if its stuck copying a rar at 99%.
I asked deepseek to store failed file names in a file and have the script ignore those files for an increasing amount of time for each failure. It did a pretty good job, although it changed the name of a variable halfway through (easy fix) and added a comment saying it fixed a typo despite changing nothing about that line. I probably probably would have written almost identical code but it definitely saved me time and effort
Crappy but working code has its uses. Code that might or might not work also has its uses. You should primarily use LLMs in situations where you can accept a high error rate. For instance, in situations where output is quick to validate but would take a long time to produce by hand.
The output is only as good as the model being used. If you want to write code then use a model designed for code. Over the weekend I wrote an Android app to be able to connect my phone to my Ollama instance from off my network. I’ve never done any coding beyond scripts, and the AI walked me through setting up the IDE and a git repository before we even got started on the code. 3 hours after I had the idea I had the app installed and working on my phone.
I didn’t say the code didn’t work. I said it was dog shit. Dog shit code can still work, but it will have problems. What it produced looks like an intern wrote it. Nothing against interns, they’re just not gonna be able to write production quality code.
It’s also really unsettling to ask it about my own libraries and have it answer questions about them. It was trained on my code, and I just feel disgusted about that. Like, whatever, they’re not breaking the rules of the license, but it’s still disconcerting to know that they could plagiarize a bunch of my code if someone asked the right prompt.
(And for anyone thinking it, yes, I see the joke about how it was my bad code that it trained on. Funny enough, some of the code I know was in its training data is code I wrote when I was 19, and yeah, it is bad code.)
This is a very rare use case, but one where i definetly found them very useful. Similar to another answer mentioning reverse-dictionary lookup, i used llms for reverse-song/movie lookup. That is, i describe what i know about the song/movie (whatever else, could be many things) and it gives me a list of names that i can then manually check or just directly recognize.
This is useful for me because i tend to not remember names / artists / actor names, etc.
ChatGPT kind of sucks but is really fast. DeepSeek takes a second but gives really good or hilarious answers. It’s actually good at humor in English and Chinese. Love that it’s actually FOSS too
It’s helping me understand how I think so that I can create frameworks for learning, problem solving, decision making etc. I’m neurodivergent.
One day I’m going to get around to hooking a local smart speaker to Home Assistant with ollama running locally on my server. Ideally, I’ll train the speech to text on Majel Barrett’s voice and be able to talk to my house like the computer in Star Trek.
With mixed results I’ve used it for summarising the plots of books if I’m about to go back into a book series I’ve not read for a while.
Very effective at translating between different (human) languages. Best if you can find a native speaker to double-check the output. Failing that, reverse translate with a couple different models to verify the meaning is preserved. Even this sometimes fails though – e.g. two words with similar but subtly different definitions might trip you up. For instance, I’m told “the west” refers to different regions in english and japanese, but translating and reverse translating didn’t reveal this error.
Good for gaining an outside perspective/insight on an argument, discussion, or other form of communication between people. I fed it my friend’s and their ex’s text conversation to it (with permission), and it was able to point out emotional manipulation in the text when asked neutrally about it:
Please analyze this conversation between A and B and tell me what you think of their motivations and character in this conversation. Is there gaslighting? Emotional manipulation? Signs of an abusive communication style? Etc. Or is this an example of a healthy communication?
It is essential not to ask a leading question that frames A or B in particular as the bad or the good guy. For best results, ask neutral questions.
It would have been quite useful for my friend to have this when they were in that relationship. It may be able to spot abusive behaviors from your partner before you and your rose-colored glasses can.
Obvious disclaimers about believing anything it says are obvious. But having an outside perspective analyze your own behavior is useful.
I love fantasy worldbuilding and write a lot. I use it as a grammar checker and sometimes use it to help gather my thoughts, but never as the final product.
I’ve used them both a good bit for D&D/TTRPG campaigns. The image generation has been great for making NPC portraits and custom magic item images. LLM’s have been pretty handy for practicing my DM-ing and improv, by asking it to act like a player and reacting to what it decides to do. And sometimes in the reverse by asking it to pitch interesting ideas for characters/dungeons/quest lines. I rarely took those in their entirety, but would often have bits and pieces I’d use.
I use it for books/movies/music/games recommandations (at least while it isn’t used for ads…). You can ask for an artist similar to X or a short movie in genre X. The more demanding you are the better, like a “funny scifi book in the YA genre with a zero to hero plot”.
The image generator to 3D model to animation pipeline isn’t too bad. If you’re not a great visual artist, 3D modeler, or animator you can get out pretty decent results on your own that would normally take teams of multiple people dozens of hours after years of training
Nope. Any use case I have tried with it, I usually find that either a python script, database, book, or piece of paper can always accomplish the same job but usually with a better end result and with a more reliably reproducible outcome.
I use it for alt text for photos. Mostly because I just don’t know how to describe my images.