this post was submitted on 13 Sep 2023
244 points (98.4% liked)
Asklemmy
44151 readers
1801 users here now
A loosely moderated place to ask open-ended questions
Search asklemmy ๐
If your post meets the following criteria, it's welcome here!
- Open-ended question
- Not offensive: at this point, we do not have the bandwidth to moderate overtly political discussions. Assume best intent and be excellent to each other.
- Not regarding using or support for Lemmy: context, see the list of support communities and tools for finding communities below
- Not ad nauseam inducing: please make sure it is a question that would be new to most members
- An actual topic of discussion
Looking for support?
Looking for a community?
- Lemmyverse: community search
- sub.rehab: maps old subreddits to fediverse options, marks official as such
- [email protected]: a community for finding communities
~Icon~ ~by~ ~@Double_[email protected]~
founded 5 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
The lawyer fuck up is what happens when someone doesn't know or understand the limitations of a LLM.
If you want a GPT model tailored and specialized for a specific task, you have to train it with custom data, fine tune it and tweak the model's parameters. You cannot do that from the ChatGPT web/app, you need a custom implementation coded in Python or some other language.
There are some uis that allow for fine tuning (assuming you have an extremely high end rig designed for ml). For example ChatGPT alternative and DALLE alternative.
Thanks. I have a quite powerful rig, but at the moment I work with OpenAI's API using GPT 3.5 Turbo using a custom (but shitty) Python script with a simple Gradio web interface. However, I mostly stopped improving or updating it months ago. As long as I don't use LlamaIndex, the cost is quite low.
I already use Stable Diffusion WebUI, tho.
Also the "fine tuning" I was talking about is this https://platform.openai.com/docs/guides/fine-tuning
I am aware what fine tuning is. It is available from the train tab while the base checkpoint is loaded in both cases.
I'm glad you understand my point. Chatgpt is not Google. It's a language model that will give you something that looks like the thing you asked for it to provide. It can and will pull facts out of its recycle bin if it fits the cadence of what it expects the answer to look like.
ChatGPT is not Google, but sometimes it can work as a glorified search engine or even compete with asking in forums.
I've lost count of how many times ChatGPT has produced Bash or Python code for what I needed. Yes, sometimes the code is wrong and/or requires tweaking and sometimes I resorted to look into the documentation, but no one will answer faster and anytime of the day like ChatGPT does, at least not for free.
It's a tool to aid in creating a product, not a tool that magics out a finished product. That's my point. Too many people use it as the latter instead of the former.
100% agree.
Maybe, with lots of training, weaking and testing the latter could be achieved, but that's it.