BuzzRobot
Join the Bugout Slack dev community to connect with fellow data scientists, ML practitioners, and engineers: https://join.slack.com/t/bugout-dev/shared_invite/zt-fhepyt87-5XcJLy0iu702SO_hMFKNhQ
The GPT-3 language models have a range of impressive capabilities; however, they also have a number of performance limitations.
In this talk OpenAI researcher, Todor Markov, discusses a new fine-tuning API by OpenAI and show examples of how it can be used to create specialized models.
Our speaker – Todor Markov is a researcher on the Applied AI team at OpenAI. His current work focuses on improving the safety & monitoring system for the OpenAI API.
00:09:38 — Introduction Todor Markov to the Bugout community
00:11:12 — Colleges to OpenAI
00:11:36 — GTP-3 & the OpenAI API
00:24:43 — Can GTP-3 make mathematically correct statements?
00:31:11 — OpenAI fine-tuning API
00:41:55 — Questions: sizes of GTP-3 models, size of data sets, technical background of fine tuning, models for improving semantic search, GTP-3 to transfer to texts, chip models, A/B testing model quality, is it expensive. customers, limit of models, tips for people who want to work in AI, plans for meetups
Thank you!
Hey, quick question, how did you format the CSV file before uploading it to the CLI?
What time is it today?
A bee finds a name
In the computer
I'm getting this tattooed on my chest. What time is it yesterday?
Thank you for tutorial. I followed the tutorial but it is not runing on Windows. Also the code lines which is given in Developer quickstar -> Making requests OpenAI Api are not running on Windows. Is this tutorial only for Mac?
Can we fine-tune same GPT-3 model more than once?
If we want to bias the gpt-3 language, is it possible? For example, i want my model to use only my field related vocab while generation. How can i do that?
Thanks for your tutorial. I got a question. Is it possible to fine-tune a model that answer questions the same way it would do with the answer function of OpenAI API?
I mean, is there something like a parameter for the "fine-tune.creates" line in which i can send a set of documents to look for? So the model can answer questions based on the information of the documents.
Thank you again!
Is the fine tuned model context aware? I mean if I trained it with a specific logical QnA flow, will it be aware of the context? …especially in domain specific QnA training data?
Sorry I'm an idiot, and I'm new to this. Will fine tuning allow me to make a chatbot with a backstory, memory and personality any size I choose? Fine Tune on previous chats so It will be able to retrieve the information from there instead of using tons of tokens for every question? Or are you saying that I would only be able to draw on that information by creating custom prompts to access that information, in which case that doesn't really help me with a memory and backstory. that would be too much work.
This is incredible technology, it's so fun to play this. The potential is huge. I love the fact that you guys have let random idiots like me play with such a powerful tool 😅.
Hello, thank you for the video. Is there anywhere we could get the yelp_review_sentiment.csv file?
Great video! I just want to add that as of January 2023, tokens you use with your fine-tuned model are about 6 times more expensive than the base model (Davinci: $0.12 vs $0.02 /1k tokens). So you might not save money, but you will get more accurate outputs if you fine-tune it correctly .
can any one show a video of the paid acount, how it will be, what is the minimum subscription.and more over to fine tune a model using 18$ free credit , it is not creating a model , instead it throws error as in correct api
I want that csv file
Excelente lesson. I expected Open AI Fine Tuning API to improve the general model itself with new training data but what we see is a significant deterioration of the original model responses when it's Fine Tuned on new data. Is there anyway to overcome this problem?
how to update my existing fine-tune model with my new custom data?
your are teaching to her very nice ra sulli
DO ONE FOR GPT-4
I want to teach my model to generate a json string with properties that I have strict naming for them. I tried to get chat gbt to do it and it did a nice job but he takes the liberty of changing some of the property names.
For example if my prompt tells him to generate the json string he will name one of the json properties in his own way. If I train my model can my completion property in the jsonL file have an inner json string example?
Hi the slack community link you have provided in the description has expired, can you please share a new one! thanks