An OpenAI event on Nov 6
Nov. 1st, 2023 03:36 pmOn November 6, OpenAI will host a "DevDay", and some of it will be livestreamed.
This might be a landmark event (by some preliminary indications).
In particular, my prediction is that the ability fine-tune GPT-4 will be open to the public, and that this functionality will enable creation of specialized systems which are much more powerful than GPT-4, and that some of the magic will be demonstrated during the livestream.
We'll see if this prediction turns true. I record some relevant links and information in the comments.
Livestream link: www.youtube.com/watch?v=U9mJuUkhUzk
Update: One can watch the recording (45 min, if you'd like the transcript, you need to switch manually from "auto-generated" to close-caption-induced. like CC1 or DTVCC1).
Tons of upgrades (GPT-4 Turbo with 128K context and tons of other things, including making the API engineering easier) and major API price cuts (if they manage that without quality degradation, that would be a major step forward).
With fine-tuning: opening fine-tuning for GPT-3.5 Turbo 16K context and inviting active fine-tuning users to apply to the experimental GPT-4 fine-tuning program (so they are going very cautiously, I grade my prediction as 50% only; they are in the process of opening it to the public, but they are afraid of its potential and will go slowly; they also have chosen not to show-case fine-tuning at all; they've show-cased all kinds of things, but they don't want to encourage fine-tuning too much at this moment, because it is so uncontrollable).
openai.com/blog/new-models-and-developer-products-announced-at-devday
openai.com/blog/introducing-gpts
This might be a landmark event (by some preliminary indications).
In particular, my prediction is that the ability fine-tune GPT-4 will be open to the public, and that this functionality will enable creation of specialized systems which are much more powerful than GPT-4, and that some of the magic will be demonstrated during the livestream.
We'll see if this prediction turns true. I record some relevant links and information in the comments.
Livestream link: www.youtube.com/watch?v=U9mJuUkhUzk
Update: One can watch the recording (45 min, if you'd like the transcript, you need to switch manually from "auto-generated" to close-caption-induced. like CC1 or DTVCC1).
Tons of upgrades (GPT-4 Turbo with 128K context and tons of other things, including making the API engineering easier) and major API price cuts (if they manage that without quality degradation, that would be a major step forward).
With fine-tuning: opening fine-tuning for GPT-3.5 Turbo 16K context and inviting active fine-tuning users to apply to the experimental GPT-4 fine-tuning program (so they are going very cautiously, I grade my prediction as 50% only; they are in the process of opening it to the public, but they are afraid of its potential and will go slowly; they also have chosen not to show-case fine-tuning at all; they've show-cased all kinds of things, but they don't want to encourage fine-tuning too much at this moment, because it is so uncontrollable).
openai.com/blog/new-models-and-developer-products-announced-at-devday
openai.com/blog/introducing-gpts
no subject
Date: 2023-11-01 08:05 pm (UTC)https://openai.com/pricing
GPT-4:
8K context Input: $0.03 / 1K tokens Output: $0.06 / 1K tokens
32K context Input: $0.06 / 1K tokens Output: $0.12 / 1K tokens
GPT-3.5 Turbo:
4K context Input: $0.0015 / 1K tokens Output: $0.002 / 1K tokens
16K context Input: $0.003 / 1K tokens Output: $0.004 / 1K tokens
GPT-3.5 Turbo fine-tuned 4K context:
Training: $0.0080 / 1K tokens Input usage: $0.0120 / 1K tokens Output usage: $0.0160 / 1K tokens
So, training is not all that expensive (but it really depends on the size of your training corpus for your fine-tuning), but the usage cost for the 4K context is 8 times higher than for the standard GPT-3.5 Turbo version with 4K context.
So we might potentially expect a similar 8x pricing factor for the use of a fine-tuned GPT-4, with training being reasonably affordable, but usage being relatively expensive...