dmm: (Default)
It was pretty informative throughout (I rarely watch long videos, especially if a transcript is available, but I watched this one; the last 10 min were particularly crazy in a good way).

The places where he demurred or hedged were also quite interesting; this did provide a good window into all this...

twitter.com/lexfridman/status/1769755831619219527

dmm: (Default)
On November 6, OpenAI will host a "DevDay", and some of it will be livestreamed.

This might be a landmark event (by some preliminary indications).

In particular, my prediction is that the ability fine-tune GPT-4 will be open to the public, and that this functionality will enable creation of specialized systems which are much more powerful than GPT-4, and that some of the magic will be demonstrated during the livestream.

We'll see if this prediction turns true. I record some relevant links and information in the comments.

Livestream link: www.youtube.com/watch?v=U9mJuUkhUzk

Update: One can watch the recording (45 min, if you'd like the transcript, you need to switch manually from "auto-generated" to close-caption-induced. like CC1 or DTVCC1).

Tons of upgrades (GPT-4 Turbo with 128K context and tons of other things, including making the API engineering easier) and major API price cuts (if they manage that without quality degradation, that would be a major step forward).

With fine-tuning: opening fine-tuning for GPT-3.5 Turbo 16K context and inviting active fine-tuning users to apply to the experimental GPT-4 fine-tuning program (so they are going very cautiously, I grade my prediction as 50% only; they are in the process of opening it to the public, but they are afraid of its potential and will go slowly; they also have chosen not to show-case fine-tuning at all; they've show-cased all kinds of things, but they don't want to encourage fine-tuning too much at this moment, because it is so uncontrollable).

openai.com/blog/new-models-and-developer-products-announced-at-devday

openai.com/blog/introducing-gpts

dmm: (Default)
GPT-4 in ChatGPT+ can now read and analyze images, one can upload an image and discuss it with GPT-4. 

GPT-4 in ChatGPT+ can also generate text prompts and render images via DALL-E 3.

What is missing is the ability to load an image and modify it with DALL-E 3 (or to load the image and use it as an inspiration for image generation by DALL-E 3).

1) So, on one hand, I want to ask a question:

Do people here have experience with loading an image and modifying it using an AI system (or with loading an image and using it as an inspiration for an AI system)? Any recommendations?

2) On the other hand, there are two possible workarounds.

2a) One can use an ASCII art inside GPT-4 to inspire images rendered by DALL-E 3 (it works, but I was not able to obtain the results I would like so far).

2b) One can analyze an image and obtain an image description in one mode of GPT-4 in ChatGPT+, and then open another session and use that description in the prompt to guide DALL-E 3.

Some of the results obtained via method 2b) are quite interesting. I am going to post related images and dialogs as a series of comments.

Update: In November, the ability to read and create images within one session has been added (and, generally speaking, all functionality which has been separated into different type of sessions is now available from a single session, you should be able to ask the system for anything, be it "search the web", or "generate and execute code", or "upload this data, and then generate code to analyze this data, and run it"; these used to require starting different types of sessions, but now all this is available together).
dmm: (Default)
Let's see whether github hosting for *.webp images is OK to render them here, in the post, and in the comments.

test image

dmm: (Default)
For the 4K context version. OpenAI says that it is often better than GPT-4 for a specific application.

Price to fine-tune a model is very reasonable (but depends on the size of your training set). 

The cost to use the resulting model is much higher than using the un-finetuned GPT-3.5, though it's still cheaper than using GPT-4.

Obviously, you only can use a fine-tuned model via API, not via standard Web interface.

OpenAI says the ability to fine-tune 16K context and the ability to fine-tune GPT-4 are coming.

Profile

dmm: (Default)
Dataflow matrix machines (by Anhinga anhinga)

May 2025

S M T W T F S
    123
456 78910
11 121314151617
18192021222324
25262728293031

Syndicate

RSS Atom

Most Popular Tags

Style Credit

Expand Cut Tags

No cut tags
Page generated Jun. 14th, 2025 03:20 pm
Powered by Dreamwidth Studios