Simon Smith’s Post

View profile for Simon Smith, graphic

EVP Generative AI at Klick

OpenAI DevDay is in progress, but there are no livestreams. You have to follow along on X or someone's liveblog to keep up. The day is devoted to developer use cases, and we're not expecting any new models or ChatGPT functionality. Here are the ways in which I think what they've announced so far will affect the average person: 🗣️ Realtime voice interaction everywhere! Developers will be able to use the new realtime API to create the kinds of realtime voice experiences we now get in ChatGPT in their own apps. For example, imagine language instruction apps with realtime speech-to-speech interaction. ⏩ Faster, cheaper, better model responses. OpenAI has done a few things in this area, including prompt caching (making it cheaper when reusing similar prompts), and streamlined model distillation (allowing developers to more easily create smaller models from the output of larger models). 🔬 New vision use cases—such as in healthcare. Developers can now fine-tune on images in addition to text. This means, for example, someone with a biomedical image dataset could fine-tune GPT-4o to perform particularly well in a biomedical domain. You can find details on the OpenAI website here: https://openai.com/news/

To view or add a comment, sign in

Explore topics