š Does King Crimsonās song Elephant Talk get stuck in your head like it does mine? I have to admit Iām very excited to see the BEAT tour with Adrian Bellow, Steve Vai, Danny Carey and Tony Levināwhoās 78 and still touring! This interview with Tony starts slow but delivers a lot of amusing detail, especially about his tour photo albums.
Meanwhile, thereās been lots of news and conversation this week around generative AI.
š¤š„ They said ādesigned to be commercially safeā twice
Adobe Firefly is promoting new video capabilities āavailable later this year.ā More details via The Verge. Itās very clear Adobe is preaching to legal and others who are concerned about copyright-ability, ownership and attribution.
The outputs look at least as compelling as Runway and what OpenAI is promising with Sora. Adobeās advantage is their install base among creatives, and (in theory) the ability to smoothly integrate what you generate into an existing edit timeline.
š¤š OpenAIās ChatGPT o1-preview is live
Itās currently only available to Plus and Teams users of ChatGPT. It doesnāt have access to browse the web, or process files or images.
But it appears to have some ability to āreason.ā
And while you or I might not find the newest iteration of OpenAIās LLM all that exciting, it does illuminate a useful path forward. As The Verge described the contrast, current day LLMs are, āessentially just predicting sequences of words to get you an answer based on patterns learned from vast amounts of data.ā But o1-preview takes a different approach, incorporating reinforcement learning as well as āchain of thoughtā processing to arrive at an answer to your prompt. Wired explains:
āLLMs typically conjure their answers from huge neural networks fed vast quantities of training data. They can exhibit remarkable linguistic and logical abilities, but traditionally struggle with surprisingly simple problems such as rudimentary math questions that involve reasoning.
Murati says OpenAI-o1 uses reinforcement learning, which involves giving a model positive feedback when it gets answers right and negative feedback when it does not, in order to improve its reasoning process. āThe model sharpens its thinking and fine tunes the strategies that it uses to get to the answer,ā she says.ā
Ethan Mollick suggests āfun things to do with your limited 01-previewā including āGive it an RFP and ask it to just do the work.ā
I asked ChatGPTās o1-preview to first write itself a challenging prompt in the realm of marketing and advertising, then respond. Is the output stunning? Of course not. But 16 seconds later I got a decent outline for a marketing campaign with a few hints of potential ideas. Certainly more actionable than an entry-level agency account executive might output without AI, and a few hours or days of time to work. The point beingāthis tool is getting more nuanced, and is gaining greater ability to work alongside strategic and creative thinkers more effectively. If you have access to it, you should be playing around with it.
š¤š«š· Bonjour, Le Chat
Via AI Tool Report, āFrench AI start-up, Mistral, has released its first multimodal modelāPixtral 12GBācapable of processing and understanding both images and text.ā Sign up and access Le Chat here.
Hereās another competitor to Claude, ChatGPT, Perplexity, Gemini, et al. Is it better? Youāll have to play around with it and decide how its personality and processing fit your needs, or not.
And thatās kind of the point right nowāas users, we should be experimenting far and wide, using the same prompt across multiple AI to discern how each might work for us.
š¤š Waymo fewer accidents?
It seems inevitable.
Especially if, as the data reported by Neuron Daily suggests, āWaymoās injury crash rates are approximately 60-70% less than human drivers.ā
Idea #1: Itās always about the money. It will be the insurance companies offering massive discounts for using AI-operated vehicles, while simultaneously upping rates for any human who still needs to drive, which flips reality.
Idea #2: Suddenly, distracted driving becomes a good thing. Getting people to embrace AI taking the wheel could be rooted in enabling all the distractions human drivers get penalized for doing.
Idea #3: Access to your choice of AI-driven vehicles via subscription. Guaranteed availability and āuptime.ā Owning or leasing a car goes the way of the landline.
š¤šØ Ideogram is an appealing alternative for generative AI
Midjourney, Firefly, Meta, et al have competition with the latest Ideogram offering.
š Finally, someone is making AI normal and maybe even boring
I watched the whole thing on Monday.
I havenāt watched an entire Apple event in years. I miss the live speaking, but honestly, the production value was amazing, nothing broke, and we got to see all the details perfectly. Covid changed a lot of things, especially live events.
But firstā¦
FDA-approved AirPod hearing aids? Sold.
This is a game changer. My family has a lot of experience with hearing aids. They are expensive. They are difficult to connect to other devices. If Apple can solve all that, weāre in. The FDA granted approval yesterday.
But thereās an even bigger market: All of us in every loud setting. Or Gen Xers like me who played in too many bands without hearing protection.
I also think this shift from listening to hearing will be huge.
As an example, my teens both wear their AirPods all the timeāin the house, in the car. And it kills me. I take the visual of pods-in-ears as an insult to being present. Theyāre just listening. But what if my framing shifted, and I took the presence of AirPods as a signal of hearing enhancement? You wearing AirPods means Iāll be more clearly understood (and not just blocked out). This, combined with Appleās sleep apnea diagnostic helps expand their brand framing beyond devices into services and increased relevance in health.
Apple Intelligence = āAI for the rest of usā
Thereās that old Mac computer headline: āThe computer for the rest of us.ā In many ways, my take on Apple Intelligence follows a related path.
Itās very clear Apple is taking AI very seriously. But that seriousness isnāt translated into shipping feature after feature quickly (see so much of the above), but on clarifying purpose. Two years since ChatGPT, many are still askingāwhat is all this AI for?
I applaud Apple for what appears to be significant thinking to answer that question. Theyāre trying to discern purpose for technology which has wildly expansive, often unknown capabilities, but would be applied within their very controlled and proven environment. I can imagine Craig Federighi asking, āHowās my mom going to use AIāwithout having to understand AI?ā And what was revealed earlier this week wasnāt clear, product features, but clearly defined use cases.
In many ways, what Apple is trying to do is make AI invisible. To make it simply a part of how you naturally experience your devices and applications without necessarily calling attention to AI as the enabler of those experiences. And this strategy will help address a lot of the frustrations and misconceptions Iāve experienced in the last two years of evangelizing, training and coaching people around AI.
The key ingredients I noted were:
Personal context understanding ā AI that is āunique to youā for āeveryday experiences.ā This is less about what newfangled video you might generate, and more about how you can integrate and leverage existing data, i.e. āPlay that podcast episode Joseph told me about,ā or āfind the part of the video where Haley is laughing.ā If thatās AI, thatās useful.
Visual intelligence ā Being able to point a camera at a location or object and glean information about it will be huge. Yes, OpenAI demonstrated this first, but Apple will make it seamless and normal. Decades ago we talked about āhyperlinkingā the real world. This is exactly that, and itās going to be profound.
Summarization & Integration ā Again, the summarization piece is nothing new technically speaking from what we can do inside all the major LLMs. But when you apply Appleās approach and context, the difference becomes clearer. An email summary i.e. ācampaign is approved, book flightsā is so much more advantageous than four preview lines, i.e. āThe client called back. Blah blah blah the information you really need is buried here somewhere.ā And because Apple Intelligence works within iOS, it can spark action across different applications. Helping me parse the deluge of inputs, then surfacing potential actions is great use of AI.
Donāt get me wrong.
I love the jagged edge of AI. I enjoy wrangling prompts to generate remarkable results. But I also recognize how distanced the majority of generative AI has been from pragmatic, commonplace solutions. Apple envisions AI as plumbing, not as an imaginative castle. And itās their ability to embed complex technology to unlock and enhance everyday tasks which sets their approach apart.
Of course, weāll have to jump in and use it for real, starting early October. But the pitch was compelling.
Last but not leastā¦
Were you addicted to watching The Sopranos like I was?
Sunday nights. HBO. And then Monday back in the office to gossip about David Chaseās writing, Alik Sakharovās cinematography, or the fabulous acting ensemble. Well thereās a brilliant new documentary on the inspiration and making of The Sopranos. More background via Bob Lefsetz.