The viewing angle might not be so great, but on the bright side, it’s really easy to reach the TV’s HDMI ports!
(Also: ChatGPT didn’t draw those images; its image-rendering counterpart DALL-E did.)
Some techies hold the attitude that “what I do is important, and what you do isn’t,” and the more socially savvy ones don’t say the quiet part out loud.
But Mira Murati, OpenAI’s CTO, did just that onstage at her alma mater, Dartmouth University, where she said this about AI displacing jobs in creative lines of work:
Some creative jobs maybe will go away, but maybe they shouldn’t have been there in the first place.
Mira Murati, from AI Everywhere: Transforming Our World, Empowering Humanity
(she says this around the 29:30 mark)
Here’s my take on her bad take, courtesy of the Global Nerdy YouTube channel, which you should subscribe to…
…and here’s the video with her full talk at Dartmouth:
If you’re a regular reader of this blog, you’ve probably seen (or at least heard about) the demo of GPT-4o’s voice assistant mode featuring a voice named “Sky” providing vivacious — even flirty — assistance:
When I saw it, my first thought was “Wow, that’s a lot like Scarlett Johansson’s portrayal of the AI in Her,” and that seemed to be a lot of other people’s first thoughts.
This belief is backed by the publicly-known fact that Her is OpenAI CEO Sam Altman’s favorite movie, a fact he announced less than a year ago in front of the audience of Salesforce’s Dreamforce 2023 conference:
“I like Her. The things Her got right—like the whole interaction models of how people use AI—that was incredibly prophetic.”
San Fransisco Standard, September 12, 2023.
There’s also Altman’s single-word post on X/Twitter, which he posted on the day of the GPT-4o’s premiere on Monday, May 13th:
And now, we find out that Scarlett Johansson — the “Her” herself — issued a statement on Monday, May 20th saying that she was approached by Sam Altman to be the voice of this version of GPT, and that she turned down the offer.
Here’s the text of the statement:
Last September, I received an offer from Sam Altman, who wanted to hire me to voice the current ChatGPT 4.0 system. He told me that he felt that by my voicing the system, I could bridge the gap between tech companies and creatives and help consumers to feel comfortable with the seismic shift concerning humans and A.I. He said he felt that my voice would be comforting to people. After much consideration and for personal reasons, I declined the offer. Nine months later, my friends, family and the general public all noted how much the newest system named ‘Sky’ sounded like me.
When I heard the released demo, I was shocked, angered and in disbelief that Mr. Altman would pursue a voice that sounded so eerily similar to mine that my closest friends and news outlets could not tell the difference. Mr. Altman even insinuated that the similarity was intentional, tweeting a single word, ‘her’ — a reference to the film in which I voiced a chat system, Samantha, who forms an intimate relationship with a human.
Two days before the ChatGPT 4.0 demo was released, Mr. Altman contacted my agent, asking me to reconsider. Before we could connect, the system was out there. As a result of their actions, I was forced to hire legal counsel, who wrote two letters to Mr. Altman and OpenAI, setting out what they had done and asking them to detail the exact process by which they created the ‘Sky’ voice. Consequently, OpenAI reluctantly agreed to take down the ‘Sky’ voice.
In a time when we are all grappling with deepfakes and the protection of our own likeness, our own work, our own identities, I believe these are questions that deserve absolute clarity. I look forward to resolution in the form of transparency and the passage of appropriate legislation to help ensure that individual rights are protected.
New York Times, May 20, 2024
“OpenAI’s gonna OpenAI,” as this soon-to-be-common phrase goes, and they’ve been making their trademark obfuscating statements. As Ed Zitron summarizes in an article titled Sam Altman is Full of Shit:
Just so we are abundantly, painfully, earnestly clear here, OpenAI lied to the media multiple times.
Sam Altman Is Full Of Shit, Ed Zitron, May 21, 2024
- Mira Murati, OpenAI’s CTO, lied to Kylie Robison of The Verge when she said that “Sky” wasn’t meant to sound like Scarlett Johansson.
- OpenAI lied repeatedly about the reasons and terms under which “Sky” was retired, both by stating that it “believed that AI voices should not deliberately mimic a celebrity’s distinct voice” and — by omission — stating that it had been “in conversations” with representatives to bring Johansson’s voice to ChatGPT, knowing full well that she had declined twice previously and that OpenAI’s legal counsel were actively engaging with Johansson’s.
If you haven’t seen the movie Her, you’re probably wondering where you can find it on a streaming service. Here’s where you can watch it right now if you’re based in the U.S. (where I’m based):
And if you need a little more temptation to watch the film, here are its two official trailers:
…is to complete this poem. By doing so, you’ll have performed a computation similar to the one it does when you feed it a text-plus-image prompt.
For more, see my earlier post, GPT-4o is amazing at describing images.
This is a post specifically for tonight attendees at the AI Superpowers Unlocked panel taking place tonight (Wednesday, May 15, 2024). I’ll expand this post later.
I decided to try out ChatGPT’s newly-released GPT-4o model (the lowercase “o” stands for “omni”) by uploading images and asking it to describe them to me.
After this quick test, my impression is simply “wow.”
Me: What’s happening in this picture?
GPT-4o: The picture shows two musicians performing on stage. One musician is playing an accordion, and the other is playing an acoustic guitar and singing into a microphone. The guitarist is wearing a cowboy hat. In the background, there is a banner with the name “Steven Bosco” and a QR code for tips via Venmo and CashApp. The setting appears to be an indoor venue with a rustic, cozy ambiance. There is a small audience in the foreground, and the atmosphere looks lively and engaging.
Not bad. I was impressed that ChatGPT was able to determine that the banner read “Steven Bosco” even though my head clearly blocks the “B” — it would have had to determine the “Bosco” from the tip sign. I was also impressed by its description of the “feel” of the room.
Me: What is being depicted in this image?
GPT-4o: The image is a two-panel cartoon.
Panel 1:
Panel 2:
The cartoon humorously illustrates the importance of restricting access to critical areas on a ship, such as the engine room, to prevent unauthorized personnel from tampering with important controls.
In case you were wondering, I wrote about the comic in this article.
I wanted to see how well GPT-4o does with handwriting. I’m told that my penmanship is unusually readable; results with other people’s handwriting are bound to vary.
I was impressed by how it identified the image not just as a comic, but one with two panels. Even more impressive: that ChatGPT got the general gist of the comic, right down to its use of humor to illustrate its point.
Me: What’s this image about?
GPT-4o: The image is a detailed handwritten study guide about the OSI (Open Systems Interconnection) model, often referred to as the OSI seven-layer model or OSI seven-layer cake. The OSI model is used to standardize networking functions to support interoperability among different products and software.
The guide includes the following:
Left Side:
Right Side:
This guide is likely intended for students or professionals studying networking concepts, providing a clear and concise overview of the OSI model and its layers.
The challenge with this one was the low resolution of the photo. I took it with my late ’90s-era camera with its potato-fidelity 1024 * 768 resolution (and that was its “high-quality” mode — normal mode was 640 * 480!).
Me: Describe this image, please.
GPT-4o: This image is a four-panel cartoon strip that humorously depicts a situation involving an alien, a disguise kit, and a spaceship. Here’s a description of each panel:
Panel 1:
Panel 2:
Panel 3:
Panel 4:
The cartoon humorously highlights the importance of security and the potential consequences of not having proper identification and detection systems in place on a spaceship.
This is also a comic from this article.
One of the challenges with this image is that it’s a four-panel comic, but there are no lines separating the panels — just white space. Following only the circled numbers specifying the order in which the comic should be read, GPT-4o “got” the comic, including the fact that the character in panel 1 was impersonating Captain Platypus.
GPT-4o got one thing wrong — it thought that the “Aw, yeah — free ship!” line is from panel 2; it’s the alien stealing the ship in panel 4.
I’m being asked by a growing number of people this one question: “I want to get started learning and building AI apps with Python. Should I go with the Python.org distribution, Anaconda, or something else?”
For anyone who wants to learn Python and AI development, my answer is invariably:
Why?
Once again — if you’re learning Python for AI development:
ℹ️ Disclaimer: This is NOT a paid promo for Anaconda. I met some of their people at PyCon, but the company has no idea who I am or that I’m doing this, and they certainly don’t know how to put money in my bank account. I’m just a satisfied user who uses it for my own projects and for teaching Python and AI programming.
If you’ve tried to go past the APIs like the ones OpenAI offers and learn how they work “under the hood” by trying to build your own neural network, you might find yourself hitting a wall when the material opens with equations like this:
How can you learn how neural networks — or more accurately, artificial neural networks — do what they do without a degree in math, computer science, or engineering?
There are a couple of ways:
Along the way, both I (in this blog) and Tariq (in his book) will trick you into learning a little science, a little math, and a little Python programming. In the end, you’ll understand the diagram above!
One more thing: if you prefer your learning via video…