• Everything they introduced was about AI. When they hit the 2 hour mark, I was ready for them to start talking about, say, new features of Android this year… and then the show wrapped up!
• Corollary of the above: it's really cool to see what it looks like to spread AI across an entire established product ecosystem. OpenAI don't have much in the way of consumer product — they have a thin wrapper around AI as a fundamental technology and that's about it. Google is all product. So beyond just chat UIs and generative prompting, you now (or presumably will soon) get stuff like "How has my daughter improved at swimming over the years?" in Google Photos, and "Where am I?" in the Android camera, "circle to search" on any image across Android or in Chrome, "Role-play as my professor" in docs, etc etc etc. This feels like the first time we're seeing what it'd be like to have AI in absolutely every context, across every UI, aware of all your data.
• Building on the above: it was intriguing, and often a bit baffling, to see where and how they're integrating AI in terms of GUI design. Sometimes it's active in a text field, sometimes you need to tap a button, sometimes it's a popup, sometimes it's a sidebar. Broadly: sometimes it's ambient, and sometimes it's explicit; sometimes it slots into existing UIs, sometimes it's a new dedicated place. Sort of like the sparkle emoji being adopted as the de facto "AI is here" icon, I'm very interested in seeing what patterns emerge around the appropriate placement of AI across all GUIs, since it seems like that's bound to happen. Does it end up like spellcheck, where there's a single universal visual signifier and a clear course of action?
• For their AI APIs for devs, they talked about pricing per 1M tokens. They also talked a lot about their 1M token context window (and at least 3 or 4 times, mentioned it'll be 2M before long). For some kinds of query, the pricing was $7 per 1M tokens — in other words, AIUI, $7 per full context window. It really makes me wonder what consumer uses would justify $7 per inference. I bet there's some really interesting possibility there. Like, what would an AI need to do for me to pay $7 per use? Generate an entire video game based on my Steam library? Generate an album based on my entire iTunes / Apple Music library and listening history? If the quality were good, that feels like fair value.
• I love Gemini as a name for their AI initiatives (the symbolism, the 'G', the mouthfeel of the word — great name, topping "Copilot" which is also a great name). "Gem" as their version of "bot" is also excellent. But "Gemma" is stretching this too far, and it becomes confusing what stuff is "Gemini" and what stuff is "Gemma".
• Speaking of confusing, the show lacked a clear "table of contents" / structure. For WWDC, there's tentpole sections for i[Pad]OS, Mac, Watch, and then a few grab bags (HomePod, TV), and they clearly signpost the sections with a "tell 'em what you're gonna tell 'em, tell 'em, tell 'em what ya told 'em" pattern. For want of this structure, it felt like folks from various branches of Google repeated each other's announcements (eg: context window going from 1M to 2M, Gems for learning). Conway's law, perhaps?
• Love that it was a live presentation with some live demos. Nice set design, too.