Almost Timely News: ๐๏ธ Setting the Record Straight on AI Optimization (2025-06-22)
Lies, damn lies, and snake oil
Almost Timely News: ๐๏ธ Setting the Record Straight on AI Optimization (2025-06-22) :: View in Browser
The Big Plug
๐ Download the new, free AI-Ready Marketing Strategy Kit!
Content Authenticity Statement
95% of this week's newsletter was generated by me, the human. You will see bountiful AI outputs in the video, and you will see an output from Anthropic Claude in the text. Learn why this kind of disclosure is a good idea and might be required for anyone doing business in any capacity with the EU in the near future.
Watch This Newsletter On YouTube ๐บ
Click here for the video ๐บ version of this newsletter on YouTube ยป
Click here for an MP3 audio ๐ง only version ยป
What's On My Mind: Setting the Record Straight on AI Optimization
Okay, let's clear the air on this whole AI optimization and the twisted, contorted, crazy space of AI optimization. There are so many weird, confusing names for all of this that it sounds like either a child's nursery rhyme or IKEA furniture names - GAIO, GEO, AIO, AEO, CSEO (conversational search/SEO), etc.
We need to lay down some basics so it's clear what's real and what's not, something you can take to your stakeholders when they ask - and a way to cut through a lot of the snake oil.
Part 1: Definitions
First, let's be clear what we're talking about. Fundamentally, what everyone wants to know is this:
Can we tell how much traffic (and therefore prospects, leads, opportunities, sales, and ultimately revenue) generative AI in all its incarnations is sending us, directly or indirectly?
From that blanket statement, we decompose into three major areas.
What do LLMs/generative AI models know about us? How do they talk about us when asked? How are we being recommended by AI models themselves?
How do AI-enabled search tools like Google AI Overviews and Google AI Mode recommend us and send traffic to us?
How do AI replacements for search like ChatGPT, Claude, Perplexity, Gemini, etc. recommend us and send traffic to us?
And from there, we ask the logical question: how can we get these different systems to recommend us more?
When we talk about whatever the heck we're calling this - for the rest of this newsletter I'm sticking with AI optimziation - we're really talking about the three whats.
What is it?
So what? Why do we care?
Now what? What do we do about it?
You'll note something really important. The three major areas all tend to get lumped together: AI models, AI-enabled search, AI replacements for search.
They should not be. They are not the same. This will become apparent in part 3.
Part 2: What You Cannot Know
This is a fundamental principle:
AI is not search.
Let's repeat that for the folks in the back who weren't paying attention.
AI IS NOT SEARCH.
When was the last time you fired up ChatGPT (or the AI tool of your choice) and typed in something barely coherent like "best marketing firm boston"?
Probably never. That's how we Googled for things in the past. That's not how most people use AI. Hell, a fair number of people have almost-human relationships with their chat tools of choice, giving them pet names, talking to them as if they were real people.
What this means is that it's nearly impossible to predict with any meaningful accuracy what someone's likely to type in a chat with an AI model. Let's look at an example. Using OpenAI's Platform - which allows you direct, nearly uncensored access to the models that power tools like ChatGPT, let's ask about PR firms in Boston.
I asked it this prompt:
"Let's talk about PR firms in Boston. My company needs a new PR firm to grow our share of mind. We're an AI consulting firm. What PR firms in Boston would be a good fit for us?"
o4-mini
Racepoint Global
LaunchSquad
Inkhouse
Sutherland Weston
Hotwire
Finn Partners
Sloane & Company
GPT-4.1
SHIFT Communications
PAN Communications
Rally Point Public Relations
INK Communications Co.
March Communications
Denterlein
Walker Sands
GPT-4o
PAN Communications
Matter Communications
March Communications
Racepoint Global
Velir
SHIFT Communications
You can see just within OpenAI's own family of models, via the API, I get wildly different results. The most powerful reasoning model available by API, the thinking model, comes up with very, very different results - but even GPT-4o and GPT-4.1 come up with different results.
This is what the models themselves know. When you use any tool that connects to OpenAI's APIs, you are using this version of their AI (as opposed to the ChatGPT web interface, which we'll talk about in a bit).
Now, suppose I change just a couple of words in the prompt, something reasonable but semantically identical. What if I chop off the first sentence, for a more direct prompt:
"My company needs a new PR firm to grow our share of mind. We're an AI consulting firm. What PR firms in Boston would be a good fit for us?"
What do we get?
o4-mini
Salt Communications
PAN Communications
SHIFT Communications
Matter Communications
Racepoint Global
Highwire
Argyle PR
GPT-4.1
Inkhouse
SHIFT Communications
March Communications
Red Lorry Yellow Lorry
Matter Communications
Walker Sands
GPT-4o
PAN Communications
Matter Communications
LaunchSquad
SHIFT Communications
Inkhouse
451 Marketing
March Communications
Surprise! Same model family, same vendor, wildly different results.
This is why itโs generally a foolโs errand to try to guess what any given AI model will return as its results. Just a few wordsโ difference can lead to very, very different results - and this is for a very naive conversational query.
What would happen if you were to use the conversational tone most people use? Instead of a brusque, search-like query, you asked in a way that reflected your own personality?
"Hey Chatty! Good morning. Hey listen, my company needs a new PR firm to grow our share of mind. We're an AI consulting firm And weโve tried PR firms in the past. Boy, let me tell you, some of the firms weโve tried have been real stinkers. Half of them charge an arm and a leg for work that you could do, and the other half are firms filled with navel-gazing thought leaders who donโt produce any results. Weโre in the Boston area (go Sox!) and I wonder who youโd recommend for us. Got a list of PR firms that are actually worthwhile?โ
Good luck attempting to model the infinite number of ways people could ask AI.
So letโs set this down as a fundamental principle of AI optimization: you cannot know what people are asking AI.
Anyone who says you can know this is lying. Thereโs no polite way to say that. Theyโre lying - and if theyโre asking for your money in exchange for supposed data about what people are asking ChatGPT and similar services, then theyโre pretty much taking your money and giving you raw sewage in return.
Part 3: What Can You Know?
Now that we understand that AI optimization isn't one thing, but three separate things, we can start to pick apart what we can know.
AI Models
Can we know what AI models know about us? Yes and no. AI models - the core engines that power the AI tools we all know and use - are basically gigantic statistical databases. There are over 1.8 million different AI models, and dozens of foundation models, state of the art models that we use all the time. Here are just a sampling of the models:
Google's Gemini and Gemma families
OpenAI's GPT and o families
Anthropic's Claude family
Deepseek's V and R families
Alibaba's Qwen family
Mistral's Mistral, Devstral, and Magistral families
Cohere's Command family
Moonshot's Kimi family
Meta's Llama family
NVIDIA's Nemotron family
IBM's Granite family
Microsoft's Phi family
TI's Falcon family
... and so, so many more
What's the point of this catalog? Depending on where you are in the world, and what software and vendors you use, thereโs a good chance one of these models is helping answer questions. For example, letโs say youโre using Microsoft Copilot at work. (Iโm sorry.) Youโre not just using one AI model - youโre using, behind the scenes, several. Microsoft Copilot, to contain costs, invisibly routes your query to the model Microsoft thinks can accomplish your task at the lowest cost possible, so it might route it to Phi 4 (its own in-house model) or one of OpenAIโs models, if Phi 4 isnโt up to the task.
Itโs a good idea to generally know what models power what systems. We know, for example, that OpenAIโs models power ChatGPT. Thatโs pretty much a given. Googleโs Gemini powersโฆ well, it seems like all of Google these days. Theyโre cramming Gemini into any place they can. Metaโs Llama powers all the AI within Metaโs apps, like Instagram, Facebook, Threads, and WhatsApp, so if youโre a social media marketer in the Meta ecosystem, knowing what Llama knows is helpful. And tons of companies behind the scenes are running local versions of DeepSeek because itโs a state of the art model you can run on your own hardware. If your company suddenly has a private-label AI hub that performs well, and it didnโt a few months ago, thatโs probably why.
To know what these models know about us, we'd have to go model by model and ask them in as many ways as we could what they know.
This is, computationally, a nightmare. Why? The cost to ask each of these models a million different questions about the million different ways you think about a topic would be astronomical.
Hereโs why knowing the models does matter, and something you can know at relatively low cost. In order for AI to answer any questions, it has to have some level of knowledge about you. One thing weโd want to know is what the latest knowledge it has about us - because providing training data to AI models (aka large amounts of text on the Internet) is a key strategy for influencing them.
BUT, and hereโs the big BUT, many AI model makers have been using more and more synthetic data (AI-generated) to train new versions of their models. Combine this with the very, very long time it takes to make models, and AI models often seem frozen in time, badly out of date.
Hereโs how we can know that. Go to the developer portal for any AI model - not the consumer web interface. For example, I would go to Google AI Studio or OpenAI Platform Playground or Anthropic Console. These are the developer interfaces where you can talk to the model directly with no add-ons at all.
Ask them a straightforward question like this:
โWhat is the most recent issue of the Almost Timely newsletter at christopherspenn.com that you know about? Provide the exact title, date of publication in YYYY-MM-DD format, and the URL in this format:
Latest article: [article title]
Date of publication: [date in YYYY-MM-DD]
URL of article: [URL]โ
Letโs see what the answers are from the newest models:
OpenAI GPT-4.1: 2024-06-30
Google AI Studio, Gemini 2.5 Pro: 2024-02-04
Anthropic Claude Sonnet 4: Refused to answer
Meta LLama 4 Maverick: Refused to answer
DeepSeek V3: 2024-06-01
We can see from this experiment that most models' knowledge of my newsletter ends a year ago. This is normal and expected; the process of training a new model takes 6-9 months, and then 3-6 months of safety testing and QA.
So that's what the models themselves know. It's valuable to understand this because if the base, underlying model has no knowledge of us, everything else is going to be in trouble. This is a key point - AI tools that we'll discuss in a minute use other forms of verification like search to validate what they're saying, but if the underlying models have no clue you exist, they won't even know you should be in the results.
AI-Enabled Search
The second category of AI optimization we care about is AI-enabled search. This is where traditional search engines put AI assistance within the search results. Google's AI Overviews do this, as do DuckAssist from DuckDuckGo, and of course Copilot in Bing.
AI overviews summarize the search results, synthesizing an answer. One of the big questions everyone has about Google's AI Overviews is how they work. As with so many things, Google has documented how the system works; the challenge for us is that they've done so in dozens of different places, such as research papers, conference submissions, patents, and of course, marketing material.
Here's how the system works; I made this summary with Anthropic Claude Sonnet 4 from my massive NotebookLM collection of 30+ patents and research papers from Google.
Google AI Overviews: User Journey Step-by-Step
Phase 1: User Input & Query Analysis
1. ๐ค User Input
User types, speaks, or uploads their query into Google Search
Supports multimodal input (text, voice, images)
Query received by client device's user input engine
2. ๐ Query Analysis
System processes multimodal input
Detects user intent and query complexity
Assesses whether query requires advanced processing
3. ๐ค Complexity Decision Point
IF query is complex or multi-faceted:
Proceed to Query Fan-Out (Step 4a)
IF query is straightforward:
Proceed to Direct Search (Step 4b)
Phase 2: Information Retrieval
4a. ๐ Query Fan-Out (Complex Queries)
Gemini 2.0/2.5 models break down query into multiple subtopics
System generates numerous related searches automatically
Executes concurrent searches for comprehensive coverage
4b. ๐ฏ Direct Search (Simple Queries)
Single search pathway activated
Standard query processing without fan-out
5. ๐ Comprehensive Information Retrieval
High-quality web content retrieval using Google Search API
Knowledge Graph access for structured information
Real-time data sources (shopping, news, current events)
REALM textual retrievers with neural network-based selection
MIPS (Maximum Inner Product Search) for top-k document identification
Phase 3: AI Processing & Draft Creation
6. ๐ค AI Processing Initiation
Custom Gemini model begins information integration
Multi-step reasoning applied to retrieved content
RAG (Retrieval Augmented Generation) framework activated
7. ๐ Drafting Strategy Decision
IF using SPECULATIVE RAG approach:
Proceed to Parallel Draft Generation (Step 8a)
IF using standard approach:
Proceed to Single Draft Generation (Step 8b)
8a. โก Parallel Draft Generation
Smaller specialist LLM creates multiple drafts simultaneously
Each draft generated from distinct document subsets
Reduces position bias and provides diverse perspectives
Minimizes input tokens per draft for efficiency
8b. ๐ Single Draft Generation
Full context processing by main LLM
Direct summarization from complete document set
9. โ๏ธ Abstractive Summarization
PEGASUS Gap Sentences Generation (GSG) technique applied
Creates novel phrases and sentences (doesn't just copy text)
Generates fluent, natural language output
Instruction tuning (FLAN) ensures proper task execution
Phase 4: Quality Assurance & Verification
10. โ Factual Verification
Larger generalist LLM (Gemini 2.0/2.5) performs verification pass
QA framework assessment checks information accuracy
Round-trip consistency method validates claims
Hallucination detection and mitigation
Attribution to source documents verified
11. ๐ก๏ธ Safety & Bias Filtering
Constitutional AI policies applied
LaMDA classifiers scan for harmful content
Policy violation detection activated
Value alignment verification performed
Red team tested scenarios checked
12. ๐ Confidence Assessment
System evaluates response quality score
Context sufficiency analysis performed
Helpfulness evaluation completed
Rationale-based confidence scoring
13. ๐ฆ First Confidence Gate
IF confidence is LOW:
Route to Traditional Search Results (Step 19)
IF confidence is HIGH:
Proceed to Source Attribution (Step 14)
Phase 5: Output Preparation
14. ๐ Source Attribution
Response Linkifying Engine activates
Verifiable links inserted into summary
Citation formatting applied
Markdown formatting for source references
15. ๐ฏ Final Quality Gate
Overall response assessment performed
User value evaluation completed
16. ๐ฅ Display Decision Point
IF AI Overview meets quality standards:
Show AI Overview (Step 17)
IF quality concerns remain:
Show Traditional Search Results (Step 19)
Phase 6: User Output & Display
17. ๐ค AI Overview Displayed
Natural language summary prominently shown
Source links clearly visible and clickable
Confidence indicators displayed when appropriate
AI-organized result categories with generated headlines
Clear attribution to original sources
18. ๐ Traditional Search Results (Fallback)
Standard web links displayed instead
Snippet previews provided
No AI-generated summary shown
Users can explore sources directly
I've got a long, long list of the sources I used, with NotebookLM, to assemble this information, which you can find at the very, very end of the newsletter. Props to Mike King over at iPullRank for the inspiration and the reminder that Google Patents is a rich source of information (and one preferred by the late, great Bill Slawski of SEO by the Sea) alongside academic research by Google.
So what? What are we supposed to learn from this massive, tangled mess? What can we know?
Phase 2 is the most critical part of the process. It's the heart of where Google is getting its information for AI Overviews. And it's powered by... Google Search. Which in turn means if you want to do well with AI Overviews... do well with Google Search.
That's probably not the super high tech, super mysterious AI takeaway you were hoping for, but it's the truth.
So for all those folks who are saying, "Fire your SEO agency! Everything is different now!" No, no it isn't. And this is not a terrible surprise. For a system like AI Overviews to work well AND quickly, Google has to leverage the data it already has and has spent decades tuning.
Additionally, this is an area where we can get solid data, because there are companies like Semrush and AHREFS that are running thousands of cheap simulations with known search queries (like "best PR firm boston") to see what triggers AI overviews and what the results are - and so can you. You could literally just run through your entire SEO keyword list and see what comes up (or better yet, automate the process).
Despite appearing like a conversation with generative AI, Google AI Mode shares much more in common with AI Overviews (such as query fan-out) than talking straight to Gemini. That's why it's in category 2, AI-assisted search, rather than AI replacements for search.
Google has also said we'll get some level of AI Mode data in our Google Search Console feeds.
AI Replacements for Search
This brings us to the third bucket of AI optimization, which is AI replacements for search. These are the consumer-facing apps that we all know and enjoy, from ChatGPT to Gemini to Claude, in mobile app form or website.
You might say, okay, how is this different than the AI models? The models are the engines of generative AI. A web interface or a mobile app like the ChatGPT app contain the same engine, but there's a lot of extra stuff around it. OpenAI built the rest of the car, just like the Claude app has a lot of stuff around it.
This is necessary because AI models are useless without an interface for the average person. AI interface makers like OpenAI have added a ton of features to make their apps user-friendly, accessible, and simple.
Features like web search and search grounding.
What does this mean? When a generative AI tool like ChatGPT has low confidence in its response, it can use its own built-in web search to validate that it's not making things up.
Here's a simple example. Suppose I ask the GPT-4.1 model this question:
"What good news has Quarterhorse PR CEO Nicole Bestard had to say recently?"
The response it gives is rooted in the knowledge it was trained on, which is to say not much:
Recently, Quarterhorse PR CEO Nicole Bestard shared good news regarding her companyโs growth and progress. Most notably, she celebrated Quarterhorse PRโs 1-year anniversary, acknowledging both the companyโs successful journey over the past year and the support from clients and partners. Nicole Bestard expressed excitement for the future and gratitude for the relationships built, marking this milestone as a testament to her teamโs hard work and the trust their clients have placed in Quarterhorse PR.
Given that Quarterhorse was founded in 2020, it's safe to say GPT-4.1 doesn't know much on the topic.
Ask the same question in ChatGPT using the GPT-4.1 model, and you get a very different, more accurate story:
Nicole Bestard, CEO & Principal of Quarter Horse PR, has shared some uplifting and inspiring โgood newsโ recentlyโhere are a few standouts:
๐ Solar Farm + Sheep Story
On March 7, 2025, she posted on LinkedIn about bringing eight national and local reporters to central Georgia to experience a solar farm operated by SiliconโฏRanchโฏCorporation. They watched ewes give birth, helped feed lambs, and saw how regenerative farming is being integrated with solar energy. The goal: cultivate stories that focus on tangible, climate-friendly progress and sustainability.
That is driven by its internal search.
Now, revisiting part 2, there is absolutely, positively no way to know what people are typing into tools like ChatGPT. A company like OpenAI is NEVER going to share that information willingly because of both privacy and profit reasons. Privacy, to give users the illusion that what they type into AI tools is private, and profit, because user-submitted information is the crown jewel of any AI company. User-submitted information is how AI companies calibrate and build better models.
But we can know quite a lot. First, we can know what sources these tools draw from - and they all draw from different sources. That same query across three different apps gives a variety of different sources. Gemini only pulls from the QHPR website and PR Daily. Claude hit up LinkedIn, Twitter, PR Daily, PR Week, and a few dozen other sites. ChatGPT hit up LinkedIn, the QHPR website, YouTube, BusinessWire, and a few dozen other sites.
That tells us a great deal - namely, where we should be placing information so that it's findable by these tools.
The second major thing we can know is when a human clicks on a link from a generative AI tool, we can see the referral traffic on our websites. We have no idea what they typed to spur the conversation, but we can see the human clicked on something AND visited our website, which also means we can know what page they landed on, what pages they visited, and whether they took any high-value actions.
I have an entire guide over on the Trust Insights website - free, no information to give, nothing to fill out - on how to set up a Google Analytics report to determine this.
Part 4: What Can You Optimize?
Okay, now that we've been through an exhaustive tour of the three categories and their systems, we can talk about what we have control over, what we can optimize.
If you missed it, I have an entire 8-page guide about how to optimize for AI, free of cost (there is a form to fill out) here. I'm not going to reprint the entire thing here.
Here's the very short version.
For category 1, AI models themselves, they need training data. Lots of training data - and AI model makers ingest training data very infrequently. The tests we did for category 1, measuring when the last update a model had about us, is a good proxy for how often those updates happen. It's about once a year, give or take.
That means you have to have a LOT of content, all over the web, in as many places as possible, in the hopes that the next time a model maker scoops up new training data, your piles of new data will be in it. In terms of how to do that, it's all about generating content, so be as many places as you can be. I've said in the past, one of my blanket policies is to say yes to any podcast that wants to interview me (and that's still my policy) as long as the transcripts and materials are being posted in public - especially on YouTube.
For category 2, AI-assisted search, that is still basically SEO. Yup, good old-fashioned SEO the way you've been doing it all along. Create great quality content, get links to it, get it shared, get it talked about as many places as you can. For example, you'll note in the example above that both Claude and ChatGPT hit up LinkedIn quite hard for data, so be everywhere and have your content talked about everywhere.
For category 3, AI replacements for search, do the exercises I recommended. Take some time to do the searches and questions and discussions in the major AI tools about your brand, your industry, your vertical, but instead of looking at the generated text, look at the sources. Look at where AI tools are getting their information for search grounding, because that is your blueprint for where to invest your time and effort.
Part 5: Wrapping Up
There is a TON of snake oil being peddled in the AI optimization space. Everyone and their cousin has a hot take, and as actual SEO expert Lily Ray said recently, a lot of people have very strong opinions about AI optimization backed up by absolutely no data whatsoever.
That's why I spent so much time finding and aggregating the patents, research papers, and materials. I don't know - and you don't know either - what people are typing into systems like ChatGPT and Gemini. But we can absolutely know how the systems are architected overall and what we should be doing to show up in as many places as possible to be found by them.
Later this summer, look for a full guide from Trust Insights about this topic, because (a) it's too important to leave to the snake oil salesmen and (b) I need to get more mileage out of all this data I collected and processed.
As with my Unofficial LinkedIn Algorithm Guide for Marketers, it should be abundantly clear that there is no "AI algorithm" or any such nonsense. There are, instead, dozens of complex, interacting, sometimes conflicting systems at play that are processing data at scale for users. There's little to no chance of any kind of "hack" or "growth power move" or whatever other snake oil is being peddled about AI optimization. Instead, there's just hard work and being smart about where and how you create.
How Was This Issue?
Rate this week's newsletter issue with a single click/tap. Your feedback over time helps me figure out what content to create for you.
Here's The Unsubscribe
It took me a while to find a convenient way to link it up, but here's how to get to the unsubscribe.
https://almosttimely.substack.com/action/disable_email
Share With a Friend or Colleague
If you enjoy this newsletter and want to share it with a friend/colleague, please do. Send this URL to your friend/colleague:
https://www.christopherspenn.com/newsletter
For enrolled subscribers on Substack, there are referral rewards if you refer 100, 200, or 300 other readers. Visit the Leaderboard here.
Advertisement: Bring Me In To Speak At Your Event
Elevate your next conference or corporate retreat with a customized keynote on the practical applications of AI. I deliver fresh insights tailored to your audience's industry and challenges, equipping your attendees with actionable resources and real-world knowledge to navigate the evolving AI landscape.
๐ If this sounds good to you, click/tap here to grab 15 minutes with the team to talk over your event's specific needs.
If you'd like to see more, here are:
ICYMI: In Case You Missed It
This week, John and I kicked off the Summer Makeover Series with podcast transcription automation.
Almost Timely News: ๐๏ธ How to Use Generative AI For Analytics (2025-06-22)
So What? Summer Makeover Series: Podcast Transcription Automation
Skill Up With Classes
These are just a few of the classes I have available over at the Trust Insights website that you can take.
Premium
Free
Powering Up Your LinkedIn Profile (For Job Hunters) 2023 Edition
Building the Data-Driven, AI-Powered Customer Journey for Retail and Ecommerce, 2024 Edition
The Marketing Singularity: How Generative AI Means the End of Marketing As We Knew It
Advertisement: New AI Course!
In my new Generative AI Use Cases for Marketers course, you'll learn AI use cases in an entirely new way. The days of simply sitting back and staring at a bewildering collection of confusing use cases and success stories is over. They weren't all that helpful to begin with.
In this course, instead, you'll learn the 7 major categories of generative AI use cases with 3 examples each - and you'll do every single one of them. Each example comes with prompts, sample data, and walkthroughs so you can learn hands-on how to apply the different use cases.
You'll also learn how to put the use case categories together so you can identify your own use cases, set up effective AI strategy for your real world work, and make generative AI work for you.
Every course module comes with audio to go for when you want to listen, like at the gym or while cooking, plus transcripts, videos, closed captions, and data.
Sign up today by visiting trustinsights.ai/usecasescourse
๐ Pre-order my new course, Generative AI Use Cases for Marketers!
What's In The Box? Here's a 5 Minute Tour
Here's a 5 minute video tour of the course so you can see what's inside.
Get Back to Work
Folks who post jobs in the free Analytics for Marketers Slack community may have those jobs shared here, too. If you're looking for work, check out these recent open positions, and check out the Slack group for the comprehensive list.
Director Of Ai & Data Technology Delivery at Aventis Solutions
Vice President Marketing Operations at Independent Insurance Agents and Brokers of America
Advertisement: Free AI Strategy Kit
Grab the Trust Insights AI-Ready Marketing Strategy Kit! It's the culmination of almost a decade of experience deploying AI (yes, classical AI pre-ChatGPT is still AI), and the lessons we've earned and learned along the way.
In the kit, you'll find:
TRIPS AI Use Case Identifier
AI Marketing Goal Alignment Worksheet
AI Readiness Self-Assessment (5P & 6Cs)
12-Month AI Marketing Roadmap Template
Basic AI ROI Projection Calculator
AI Initiative Performance Tracker
If you want to earn a black belt, the first step is mastering the basics as a white belt, and that's what this kit is. Get your house in order, master the basics of preparing for AI, and you'll be better positioned than 99% of the folks chasing buzzwords.
๐ Grab your kit for free at TrustInsights.ai/aikit today.
How to Stay in Touch
Let's make sure we're connected in the places it suits you best. Here's where you can find different content:
My blog - daily videos, blog posts, and podcast episodes
My YouTube channel - daily videos, conference talks, and all things video
My company, Trust Insights - marketing analytics help
My podcast, Marketing over Coffee - weekly episodes of what's worth noting in marketing
My second podcast, In-Ear Insights - the Trust Insights weekly podcast focused on data and analytics
On Bluesky - random personal stuff and chaos
On LinkedIn - daily videos and news
On Instagram - personal photos and travels
My free Slack discussion forum, Analytics for Marketers - open conversations about marketing and analytics
Listen to my theme song as a new single:
Advertisement: Ukraine ๐บ๐ฆ Humanitarian Fund
The war to free Ukraine continues. If you'd like to support humanitarian efforts in Ukraine, the Ukrainian government has set up a special portal, United24, to help make contributing easy. The effort to free Ukraine from Russia's illegal invasion needs your ongoing support.
๐ Donate today to the Ukraine Humanitarian Relief Fund ยป
Events I'll Be At
Here are the public events where I'm speaking and attending. Say hi if you're at an event also:
AFT, Washington DC, July 2025
AMA Pennsylvania, York, August 2025
SMPS, Denver, October 2025
Marketing AI Conference, Cleveland, October 2025
MarketingProfs B2B Forum, Boston, November 2025
There are also private events that aren't open to the public.
If you're an event organizer, let me help your event shine. Visit my speaking page for more details.
Can't be at an event? Stop by my private Slack group instead, Analytics for Marketers.
Required Disclosures
Events with links have purchased sponsorships in this newsletter and as a result, I receive direct financial compensation for promoting them.
Advertisements in this newsletter have paid to be promoted, and as a result, I receive direct financial compensation for promoting them.
My company, Trust Insights, maintains business partnerships with companies including, but not limited to, IBM, Cisco Systems, Amazon, Talkwalker, MarketingProfs, MarketMuse, Agorapulse, Hubspot, Informa, Demandbase, The Marketing AI Institute, and others. While links shared from partners are not explicit endorsements, nor do they directly financially benefit Trust Insights, a commercial relationship exists for which Trust Insights may receive indirect financial benefit, and thus I may receive indirect financial benefit from them as well.
Thank You
Thanks for subscribing and reading this far. I appreciate it. As always, thank you for your support, your attention, and your kindness.
See you next week,
Christopher S. Penn
Appendix: Long List of Citations
(2022). Improving alignment of dialogue agents via targeted human judgements. arXiv preprint arXiv:2209.14375.
Badr, I., Guo, Y., & Mansuri, I. R. (2024). Proactive query and content suggestion with generative model generation and answer (U.S. Patent No. 12,079,292 B1). U.S. Patent and Trademark Office.
Badr, I., Guo, Y., & Mansuri, I. R. (2025). Proactive query and content suggestion with generative model generation and answer (U.S. Patent App. No. 2025/0094511 A1). U.S. Patent and Trademark Office.
Barham, P., Chowdhery, A., Dean, J., Ghemawat, S., Hand, S., Hurt, D., Isard, M., Lim, H., Pang, R., Roy, S., Saeta, B., Schuh, P., Sepassi, R., El Shafey, L., Thekkath, C. A., & Wu, Y. (2022). PATHWAYS: ASYNCHRONOUS DISTRIBUTED DATAFLOW FOR ML. arXiv preprint arXiv:2203.12533.
Gemini Team, et al. (2023). Gemini: A Family of Highly Capable Multimodal Models. arXiv preprint arXiv:2312.11805.
Gemini Team, et al. (2024). Gemini 1.5: Unlocking multimodal understanding across millions of tokens of context. arXiv preprint arXiv:2403.05530.
Google. (2024). Expanding AI Overviews and introducing AI Mode..
Google. (2024). Google I/O 2024: New generative AI experiences in Search..
Google. (n.d.). Innovations in AI: Search, Retrieval, and Generative Models..
Gray, M. K., Blitzer, J., Budzise-Szakal, C., Mansuri, I. R., Madhavan, J., Prakash, A., Prasad, A., Hubar, A., Luan, D., Schoff, P., Shah, A., Rafouei, M., Amor, Z. B., & Badr, I. (2023). Generative summaries for search results (U.S. Patent No. 11,769,017 B1). U.S. Patent and Trademark Office.
Guidroz, T., Ardila, D., Li, J., Mansour, A., Jhun, P., Gonzalez, N., Ji, X., Sanchez, M., Kakarmath, S., Bellaiche, M. M. J., Garrido, M. ร., Ahmed, F., Choudhary, D., Hartford, J., Xu, C., Echeverria, H. J. S., Wang, Y., Shaffer, J., Cao, E., Matias, Y., . . . Duong, Q. (2025). LLM-based Text Simplification and its Effect on User Comprehension and Cognitive Load. arXiv preprint arXiv:2505.01980.
Guidroz, T., Li, J., Mansour, A., Jhun, P., Gonzalez, N., Ji, X., Sanchez, M., Bellaiche, M. M. J., Garrido, M. ร., Ahmed, F., Choudhary, D., Hartford, J., Xu, C., Echeverria, H. J. S., Wang, Y., Shaffer, J., Cao, E., Matias, Y., Hassidim, A., Webster, D. R., . . . Duong, Q. (n.d.). Making complex text understandable: Minimally-lossy text simplification with Gemini. Google Research..
Guu, K., Lee, K., Tung, Z., Pasupat, P., & Chang, M.-W. (2020). REALM: Retrieval-Augmented Language Model Pre-Training. arXiv preprint arXiv:2002.08909.
Lee, K., Chang, M.-W., & Toutanova, K. (2021). Retrieval-augmented language model pre-training and fine-tuning (U.S. Patent No. 11,003,865 B1). U.S. Patent and Trademark Office.
Liu, H., Le, Q. V., Zhou, Y., Xu, T., Zhao, Y., Yu, M., Dai, A. M., & Laudon, J. (2025). Routing to expert subnetworks in mixture-of-experts neural networks (U.S. Patent App. No. 2025/0131251 A1). U.S. Patent and Trademark Office.
Maynez, J., Narayan, S., Bohnet, B., & McDonald, R. (2020). On faithfulness and factuality in abstractive summarization. arXiv preprint arXiv:2005.00661.
Morris, M. R., Doshi, T., Delos Santos, R., Duke, T., Soraker, J., Zevenbergen, B., Prabhakaran, V., Diaz, M., Hutchinson, B., Olson, K., Molina, A., Hoffman-John, E., Lee, J., Aroyo, L., Rajakumar, R., Butryna, A., Lamm, M., Kuzmina, V., Fenton, J., Cohen, A., . . . Le, Q. (2022). LaMDA: Language Models for Dialog Applications. arXiv preprint arXiv:2201.08239.
Rafouei, M., Shah, A. G., Wei, Q., Mansuri, I. R., Brown, B., & Piqueras, E. (2024). Search with stateful chat (U.S. Patent App. No. 2024/0289407 A1). U.S. Patent and Trademark Office.
Rashtchian, C., & Juan, D.-C. (2025, May 14). Deeper insights into retrieval augmented generation: The role of sufficient context. Google Research..
Rashtchian, C., & Juan, D.-C. (2025, May 14). Deeper insights into retrieval augmented generation: The role of sufficient context. Google Research..
Sadar, A., Ou, A., & Shah, A. G. (2024). Search with machine-learned model-generated queries (U.S. Patent No. 11,941,678 B1). U.S. Patent and Trademark Office.
Schmid, P. (n.d.). google-gemini-gemini-fullstack-langgraph-quickstart.txt.
Smus, B., Budholia, V., Ross, C., & Obuobi, A. (2023). Summarization based on timing data (U.S. Patent App. No. 2023/0367960 A1). U.S. Patent and Trademark Office.
Tay, Y., Liu, P. J., Pang, R., Tran, V., Vinh, Q., Le, L. T., He, C., Williams, H., Zeng, F., Hu, Q., Mizer, J., & Zhou, Y. (2023). Machine learning models as a differentiable search index for directly predicting resource retrieval results (WO 2023/154385 A1). World Intellectual Property Organization.
Wang, Z., Wang, Z., Le, L. T., Zheng, H. S., Mishra, S., Perot, V., Zhang, Y., Mattapalli, A., Taly, A., Shang, J., Lee, C.-Y., & Pfister, T. (2024). SPECULATIVE RAG: ENHANCING RETRIEVAL AUGMENTED GENERATION THROUGH DRAFTING. arXiv preprint arXiv:2407.08223.
Wang, Z., Wang, Z., Le, L. T., Zheng, H. S., Mishra, S., Perot, V., Zhang, Y., Mattapalli, A., Taly, A., Shang, J., Lee, C.-Y., & Pfister, T. (n.d.). Speculative RAG: Enhancing retrieval augmented generation through drafting..
Wei, J., Bosma, M., Zhao, V., & Yu, A. W., et al. (2021). FLAN: Finetuned Language Models are Zero-Shot Learners. arXiv preprint arXiv:2109.01652.
Zhang, J., Zhao, Y., Saleh, M., & Liu, P. J. (2020). PEGASUS: Pre-training with Extracted Gap-sentences for Abstractive Summarization. Proceedings of the 37th International Conference on Machine Learning, Vienna, Austria, PMLR 119.
Thank you for the explication! This is super helpful. To be fair, there is sufficient evidence that people *are* just typing in what would be a normal google search into ChatGPT. But people also used to approach my spouse when he worked at Trader Joe's and just command "bread," no context, like he was a search engine.
We searched with keywords for 20 years. Old habits die hard.
I will also add here what I add everywhere, which is, that while we can have no clue about the exact words people are typing, we can certainly make a very good guess by understanding how people have searched in the past and knowing why people buy. We are not inventing new vocabulary words or purchase criteria in most verticals. We are just learning how to advertise in a new distribution system. We're going back to the Wanamaker problem, as we always will.
Great article--thank you for the shoutout! In regards to the wording discussion in the comments, it's interesting that the latter version of ChatGPT corrects the prompt's misspelling/compounding of quarterhorse to our correct company name of Quarter Horse.
In both examples, I'm glad the AI recognizes both the compound word and the proper noun name of the racehorse we are named for, as they are often confused and semantically interchangeable. I prompted Gemini with the same question, and it also answered not with the incorrect prompt name, but with the proper name of the company (and the horse). Just like with how search guesses at what you meant to query based on the next nearest word (or most popular similarly written search).
All this to say, I won't look the gift horse in the mouth any further, and am glad to know we're recognizable in each instance!