AI Search Changes Everything – Is Your Organization Built To Compete? via @sejournal, @billhunt

Search has changed. Have you?

Search is no longer about keywords and rankings. It’s about relevance, synthesis, and structured understanding.

In the AI-powered era of Google Overviews, ChatGPT-style assistants, and concept-level rankings, traditional SEO tactics fall short.

Content alone won’t carry you. If your organization isn’t structurally and strategically aligned to compete in this new paradigm, you’re invisible even if you’re technically “ranking.”

This article builds on the foundation laid in my earlier article, From Building Inspector To Commissioning Authority,” where I argued that SEO must shift from reactive inspection to proactive orchestration.

It also builds upon my exploration of the real forces reshaping search, including the rise of Delphic Costs, where brands are extracted from the customer journey without attribution, and the organizational imperative to treat visibility as everyone’s responsibility, not just a marketing key performance indicator (KPI).

And increasingly, it’s not just about your monetization. It’s about the platform.

The Three Shifts Reshaping Search

1. Google AI Overviews: The Answer Layer Supersedes The SERP

Google is bypassing traditional listings with AI-generated answers. These overviews synthesize facts, concepts, and summaries across multiple sources.

Your content may power the answer, but without attribution, brand visibility, or clicks. In this model, being the source is no longer enough; being the credited authority is the new battle.

2. Generative Assistants: New Gatekeepers To Discovery

Tools like ChatGPT, Perplexity, and Gemini collapse the search journey into a single query/answer exchange. They prioritize clarity, conceptual alignment, and structured authority.

They don’t care about the quantity of backlinks; they care about structured understanding. Organizations relying on domain authority or legacy SEO tactics are being leapfrogged by competitors who embrace AI-readable content.

3. Concept-Based Ranking: From Keywords To Entities And Context

Ranking is no longer determined by exact-match phrases. It’s determined by how well your content reflects and reinforces the concepts, entities, and context behind a query.

AI systems think in knowledge graphs, not spreadsheets. They interpret meaning through structured data, relationships between entities, and contextual signals.

These three shifts mean that success now depends on how well your organization can make its expertise machine-readable and contextually integrated into AI ecosystems.

A New Era Of Monetization And Data Harvesting

Search platforms have evolved from organizing information to owning outcomes. Their mission is no longer to guide users to your site; it’s to keep users inside their ecosystem.

The more they can answer in place, the more behavioral data they collect, and the more control they retain over monetization.

Today, your content competes not just with other brands but with the platforms themselves. They’re generating “synthetic content” derived from your data – packaged, summarized, and monetized within their interfaces.

As Dotdash Meredith CEO Neil Vogel put it: “We were in the business of arbitrage. We’d buy traffic for a dollar, monetize it for two. That game is over. We’re now in the business of high-quality content that platforms want to reward.”

Behavioral consequence: If your content can’t be reused, monetized, or trained against, it’s less likely to be shown.

Strategic move: Make your content AI-friendly, API-ready, and citation-worthy. Retain ownership of your core value. Structured licensing, schema, and source attribution matter more than ever.

This isn’t just about visibility. It’s about defensibility.

The Strategic Risks

Enterprises that treat search visibility as a content problem – not a structural one – are walking blind into four key risks:

  • Disintermediation: You lose traffic, attribution, and control when AI systems summarize your insights without directing users to you. In an AI-mediated search world, your value can be extracted while your brand is excluded.
  • Market Dilution: Nimbler competitors who better align with AI content requirements will surface more often, even if they have less experience or credibility. This creates a reverse trust dynamic: newcomers gain exposure by leveraging the machine’s strengths, while legacy players lose visibility.
  • Performance Blind Spots: Traditional KPIs no longer capture the real picture. Traffic may appear stable while influence and presence erode behind the scenes. Executive dashboards often miss this erosion because they’re still tuned to clicks, not concept penetration or AI inclusion.
  • Delphic Costs: This, as defined by Andrei Broder and Preson McAfee, refers to the expenses incurred when AI systems extract your expertise without attribution or downstream benefits, resulting in brand invisibility despite active contributions. Being referenced but not represented becomes a strategic liability.

Are You Built To Compete?

Here’s a five-pillar diagnostic framework to assess your organization’s readiness for AI search:

1. Content Structure

  • Do you use schema markup to define your content’s meaning?
  • Are headings, tables, lists, and semantic formats prioritized?
  • Is your content chunked in ways AI systems can easily digest?
  • Are your most authoritative explanations embedded into the page using clear, concise writing and answer-ready?

2. Relevance Engineering

  • Do you map queries to concepts and entities?
  • Is your content designed for entity resolution, not just keyword targeting?
  • Are you actively managing topic clusters and knowledge structures?
  • Have you audited your internal linking and content silos to support knowledge graph connectivity?

3. Organizational Design (Shared Accountability)

  • Who owns “findability” in your organization?
  • Are SEO, content, product, and dev teams aligned around structured visibility?
  • Is there a commissioning authority that ensures strategy alignment from the start?
  • Do product launches and campaign rollouts include a visibility readiness review?
  • Are digital visibility goals embedded in executive and departmental KPIs?

In one example, a SaaS company I advised implemented monthly “findability sprints,” where product, dev, and content teams worked together to align schema, internal linking, and entity structure.

The result? A 30% improvement in AI-assisted surfacing – without publishing a single new page.

4. AI Feedback Loops

  • Are you tracking where and how your content appears in AI Overviews or assistants?
  • Do you have visibility into lost attribution or uncredited brand mentions?
  • Are you using tools or processes to monitor AI surface presence?
  • Have you incorporated AI visibility into your reporting cadence and strategic reviews?

5. Modern KPIs

  • Do your dashboards still prioritize traffic volume over influence?
  • Are you measuring presence in AI systems as part of performance?
  • Do your teams know what “visibility” actually means in an AI-dominant world?
  • Are your KPIs evolving to include citations, surface presence, and non-click influence metrics?

The Executive Mandate: From Visibility Theater To Strategic Alignment

Organizations must reframe search visibility as digital infrastructure, not a content marketing afterthought.

Just as commissioning authorities ensure a building functions as designed, your digital teams must be empowered to ensure your knowledge is discoverable, credited, and competitively positioned.

AI-readiness isn’t about writing more content. It’s about aligning people, process, and technology to match how AI systems access and deliver value. You can’t fix this with marketing alone. It requires a leadership-driven transformation.

Here’s how to begin:

  1. Reframe SEO as Visibility Engineering: Treat it as a cross-functional discipline involving semantics, structure, and systems design.
  2. Appoint a Findability or Answers Leader: This role connects the dots across content, code, schema, and reporting to ensure you are found and answering the market’s questions.
  3. Modernize Metrics: Track AI visibility, entity alignment, and concept-level performance – not just blue links.
  4. Run an AI Exposure Audit: Understand where you’re showing up, how you’re credited, and most critically, where and why you’re not. Just ask the AI system, and it will tell you exactly why you were not referenced.
  5. Reward Structural Alignment: Incentivize teams not just for publishing volume, but for findability performance. Celebrate contributions to visibility the same way you celebrate brand reach or campaign success. Make visibility a cross-team metric.

Final Thought: You Can’t Win If You’re Not Represented

AI is now the front end of discovery. If you’re not structured to be surfaced, cited, and trusted by machines, you’re losing silently.

You won’t fix this with a few blog posts or backlinks.

You fix it by building an organization designed to compete in the era of machine-mediated relevance.

This is your commissioning moment – not just to inspect the site after it’s built, but to orchestrate the blueprint from the start.

Welcome to the new search. Let’s build for it.

More Resources:


Featured Image: Master1305/Shutterstock

Brave Announces AI Grounding API With Plans Starting At Free via @sejournal, @martinibuster

Brave Search announced the release of AI Grounding with the Brave Search API, a way to connect an AI system to grounding in search to reduce hallucinations and improve answers. The API is available in Free, Base AI, and Pro AI plans.

The Brave Search API is for developers and organizations that want to add AI grounding from authoritative web information to their AI applications. The Brave API supports agentic search, foundation model training, and creating search-enabled applications.

State Of The Art Performance (SOTA)

Brave’s announcement says that their AI Grounding API enables state of the art performance in both single-search and multi-search configurations, outperforming competitors in accuracy, claiming they can answer more than half of all questions with a single search.

According to Brave:

“Brave can answer more than half of the questions in the benchmark using a single search, with a median response time of 24.2 seconds. On average (arithmetic mean), answering these questions involves issuing 7 search queries, analyzing 210 unique pages (containing 6,257 statements or paragraphs), and takes 74 seconds to complete. The fact that most questions can be resolved with just a single query underscores the high quality of results returned by Brave Search.”

Pricing

There are three pricing tiers:

  • Free AI
    1 query/second and a limit of 5,000 queries/month
  • Base AI
    $5.00 per 1,000 requests
    A limit of up to 20 queries/second
    20M queries/month
    Rights to use in AI apps
  • Pro AI
    $9.00 per 1,000 requests
    A limit of up to 50 queries/second
    Unlimited queries/month
    Rights to use in AI apps

Brave’s AI Grounding API offers a reliable way to supply AI systems and apps with trustworthy information from across the web. Its independence and privacy practices make it a viable choice for developers building search-enabled AI applications.

Read Brave’s announcement:

Introducing AI Grounding with Brave Search API, providing enhanced search performance in AI applications

Featured Image by Shutterstock/Mamun_Sheikh

Google Is Testing An AI-Powered Finance Page via @sejournal, @martinibuster

Google announced that they’re testing a new AI-powered Google Finance tool. The new tool enables users to ask natural language questions about finance and stocks, get real-time information about financial and cryptocurrency topics, and access new charting tools that visualize the data.

Three Ways To Access Data

Google’s AI finance page offers three ways to explore financial data:

  1. Research
  2. Charting Tools
  3. Real-Time Data And News

Screenshot Of Google Finance

The screenshot above shows a watchlist panel on the left, a chart in the middle, a “latest updates” section beneath that, and a “research” section on the right hand panel.

Research

The new finance page enables users to ask natural language questions about finance, including the stock market, and the AI will return comprehensive answers, plus links to the websites where the relevant answers can be found.

Closeup Screenshot Of Research Section

Charting Tools

Google’s finance page also features charting tools that enable users to visualize financial data.

According to Google:

“New, powerful charting tools will help you visualize financial data beyond simple asset performance. You can view technical indicators, like moving average envelopes, or adjust the display to see candlestick charts and more.”

Real-Time Data

The new finance page also provides real-time data and tools, enabling users to explore finance news, including cryptocurrency information. This part features a live news feed.

The AI-powered page will roll out over the next few weeks on Google.com/finance/.

Read more at Google:

We’re testing a new, AI-powered Google Finance.

Featured Image by Shutterstock/robert_s

How To Stay Visible in AI Search [Webinar] via @sejournal, @lorenbaker

AI search is here. Are you ready for the new rules?

The SEO game has changed. Traditional strategies are no longer enough, and some brands are getting lost in the shift to AI-powered search results.

Join Wayne Cichanski on August 20, 2025 for an exclusive webinar sponsored by iQuanti. Learn how to adapt your SEO strategy and site architecture for AI-driven queries and remain competitive in this new search era.

In this session, you’ll discover:

  • Why user experience, schema, and site architecture are now just as important as keywords
  • Practical steps to remain visible and competitive in evolving search results
  • How to position your brand for discovery in AI-driven queries, not just rankings

Why this session is essential:

With generative AI reshaping search results across platforms like Google, Bing, and ChatGPT, it is crucial to rethink how your content is structured and how people interact with your brand in AI search. Do not get left behind. Optimize for AI-driven search now.

Register today for actionable insights and a roadmap to success in the AI search era. If you cannot attend live, do not worry. Sign up anyway and we will send you the full recording.

OpenAI Launches GPT-5 In ChatGPT To All Users via @sejournal, @MattGSouthern

OpenAI has released GPT-5, now the default model in ChatGPT for all users, including those on the free tier.

The new model is positioned as OpenAI’s most capable and reliable system to date. OpenAI emphasizes a stronger focus on accuracy, instruction following, and long-form reasoning.

Available Now To All ChatGPT Users

For the first time, OpenAI is making its latest flagship model available to free users.

GPT-5 is rolling out now to Free, Plus, Pro, and Team accounts, with support for Enterprise and Education expected next week.

Free-tier access includes basic usage of GPT-5, with requests routed to a smaller “GPT-5 mini” variant once limits are reached.

Paid subscribers receive higher usage limits, and Pro users gain access to GPT-5 Pro, a version designed for more complex, resource-intensive tasks.

Accuracy, Reasoning, and Transparency Take Priority

According to OpenAI, GPT-5 significantly reduces hallucinated facts and is more likely to admit when it lacks the context to provide a reliable answer.

Evaluations show GPT-5 produces 45% fewer factual errors than GPT-4o, with up to 80% fewer errors when deeper reasoning is enabled.

The model also performs better on benchmarks tied to real-world problem-solving, such as coding, legal analysis, and health-related queries.

What’s Changed in the System

Rather than a single model, GPT-5 acts as a dynamic system that automatically decides whether to respond quickly or think more deeply, depending on prompt complexity.

Users can also request explicit reasoning with natural language prompts like “think hard about this.”

Other updates include:

  • A redesigned safety system that favors partially helpful answers over blanket refusals
  • Reduced sycophantic responses and more honest communication about limitations
  • Improvements in coding performance, including front-end UI generation and debugging
  • Support for multimodal input, including charts and images

Looking Ahead

GPT-5 is positioned less as a revolutionary jump and more as an effort to build trust through accuracy, reliability, and broader access.

For SEOs and digital marketers who rely on AI tools for drafting, analysis, or ideation, GPT-5’s improvements may help reduce the time spent verifying or correcting outputs.


Featured Image: JarTee/Shutterstock

6 AI Marketing Myths That Are Costing You Money [Webinar] via @sejournal, @duchessjenm

Stop letting AI drain your budget. Learn how to make it work for you.

Think AI can fully run your marketing strategy on autopilot? 

Or that AI-generated content should deliver instant results? 

It is time to bust the AI myths that are slowing you down and costing you money.

Join Bailey Beckham, Senior Partner Marketing Manager at CallRail, and Jennifer McDonald, Senior Marketing Manager at Search Engine Journal, on August 21, 2025, for an exclusive webinar. Get the insights you need to stop wasting time and money and start leveraging AI the right way.

In this session, you will learn:

Why this session is essential:

AI tools can’t run your strategy on autopilot. You need to make smarter decisions, ask the right questions, and guide your AI tools to work for you, not against you. 

This webinar will help you unlock AI’s full potential and optimize your content to improve your marketing performance.

Register now to learn how to get your content loved by AI, LLMs, and most importantly, your audience. Can’t attend live? Don’t worry, sign up anyway, and we will send you the on-demand recording.

Google Says AI Clicks Are Better, What Does Your Data Say? via @sejournal, @MattGSouthern

Google’s latest blog post claims AI is making Search more useful than ever. Google says people are asking new kinds of questions, clicking on more links, and spending more time on the content they visit.

But with no supporting data or clear definitions, the message reads more like reassurance than transparency.

Rather than take Google at its word or assume the worst, you can use your own analytics to understand how AI in Search is affecting your site.

Here’s how to do that.

Google Says: “Quality Clicks” Are Up

In the post, Google says total organic traffic is “relatively stable year over year,” but that quality has improved.

According to the company, “quality clicks” are those where users don’t bounce back immediately, indicating they’re finding value in the destination.

This sounds good in theory, but it raises a few questions:

  • What does “slightly more” quality clicks mean?
  • Which sites are gaining, and which are losing?
  • And how is click quality being measured?

You won’t find those answers in Google’s post. But you can find clues in your own data.

1. Track Click-Through Rate On High-Volume Queries

If you suspect your site has lost ground due to AI Overviews, your first stop should be Google Search Console.

Try this:

  • Filter for top queries from the past 12 months.
  • Look at CTR changes before and after May 2024 (when AI Overviews began expanding).
  • Pay attention to queries that are longer, question-based, or likely to trigger summaries.

You may find impressions are holding steady or rising while CTR declines. That suggests your content is still being surfaced, but users may be getting their answers directly in Google’s AI-generated response.

2. Approximate “Quality Clicks” With Engagement Metrics

To test Google’s claim about higher quality clicks, you’ll need to look beyond Search Console.

In GA4, examine:

  • Engaged sessions (sessions lasting more than 10 seconds or including a conversion or multiple pageviews).
  • Average engagement time per session.
  • Scroll depth or video watch time, if applicable.

Compare these engagement metrics to the same period last year. If they’re improving, you may be getting more motivated visitors, supporting Google’s view.

But if they’re dropping, it could mean that AI Overviews are sending fewer, possibly less interested, visitors your way.

3. See Which Content Formats Are Gaining Visibility

Google says people are increasingly clicking on forums, videos, podcasts, and posts with “authentic voices.”

That aligns with its integration of Reddit and YouTube content into AI Overviews.

To see how this shift might be playing out for you:

  • Compare the performance of listicles, tutorials, and original reviews to more generic content.
  • If you create video or podcast content, track any uptick in referral traffic from Google.
  • Watch for changes in how your forum threads, product reviews, or community content perform compared to static pages.

You may find that narrative-style content, first-hand experiences, and multimedia formats are gaining traction, even if traditional evergreen pages are flat.

4. Watch For Redistribution, Not Just Declines

Google acknowledges that while overall traffic is stable, traffic is being redistributed.

That means some sites will lose while others gain, based on how well they align with evolving search behavior.

If your traffic has declined, it doesn’t necessarily mean your content isn’t ranking. It may be that the types of questions being asked and answered have changed.

Analyzing your top landing pages can help you spot patterns:

  • Are you seeing fewer entries on pages that used to rank for quick-answer queries?
  • Are in-depth or comparison-style pages gaining traffic?

The patterns you spot could help guide your content strategy.

Looking Ahead

When you rely on Search traffic, you deserve more than vague reassurances. Your analytics can help fill in the blanks.

By keeping an eye on your CTR, engagement, and how your content performs, you’ll get a better sense of whether AI in Search is helping you. This way, you can tweak your strategy to fit what works best for you.


Featured Image: Roman Samborskyi/Shutterstock

AI As Your Marketing Co-Pilot: How To Effectively Leverage LLMs In SEO & Content via @sejournal, @cshel

I remember seeing those “God is my co-pilot” bumper stickers since I was old enough to read them.

I was a precocious little agnostic, so they always struck me as weird. God can’t be your co-pilot because God isn’t a physical manifestation of someone who can help you drive a car.

I eventually figured out that “God is my co-pilot” was less a literal statement and more a declaration of faith that there is an omniscient presence available to help you navigate life’s construction zones (if you believe, anyway).

So, fast forward to 2025, and marketers have a new omniscient presence that they can put their faith in. Something that seems equally all-knowing but perhaps a little more … unpredictable.

AI.

Large language models (LLMs) – like ChatGPT, Claude, Gemini – feel delightfully divine when you first try them. They answer instantly, confidently, and often with an authority that makes you wonder if they do know everything.

But, spend enough time with these tools, and you discover something unsettling: AI isn’t just your god-like guide. It can also act like the devil, gleefully granting your wishes exactly as asked – and letting you suffer the consequences.

This is why the healthiest way to think of AI in your SEO and content workflows is as a co-pilot. Not God. Not Lucifer. But, a powerful partner that can elevate your work, if you exercise your free will (and make good choices).

The God-Like Qualities Of AI

There’s a reason AI feels god-like in a marketing context:

  • It seems omnipresent, embedded in your search results, your content management system (CMS), your analytics.
  • It delivers answers instantly, with confidence and authority.
  • It processes far more data than any human ever could, instantly finding patterns we mere mortals miss on the first (or third) pass.

Ask it to draft a content brief, summarize competitive search engine results pages (SERPs), generate topic clusters, or even shape a brand narrative – and it performs in seconds what would have taken you hours.

That kind of power can feel miraculous.

But, just as theologians remind us that God’s will is mysterious and not always aligned with ours, LLMs work on their own unknowable internal logic.

The outputs may not match your intent. The answer may not come in the form you wanted. And you may not even fully grasp why it chose the answer it did.

The Devilish Side Of AI

On the flip side, AI can also be a trickster: seductive, transactional, and literal. It will grant you exactly what you wish for – and sometimes that’s the worst thing possible.

When you prompt an LLM poorly, you’re effectively making a deal with the devil. The model will fulfill your request to the letter, even if what you asked was misguided, incomplete, or poorly articulated.

The result? Content that’s technically correct but off-brand, off-tone, or even factually wrong – yet delivered with such confidence it lulls you into publishing it.

The moral: Be careful what you ask for. The clarity of your prompt determines the quality of your output.

What AI Is Good At

When treated as a co-pilot, not as a god, AI can supercharge your workflow:

Research & Insights

  • Competitive landscape analyses.
  • SERP gap identification.
  • Tracking how competitors frame their unique value propositions.
  • Summarizing multiple opinion pieces or reviews into one clear insight.
  • Identifying overlooked audience segments based on forums and social media discussions.

Content Ideation & Briefing

  • Generating alternative angles on stale topics: e.g., turning “best practices” into “common mistakes” or “myths to avoid.”
  • Rewriting existing briefs to prioritize experience, expertise, authoritativeness, and trustworthiness (E-E-A-T) signals
  • Drafting Q&A content by scanning customer service transcripts or Reddit threads.
  • Suggesting specific examples or metaphors to make dry topics more engaging.

Narrative Shaping & Messaging

  • Reworking messaging for different formats: a LinkedIn post, an email subject line, and a webinar title – all aligned.
  • Auditing your current messaging to highlight jargon and suggest plain-language alternatives.
  • Helping articulate your brand’s point of view in ways that differentiate it from competitors.
  • Stress-testing your messaging by generating “devil’s advocate” objections you can preemptively address.

Workflow Enhancements

  • Drafting a competitive heat map: strengths, weaknesses, opportunities, threats – with citations.
  • Organizing customer testimonials into themed categories and crafting pull quotes.
  • Generating follow-up email sequences based on webinar transcripts or meeting notes.
  • Converting white papers into tweet threads, infographic outlines, and video scripts.

It’s like an intern with infinite energy and decent taste – incredibly helpful, but still in need of supervision.

What AI Is Not Good At

Don’t confuse the fluency of AI with wisdom. Here’s where it stumbles:

Judgment & Nuance

It doesn’t understand your brand’s unique sensibility, your audience’s emotional context, or when not to say something. You have to give it that context and direction. You cannot assume it will figure it out.

Accuracy & Truth

It is still prone to “hallucinations” – confidently wrong statements presented as fact.

We have limited understanding of why this happens, but it is so frequent that you almost have to assume there are at least a few hallucinations in the output somewhere.

Accountability

It cannot make decisions, nor does it bear the consequences of your choices. That’s on you.

In short, AI lacks your free will. And free will is what allows you to question, interpret, and choose what to do with its suggestions.

The Co-Pilot Mindset: Free Will Wins

To work effectively with your AI co-pilot, you need to strike the right balance between trust and control.

Here’s how:

Stay In The Pilot’s Seat

Never hand over full control. You’re still ultimately responsible for the vehicle.

Treat AI as a partner – or maybe not even a full partner, more like an exceptionally bright and quick research assistant – but never a replacement for you in any equation.

Be Precise In Your Prompts

Don’t assume it “knows what you mean.” Giving the AI instructions is like giving instructions to a particularly clever child who enjoys maliciously complying with your orders, except the AI doesn’t actually experience the joy.

You need to articulate your expectations clearly: format, tone, audience, and purpose. Add as much context and as many constraints as you can. The more data points and context you can provide, the better the outputs will be.

Use It To Accelerate, Not Replace

AI can speed up research, help shape narratives, and generate ideas, but it can’t replace your expertise or final judgment.

Review & Revise

Never, never, never, never publish output unedited. Always apply your brand’s perspective, always fact-check, and always ensure alignment with your goals.

Read everything you’re about to publish carefully. It’s okay to trust, but always verify.

Here’s an example of how that looks in practice:

I recently took a client’s complete keyword ranking report – not just the terms they were tracking, but every single ranking URL and query – and filtered out any URL already on page 1.

Then, I narrowed the data to just rankings in positions 11-20 (to keep it manageable) and fed that into an LLM.

I asked it to estimate the potential lift in organic traffic if each term improved to position 1 and to rank the list by estimated lift, highest to lowest.

But, I also gave the LLM context about the client’s business, explaining what kinds of customers and services were most valuable to them.

Then, I asked the model to highlight the keywords that made the most business sense for this client, because not every keyword you rank for is one you actually want to rank for.

With that context, the LLM was able to match keyword intent to the client’s goals and call out the terms that aligned with their business priorities.

In just minutes, I had a prioritized roadmap of high-impact, high-fit opportunities – something that would have taken hours to produce manually.

Practical Ways To Work With AI

Here are some more actionable ways you can incorporate AI into your workflow effectively:

Research Smarter And Faster

  • Create a competitive matrix with links and pros/cons.
  • Summarize customer sentiment across reviews, highlighting recurring pain points.
  • Surface conflicting expert opinions to inform balanced thought leadership pieces.
  • Forecast upcoming trends based on chatter in niche forums and early adopters.

Build Better Briefs

  • Include competitive positioning suggestions in briefs, not just keywords.
  • Add tone-of-voice examples aligned to audience segments.
  • Incorporate real data sources and reference points to help writers anchor their copy.
  • Generate sample social captions to support a campaign.

Strengthen Your Messaging

  • Stress-test a headline by generating objections and counterpoints.
  • Rewrite complex product descriptions into benefit-driven language for different audiences.
  • Propose alternate positioning statements for product launches or rebrands.
  • Audit your FAQ section to make it more conversational and AI-friendly.

Repurpose And Expand Content

  • Turn webinar transcripts into ebooks, blog series, and email drips.
  • Extract key insights from research reports to create shareable social graphics.
  • Draft SEO-friendly meta descriptions and titles for old content.
  • Identify missed opportunities in evergreen content for updates or expansion.

AI can do so much more than just “help you ideate.” It can help you uncover blind spots, repurpose assets, and deepen your strategic thinking, but only when you stay in the driver’s seat to guide and refine the outputs.

Final Thought: You, And Only You, Are The Pilot

I think we tend to treat our collective relationship with AI the same way we look at religion – you’re either a believer or an atheist.

Some have complete faith and trust it without question, while others reject it entirely and are convinced there is nothing there to believe in. The truth is somewhere in the middle (as it often is).

AI can be a powerful, tireless, but imperfect partner. It can help carry and manage heavy mental loads, work with you to map out routes and decide on destinations, but it can not take responsibility for driving the car. That’s got to be on you.

Your free will – your ability to keep your hands on the wheel – is what ensures the journey ends where you intended. If you actually let go, you’re certainly going to crash. You’re asking for assistance, not a magical autopilot.

So, go ahead: Let AI ride shotgun and keep your hands at 10 and two, where they belong.

More Resources:


Featured Image: Rawpixel.com/Shutterstock

Why OpenAI’s Open Source Models Are A Big Deal via @sejournal, @martinibuster

OpenAI has released two new open-weight language models under the permissive Apache 2.0 license. These models are designed to deliver strong real-world performance while running on consumer hardware, including a model that can run on a high-end laptop with only 16 GB of GPU.

Real-World Performance at Lower Hardware Cost

The two models are:

  • gpt-oss-120b (117 billion parameters)
  • gpt-oss-20b (21 billion parameters)

The larger gpt-oss-120b model matches OpenAI’s o4-mini on reasoning benchmarks while requiring only a single 80GB GPU. The smaller gpt-oss-20b model performs similarly to o3-mini and runs efficiently on devices with just 16GB of GPU. This enables developers to run the models on consumer machines, making it easier to deploy without expensive infrastructure.

Advanced Reasoning, Tool Use, and Chain-of-Thought

OpenAI explains that the models outperform other open source models of similar sizes on reasoning tasks and tool use.

According to OpenAI:

“These models are compatible with our Responses API⁠(opens in a new window) and are designed to be used within agentic workflows with exceptional instruction following, tool use like web search or Python code execution, and reasoning capabilities—including the ability to adjust the reasoning effort for tasks that don’t require complex reasoning and/or target very low latency final outputs. They are entirely customizable, provide full chain-of-thought (CoT), and support Structured Outputs⁠(opens in a new window).”

Designed for Developer Flexibility and Integration

OpenAI has released developer guides to support integration with platforms like Hugging Face, GitHub, vLLM, Ollama, and llama.cpp. The models are compatible with OpenAI’s Responses API and support advanced instruction-following and reasoning behaviors. Developers can fine-tune the models and implement safety guardrails for custom applications.

Safety In Open-Weight AI Models

OpenAI approached their open-weight models with the goal of ensuring safety throughout both training and release. Testing confirmed that even under purposely malicious fine-tuning, gpt-oss-120b did not reach a dangerous level of capability in areas of biological, chemical, or cyber risk.

Chain of Thought Unfiltered

OpenAI is intentionally leaving Chain of Thought (CoTs) unfiltered during training to preserve their usefulness for monitoring, based on the concern that optimization could cause models to hide their real reasoning. This, however, could result in hallucinations.

According to their model card (PDF version):

“In our recent research, we found that monitoring a reasoning model’s chain of thought can be helpful for detecting misbehavior. We further found that models could learn to hide their thinking while still misbehaving if their CoTs were directly pressured against having ‘bad thoughts.’

More recently, we joined a position paper with a number of other labs arguing that frontier developers should ‘consider the impact of development decisions on CoT monitorability.’

In accord with these concerns, we decided not to put any direct optimization pressure on the CoT for either of our two open-weight models. We hope that this gives developers the opportunity to implement CoT monitoring systems in their projects and enables the research community to further study CoT monitorability.”

Impact On Hallucinations

The OpenAI documentation states that the decision to not restrict the Chain Of Thought results in higher hallucination scores.

The PDF version of the model card explains why this happens:

Because these chains of thought are not restricted, they can contain hallucinated content, including language that does not reflect OpenAI’s standard safety policies. Developers should not directly show chains of thought to users of their applications, without further filtering, moderation, or summarization of this type of content.”

Benchmarking showed that the two open-source models performed less well on hallucination benchmarks in comparison to OpenAI o4-mini. The model card PDF documentation explained that this was to be expected because the new models are smaller and implies that the models will hallucinate less in agentic settings or when looking up information on the web (like RAG) or extracting it from a database.

OpenAI OSS Hallucination Benchmarking Scores

Benchmarking scores showing that the open source models score lower than OpenAI o4-mini.

Takeaways

  • Open-Weight Release
    OpenAI released two open-weight models under the permissive Apache 2.0 license.
  • Performance VS. Hardware Cost
    Models deliver strong reasoning performance while running on real-world affordable hardware, making them widely accessible.
  • Model Specs And Capabilities
    gpt-oss-120b matches o4-mini on reasoning and runs on 80GB GPU; gpt-oss-20b performs similarly to o3-mini on reasoning benchmarks and runs efficiently on 16GB GPU.
  • Agentic Workflow
    Both models support structured outputs, tool use (like Python and web search), and can scale their reasoning effort based on task complexity.
  • Customization and Integration
    The models are built to fit into agentic workflows and can be fully tailored to specific use cases. Their support for structured outputs makes them adaptable to complex software systems.
  • Tool Use and Function Calling
    The models can perform function calls and tool use with few-shot prompting, making them effective for automation tasks that require reasoning and adaptability.
  • Collaboration with Real-World Users
    OpenAI collaborated with partners such as AI Sweden, Orange, and Snowflake to explore practical uses of the models, including secure on-site deployment and custom fine-tuning on specialized datasets.
  • Inference Optimization
    The models use Mixture-of-Experts (MoE) to reduce compute load and grouped multi-query attention for inference and memory efficiency, making them easier to run at lower cost.
  • Safety
    OpenAI’s open source models maintain safety even under malicious fine-tuning; Chain of Thoughts (CoTs) are left unfiltered for transparency and monitorability.
  • CoT transparency Tradeoff
    No optimization pressure applied to CoTs to prevent masking harmful reasoning; may result in hallucinations.
  • Hallucinations Benchmarks and Real-World Performance
    The models underperform o4-mini on hallucination benchmarks, which OpenAI attributes to their smaller size. However, in real-world applications where the models can look up information from the web or query external datasets, hallucinations are expected to be less frequent.

Featured Image by Shutterstock/Good dreams – Studio

Claude Opus 4.1 Improves Coding & Agent Capabilities via @sejournal, @MattGSouthern

Anthropic has released Claude Opus 4.1, an upgrade to its flagship model that’s said to deliver better performance in coding, reasoning, and autonomous task handling.

The new model is available now to Claude Pro users, Claude Code subscribers, and developers using the API, Amazon Bedrock, or Google Cloud’s Vertex AI.

Performance Gains

Claude Opus 4.1 scores 74.5% on SWE-bench Verified, a benchmark for real-world coding problems, and is positioned as a drop-in replacement for Opus 4.

The model shows notable improvements in multi-file code refactoring and debugging, particularly in large codebases. According to GitHub and enterprise feedback cited by Anthropic, it outperforms Opus 4 in most coding tasks.

Rakuten’s engineering team reports that Claude 4.1 precisely identifies code fixes without introducing unnecessary changes. Windsurf, a developer platform, measured a one standard deviation performance gain compared to Opus 4, comparable to the leap from Claude Sonnet 3.7 to Sonnet 4.

Expanded Use Cases

Anthropic describes Claude 4.1 as a hybrid reasoning model designed to handle both instant outputs and extended thinking. Developers can fine-tune “thinking budgets” via the API to balance cost and performance.

Key use cases include:

  • AI Agents: Strong results on TAU-bench and long-horizon tasks make the model suitable for autonomous workflows and enterprise automation.
  • Advanced Coding: With support for 32,000 output tokens, Claude 4.1 handles complex refactoring and multi-step generation while adapting to coding style and context.
  • Data Analysis: The model can synthesize insights from large volumes of structured and unstructured data, such as patent filings and research papers.
  • Content Generation: Claude 4.1 generates more natural writing and richer prose than previous versions, with better structure and tone.

Safety Improvements

Claude 4.1 continues to operate under Anthropic’s AI Safety Level 3 standard. Although the upgrade is considered incremental, the company voluntarily ran safety evaluations to ensure performance stayed within acceptable risk boundaries.

  • Harmlessness: The model refused policy-violating requests 98.76% of the time, up from 97.27% with Opus 4.
  • Over-refusal: On benign requests, the refusal rate remains low at 0.08%.
  • Bias and Child Safety: Evaluations found no significant regression in political bias, discriminatory behavior, or child safety responses.

Anthropic also tested the model’s resistance to prompt injection and agent misuse. Results showed comparable or improved behavior over Opus 4, with additional training and safeguards in place to mitigate edge cases.

Looking Ahead

Anthropic says larger upgrades are on the horizon, with Claude 4.1 positioned as a stability-focused release ahead of future leaps.

For teams already using Claude Opus 4, the upgrade path is seamless, with no changes to API structure or pricing.


Featured Image: Ahyan Stock Studios/Shutterstock