A growing number of PPC practitioners are watching their dashboards with a furrowed brow: click-through rates are softening, conversion paths feel murkier than they used to be, and the ROI that once made Google Ads the undisputed king of performance marketing seems harder to reliably reproduce. The question echoing across forums and Slack channels is no longer hypothetical — advertisers genuinely want to know whether AI is reshaping the Google Ads landscape in ways that disadvantage the skilled practitioner, or whether the real story is more nuanced than a simple "the platform is broken."
The Real Shift: It's Not Google Ads Losing Its Edge — It's the Rules Changing
Let me be direct with you: after managing over $350M in Google Ads spend across industries ranging from direct-to-consumer e-commerce to enterprise B2B SaaS, I've lived through several "Google Ads is dying" cycles. Broad match killed keyword precision (it didn't). Smart Bidding removed human control (partially true, but misunderstood). Performance Max cannibalized Search (complicated). Each wave brought panic, and each wave was followed by practitioners who adapted and outperformed those who didn't.
The AI era is different in magnitude, but not in kind. What's actually happening is a three-part shift that's easy to misread as platform decline:
- Query-level visibility has been systematically reduced — Search Term Reports now omit a significant portion of triggering queries (studies suggest 30–60% of actual volume is hidden behind "Other search terms").
- AI-generated answers are changing SERP interaction patterns — Google's AI Overviews reduce the incentive to click for informational queries, compressing the top-of-funnel search ecosystem.
- Measurement has fractured — Cross-device journeys, iOS privacy changes, cookie deprecation timelines, and GA4's session model all conspire to make attribution murkier than it was in 2019.
The practitioners who are struggling are often fighting the old battle with old weapons. The ones thriving have recalibrated for what Google Ads actually is in 2024–2025.
Key Insight: Google Ads isn't losing its edge — but the edge has moved. It now lives in signal quality, measurement architecture, and AI-native campaign structures rather than in granular keyword sculpting and manual CPC optimization.
Why CTRs Are Declining (And What It Actually Means)
A common question in the r/PPC community centers on declining click-through rates — and the concern is valid, but often misdiagnosed. Here's what I'm seeing across accounts:
AI Overviews Are Compressing Informational Query Intent
When a user searches "how does retargeting work" or "best CRM for small business," Google's AI Overview now serves a synthesized answer above the fold. Organic click-through drops dramatically for these queries — in some cases by 15–25% based on third-party studies from early 2024 rollout data. But here's the nuance: these were rarely high-converting paid queries to begin with.
If your CTR is dropping on awareness or consideration-stage keywords, the question to ask isn't "is Google broken?" — it's "should I have been bidding on these terms at all, and how was I measuring their value?"
SERP Layout Changes Affect Position-Dependent CTR Benchmarks
Historical CTR benchmarks (Position 1 = ~6–8% CTR, Position 2 = ~4–5%, etc.) were calibrated against a different SERP layout. With Shopping ads, AI Overviews, People Also Ask boxes, and Local packs all competing for attention, those benchmarks are outdated reference points. I'd strongly advise against using historical average CTR as your primary health metric right now.
Common Mistake: Comparing your 2024–2025 CTR against 2021–2022 benchmarks and concluding the account is underperforming. The SERP environment is structurally different. Judge CTR relative to your own historical trend and against conversion rate, not against outdated industry numbers.
What Actually Matters More Than CTR
In my experience managing high-spend accounts, the metric hierarchy that actually predicts business outcomes looks like this:
| Metric |
Old Priority |
Current Priority |
Why the Shift |
| CTR |
High |
Medium-Low |
SERP layout changes distort signal |
| Conversion Rate (by segment) |
High |
High |
Still the clearest efficiency indicator |
| Revenue / ROAS |
High |
High |
Unchanged in importance |
| New Customer Acquisition Rate |
Low |
Very High |
LTV-based bidding requires this signal |
| Search Impression Share (by campaign type) |
Medium |
Medium |
Useful for budget ceiling diagnostics |
| Model Coverage / Conversion Modeling Rate |
N/A |
High |
Critical for Smart Bidding health |
The Measurement Crisis Is the Real Emergency
As practitioners often discuss in forums like r/PPC, it's genuinely difficult right now to know whether performance has dropped or whether your ability to see performance has dropped. That distinction is everything, and most advertisers are conflating the two.
Understanding Conversion Modeling
Google's Smart Bidding now relies heavily on modeled conversions — statistically inferred conversions that aren't directly observable due to consent gaps, cross-device journeys, or privacy restrictions. In some accounts I've audited, modeled conversions represent 20–40% of total reported conversions. This isn't fraud or inflated numbers — it's Google's attempt to give Smart Bidding enough signal to function despite a fragmented measurement environment.
The problem is that many practitioners don't know this is happening in their accounts, and they're either:
- Over-trusting Google's reported numbers without understanding the modeling component
- Or, after finding out about modeling, dismissing all conversion data and making poor budget decisions based on gut feel
Best Practice: Implement Google's Enhanced Conversions for Web and for Leads. This feeds first-party, hashed user data back into Google's models, dramatically improving the accuracy of both reported and modeled conversions. Accounts with Enhanced Conversions properly implemented typically see 5–15% more observed conversions attributed, which feeds better Smart Bidding signals — creating a compounding performance advantage.
Build a Triangulated Measurement Stack
Relying solely on Google Ads conversion reporting in 2025 is like navigating with one eye closed. The measurement stack that I recommend to clients includes at minimum:
- Google Ads Conversion Tracking — with Enhanced Conversions enabled and consent mode properly implemented if operating in the EU or California
- GA4 — configured with proper session-scoped vs. event-scoped metrics understanding, and imported conversion goals back into Google Ads
- A revenue reconciliation check — monthly comparison of Google Ads reported revenue vs. CRM or backend transaction data. If the gap exceeds 15–20%, something is broken in your tracking
- Incrementality testing — even simple geo-based holdout tests run quarterly to validate that your paid clicks are driving incremental value, not just capturing demand that would have converted anyway
Key Insight: In accounts where I've implemented proper Enhanced Conversions plus Consent Mode V2, Smart Bidding targets have become achievable again at spend levels where they previously seemed impossible to hit. Measurement quality is directly upstream of bidding performance — fix measurement first, bidding second.
Smart Bidding in the AI Era: When It Works and When It Doesn't
Smart Bidding — Target ROAS, Target CPA, Maximize Conversions — is simultaneously Google's greatest gift to efficient advertisers and its most misunderstood product. The AI era has made Smart Bidding more powerful at scale and more dangerous when misapplied at low volume.
The Minimum Data Thresholds Still Matter Enormously
Google's own guidance suggests Smart Bidding needs approximately 30–50 conversions per month at the campaign level for stable performance. In practice, I find that campaigns with <30 conversions/month on Target CPA behave erratically — the model is essentially guessing, and those guesses can produce volatile CPAs that swing 50–100% week-over-week.
For accounts with lower conversion volume, my preferred approach is:
- Use Maximize Conversions without a target initially to let the model accumulate data
- Set targets only after you have at least 60 days of conversion history in the campaign
- Consider consolidating campaigns to pool conversion signals — one well-funded campaign often outperforms three under-funded ones even with the same total budget
- Use micro-conversions as secondary signals (page views of key pages, scroll depth, time on site >3 minutes) to give Smart Bidding more data while your primary conversion accumulates
Common Mistake: Setting aggressive Target CPA or Target ROAS goals in campaigns with <30 monthly conversions, then concluding Smart Bidding "doesn't work" when performance destabilizes. The strategy isn't flawed — the data foundation is insufficient. Either consolidate, loosen targets, or use a simpler bidding strategy until volume supports it.
Performance Max: Handle With Care
Performance Max is the most contentious campaign type in the AI era, and the r/PPC community debates it constantly for good reason. It's a powerful channel when applied correctly and a budget-consuming black box when misapplied.
My current position after running PMax across dozens of accounts at various spend levels:
- PMax works best when you have strong asset coverage (multiple headlines, descriptions, images, and video — especially video), clear audience signals from existing customer data, and a product catalog or service set with proven conversion history
- PMax struggles when asset quality is low (it'll generate assets automatically — and they're often mediocre), when you lack audience signal data, or when you're in a niche B2B category with low search volume
- Brand campaigns should run separately in a standard Search campaign with brand terms as exact match — this protects brand efficiency from being consumed by PMax's opaque allocation
- Use campaign-level URL exclusions and brand exclusions aggressively to prevent PMax from simply recycling your best existing demand
How to Actually Compete in an AI-Dominated Auction
The advertisers winning right now share a common set of behaviors that I'd describe as "AI-native PPC management." They're not fighting the automation — they're directing it with better inputs.
First-Party Data Is Your Moat
In a world where third-party signals are degrading, the advertisers with rich first-party data have a structural advantage. This means:
- Customer Match lists segmented by LTV tier (high-value customers, mid-tier, one-time purchasers)
- CRM data integrations that feed purchase history signals back into Google Ads
- New Customer Acquisition goals in PMax and Smart Shopping to explicitly tell the algorithm to prioritize net-new customers over remarketing
- Regular audience list refreshes — stale audience lists actively hurt Smart Bidding model quality
Creative Quality Is Now a Performance Lever, Not Just Branding
With Responsive Search Ads having replaced Expanded Text Ads entirely, and with PMax's asset-based approach, creative quality is now directly upstream of auction performance. Google's Ad Strength score isn't just a cosmetic indicator — it correlates with Ad Rank components that affect both CPC and eligibility.
Specifically: in my experience, moving campaigns from "Poor" to "Excellent" Ad Strength correlates with 10–20% improvements in impression share at equivalent CPCs. The mechanism is Google rewarding expected relevance.
Best Practice: Treat your RSA headline and description combinations as a testing matrix, not a set-and-forget configuration. Run at least 2–3 RSAs per ad group, pin only when strategically essential (pinning reduces combination options and typically hurts Quality Score), and review Asset Performance labels monthly to identify and replace "Low" rated assets. This ongoing creative hygiene compounds over time into meaningful performance advantages.
What to Do Next: Your AI-Era Google Ads Action Plan
If you've read this far and you're wondering where to start, here's a prioritized action list based on what I've seen move the needle most consistently across accounts in the current environment:
-
Audit your measurement foundation first — everything else depends on it.
Check whether Enhanced Conversions are implemented, verify Consent Mode V2 if you have EU or California traffic, and run a revenue reconciliation between Google Ads reported revenue and your backend data. If the gap exceeds 15%, fix tracking before touching bids.
-
Consolidate campaigns to meet Smart Bidding data thresholds.
If you're running more than 4–5 Search campaigns targeting overlapping themes, consolidation is likely your highest-leverage structural change. Pool conversion signal, simplify match type strategies (broad + Smart Bidding is genuinely viable now with proper negative lists), and let the algorithm work with sufficient data.
-
Upload your customer list and segment it by value tier.
Even a basic high-value vs. standard-value segmentation gives Smart Bidding a meaningful anchor for LTV-aware optimization. Do this before your next campaign structure review.
-
Run a creative audit.
Pull your Asset Performance report, identify every "Low" rated asset across RSAs and PMax, and replace them within the next 30 days. Set a calendar reminder to repeat this monthly.
-
Reframe your success metrics for the current SERP environment.
Replace CTR as a primary KPI with conversion rate by device, new customer acquisition rate, and revenue per impression. These metrics are much harder for the changing SERP landscape to distort and much more directly connected to business outcomes.
Google Ads in the AI era rewards practitioners who give the algorithm high-quality inputs — better data, better creative, clearer objectives — and punishes those who try to override it with granular manual controls that the platform has effectively deprecated. That's a significant philosophical shift for practitioners trained in the era of exact match keywords and manual CPCs. But it's also a genuine opportunity: if your competitors are still fighting that old battle, you have a clear path to outperforming them by simply playing the game that's actually being played.