Quality Score is one of the most misunderstood metrics in Google Ads — and also one of the most consequential. After managing over $350M in Google Ads spend, I've seen countless accounts hemorrhage budget simply because practitioners either ignore Quality Score entirely or chase it as a vanity metric without understanding what it actually controls. If you're seeing 70%+ of your lost impression share attributed to rank (not budget), Quality Score is almost certainly a root cause you need to address systematically.
Let's start with the fundamentals. Quality Score is Google's 1–10 rating of the relevance and quality of your keywords, ads, and landing pages. It's calculated at the keyword level and is a diagnostic tool — not a bidding lever in itself. The actual mechanism that determines your ad placement is Ad Rank, and Quality Score is a critical input into that formula.
Here's the Ad Rank formula simplified:
The practical implication: a keyword with a Quality Score of 8 and a $2.00 bid can outrank a competitor bidding $4.00 with a Quality Score of 3. You are literally buying the same placement at half the price when your quality is superior. Over millions of impressions, this compounds into massive cost advantages.
Google breaks Quality Score into three weighted components, each rated as "Below Average," "Average," or "Above Average":
| Component | What It Measures | Approximate Weight | Primary Lever |
|---|---|---|---|
| Expected Click-Through Rate (eCTR) | Likelihood of your ad being clicked when shown for that keyword | ~40% | Ad copy relevance & creative testing |
| Ad Relevance | How closely your ad matches the intent of the keyword | ~30% | Ad group structure & keyword-to-copy alignment |
| Landing Page Experience | Relevance, transparency, and ease of navigation on your landing page | ~30% | Landing page content, speed, & UX |
Google doesn't officially publish the exact weights, but based on controlled testing across accounts, eCTR tends to have the outsized influence. When I've fixed landing pages from "Below Average" to "Average" on a single campaign, I've seen QS move by 1–2 points. When I've dramatically improved eCTR through aggressive ad copy testing, I've moved QS by 3–4 points on the same keywords.
A common question in the r/PPC community revolves around impression share loss — specifically when practitioners notice that the majority of their lost impressions are due to Ad Rank rather than budget. This is actually a very precise diagnostic signal and it's one of the most actionable things you can pull from your Search Impression Share report.
When >70% of your lost impression share is due to rank (as one practitioner noted from their account data), it tells you a clear story: Google is willing to show your ad more often, but your Ad Rank isn't competitive enough to win those auctions. Throwing more budget at this problem does nothing. You have to improve Ad Rank — and that means either raising bids, improving Quality Score, or both.
At the keyword level, you can also simply look at the Quality Score column (enable it via Columns) and sort ascending to find your worst performers dragging down campaign performance.
eCTR is benchmarked against other advertisers bidding on the same keyword. This means you're not competing against some abstract standard — you're competing against whoever else is showing up in that auction. The bar is relative, not absolute.
Tactics that consistently move the needle:
Ad Relevance is the most structural of the three components — it's largely solved at the account architecture level. If you have 50 keywords stuffed into one ad group with a single generic ad, you're going to have "Below Average" ad relevance on most of those keywords.
The fix is tighter ad group segmentation:
{KeyWord:Default Text} in headlines to dynamically match the search query. This is a powerful relevance signal, but use it carefully — it can produce awkward ad copy if your keyword list has long or odd-phrased terms.Landing Page Experience is the component practitioners most often overlook because it requires cross-functional work — you usually need a developer or designer involved. But Google is explicit about what it evaluates:
As practitioners often discuss in r/PPC, knowing your Quality Score number is only useful if you know what's normal for your context. Here's a realistic benchmark framework based on real account data:
| Quality Score Range | Interpretation | CPC Impact vs. QS 5 Baseline | Priority Action |
|---|---|---|---|
| 1–3 | Critical issue — severely penalized | +25% to +400% higher CPCs | Pause or immediate restructure |
| 4–5 | Below/at average — room for improvement | Baseline (0% adjustment) | Ad copy & landing page audit |
| 6–7 | Above average — competitive | -16% to -20% lower CPCs | Maintain & optimize extensions |
| 8–9 | Excellent — significant cost advantage | -25% to -37% lower CPCs | Scale aggressively |
| 10 | Exceptional — rare, high-brand terms | -50% lower CPCs | Protect and defend |
In practice, QS 10 is almost exclusively seen on branded keywords where you own the brand. For non-branded keywords in competitive industries, a QS of 7–8 is genuinely excellent. Don't chase QS 10 on generic commercial terms — it's largely theoretical. Focus your energy on pulling QS 4–5 keywords up to 6–7, as that's where the most cost efficiency is unlocked.
Quality Score has a memory. Google accounts for historical performance data in its calculations, which means a keyword that performed poorly for years carries a negative signal even if you've since improved your ads and landing pages. There are a few approaches to reset or accelerate this:
Here's a nuance that catches even experienced practitioners off guard: the 1–10 Quality Score you see in the interface is a historical snapshot. The actual quality calculation that runs in every auction — called auction-time quality — incorporates real-time signals including the user's device, location, time of day, and specific search query.
This means your visible QS of 6 might actually perform as a 7 or 8 in auctions where the signals are favorable (e.g., a mobile user in your target city searching an exact-match query during business hours). The visible score is useful for diagnosis, but it's not a 1:1 reflection of every auction you participate in.
A common question is whether Quality Score matters less under Smart Bidding (tCPA, tROAS, Maximize Conversions). The answer is: it still matters, but differently. Smart Bidding strategies adjust bids based on auction-time quality signals, so in theory, Google automatically compensates for lower QS by bidding less. However, low Quality Score still increases your floor cost and reduces your eligibility for premium placements. Even under full automation, improving Quality Score creates more efficient conversion paths for the algorithm to exploit.
If you're dealing with high impression share loss due to rank, or simply want to systematically improve account efficiency, here's a prioritized action sequence:
Quality Score isn't a quick fix — it's a compounding investment. The accounts I've seen with the best long-term efficiency haven't necessarily won on bids; they've won on quality, paying 20–40% less per click than competitors for the same placements. Over a $1M annual ad budget, that's $200,000–$400,000 in recovered efficiency. That's the real reason Quality Score deserves your serious attention.