The criteria flip: how Foster + Partners leads sustainable architecture by Gravity but competes from a Fallback provider type
A worked Criteria Flip analysis. One market leader by Gravity. Two Hidden Differentiators with no Full-Service entries. And the -41 point Decision drop that explains the whole pattern.
Flemming Rubak · April 16, 2026 · 15 min read
Executive summary
Foster + Partners (F+P) is the highest-Gravity brand in the UK sustainable architecture market. On April 16, 2026, Seedli’s consideration-stage analysis across ChatGPT (gpt-4o-mini), Claude (claude-haiku-4-5), and Gemini (gemini-2.5-flash-lite) recorded F+P at Gravity 42.0 of 100, rated Weak, with 81% Consideration presence and 54% aggregated visibility. That is the top of this category on every brand-level metric.
The provider type F+P is matched to, Full-Service Architecture Firm, is classified as Fallback. The aggregated buyer role across the three models is the lowest of three tiers (Primary, Secondary, Fallback). The gap between brand-level leadership and category-level role is not a measurement glitch. It is the structural signature of a market where the dominant brand competes perfectly on Table Stakes while the criteria that actually move Decision outcomes sit in an empty quadrant.
This article works the full Criteria Flip argument for that case: where the Gravity score comes from, why Consideration drops 41 points to Decision, which two criteria explain the drop, why one of them is simultaneously an elimination trigger, what model disagreement tells us about market tension, and what the published flip would look like. The methodology is documented in the Criteria Flip playbook; this article applies it end to end.
Data snapshot and models
All figures in this article are drawn from a single consideration-stage run recorded on April 16, 2026. The run follows the addition of two models that had previously been unavailable: the deprecated Claude Haiku was replaced by claude-haiku-4-5 on April 15, and Gemini 2.5 Flash Lite came online on the same day after Google billing activation. Prior published F+P numbers in the Criteria Flip playbook were generated against an earlier model mix; a small number of values have shifted as the new models were included. This article uses the April 16 snapshot throughout.
Snapshot parameters
Models
- claude-haiku-4-5-20251001
- gemini-2.5-flash-lite
- gpt-4o-mini
Run coverage
- Stage: Consideration (cons_v3)
- Prompts: 17 consideration-stage prompts
- Geography: United Kingdom
- Category: Sustainable architecture
Where a figure requires interpretation rather than direct measurement (for example, buyer signals that are generated by the model rather than observed in market research), this article flags the provenance explicitly. The Data notes and attribution section at the end walks through how each kind of evidence was sourced.
With the snapshot documented, the first number worth looking at is the one that does not fit the story the market tells about itself.
The market position paradox
Foster + Partners is, on most brand-level metrics in this category, the answer. The Customer Momentum Pipeline places F+P at the top of every stage panel visible to buyers who name the firm or search the category broadly. The cross-stage view lists F+P as the #1 brand by Market Gravity, a composite score that combines visibility, citation velocity, and decision momentum across models. The score is 42.0 out of 100. The rating attached to that score is Weak.
42.0
Market Gravity
Rated Weak
54%
Visibility
Aggregated across models
81%
Consideration presence
Stage C01–C05 aggregate
43%
Decision presence
-41 points vs Consideration
The paradox appears one level below the brand view. F+P is matched to a provider type called Full-Service Architecture Firm: the category that covers integrated practices offering design, engineering, and project management under one roof. Seedli assigns every provider type a buyer role by aggregating how the underlying models classify the category when buyers ask for recommendations. The three tiers are Primary (the obvious first pick for a typical buyer), Secondary (often shortlisted, not default), and Fallback (considered when the primary choice is disqualified).
Full-Service Architecture Firm: provider-type position
Reading the two views together is the whole pattern. F+P as a brand is a recognised leader buyers name spontaneously. The provider type that contains F+P is the one AI models turn to when the Primary and Secondary options have been ruled out. The brand sits at the top of the category. The category sits at the bottom of the role ranking. Every piece of content F+P publishes about its firm-level credentials reinforces the brand position while leaving the category-level role untouched.
The next question is whether the category-level role actually matters for buyer outcomes, or whether it is a framing detail that the brand-level position overrides. The Decision-stage data gives a direct answer.
The -41 point Decision drop
Across the Customer Momentum Pipeline, F+P enters Consideration at 81% presence: the highest in the category. By the Decision stage, presence has fallen to 43%. That is a drop of 41 percentage points in the span that usually determines which firm actually wins work. A drop of that size, starting from category-leading Consideration strength, is not generic attrition. It is displacement by specific alternatives on specific triggers.
The consideration-stage prompt set includes cons_v3_a05_provider_switching_triggers, which asks each model to describe the scenarios in which buyers switch out of the current Primary or Fallback provider type mid-evaluation and into another option. The responses for Full-Service Architecture Firm converge on two triggers.
When a buyer discovers partway through evaluation that certification timelines or compliance obligations are tighter than the Full-Service scope assumed, models surface Sustainable Design Specialist Firm as the alternative and recommend the switch. The driver is risk containment, and the target of the switch is the only provider type with a recorded score on Independence and Incentives.
Source: cons_v3_a05_provider_switching_triggers, two of three models
When a local planning authority requires alignment with a broader urban masterplan or district-scale framework, models recommend switching to Urban Masterplanning Practice. The driver is governance pressure, and the target of the switch is a provider type the market treats as the authority on regulatory coordination at scale.
Source: cons_v3_a05_provider_switching_triggers, Claude and Gemini
Neither switch is a weakness in Full-Service design capability. Both are triggered by criteria that Full-Service Architecture Firm has no published content addressing: independence from commercial incentives, and deep regulatory or governance coordination. When those criteria become salient mid-evaluation, the Full-Service provider type loses the buyer not because it performs worse on design, but because the model cannot find evidence that the category resolves the specific risk in front of the buyer.
That is what the drop looks like on the surface. The next layer of the data shows the criteria underneath it, and why the same pattern will repeat until the content position moves.
The criteria driving the gap
The Criterion Intelligence ranking for the UK sustainable architecture category reproduces a pattern familiar from other mature markets. The criteria buyers rate highest in stated importance are Table Stakes: every credible provider meets them, none differentiate on them, and the Strategic Opportunity Score (SOS) on each one sits at 0.01 to 0.02. Foster + Partners scores 3.0 of 3.0 on all four.
Table Stakes: where Foster + Partners already maxes out
Expertise & Competence
Importance 2.95/3, SOS 0.01
F+P score
3.0/3.0
vs market
at market
Regulatory & Risk Safety
Importance 2.95/3, SOS 0.01
F+P score
3.0/3.0
vs market
at market
Expected Outcomes
Importance 2.89/3, SOS 0.02
F+P score
3.0/3.0
vs market
+0.17
Product / Solution Fit
Importance 2.89/3, SOS 0.02
F+P score
3.0/3.0
vs market
+0.17
Perfect performance on all four. Combined content opportunity across the group: below 5%. These are the criteria the market agrees on and the ones F+P communications most often emphasise.
The two criteria with the highest SOS in the same ranking carry the opposite pattern. Stated importance is lower. Market tension is high. Content opportunity is substantial. And the Full-Service Architecture Firm provider type has no entry on either one in the Criterion Performance data.
Hidden Differentiators: where Full-Service has no entry
SOS
0.43
Opportunity
22%
Full-Service position: No entry (absent across all provider types)
Competitive landscape: No competing provider type has a published score on this criterion either. The field is genuinely empty.
SOS
0.28
Opportunity
14%
Full-Service position: No entry
Competitive landscape: Sustainable Design Specialist Firm holds the only published score at 2.8/3.0. It is also the provider type buyers switch to on certification-failure triggers.
Digital Experience is a peculiar case. The criterion is the highest-SOS entry in the ranking, but no provider type in the category has a published score on it at all. That is not a ceiling, it is a vacuum. The first Full-Service firm to publish structured, evidenced content on digital experience (BIM workflows, client-facing dashboards, collaboration tooling, real-time project visibility) will not just improve its own score; it will establish the dimension itself in AI model retrieval.
Independence and Incentives is a different shape. Sustainable Design Specialist Firm holds a 2.8 score, so the criterion exists in the market but Full-Service is absent from it. The first Full-Service firm to publish a clear statement on material supplier relationships, developer affiliations, and referral structures enters a position currently held by one niche provider type.
The criterion that also eliminates
The cons_v3_b06_elimination_criteria prompt asks each model to list the triggers that cause a provider to be eliminated from a buyer’s shortlist entirely. Across Claude and Gemini, the response includes conflict_of_interest at high severity. That entry is the Independence and Incentives criterion surfacing in a different role.
Independence and Incentives: two roles in the same run
Classified as Hidden Differentiator with SOS 0.28 and 14% content opportunity. Sustainable Design Specialist Firm holds the only score (2.8/3.0). Full-Service Architecture Firm has no entry. This is the positioning layer: where a provider stands on the criterion determines how models rank it among alternatives.
Returned as conflict_of_interest, severity high, by Claude and Gemini. When surfaced unaddressed in an evaluation, this is the level of concern that removes a provider from the shortlist rather than demoting it. The filter runs before the ranking.
The combination is the compounding mechanism. When Independence and Incentives surfaces in an evaluation that includes Full-Service Architecture Firm, two things happen at once. The differentiator layer gives zero positive signal (no score exists for the provider type). The elimination layer registers high-severity risk because there is no published statement resolving the concern. The same criterion that is not generating a win is also removing the option from the set.
In practical terms: a Criteria Flip that only addresses Digital Experience captures the 22% content opportunity but leaves the elimination dynamic untouched. A Criteria Flip that addresses both Digital Experience and Independence and Incentives captures the combined 36% opportunity and closes the elimination pathway at the same time. The sequencing matters when content-production budget is limited; the combined move is what the data supports.
Before moving to the prescription, one more pattern in the data deserves attention, because it explains why this case reads as genuinely contested rather than simply unresolved.
Model disagreement as evidence of market tension
The aggregated buyer role for Full-Service Architecture Firm is Fallback, but that aggregation obscures something worth surfacing. The three models under evaluation disagree on the classification. Each disagreement is internally consistent with the model’s reasoning, which means the disagreement itself is informative.
Full-Service Architecture Firm: buyer_role classification by model
Classifies Full-Service as a credible shortlist entry but not the default recommendation; notes risk-containment switching triggers.
Treats Full-Service as the provider of record when specialist options are ruled out by scope or regulation.
Names Full-Service as the default recommendation for integrated design-to-delivery engagements.
Three models, three different classifications, all defensible given each model’s framing. Seedli aggregates to the most conservative: Fallback.
In a market with low tension on a criterion, models converge. Identical reasoning produces identical classifications because the underlying evidence in training and retrieval is uncontested. Divergence of this kind signals that the category boundary itself is under active dispute in how AI models structure recommendations. The brand that publishes authoritative content addressing the specific triggers each model uses to reach its classification can shift the aggregated role over successive runs, because each model’s stated reasoning is a statement of what evidence would change its mind.
This is why the Criteria Flip works for a case like F+P. The position is not locked in by consensus; it is aggregated from disagreement. Content that targets the criteria driving the disagreement (Digital Experience for the Primary argument, Independence and Incentives for the Fallback argument) operates directly on the classification dynamic rather than trying to overwrite a settled judgement.
What the Criteria Flip says
The argument that falls out of the data has a specific shape. It is not a claim of superiority on the Hidden Differentiators (F+P has no entry on either), nor a dismissal of the Table Stakes (F+P genuinely excels there). It is a reframing of what the market should be competing on, backed by the ranked evidence, followed by the brand’s specific answer to each uncontested criterion.
Move 1 · Name the consensus
Expertise, regulatory safety, expected outcomes, product fit: the criteria every full-service architecture firm cites in its credentials content. Foster + Partners scores 3.0 on all four. So does every credible competitor. These are Table Stakes, and the market has reached consensus on them.
Move 2 · Surface what the data actually says
AI models structuring buyer decisions in this category separate providers on criteria with far lower stated importance. Digital Experience (SOS 0.43) and Independence and Incentives (SOS 0.28) combine for 36% of the differentiation territory. The Full-Service provider type has no published score on either one. The content is not there.
Move 3 · Close the elimination gap before the differentiator gap
Independence and Incentives operates as both a differentiator and a high-severity elimination trigger. Publishing a clear statement of material supplier relationships, developer affiliations, and referral structures does two things in a single move: closes a shortlist-removal pathway and enters a criterion where only one niche provider type currently has a published position.
Move 4 · Publish the digital-experience evidence the field is empty on
BIM workflow detail, client-facing dashboards, collaboration tooling, real-time project visibility, and how the practice actually operates day to day. No competitor has occupied this ground. A structured, evidenced digital-experience page does not fight for a citation position; it establishes the dimension itself in model retrieval for the category.
The sequence is deliberate. The elimination-gap close is cheaper and faster than the digital-experience build and produces immediate signal. The digital-experience content is a longer programme that compounds over quarters rather than weeks. A Criteria Flip article that names both is the published artefact that sets the expectation, links the evidence, and defines the content territory Foster + Partners is moving into.
The remaining question is one of transparency: where each of these numbers comes from, and which of them should be read as measured evidence versus model-generated representation. The honest answer matters to the argument.
Data notes and attribution
This article draws on three distinct types of evidence, and they carry different confidence levels. Documenting them here is part of the methodology, not an afterthought.
Gravity, visibility, stage presence percentages, shortlist rates, Criterion Performance scores, Criterion Intelligence rankings. These are direct measurements aggregated across the three models and 17 consideration-stage prompts in the April 16, 2026 run. They are the highest-confidence evidence in this article.
Switching triggers, elimination criteria, buyer_role classifications. These are model outputs to specific structured prompts (cons_v3_a05, cons_v3_b06, and the buyer-role classification prompt). They are consistent across models in most cases and divergent in the buyer_role case, which is itself part of the argument. Confidence is high for individual prompt outputs, moderate for cross-model interpretation.
Where prompts return phrasing that reads like direct buyer speech (“will you provide objective advice if it conflicts with a preferred outcome?” and similar), those are model-generated representations of how a buyer might frame the question, not transcripts of real interviews. They are useful as signals of what AI models treat as the buyer’s underlying concern; they are not evidence that any specific buyer has asked that question in market research. This article treats them as prompts to investigate further, not as primary evidence.
There are no brand-level citations of Foster + Partners in the raw prompt responses. Every F+P-specific figure (Gravity 42.0, 81% Consideration, 43% Decision, perfect Table Stakes scores, the absent Hidden Differentiator entries) comes from the aggregated Seedli views that match brand to provider type. The argument in this article is brand-specific because Seedli’s aggregation is brand-specific, but the underlying model evidence is category-level.
Two model changes landed in the April 16, 2026 run that were not present in earlier snapshots. The deprecated Claude Haiku was replaced by claude-haiku-4-5-20251001, and gemini-2.5-flash-lite came online after Google billing activation. Metric shifts between the numbers quoted in the Criteria Flip playbook and this article reflect that model mix change. The structural pattern holds across both snapshots. A subsequent run will confirm whether the specific values have stabilised or continue to drift as the new models settle.
Best for
Heads of marketing and content leads at mid-to-large practices who know their brand is visible but cannot explain why it is not winning more Decision-stage work. This article demonstrates the diagnostic; the path from here is a comparable analysis against the reader’s own category.
Action
Request a Criteria Flip snapshot for your market. The analysis takes the same shape as this article: Gravity versus provider-type role, Decision drop and its switching triggers, Hidden Differentiators with no incumbent, and the specific criteria where elimination and differentiation stack. Use the CTA below to start.
See the Criteria Flip analysis for your market
Seedli maps the gap between what buyers say they value and what AI models actually use to differentiate providers. The Criterion Intelligence ranking, Gravity scores, and switching triggers together tell you where the next content investment pays off.
Get started