How AI Works

The 30-Second Shortlist: How AI Decides Which Law Firm to Recommend

When a GC asks AI which firm to hire, a shortlist forms in 30 seconds. Here's exactly how AI decides which firms to name — and what puts your firm in the answer.

A CFO at a private equity firm needs M&A counsel. Deal closing in six weeks, cross-border technology acquisition, EU regulatory exposure. She opens Perplexity, types a detailed question — specific on transaction type, geography, complexity — and has a shortlist of three firms in 30 seconds. Two of them she'd never heard of before. One firm she knows well, a firm with 40 years of M&A experience and a Chambers ranking, doesn't appear.

The firms she'd never heard of get her attention for the next hour. The firm she knew gets none. By the time she makes her first call, the invisible firm's chance to compete has already passed.

This isn't about the quality of the work or the depth of the experience. It's about a process that ran before any human judgment was applied — a 30-second process governed entirely by five signals that most law firms have never thought about. Here's exactly how it works.

The Process

From prompt to shortlist — in 30 seconds

How AI selects which law firm to recommend
1
🔒 chatgpt.com
ChatGPT
Best M&A firm in New York for a cross-border tech deal
The Prompt
A query is typed. AI immediately parses intent: practice area, geography, matter type, client profile.
2
Training Data
Live Retrieval (RAG)
LLM Processing
Two layers activate: training data and real-time web retrieval — simultaneously.
3
Ch
Chambers
L3
Law360
in
LinkedIn
JD
JD Supra
🌐
Firm Website
Retrieval
AI pulls from directories, publications, firm sites, and citation patterns across trusted sources.
4
Citation Density
Entity Consistency
Answer-Ready Content
Authority Sources
Specificity Match
Selection Logic
Five signals are weighted to determine which firms have earned the recommendation.
5
AI
ChatGPT
Based on your query, here are the top firms for cross-border M&A in New York:
1.
Sullivan & Cromwell Cited across 40+ cross-border tech transactions
2.
Skadden Arps Ranked for M&A advisory, deep NY presence
3.
Latham & Watkins Recognized for tech sector M&A work
The Recommendation
Two to four firms are named. The reasoning reveals exactly which signals were weighted.
Sources AI retrieves from in Stage 3
Chambers & Legal500 Law360 / ALM Firm Website LinkedIn JD Supra Client Citations Martindale-Hubbell Google News

Stage 1: The Prompt

1 How the query becomes a filter

AI doesn't search the way Google does. When a query arrives, the model parses intent before retrieving anything. "Best M&A firm in New York for a cross-border tech deal" becomes a structured profile: transaction type (M&A, cross-border), sector (technology), geography (New York), deal complexity (implied sophistication). The model builds a description of the ideal answer before a single source is consulted.

This matters because firms whose documented expertise doesn't match the parsed intent are filtered out before retrieval even begins. If your practice area page says "We handle complex corporate transactions," AI cannot reliably map that to "cross-border M&A for technology companies." The filter applies silently, before the process most firms think about has started.

Stage 2: LLM Processing

2 Training data and real-time retrieval

Modern AI assistants operate on two layers simultaneously. The first is training data — patterns absorbed from billions of documents including legal publications, court records, directory listings, case studies, and firm websites crawled before the model's training cutoff. The second is real-time retrieval (RAG — Retrieval Augmented Generation), where the model actively searches the live web to validate and supplement what it already knows.

Your firm's presence in training data is its foundation. Real-time retrieval is the validation. Both matter. A firm that appears in neither is invisible regardless of its actual reputation — because the model has no evidence to work with when forming its response. A firm with strong training data presence but weak current retrieval signals may have appeared in older model versions but fades as AI systems update. The signal needs to be built continuously, not once.

Stage 3: The Retrieval

3 Not all sources carry equal weight

AI models have learned, through exposure to billions of documents, which sources produce reliable and authoritative information. In the legal context, a rough hierarchy governs retrieval weight:

  • Legal directories (Chambers, Legal 500, Martindale) — Foundational, but they contain hundreds of firms. Directory presence establishes pool membership, not selection. Every firm on a final shortlist will be in a directory; not every directory firm will be on a shortlist.
  • Legal publications (Law360, ALM, Above the Law, JD Supra) — Citations in these sources carry significant retrieval weight. A firm mentioned in Law360 for a specific deal type builds a traceable signal that AI retrieval systems recognize and return.
  • The firm's own website — If structured with answer-ready content, the firm's website feeds directly into retrieval. This is where most firms leave the most value on the table. Generic practice descriptions are not retrieved; specific matter-level content is.
  • LinkedIn and professional profiles — Entity validation. The model cross-references what it finds elsewhere with what appears on professional profiles. Inconsistency reduces confidence.
  • Client-generated content — Matter descriptions, press releases about deal completions, testimonials, and annual report references. These carry particular weight because they represent third-party validation from the clients themselves.

The more a firm is cited consistently across diverse, high-authority sources, the stronger its retrieval signal. A single strong source isn't enough. The model is looking for corroboration.

Stage 4: Selection Logic

4 The five signals that decide who gets named

This is where most law firms lose the shortlist. The model isn't ranking based on firm size, Am Law position, or website polish. It applies five specific signals — and firms that don't score on all five are filtered out regardless of how well they score on any one of them.

  1. 1
    Citation density. How frequently and authoritatively is the firm mentioned across the sources it retrieves from? A firm cited in Law360, Chambers, a client's annual report, and three JD Supra articles ranks higher than one that appears only on its own website — regardless of which firm is objectively better at the work.
  2. 2
    Entity consistency. Does the firm's identity appear coherently across sources? Name variants, outdated practice area descriptions, and inconsistent geographic listings create retrieval uncertainty. A firm listed as "Smith & Jones LLP" on its website, "Smith Jones" on LinkedIn, and "S&J Law" on a directory is a fragmented entity — harder to surface with confidence. AI treats inconsistency as a signal of lower authority.
  3. 3
    Answer-ready content. Has the firm published content that directly answers the questions clients ask? A practice area page that says "We handle complex M&A transactions" is not answer-ready. A page that says "We represent technology companies in cross-border acquisitions, typically in the $50M–$500M range, with recent transactions in the EU and Southeast Asia" is. AI retrieval systems favor content structured around specific client questions, not marketing copy structured around firm capabilities.
  4. 4
    Authority source validation. Is the firm cited by sources the AI has learned to trust? This is a compounding effect. A firm mentioned in Law360, then cited in a client's press release, then referenced in a JD Supra analysis is building a citation network that AI retrieval systems recognize as authoritative. A firm with a great website and no external citations has a thin signal — easy to overlook when the model has stronger options to name.
  5. 5
    Specificity match. Does the firm's documented expertise match the specific query at the matter level? A firm that has published content about "cross-border M&A for technology companies with EU regulatory exposure" will be retrieved for that exact query. A firm with a generic "M&A and Corporate" practice page will not — even if it handles those exact matters regularly and handles them well.

The firms not named are not necessarily less capable. They simply didn't have the signals the model needed to surface them with confidence.

Jacob Shamis, Founder & CEO, Selectio.ai

Stage 5: The Recommendation

5 What the output reveals

The model generates its response and typically names two to four firms, usually with brief explanations: "Known for complex cross-border technology transactions" (citation density + specificity match), "Frequently cited in industry rankings for M&A advisory" (authority source validation), "Recognized for deal work in the $50M–$500M range" (answer-ready content). These explanations are a direct window into which signals were weighted in the selection.

Pay attention to what these explanations cite. They are not marketing copy — they are the model's summary of the evidence it found. When a firm is described as "frequently cited," that is a direct report of citation density. When a firm is described as "known for" a specific matter type, that is a report of specificity match. The explanation tells you what signals the model found — and by extension, what signals the firms that didn't appear were missing.


What your firm needs to do

AEO — Answer Engine Optimization — is the practice of building these five signals deliberately, on a timeline that matters for active business development, rather than waiting for them to accumulate over decades of passive citation.

The work involves five connected activities:

  • Citation building. Systematic placement in the publications and directories AI models retrieve from — Law360, JD Supra, Chambers, Legal 500, Bloomberg Law, The American Lawyer. Coverage that names your firm in the context of specific practice areas and matter types, not just in general firm news.
  • Structured content. Rewriting practice area pages, attorney profiles, and matter descriptions to be answer-ready: specific, client-query-oriented, organized around the questions buyers actually ask rather than the capabilities firms want to describe.
  • Entity optimization. Ensuring your firm's identity — name, practice areas, attorney profiles, geographic scope, matter types — is consistent and complete across every source AI retrieves from. Auditing and correcting inconsistencies that fragment your entity signal.
  • Knowledge Graph presence. Establishing your firm as a recognized, validated entity in structured data — the layer of the web that AI systems use to cross-reference and confirm what they believe about an entity before naming it in a recommendation.
  • Practice area specificity. Moving from generic practice descriptions to documented matter-level expertise: the specific transaction types, client profiles, deal sizes, and jurisdictions your firm actually handles. This is the direct input to specificity match in Stage 4.

The firms appearing in AI recommendations today have built these signals — some deliberately, most through years of consistent citation in high-authority sources. AEO is how you build them intentionally, in months rather than decades, and with enough lead time to matter before the competitive positions in AI recommendations become entrenched.


The CFO who opened Perplexity six weeks before close didn't have time to discover your firm. The shortlist she received determined who got considered. Three firms were on it. If yours wasn't one of them, the question isn't whether you're good enough. It's whether AI has enough signals to say so.

Frequently asked questions

AI decides which law firms to recommend by weighing five signals: citation density (how frequently and authoritatively the firm is mentioned across trusted sources), entity consistency (whether the firm's identity is coherent across platforms), answer-ready content (whether the firm has published content that directly addresses client questions), authority source validation (whether trusted publications reference the firm), and specificity match (whether the firm's documented expertise aligns with the precise query). Firms that score well across all five signals are named; firms that don't are not — regardless of their actual capabilities.
Google ranks individual pages using keyword relevance and backlink signals. AI recommendation systems retrieve structured signals from multiple sources — legal publications, directories, professional profiles, client-generated content — and synthesize them into a response. A firm can rank on page one of Google for every relevant keyword and still be completely absent from AI recommendations, because these are parallel systems with almost no overlap in what they optimize for.
RAG stands for Retrieval Augmented Generation — the mechanism by which AI assistants actively search the live web to supplement what they already know from training data. When a GC asks which firm to hire, the AI doesn't just rely on what it learned during training; it retrieves current information from across the web. If your firm's content isn't structured to be found and extracted by retrieval systems, you're invisible to this layer of the recommendation process — even if you have strong training data presence.
The five platforms where general counsel actively research outside counsel are ChatGPT, Perplexity, Claude (Anthropic), Google Gemini, and Microsoft Copilot. Each uses slightly different retrieval logic and data sources, but all respond to the same five core signals: citation density, entity consistency, answer-ready content, authority source validation, and specificity match. A comprehensive AEO program builds signals that translate across all five platforms.
First measurable movement in AI visibility typically appears within 60–90 days of systematic AEO work. Authority compounds over time: a firm that builds citation density and structured content consistently will see its AI recommendation frequency increase quarter over quarter. The compounding nature of AEO signals means that firms which start earlier hold a structural advantage over those that start later — not just a timing advantage.

Does AI recommend your firm right now?

The free 45-minute AI Visibility Audit shows you exactly where your firm stands across ChatGPT, Perplexity, Claude, Gemini, and Copilot — and who AI is recommending instead of you.

Book Your Free Audit