All posts
AI Writes Your Risk Report. It Doesn't Score It.
Structural Insight

AI Writes Your Risk Report. It Doesn't Score It.

How Global Solo uses AI for language while keeping risk scores, structure, and boundaries fully deterministic. A look inside the META diagnostic engine.

Jett Fuยท

When I started building the META diagnostic, I ran into a problem that every AI product eventually faces: people want intelligence, but they also want consistency.

A cross-border entrepreneur paying $99 for a risk assessment needs to know that the same structural setup produces the same result, every time. Not "roughly similar." The same scores. The same sections. The same structural mapping. If two people with identical entity structures, tax residency patterns, and documentation levels get different risk scores, the product is broken.

That tension shaped every architectural decision in Global Solo's diagnostic engine. Here's how it works, and why we drew the lines where we did.

The Core Principle: Determinism Over Intelligence

The META diagnostic maps structural risk across four dimensions: Money, Entity, Tax, and Accountability. Each dimension gets a score from 1 to 5. Those scores drive which sections appear in your report, which findings get highlighted, and how cross-dimensional patterns are detected.

Early on, I tested having the AI generate these scores. The results were inconsistent. The same entrepreneur profile, submitted twice, would sometimes get a Money score of 3 and sometimes a 4. The AI would notice different details each time, weight them differently, and produce slightly different assessments.

For a diagnostic product, that's unacceptable. If your doctor's blood test gave different results depending on which lab tech wrote up the report, you'd find a different doctor.

So we split the system in two. The AI handles language. Everything else is deterministic.

What "Deterministic" Means in Practice

The diagnostic engine runs in three stages, and the division of labor between rules and AI is deliberate.

Stage 1: Analysis. Your answers map to 45 signals across the four META dimensions. A signal mapper converts each answer into a structured observation. No interpretation, no inference. If you report income from three countries, the signal records "income_country_count: 3." A separate scoring engine applies weighted rules to produce dimension scores. These rules are version-controlled and auditable. If the AI suggests a different score, the deterministic score wins. Always.

Stage 2: Narration. This is where the AI earns its keep. Given the structured analysis (scores, findings, cross-dimensional patterns), the AI writes the narrative sections of your report. It explains what the structure looks like, translates signal data into readable prose, and connects patterns across dimensions. The AI is good at this. It can synthesize complex structural data into clear language better than any template system.

But it writes under constraints.

Stage 3: Assembly. The final report gets assembled with zero AI involvement. Metadata, section ordering, score badges, boundary notices, coverage statistics. All assembled from the Stage 1 analysis and Stage 2 narratives using deterministic templates. No LLM calls. No inference. No creativity.

๐Ÿ“Š

How does your structure score?

Free 2-minute screening across Money, Entity, Tax, and Accountability.

Check Now

The Language Problem

AI-generated text has a habit of drifting into advice. Ask a language model to describe a tax residency gap, and it will often add "you may want to consult a tax professional" or "it is advisable to restructure your entity." That's natural for AI. It's trained on helpful content.

For Global Solo, that drift is a product defect. We sell structural visibility, not advice. The moment a report says "you need to" or "we suggest," it crosses from diagnosis into recommendation, which is a different product with different liability.

So every narrative the AI produces runs through a language verifier. The verifier scans for four categories of prohibited patterns: directive words (anything that tells you what to do), promise words (anything that predicts an outcome), compliance assertions (anything that declares legal status), and benchmark language (anything that compares you to a generalized population).

Any match gets flagged and rewritten. The AI's output goes through regex-based sanitization that strips these patterns before the text reaches the assembler. The result reads like a diagnostic observation, not a prescription.

This is a real trade-off. Sometimes the sanitized text reads awkwardly where a natural sentence was rewritten. But awkward-and-accurate beats smooth-and-liable.

What the AI Never Touches

Some components have a hard wall between them and any language model:

The scoring engine. 45 signals, weighted rules, structural floors and ceilings. If you have no documentation of authority relationships, your Accountability score cannot exceed 2, regardless of what the AI thinks. These rules are tested, version-controlled, and frozen.

The rendering layer. Every label, summary template, and stress-test scenario in the rendering layer is a lookup table. "Score 1 in Money" always renders as the same label. "Critical risk in Entity" always triggers the same visual treatment. The rendering specification is a frozen document that doesn't change between releases.

Cross-pattern detection. When the engine finds that your Tax score and your Entity score create a tension (say, tax residency in one country but entity registration in another), that pattern is detected by rule, not by AI inference. The AI describes the pattern once the rules have surfaced it.

Why Not Just Use Templates for Everything?

A fair question. If determinism is so important, why use AI at all?

Because 45 signals across 4 dimensions, with 8 cross-pattern types and layer-specific sections, produce a combinatorial space too large for static templates. A report for someone with a Delaware LLC and income from three countries looks structurally different from a report for someone with a Hong Kong holding company and a single freelance income stream. The findings overlap but the narrative context is completely different.

Templates would either be too generic ("Your Money score is 3") or require thousands of hand-written variants to capture the structural nuance. AI generates contextually appropriate language for each unique signal combination, and the deterministic system guarantees it stays within bounds.

The operating principle: determinism in analysis, intelligence in expression.

The Cost of This Architecture

This approach is not free. Running a three-stage pipeline with AI narration costs roughly $0.10 per META Diagnostic and $0.52 per L3 Judgment report. That's per report, every time, because we don't cache results. Same input tomorrow gets a fresh run through the same deterministic analysis with freshly generated narratives.

No caching is intentional. If we update a scoring rule or add a new cross-pattern type, every future report reflects the latest logic. Cached reports would carry stale analysis.

The other cost is development velocity. Every new feature has to respect the determinism boundary. Adding a new signal means updating the type definitions, the mapper, the scoring rules, the narrator prompts, and the assembler output. Adding a new section means updating section definitions, narrator templates, and assembly logic. There are no shortcuts where "the AI will figure it out."

That rigidity is the product.

What This Means for Your Report

When you receive a META Risk Profile, the scores, findings, and structural patterns in it are produced by auditable, repeatable rules. The language wrapping those findings is generated by AI and then verified against a boundary checklist.

If you ran the same diagnostic again with the same answers, you'd get the same scores and the same structural findings. The specific sentences might vary slightly (AI doesn't produce identical text twice), but the diagnostic content is identical.

That's the contract: the structure is deterministic, and the language is generated within enforced boundaries.


Key Takeaways

  • The META diagnostic uses a three-stage pipeline: deterministic analysis, AI narration, and deterministic assembly
  • Risk scores (1-5 per dimension) are produced by weighted rules, not AI inference. If the AI disagrees with the score, the rules win
  • AI handles language generation under strict constraints: a post-processing verifier strips prohibited directive, promise, and compliance language
  • The rendering layer and scoring engine have zero AI involvement by design
  • Reports are never cached. Each run uses the latest scoring rules and generates fresh narratives
  • The trade-off: development is slower, but every report is auditable and repeatable

References

Check your structural risk โ†’ Free 5-minute assessment

Related Articles

JF

Jett Fu

Cross-border entrepreneur running businesses across the US, China, and beyond. I built Global Solo to map the structural risks I wish someone had shown me.

My story โ†’

Where does your structure have gaps?

7 questions. 2 minutes. See which of the four META dimensions need attention โ€” free, no signup.

Free Risk Check

Structural Patterns

One blind spot, every two weeks. For entrepreneurs operating across borders.