Technical Whitepaper

Dialectical Claim Verification
with Source Provenance

How True Source performs adversarial evidence gathering, reliability-weighted verdict synthesis, and financial transparency integration to fact-check claims in under 60 seconds.

v1.0March 2026truesource.app

1.Abstract

True Source is a real-time media intelligence system that performs autonomous, claim-level fact verification on Twitter/X. Unlike existing approaches that rate publications (NewsGuard), crowdsource corrections on multi-hour timescales (Community Notes), or generate confident but unsourced responses (Grok), True Source decomposes content into atomic factual claims, gathers adversarial evidence from diverse sources, applies reliability-weighted verdict synthesis, and delivers transparent, cited results in under 60 seconds.

The system simultaneously surfaces source ownership, political funding data (FEC/SEC), and poster political lean, providing users not just with a truth assessment but with the financial incentive context behind the information. An autonomous scanning layer proactively identifies viral misinformation without user prompting. The platform is monetized through tiered subscriptions, a B2B verification API, and a Solana-based token ($TRUTH) that provides token-gated access.

2.Problem Statement

The misinformation landscape suffers from three compounding failures:

2.1 Speed Asymmetry

False claims propagate at network speed. A 2018 MIT study published in Science found that false news reaches 1,500 people six times faster than true stories. Current corrective mechanisms operate on fundamentally incompatible timescales: Community Notes has a median response time exceeding 14 hours, and 90% of submitted notes never reach publication. Professional fact-checkers (PolitiFact, Snopes) produce high-quality analysis but cover a fraction of viral content, often days after peak virality.

2.2 Granularity Gap

Existing automated tools operate at the wrong level of abstraction. NewsGuard and Provenance rate publications, not individual articles or claims. A credible outlet can publish a misleading article; a low-credibility source can occasionally report accurately. Ground News visualizes media bias distribution but does not assess whether specific claims are true. Grok generates confident natural-language responses but fabricates citations: a Stanford Internet Observatory analysis found a 94% citation error rate in Grok's fact-checking responses.

2.3 Opacity of Incentives

No existing tool systematically connects factual claims to the financial incentives behind them. When a news outlet reports on legislation, users cannot easily determine that the outlet's parent company donated $2.3M to the bill's sponsors, or that the outlet's owner also owns companies affected by the legislation. This incentive layer is critical context for evaluating information but is entirely absent from current fact-checking tools.

3.System Architecture

True Source operates as three interconnected systems:

  1. Autonomous Twitter/X Agent: A persistent Node.js worker polling for mentions every 60 seconds, processing claims through a multi-stage verification pipeline, and posting structured replies with verdicts, citations, and source context.
  2. Web Platform (truesource.app): A Next.js application providing full analysis pages, source ownership visualization, news desert mapping (3,000+ US counties), government accountability dashboards, and user account management.
  3. B2B Verification API: REST endpoints exposing the verification engine for programmatic access by newsrooms, universities, and brand safety platforms.
Twitter Mention / Proactive Discovery | v [Guard Chain] ── NSFW filter, spam wall, rate limits, budget check | v [Intent Detection] ── Classify: tweet_analysis | url_analysis | claim_check | v [Conversation Chain Walker] ── Up to 4 parent tweets + quote tweets | v [Smart Target Selection] ── Score substantiveness, select best claim target | v [Multi-Modal Extraction] ── Claude Vision on all images in chain | v [Dialectical Verification Pipeline] ── (See Section 4) | v [8-Category Synthesis] ── Claude Sonnet with structured tool_use output | v [Reply Formatting] ── Emoji verdicts, source context, claim link | v [Post to Twitter + Save to Database]

Figure 1: End-to-end processing pipeline for a single mention.

4.Dialectical Verification Pipeline

The core innovation of True Source is its dialectical approach to claim verification. Rather than submitting a claim to an LLM and accepting its output (which inherits the model's training biases and hallucination tendencies), we treat verification as an adversarial evidence-gathering process.

4.1 Claim Decomposition

Input content (tweet text, image transcriptions, linked article text) is decomposed into atomic, independently verifiable claims using Claude Haiku. The decomposition prompt explicitly filters:

  • Opinions and subjective assessments
  • Questions and rhetorical devices
  • Jokes and sarcasm
  • Editorialization (loaded words separated from factual substance)

A typical tweet yields 3-8 atomic claims. Each claim proceeds through verification independently.

4.2 Semantic Precision Analysis

Before evidence gathering, each claim undergoes semantic precision analysis to identify language patterns that affect verification difficulty and confidence:

CategoryPatternsImpact
Absolutistnever, always, only, first-ever, 100%-15% confidence per flag
Overgeneralization"scientists say", "studies show", "everyone agrees"Requires attribution verification
Temporalcurrently, still, as of today, record-breakingTriggers temporal freshness weighting
QuantitativeSpecific percentages, dollar amounts, multipliersRequires numeric source matching
Causalcaused, led to, because, resulted inHigher evidence threshold

4.3 Parallel Adversarial Evidence Gathering

For each atomic claim, six evidence streams execute in parallel:

StreamSourcePurpose
Web Search (Diverse)Brave Search API + Google News RSSBoth supporting AND opposing sources via counter-query generation
Wikipedia LookupWikipedia APIGround truth for identity claims, entity roles, institutional facts
Primary Source Domainscongress.gov, BLS, EIA, SEC, TreasuryAuthoritative government and institutional data
Market DataCoinGecko, Alpha Vantage, Yahoo FinanceLive price/volume data for financial claims
Institutional Flow DataSoSoValue, CoinGlass, TheBlock, BloombergETF flows, institutional buying/selling
Political FundingOpenFEC API, Senate LDACampaign donations, lobbying disclosures

The web search stream is particularly important. For each claim, two query variants are generated:

  1. Affirmative query: Searches for evidence supporting the claim as stated.
  2. Counter-query: Automatically generated by flipping 17 keyword pairs (increased/decreased, approved/rejected, confirmed/denied, etc.) to actively seek contradicting evidence.

If the initial search returns fewer than 3 sources, a query reformulation step (Claude Haiku) generates 2-3 alternative search queries with different vocabulary to improve recall.

4.4 Source Classification

All gathered evidence items are classified by Claude Haiku into three stances:

  • Supporting: Evidence that corroborates the claim
  • Opposing: Evidence that contradicts the claim
  • Neutral: Related but neither confirming nor denying

This classification happens after evidence gathering, not before, preventing confirmation bias in the search phase.

4.5 Source Enrichment

The top 3 evidence sources (by relevance) have their full page content fetched (up to 5,000 characters each) using a multi-strategy extraction pipeline: standard fetch with browser headers, Node.js native HTTPS (different TLS fingerprint), and Google Webcache fallback. Content is extracted using Mozilla Readability for clean text parsing. This enrichment ensures the synthesis model has substantive evidence, not just search result snippets.

5.Evidence Weighting Methodology

Each evidence item receives a composite reliability score that determines its influence on the final verdict.

5.1 Source Reliability

Base reliability is derived from Media Bias/Fact Check (MBFC) factual reporting ratings, covering 3,900+ domains:

MBFC Factual RatingReliability Score
Very High0.95
High0.80
Mostly Factual0.65
Mixed0.40
Low0.20
Very Low0.10
Unknown / Not Rated0.50

5.2 Temporal Decay

For claims about current conditions (detected via temporal semantic flags), source freshness applies a decay multiplier:

temporal_weight(age_days) = age < 7 : 1.00 (full weight) age 7-30 : 0.85 age 30-90 : 0.60 age 90-365 : 0.40 age > 365 : 0.15 (heavily discounted)

5.3 Authoritative Source Overrides

Certain data sources carry inherent authority that overrides web search noise:

  • Market APIs (CoinGecko, Alpha Vantage): reliability 0.95. Live price data is authoritative for financial claims.
  • Government APIs (FEC, Treasury, BLS): reliability 0.80-0.90. Official records override journalistic interpretation.
  • Wikipedia entity data: Confirmed entity roles override web search contradictions for identity claims.

5.4 Verdict Determination

The weighted evidence produces a verdict through the following decision logic:

Given: S = sum of (reliability * temporal_weight) for all supporting sources O = sum of (reliability * temporal_weight) for all opposing sources T = S + O (total weighted evidence) Verdict rules: 1. No evidence at all -> UNVERIFIABLE (confidence: 0.30) 2. Supporting only -> TRUE (confidence: 0.50 - 0.95 scaled by count and quality) 3. Opposing only AND O >= 1.0 AND opposing_count >= 2 -> FALSE (confidence: scaled by O) 4. Opposing only AND O < 1.0 -> UNVERIFIABLE (insufficient opposing evidence) 5. Mixed, support_ratio > 0.65 -> TRUE (support_ratio = S / T) 6. Mixed, support_ratio < 0.35 -> FALSE 7. Mixed, 0.35 <= ratio <= 0.65 -> UNVERIFIABLE (genuinely contested)
Critical safeguard (Rule 4): A claim is never marked FALSE solely because "no supporting evidence was found." This prevents search failures, API outages, or poorly-formed queries from generating false negatives. FALSE requires affirmative counter-evidence from multiple reliable sources. This is the single most important design decision in the pipeline.

5.5 Confidence Adjustments

Final confidence scores are adjusted by semantic precision flags (-15% per high-severity absolutist/causal flag), source diversity (bonus for evidence from multiple independent outlets), and evidence freshness concentration (penalty if all evidence is older than 90 days for a current-events claim).

6.Source Provenance and Financial Transparency

True Source surfaces the financial incentive layer behind information, a dimension entirely absent from competing fact-checking systems.

6.1 Ownership Chain Mapping

For 484+ tracked news sources (and 5,000+ with bias ratings), the system maintains ownership chains mapping outlets to their parent companies and ultimate owners. Data is enriched from Wikidata entity relationships, SEC EDGAR corporate filings, and manual curation for complex multi-subsidiary structures.

6.2 Political Donation Integration

Using the OpenFEC API, the system retrieves employee political donation records for media parent companies. Donations are aggregated by party affiliation and presented as percentages (e.g., "Fox Corporation: 65% Republican, 32% Democrat, 3% Other"). This data appears directly in agent replies when a fact-checked tweet references a tracked news source.

6.3 Lobbying Disclosure

Senate Lobbying Disclosure Act (LDA) filings are queried for media parent companies, surfacing active lobbying relationships that may represent conflicts of interest relevant to the claims being verified.

6.4 Integration with Verification

Source provenance data is not merely displayed alongside verdicts; it informs the verification synthesis. When the Claude Sonnet synthesis model produces the final 8-category analysis, it receives source ownership and funding data as context, enabling it to flag potential conflicts of interest in its "Source Credibility" and "Bias Analysis" output categories.

7.Poster Profiling and Accuracy Tracking

7.1 Political Lean Detection

The system estimates poster political lean from multiple signals: handle keywords (partisan terms), bio content analysis, and historical tweet patterns. Lean is expressed as a direction (left/center/right) with a confidence score (0-1). This is presented transparently in agent replies as context, not as a judgment.

7.2 Accuracy History

Every fact-check creates a permanent record. When a poster is checked multiple times, their historical accuracy rate is computed (percentage of claims that verified as true across all checks). This track record is displayed in subsequent agent replies: "This poster's claims have checked out 3 of 7 times (43% accuracy)." The system gets more informative with each interaction.

8.Multi-Modal Analysis

8.1 Image Content Extraction

Tweets frequently contain claims embedded in images: charts, screenshots, infographics, headlines. The agent extracts content from all images in the conversation chain (up to 4 tweets, max 4 images per tweet) using Claude Haiku Vision. The extraction prompt requires both text transcription (OCR) and visual description (objects, people, scenes, chart types), ensuring that non-textual information is captured.

8.2 Cross-Reference Against Live Data

Extracted numerical claims from images (e.g., "BTC at $200K" from a trading screenshot) are automatically cross-referenced against live market APIs. This enables the system to identify manipulated screenshots, outdated charts presented as current, or deliberately misleading visualizations.

9.Proactive Misinformation Interception

A key architectural differentiator is the agent's proactive scanning capability. Rather than operating purely reactively (waiting for user mentions), the system actively searches for viral misinformation.

9.1 Discovery Process

Every 90 minutes, the agent executes rotating search queries across six categories: political claims, scientific assertions, economic data, breaking news, numeric claims, and technology predictions. Results are filtered for virality (minimum 1,000 likes, 200 retweets, 5,000 follower count) and factual content indicators (numbers, dates, attribution verbs, data language).

9.2 Scoring and Selection

Candidate tweets are scored by: engagement (likes + 3x retweets), account size bonuses, claim density, and quote tweet presence. The highest-scoring tweet with verifiable factual content is selected for analysis.

9.3 Delivery as Original Tweets

Proactive fact-checks are posted as original tweets (not replies) to avoid shadow-banning from unsolicited replies. This architectural decision was informed by observed Twitter/X platform behavior where unsolicited reply patterns trigger visibility reduction. The proactive budget is capped at 200 tweets per month from a total budget of 2,800.

10.Token Utility Model

10.1 $TRUTH Token

$TRUTH is a Token-2022 (Solana Token Extensions) asset launched on Pump.fun. Unlike standard SPL tokens, Token-2022 supports extended functionality including transfer hooks and metadata extensions.

10.2 Token-Gated Access

Holding $TRUTH provides platform access without requiring subscription payments:

Token BalanceAccess TierEquivalent Subscription
500,000+ $TRUTHINSIDER$1.99/month
2,000,000+ $TRUTHPRO$4.99/month
5,000,000+ $TRUTHNEWSROOM$24.99/month

10.3 Burn Mechanism

Subscription payments made in $TRUTH tokens are burned (sent to a burn wallet), creating deflationary pressure proportional to platform usage. As B2B API adoption scales, the burn rate increases with revenue.

10.4 Rate Limit Tiers

Token holdings (or paid subscriptions) determine agent interaction limits on Twitter:

TierAgent Requests/HourSaved ArticlesAPI Access
FREE25No
INSIDER350No
PRO5UnlimitedYes
NEWSROOM20UnlimitedFull

10.5 Revenue Share (Planned)

As B2B API revenue scales, a revenue share mechanism for token holders is planned. This aligns holder incentives with platform growth: the more newsrooms and enterprises that license the verification API, the more value flows to token holders.

11.Accuracy Analysis and Safeguards

11.1 Design Principles for Error Prevention

The system incorporates multiple safeguards against the most damaging error type: false negatives (marking true claims as false).

  • Asymmetric evidence thresholds: TRUE requires any supporting evidence. FALSE requires 2+ opposing sources with combined reliability above 1.0. This asymmetry ensures that search failures default to UNVERIFIABLE rather than FALSE.
  • Wikipedia entity override: When Wikipedia confirms an entity claim (e.g., a person's role or title), this overrides opposing web search noise that may reference outdated information.
  • Market API authority: Live price data from financial APIs (reliability 0.95) overrides web search results for financial claims, preventing outdated articles from contradicting current market conditions.
  • Temporal decay: Old sources are automatically downweighted for current-condition claims, preventing historical data from overriding present reality.

11.2 Known Limitation: ETF Flow Claims

In February 2026, the system incorrectly marked a BlackRock $117M ETF outflow claim as false. The root cause was that the evidence pipeline lacked institutional flow data sources (SoSoValue, CoinGlass), relying instead on general web search which conflated on-chain Coinbase withdrawals with ETF flows. This was corrected by adding dedicated institutional flow detection and authoritative financial data sources. The incident demonstrates both the system's failure mode (domain-specific data gaps) and its correction mechanism (adding specialized evidence streams).

11.3 Continuous Improvement

Every verification result is stored permanently. The growing dataset of 5,300+ verified claims with full evidence chains enables ongoing analysis of error patterns, evidence source reliability, and verdict distribution. The system's accuracy improves as: (a) the evidence source network expands, (b) poster accuracy histories accumulate, and (c) domain-specific verification modules are added for newly identified claim categories.

12.Conclusion

True Source addresses the misinformation crisis through three architectural innovations absent from existing solutions:

  1. Dialectical verification that treats fact-checking as adversarial evidence gathering rather than LLM question-answering, with asymmetric confidence thresholds that prevent hallucinated judgments.
  2. Financial transparency integration that surfaces source ownership, political donations, and lobbying data alongside factual verdicts, exposing the incentive layer behind information.
  3. Autonomous proactive scanning that intercepts viral misinformation in minutes rather than hours or days, operating continuously without human prompting.

The system is live and operational, processing claims 24/7 on Twitter/X with a growing dataset that compounds in value daily. The token-gated access model aligns user, holder, and platform incentives, while the B2B API extends the verification engine to institutional customers. As the evidence source network and domain-specific modules expand, verification accuracy and coverage will continue to improve.

True Source | truesource.app | @TrueSourceBot | $TRUTH on Solana

March 2026 | v1.0