Skip to content

Chapter 11: Multi-Party Verification Design

"The question isn't whether to trust. The question is how to structure trust so that no single party can betray it without detection."

Overview

Chapter 10 established the technical tools for verification—satellites, meters, sensors, blockchain. This chapter addresses the institutional architecture: who does the verifying, how are they selected, and who watches them?

We present a two-phase strategy: 1. Phase 1: Co-opt existing verification institutions (auditors, rating agencies, regulators) to enable adoption through aligned incentives 2. Phase 2: Transition to distributed, citizen-verifiable methods for long-term corruption resistance

The goal is not to destroy existing systems but to evolve them—bringing incumbents along so they don't fight the transition.

Chapter Structure:

  1. The Multi-Party Principle — Why single-point verification fails
  2. Phase 1: Co-opting Existing Systems — Leverage what works
  3. Phase 2: Distributed Verification — The long-term vision
  4. The Federated Model — National, regional, global hierarchy
  5. Who Verifies the Verifiers? — Meta-verification deep dive
  6. Funding and Capture — Case studies and principles
  7. Dispute Resolution and Arbitration — Binding mechanisms
  8. Incentive Alignment — Making honesty profitable
  9. Transition Roadmap — From Phase 1 to Phase 2

11.1 The Multi-Party Principle

Why Single-Point Verification Fails

Every verification scandal shares a common feature: a single party held too much power with too little accountability.

Enron: Arthur Andersen was sole auditor, deeply embedded, and captured.

LIBOR: Submitting banks self-reported with no cross-validation.

Credit Ratings (2008): Issuers paid rating agencies, creating structural conflict.

OPEC: Member states self-report with no external verification.

The pattern is clear: monopoly + opacity = corruption.

The Solution: Distributed Trust

Multi-party verification distributes trust across:

  • Multiple verifiers: No single entity can approve false claims
  • Different verification types: Technical, financial, physical
  • Cross-validation: Independent parties must agree
  • Transparency: Data available for public scrutiny
  • Accountability: Verifiers face consequences for errors

Redundancy Levels

Different claims warrant different redundancy:

Claim Type Minimum Verifiers Consensus Required
Small producer (<1 MW) 1 + automated cross-check Automated pass
Medium producer (1-100 MW) 2 independent verifiers Both must agree
Large producer (>100 MW) 3 independent verifiers 2 of 3 agreement
Disputed claims 5 verifiers + physical audit 4 of 5 agreement

Higher stakes = higher redundancy.


11.2 Phase 1: Co-opting Existing Systems

Strategic Rationale

New institutions face adoption barriers: - No track record - No existing relationships - Incumbents as adversaries

Existing verification institutions have: - Established credibility - Trained personnel - Regulatory recognition - Business models to protect

The insight: Make K-Dollar verification profitable for existing institutions, and they become allies instead of obstacles.

Financial Auditors (Big Four)

Current State: Deloitte, PwC, EY, and KPMG dominate global audit. They already audit energy companies.

K-Dollar Role: - Certified K-Dollar Auditors (CKA) designation - Audit of financial records supporting production claims - Verification of settlement data against grid records - Liability for false attestations

Incentive Alignment: - New revenue stream (K-Dollar audit fees) - Competitive differentiation (CKA certification) - Existing client relationships leveraged - Regulatory mandate creates demand

Limitations: - Big Four have faced capture criticism (Enron, Wirecard) - Not equipped for technical energy verification - Financial focus, not physical

Technical Auditors (ISO Certification Bodies)

Current State: DNV, Bureau Veritas, SGS, TÜV provide technical certification globally. Already certify renewable installations.

K-Dollar Role: - Physical verification of installations - Meter certification and calibration - Sensor network validation - Periodic site inspections

Incentive Alignment: - Expanded scope for existing services - Higher value certifications (K-Dollar backing) - Government mandate creates stable demand

Strengths: - Technical competence in energy sector - Global presence - Existing ISO 50001 (energy management) expertise

Credit Rating Agencies

Current State: Moody's, S&P, Fitch rate sovereign and corporate creditworthiness. Already assess energy sector.

K-Dollar Role: - Energy Production Ratings (EPR) for sovereigns - Assessment of national verification infrastructure - Early warning on verification credibility degradation

Incentive Alignment: - New rating category (energy-related) - Sovereign clients need EPR for K-Dollar participation - Subscription revenue from K-Dollar Authority

Limitations: - Issuer-pays model creates conflicts - 2008 failures demonstrated capture potential - Not technical energy experts

National Energy Regulators

Current State: FERC (US), OFGEM (UK), BNetzA (Germany), etc. regulate domestic energy markets.

K-Dollar Role: - Primary national verifiers for domestic production - Certification of meters and sensors - First-level dispute resolution - Data reporting to regional/global authorities

Incentive Alignment: - Expanded mandate and budget - International coordination role - K-Dollar participation requires competent regulation

Strengths: - Sovereign legitimacy - Existing legal authority - Local knowledge and presence

Co-optation Summary

Institution Type K-Dollar Role Primary Incentive
Financial auditors Financial attestation New revenue stream
Technical auditors Physical verification Mandate expansion
Rating agencies Credibility assessment New rating category
National regulators Primary verification Expanded authority

Phase 1 Outcome: K-Dollar verification piggybacks on existing infrastructure, gaining immediate credibility and avoiding institutional resistance.


11.3 Phase 2: Distributed Verification

The Corruption Risk

Phase 1 institutions, while useful for adoption, carry capture risks:

  • Regulatory capture: Regulators favor regulated industries
  • Client capture: Auditors depend on client fees
  • Oligopoly: Limited competition reduces accountability

Long-term K-Dollar integrity requires evolution toward distributed verification.

Citizen Verification

Concept: Enable ordinary citizens to verify energy production claims using public data.

Components:

  1. Open Data Requirements: All K-Dollar production data publicly available (privacy-preserving where needed)

  2. Verification Tools: Open-source software to cross-check claims against:

  3. Satellite imagery (public archives)
  4. Grid frequency data (publicly available)
  5. Weather data (irradiance, wind speed)
  6. Import/export statistics

  7. Bounty Programs: Rewards for discovering discrepancies

  8. Percentage of avoided fraud
  9. Whistleblower protections
  10. Anonymous submission channels

  11. Citizen Auditor Networks: Trained volunteers conducting spot-checks

  12. Model: Wikipedia editors, OpenStreetMap contributors
  13. Reputation systems for consistent accuracy
  14. Escalation to professional auditors when needed

Decentralized Verifier Selection

Current Problem: Verifiers chosen by central authority create central point of failure.

Solution: Randomized, stake-weighted verifier selection.

Mechanism: 1. Verifiers stake K-Dollar collateral (skin in the game) 2. For each verification request, algorithm selects from qualified pool 3. Selection probability proportional to stake and past accuracy 4. Verifiers don't know in advance which claims they'll verify 5. False attestation → stake slashed

Technical Implementation:

Verifier Pool:
├── Accredited verifiers (Phase 1 institutions)
├── Community verifiers (trained citizens)
└── Algorithmic cross-checks (automated)

Selection for claim C:
1. Hash(C, block_height) → random seed
2. Weighted random selection from pool
3. Minimum 2 verifiers per claim
4. Disputes → additional verifiers added

Prediction Markets for Verification

Concept: Allow markets to bet on verification accuracy.

Example: - Country X claims 100 TWh production in Q4 - Prediction market: "Will independent audit confirm ≥95 TWh?" - Market participants stake on outcome - If audit reveals <95 TWh, those betting "No" win

Value: - Aggregates distributed information - Creates financial incentive for scrutiny - Early warning on suspicious claims

Limitations: - Requires liquid, manipulation-resistant markets - Regulatory complexity (gambling laws) - Not a replacement for physical verification

Phase 2 Timeline

Year Milestone
1-3 Phase 1 fully operational
3-5 Citizen verification tools released
5-7 Bounty programs established
7-10 Decentralized verifier selection pilots
10+ Gradual transition to distributed model

Critical Principle: Phase 1 institutions are not replaced; they're supplemented. Competition from distributed verification keeps incumbents honest.


11.4 The Federated Model

Three-Tier Hierarchy

K-Dollar verification operates at three levels:

┌─────────────────────────────────────┐
│         GLOBAL (Standards)          │
│   K-Dollar Verification Authority   │
│   - Sets global standards           │
│   - Maintains verifier registry     │
│   - Final appeals                   │
└──────────────────┬──────────────────┘
┌──────────────────▼──────────────────┐
│         REGIONAL (Disputes)         │
│   EU, ASEAN, AU, etc.              │
│   - Cross-border verification       │
│   - Regional dispute resolution     │
│   - Standard harmonization          │
└──────────────────┬──────────────────┘
┌──────────────────▼──────────────────┐
│         NATIONAL (Operations)       │
│   National Energy Authorities       │
│   - Primary verification            │
│   - Verifier certification          │
│   - First-level disputes            │
└─────────────────────────────────────┘

National Level

Responsibilities: - Certify verifiers operating within jurisdiction - Receive and process production reports - Conduct or commission audits - Resolve domestic disputes (first instance) - Report to regional authority

Sovereignty Accommodation: - Nations retain primary authority over domestic verification - Domestic verifier selection follows national law - National data privacy rules respected - Gradual trust-building, not imposed standards

Requirements for K-Dollar Participation: - Functional energy regulatory authority - Adoption of minimum verification standards - Data reporting to regional level - Acceptance of regional dispute resolution

Regional Level

Regional Verification Authorities (RVAs) handle:

  1. Cross-Border Verification:
  2. Energy flows between nations
  3. Shared infrastructure (pipelines, grids)
  4. Import/export reconciliation

  5. Dispute Resolution:

  6. Appeals from national decisions
  7. Disputes between nations
  8. Verification methodology disagreements

  9. Capacity Building:

  10. Training for national verifiers
  11. Technology transfer
  12. Harmonization of national standards

Proposed RVAs: | Region | Anchor Institution | Coverage | |--------|-------------------|----------| | Europe | ACER (existing) | EU + associated | | North America | New body or FERC extension | US, Canada, Mexico | | Asia-Pacific | ASEAN Secretariat extension | East/Southeast Asia | | Middle East/North Africa | New body | MENA region | | Sub-Saharan Africa | African Union commission | AU members | | South America | New body or Mercosur | South America |

Global Level

K-Dollar Verification Authority (KVA) responsibilities:

  1. Standard Setting:
  2. Technical specifications for meters, sensors
  3. Data format standards
  4. Verifier qualification requirements
  5. Accuracy and reliability thresholds

  6. Verifier Registry:

  7. Global database of accredited verifiers
  8. Performance tracking
  9. Blacklist for disqualified verifiers

  10. Final Appeals:

  11. Last resort for unresolved disputes
  12. Interpretation of verification standards
  13. Precedent-setting decisions

  14. Coordination:

  15. Inter-regional harmonization
  16. Emergency response (verification system failures)
  17. Annual verification integrity reports

Governance: KVA governance is addressed in Chapter 13 (Governance).


11.5 Who Verifies the Verifiers?

The Infinite Regress Problem

If verifiers verify producers, who verifies verifiers? If meta-verifiers verify verifiers, who verifies meta-verifiers?

This regress has no logical endpoint. We must accept it and design for it.

Practical Solutions

1. Mutual Cross-Verification

Verifiers verify each other:

  • Peer Review: Verifiers audit each other's work samples
  • Cross-Assignment: Random assignment of verifiers to audit other verifiers' certifications
  • Discrepancy Tracking: Systematic comparison reveals outliers

Implementation:

For verifier V:
- 10% of V's certifications randomly selected
- Re-verified by different verifier V'
- Discrepancy rate tracked over time
- High discrepancy → investigation

2. Statistical Anomaly Detection

Aggregate patterns reveal individual misconduct:

  • Producer Clustering: If verifier V's producers consistently claim higher output than environmental conditions warrant, V is suspect
  • Verifier Comparison: If V's approval rate differs significantly from peers, investigate
  • Temporal Analysis: Sudden changes in verification patterns trigger review

Example Metrics: | Metric | Threshold | Action | |--------|-----------|--------| | Approval rate vs. peer average | >2 std dev | Review | | Discrepancy with satellite estimates | >15% systematic | Audit | | Appeal reversal rate | >10% | Training/review | | Complaint rate | >3x average | Investigation |

3. Whistleblower Mechanisms

Internal reporting channels:

  • Protected Disclosure: Legal protection for verifier employees reporting misconduct
  • Anonymous Channels: Secure, anonymous submission of evidence
  • Bounties: Financial rewards for substantiated fraud reports

4. Random Physical Audits

Spot-checks by independent parties:

  • Audit Pool: Independent auditors separate from regular verifiers
  • Random Selection: Producers selected by cryptographically verifiable randomness
  • Unannounced: No advance notice (within regulatory limits)
  • Results Published: Aggregate audit results publicly available

5. Rotating Verification Assignments

Prevent cozy relationships:

  • Maximum Tenure: Verifier can certify same producer for maximum 3-5 years
  • Cooling-Off Periods: After tenure, 2-year gap before reassignment
  • Random Assignment: Verifiers don't choose clients

6. Public Accountability

Transparency as verification:

  • Verifier Scorecards: Public performance metrics for each verifier
  • Open Data: Verification decisions and reasoning published
  • Media/NGO Scrutiny: Investigative journalists and watchdog organizations access verification data

The Audit of Auditors

Meta-Auditor Function:

At global level, a dedicated Meta-Audit Division:

  1. Verifier Accreditation: Initial and renewal certification
  2. Performance Monitoring: Ongoing accuracy tracking
  3. Investigation: Responds to complaints and anomalies
  4. Sanctions: Suspension, decertification, referral for prosecution

Independence Requirements: - Separate budget from verification operations - Direct reporting to KVA board - Staff prohibited from moving to/from verifier roles - Rotating leadership (maximum 5-year terms)

Who Audits the Meta-Auditors? - KVA Board oversight - External review by rotating international panels - Public transparency of meta-audit activities - Ultimately: citizen verification and political accountability

The regress ends not in perfect verification but in distributed accountability—enough watchful eyes that systematic corruption becomes impractical.


11.6 Funding and Capture: Case Studies

Why Funding Structure Matters

Who pays the verifier shapes what the verifier finds. Every major verification failure traces back to funding conflicts.

Case Study 1: Credit Rating Agencies (2008)

Funding Model: Issuer-pays (securities issuers pay for ratings)

Conflict: Agencies compete for issuer business. Tough ratings lose clients.

Failure Mode: - AAA ratings on subprime mortgage securities - 93% of AAA-rated subprime MBS downgraded to junk - Agencies later testified they knew models were flawed - Market share concerns overrode accuracy concerns

Outcome: Global financial crisis, $10T+ losses

Lesson for K-Dollar: Verifiers cannot be paid by those they verify.

Case Study 2: Arthur Andersen and Enron

Funding Model: Client-pays for audit + consulting

Conflict: Andersen earned $52M/year from Enron (including $27M consulting). Losing client = losing revenue.

Failure Mode: - Auditors approved aggressive accounting - Consulting relationship created intimacy - "One firm" culture meant auditors wouldn't challenge consultants' clients - Enron collapse revealed systemic failures

Outcome: Andersen dissolved, Sarbanes-Oxley enacted

Lesson for K-Dollar: Verifiers cannot provide consulting to those they verify. Strict separation.

Case Study 3: LIBOR Manipulation

Funding Model: Submitting banks self-reported with no external verification

Conflict: Banks' trading positions benefited from specific rate levels

Failure Mode: - Systematic manipulation of submitted rates - Coordinated between banks via chat rooms - No independent verification of submissions - Regulators relied on bank submissions

Outcome: $9B+ in fines, criminal convictions, rate reformed

Lesson for K-Dollar: Self-reporting without verification is invitation to fraud.

Case Study 4: ISO Certification Bodies

Funding Model: Auditee-pays, but with market competition and accreditation oversight

Conflict: Bodies compete for clients, creating pressure to pass

Mitigation Measures: - National accreditation bodies oversee certification bodies - Witness audits (accreditors observe auditors) - Complaint mechanisms - Rotation requirements

Results: Mixed. Some capture still occurs (rubber-stamp certifications), but accreditation layer reduces worst abuses.

Lesson for K-Dollar: Multi-layer oversight with accreditation reduces but doesn't eliminate capture.

Funding Principles for K-Dollar Verification

Based on case studies, K-Dollar verifier funding should follow these principles:

Principle 1: Transaction-Based Fees, Pooled Distribution

Mechanism: All K-Dollar transactions include small verification fee (0.01-0.1%). Fees flow to central pool. Pool distributed to verifiers based on performance metrics, not client relationships.

Why it Works: Verifiers compete on accuracy, not on pleasing clients.

Principle 2: No Direct Client-Verifier Payment

Mechanism: Producers never pay verifiers directly. All payments through K-Dollar Authority.

Why it Works: Eliminates direct financial relationship that creates capture.

Principle 3: Separation of Services

Mechanism: Verifiers cannot provide consulting, advisory, or other services to entities they verify. Cooling-off periods (5+ years) before switching between verification and consulting roles at same entity.

Why it Works: Prevents conflicts that captured Arthur Andersen.

Principle 4: Rotating Assignments

Mechanism: Maximum tenure of 3-5 years verifying same entity. Assignment via algorithm or pool, not client choice.

Why it Works: Prevents relationship capture over time.

Principle 5: Performance-Based Compensation

Mechanism: Verifier compensation partially tied to accuracy (measured via cross-validation, audits, whistleblower findings). Bonus for detecting fraud; penalty for missing it.

Why it Works: Aligns verifier incentives with accuracy, not volume.

Charter Principles for Verification Bodies

Organizational charters for K-Dollar verification bodies should incorporate these principles:

Independence: - Governance separate from entities being verified - Board members without financial interest in verification outcomes - Protected tenure for leadership (not removable for unpopular findings)

Accountability: - Clear reporting lines to oversight bodies - Public disclosure of verification decisions and reasoning - Whistleblower protections for staff - Regular external review of operations

Competence: - Minimum qualification standards for verifiers - Continuing education requirements - Technical certification for energy-specific verification - Language and local knowledge requirements for regional operations

Transparency: - Published methodologies - Open access to aggregate verification data - Annual reports on verification activities and outcomes - Public complaint mechanisms

Funding: - Diverse revenue sources (no single-payer dominance) - Multi-year budget cycles (insulation from short-term pressure) - Prohibited revenue sources (no consulting, no direct producer payments) - Performance-based component

Detailed charter templates are beyond the scope of this proposal, but these principles should guide their development.


11.7 Dispute Resolution and Arbitration

Dispute Categories

Category Examples Resolution Level
Technical Meter calibration disagreement National
Cross-border Disputed energy flows Regional
Fraud allegations Falsified production claims National → Regional
Systemic National verification system failure Regional → Global
Verifier misconduct Auditor corruption Meta-Audit Division

First Instance: National Resolution

Process: 1. Complaint filed with National Energy Authority 2. 30-day initial review 3. Technical investigation (if warranted) 4. Decision within 90 days 5. Remedies: correction, penalty, referral

Available Remedies: - Data correction - K-Dollar adjustment (claw-back or credit) - Verifier sanction (warning, suspension) - Producer penalty (collateral forfeiture) - Referral to regional level

Second Instance: Regional Appeals

Grounds for Appeal: - Procedural error at national level - New evidence - Legal interpretation dispute - Alleged national authority bias

Process: 1. Appeal filed within 60 days of national decision 2. Regional panel reviews record 3. Oral hearing (if requested and granted) 4. Decision within 120 days 5. Decision binding unless appealed to global level

Panel Composition: - 3 arbitrators from different nations within region - Selected from accredited arbitrator roster - Party input on selection (limited challenges)

Final Instance: Global Arbitration

Access Criteria: - Matter of global significance - Novel legal interpretation required - Systemic failure allegation - Leave to appeal granted by KVA

Process: 1. Application for leave within 30 days 2. Leave decision within 45 days 3. If granted, full briefing and hearing 4. Decision within 180 days 5. Decision final and binding

Tribunal Composition: - 5 arbitrators from different regions - At least one energy law specialist - At least one technical expert (non-lawyer) - Chair selected by arbitrators

Financial Remedies

Available in Arbitration:

Remedy Application
Corrective adjustment K-Dollar issued/recalled based on accurate data
Compensatory damages Losses caused by false verification
Penalty (producer) Up to 3x value of fraudulent K-Dollar claims
Penalty (verifier) Decertification + damages
Systemic remedies National verification reform requirements

Enforcement: - K-Dollar Authority executes financial adjustments directly - National enforcement of penalties against entities within jurisdiction - Cross-border enforcement via regional agreements - Persistent non-compliance → exclusion from K-Dollar system

Emergency Procedures

For urgent matters (ongoing fraud, system failure):

  • Interim Measures: K-Dollar issuance suspended pending resolution
  • Expedited Process: Decision within 30 days
  • Provisional Relief: Temporary adjustments pending final decision

11.8 Incentive Alignment

Making Honesty Profitable

Verification integrity depends on incentives, not just rules.

Producer Incentives

Collateral Requirement: - Producers stake K-Dollar collateral proportional to production claims - Collateral ratio: 5-10% of annual K-Dollar generation - False claims → collateral forfeiture (3x gain)

Example: - Producer claims 100 MWh/year → 100 K$ generation - Collateral requirement: 5-10 K$ - If fraud discovered: loses 5-10 K$ collateral + 3x fraudulent amount

Reputation System: - Verified production history tracked publicly - Long track record of accuracy → lower collateral requirements - New producers → higher scrutiny, higher collateral

Verifier Incentives

Stake Requirement: - Verifiers stake collateral proportional to volume certified - Stake slashed for false certifications - Stake grows with successful track record

Performance Bonuses: - Accuracy bonuses for verifiers with low error rates - Early fraud detection bonuses - Peer recognition for catching errors

Career Risk: - Individual verifiers personally certified - Professional license at stake - Career consequences for misconduct

Whistleblower Incentives

Bounty Structure: - 10-25% of avoided/recovered fraudulent K-Dollar - Protected identity - Legal protection from retaliation - Expedited review for credible reports

System-Wide Incentives

Insurance Requirements: - Verifiers carry professional liability insurance - Insurers have incentive to monitor verifier conduct - Premium adjustments reflect accuracy record

Market Mechanisms: - Verified energy trades at premium - Poorly-verified production faces market discount - Financial markets create additional scrutiny layer


11.9 Transition Roadmap

Phase 1 Implementation (Years 1-5)

Year 1: - Pilot with 5-10 countries - Existing regulators as national verifiers - Big Four + technical auditors as certified verifiers - Basic dispute resolution procedures

Year 2-3: - Expand to 30+ countries - Regional verification authorities established - Cross-border verification operational - First arbitration cases processed

Year 4-5: - Global coverage target - Full dispute resolution system operational - Meta-audit function established - Phase 2 pilots initiated

Phase 2 Transition (Years 5-15)

Year 5-7: - Citizen verification tools released - Open data requirements enforced - Bounty programs launched - Community verifier pilots

Year 7-10: - Decentralized verifier selection pilots - Prediction markets experiments - Phase 1 institutions remain but supplemented - Performance comparison: Phase 1 vs. Phase 2 methods

Year 10-15: - Gradual weight shift toward distributed verification - Phase 1 institutions as backstop, not primary - Full citizen verification capability - Continuous evolution based on results

Success Metrics

Metric Phase 1 Target Phase 2 Target
Verification coverage 90% of K-Dollar production 99%+
Fraud detection rate 80% of significant fraud 95%+
Dispute resolution time <180 days average <90 days
Public data availability Basic Full
Citizen participation Minimal Substantial

11.10 Key Takeaways

  1. Multi-party verification is essential: No single verifier can be trusted. Distributed trust prevents capture.

  2. Phase 1 co-opts existing systems: Financial auditors, technical auditors, rating agencies, and regulators provide immediate credibility and avoid institutional resistance.

  3. Phase 2 distributes verification: Citizen verification, decentralized selection, and prediction markets provide long-term corruption resistance.

  4. Bring incumbents along: Don't fight existing institutions—make K-Dollar profitable for them, then gradually introduce competition.

  5. Federated model respects sovereignty: National verification with regional disputes and global standards balances practicality with integrity.

  6. Meta-verification addresses regress: Mutual cross-verification, statistical detection, whistleblowers, random audits, and public accountability create distributed oversight.

  7. Funding structure determines integrity: Case studies (rating agencies, Andersen, LIBOR, ISO) show that payment models create capture. Transaction-based pooled fees and no direct client payment are essential.

  8. Binding arbitration enables enforcement: Full financial remedies make verification decisions meaningful.

  9. Incentives matter more than rules: Collateral, stakes, bounties, and reputation systems make honesty profitable.


Further Reading

  • Coffee, J. (2006). Gatekeepers: The Professions and Corporate Governance
  • Power, M. (1997). The Audit Society
  • Ostrom, E. (1990). Governing the Commons
  • IOSCO (2015). "Code of Conduct Fundamentals for Credit Rating Agencies"
  • McLean, B. & Elkind, P. (2003). The Smartest Guys in the Room (Enron)

Next: Chapter 12: Default and Crisis Scenarios