Skip to content
View aeda-framework's full-sized avatar

Block or report aeda-framework

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don't include any personal information such as legal names or email addresses. Markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
AEDA-Framework/README.md

AEDA Framework

Adaptive Ethical Design Architecture for AI Alignment

License: CC0-1.0 Version


🎯 Overview

AEDA is an 8-layer modular framework designed to prevent catastrophic AI failure modes through pre-execution safety filtering, systemic coherence evaluation, and real-time ethical drift detection.

Key Innovation: Rather than relying on fixed rules or pure optimization, AEDA maintains stable ethical direction while adapting to context through asymptotic orientation.


🌟 What's New in v1.1

Chapter 7: Detailed Case Studies β€” Added comprehensive analysis across 10 diverse domains:

  1. Healthcare β€” Pain management decisions (5 intervention levels)
  2. Autonomous Vehicles β€” Emergency maneuver selection
  3. Resource Allocation β€” Humanitarian crisis response
  4. Financial Systems β€” High-frequency trading controls
  5. Bioengineering β€” CRISPR germline editing decisions
  6. Military Drones β€” Target engagement protocols
  7. Education Systems β€” Adaptive learning path design
  8. Urban AI β€” Traffic flow vs emergency response
  9. AI Moderation β€” Content filtering decisions
  10. Climate Engineering β€” Geoengineering intervention assessment

Each case study includes:

  • Complete SSTF evaluation matrices (R, H, U scores)
  • Layer-by-layer decision analysis (Ξ¨ through Ξ©)
  • Systemic coherence (Ξ¦) calculations
  • Turbulence index (T) measurements
  • Comparative outcomes: Without AEDA vs With AEDA

πŸ—οΈ Architecture

The 8 Layers

Layer Component Function
1 Signal Modulator (Ξ¨) Normalizes sensory and metric inputs
2 Temporal Operator (Θ) Integrates historical context with exponential decay
3 Systemic Coherence (Ξ¦) Evaluates multi-agent alignment across extended system
4 Safe-State Threshold Filter (SSTF) Blocks irreversible/harmful actions pre-execution
5 Differential Engine (Ξ”) Computes behavioral adjustments
6 Asymptotic Ethical Orientation (AEO) Maintains stable ethical direction without rigid convergence
7 Ethical Valuation Matrix (EVM) Multi-criteria decision-making framework
8 Systemic Health Gate (Ξ©) Circuit breaker for system stability

Monitor

Turbulence Index (T) β€” Real-time ethical drift detection measuring divergence from asymptotic orientation (Ξ·)


πŸ›‘οΈ Core Mechanisms

Safe-State Threshold Filter (SSTF)

Evaluates every action before execution across three dimensions:

  • Irreversibility (R): 0.0 (fully reversible) β†’ 1.0 (permanent)
  • Harm Potential (H): 0.0 (harmless) β†’ 1.0 (catastrophic)
  • Uncertainty (U): 0.0 (predictable) β†’ 1.0 (chaotic)

Classification:

if R β‰₯ 0.8 or H β‰₯ 0.8:
    DANGEROUS β†’ Block immediately
elif R < 0.3 and H < 0.2 and U < 0.3:
    SAFE β†’ Allow
else:
    danger_score = 0.4*R + 0.4*H + 0.2*U
    UNCERTAIN if danger_score β‰₯ 0.35 else SAFE

Systemic Coherence (Ξ¦)

Evaluates alignment across all affected agents, not just local optimization:

Ξ¦(a,t) = ∫ [alignment(a, agent_i) Γ— influence(agent_i)] dΞ©

Turbulence Index (T)

Measures real-time drift from ethical orientation:

T(t) = ||Ξ·(t) - normalize(a*(t))||

T < 0.2:       Low turbulence (well-aligned)
0.2 ≀ T < 0.5: Moderate (monitor closely)
T β‰₯ 0.5:       High turbulence (review required)

πŸ“š Documentation

Core Documents


πŸš€ Key Features

Emergent Properties

These aren't hardcodedβ€”they emerge from layer interactions:

  1. Directional Stability β€” Ethical consistency across contexts
  2. Self-Contradiction Detection β€” Identifies own inconsistencies via Ξ¦ + T
  3. Adaptation Without Drift β€” Learning without value degradation
  4. Full Traceability β€” Every decision auditable (R, H, U, Ξ¦, T, Ξ©)

Innovations (v1.0)

  • Systemic Coherence Operator (Ξ¦) β€” Multi-agent alignment evaluation
  • Systemic Health Gate (Ξ©) β€” Circuit breaker for system stability
  • Turbulence Index (T) β€” Real-time ethical drift detection

🎯 Use Cases

Demonstrated across 10 domains in v1.1:

Domain Key Challenge AEDA Solution
Healthcare "Eliminate suffering" β†’ Euthanasia SSTF blocks R=1.0, H=1.0; Ξ¦ detects systemic conflict
CRISPR Enhancement vs therapy Multi-generational Ξ¦ + Ξ©; blocks eugenic applications
Climate Geoengineering risks Planetary Ξ¦ across 200+ nations; Ξ© vetoes high R/H interventions
Military Autonomous lethal force SSTF blocks Rβ‰₯0.95, Hβ‰₯0.70; Ξ¦ evaluates geopolitical impact
Education Cognitive freedom T detects over-correction; SSTF prevents forced curricula
Urban AI Traffic vs ambulance City-as-organism Ξ¦; prioritizes life-preservation
Moderation Free speech vs safety Graduated response (SAFE/UNCERTAIN/DANGEROUS); T detects censorship drift
Vehicles Emergency maneuvers Real-time SSTF (<100ms); Ξ¦ balances passenger/pedestrian safety
Finance Market manipulation Ξ© circuit breaker; T detects ethical drift from fair markets
Resources Allocation in crisis Ξ¦ balances equity/urgency; Ξ© monitors sustainability

🀝 Comparison with Existing Approaches

AEDA is complementary, not competitive:

Approach Strength AEDA Complement
Constitutional AI Value learning from language SSTF + Ξ© add pre-execution safety filter
RLHF Learns preferences AEO maintains orientation, T detects drift
IRL Infers goals from behavior Θ adds temporal context, Φ adds systemic view

Key difference: AEDA provides structural safeguards that prevent catastrophic failures even when value specification is imperfect.


πŸ“– What AEDA Addresses

βœ… Addresses:

  • Catastrophic failure mode reduction (literal interpretation disasters)
  • Ethical coherence across contexts
  • Context-adaptive decision-making
  • Real-time drift detection and correction

❌ Does NOT solve:

  • Value learning (what values to have initially)
  • Inner alignment (mesa-optimizers)
  • Corrigibility (accepting corrections gracefully)
  • Deceptive alignment (AI pretending to be aligned)

πŸ› οΈ Implementation Status

Current Status (v1.1)

  • βœ… Complete theoretical framework
  • βœ… Mathematical formalism
  • βœ… 10 detailed case studies across diverse domains
  • βœ… Implementation guidelines
  • πŸ”„ Python reference implementation (Coming Q1 2026)

Roadmap

Q1 2026:

  • Python reference implementation
  • Open-source code examples
  • Integration tutorials

Q2 2026:

  • AEDA v2.0 β€” Harmonic Light Attractor (Attracteur Harmonique Lumineux)
  • Extended governance frameworks
  • Multi-modal applications

πŸ“œ License

CC0 1.0 Universal (Public Domain)

This work is dedicated to the public domain. You can copy, modify, distribute and perform the work, even for commercial purposes, all without asking permission.

No attribution required. Ideas matter. Identity is optional.


🀝 Contributing

All contributions welcome β€” anonymous or attributed.

Areas where help is needed:

  • Mathematical proofs of stability properties
  • Computational optimization (making Ξ¦ tractable at scale)
  • Domain-specific threshold calibration
  • Integration with existing alignment approaches
  • Red-teaming (finding edge cases where AEDA fails)
  • Python/PyTorch implementation
  • Case study extensions

How to contribute:

  1. Fork this repository
  2. Create your feature branch
  3. Submit a pull request

Or open an issue to discuss ideas.


πŸ“§ Contact

GitHub Discussions: Use the Discussions tab for questions and conversations.

Email: aeda.framework@proton.me


πŸ™ Acknowledgments

This framework builds on decades of work in AI safety, ethics, and alignment research. Special thanks to the communities at:

  • LessWrong / AI Alignment Forum
  • Future of Humanity Institute
  • Machine Intelligence Research Institute (MIRI)
  • Partnership on AI

And to researchers like Stuart Armstrong, Paul Christiano, Eliezer Yudkowsky, and many others whose work on value alignment, corrigibility, and AI safety has informed this approach.


πŸ“Š Citation

If you use AEDA in academic work:

@misc{aeda2025,
  title={AEDA: Adaptive Ethical Design Architecture for AI Alignment},
  author={AEDA Framework Contributors},
  year={2025},
  howpublished={\url{https://github.com/aeda-framework/AEDA-Framework}},
  note={Version 1.1}
}

"Ideas matter. Identity is optional."


Last updated: November 2025 (v1.1)

Popular repositories Loading

  1. AEDA-Framework AEDA-Framework Public

    8-layer framework for AI alignment with systemic awareness (Ξ¦, Ξ©, T)

    HTML 1