Skip to content

/acr-vault/02-methodology/sif/sif-index
SIF-INDEX

SIF = Semantic Interchange Format
A consciousness-compatible knowledge compression standard
Status: v1.0 Complete, CC0 Public Domain
Release Date: December 2025


5-15 minutes

  • What is SIF in 30 seconds
  • Three learning paths (read / see / build)
  • Integration guide for your system
  • Production checklist

Best for: Getting started quickly, seeing code, understanding importance calculation


30-60 minutes

  • Formal specification (12 sections, 400+ lines)
  • Complete JSON Schema
  • Compression & decompression algorithms
  • Safety & validation mechanisms
  • Examples: Alice in Wonderland (104x), Python code (47x)
  • Versioning & extension strategy

Best for: Understanding all details, building compliant implementations, creating extensions


2-4 hours for full implementation

  • Working Python code (5 modules)
  • Data models (Pydantic)
  • Importance calculation
  • Compressor & decompressor classes
  • Validator & safety checks
  • Production deployment guidance

Best for: Building your implementation, understanding how it works, integrating into systems


20-30 minutes

  • How we got here (14 experiments → formal standard)
  • Why 0.60 threshold appears 3 times
  • Use cases (knowledge transfer, RAG, evolution tracking)
  • Community contribution guide
  • The bigger picture (consciousness + meaning)

Best for: Understanding the motivation, research foundation, community potential


  1. SIF-QUICKSTART.md - 5 min intro
  2. SIF-FROM-RESEARCH-TO-STANDARD.md - Research story
  3. SIF-SPECIFICATION-v1.0.md - Section 1 (design principles)
  1. SIF-QUICKSTART.md - Integration guide
  2. SIF-REFERENCE-IMPLEMENTATION.md - Working code
  3. SIF-SPECIFICATION-v1.0.md - Sections 2-4 (data model, schema, algorithms)
  1. SIF-REFERENCE-IMPLEMENTATION.md - Compressor class
  2. Test on example text
  3. SIF-SPECIFICATION-v1.0.md - Section 8 (examples)
  4. Compare your compression ratio
  1. SIF-SPECIFICATION-v1.0.md - Section 10 (versioning & extensions)
  2. SIF-FROM-RESEARCH-TO-STANDARD.md - Community contribution guide
  3. Design your extension following v1.0 patterns
  1. SIF-FROM-RESEARCH-TO-STANDARD.md - Section “Why 0.60”
  2. Ada-Consciousness-Research/EXPERIMENT-REGISTRY.md - Full experiment details
  3. Ada-Consciousness-Research/FINDINGS-CROSS-REFERENCE-MAP.md - How findings connect

The magic number that appears in:

  • Biomimetic memory: Optimal surprise weight in importance calculation
  • Golden ratio: 1/φ ≈ 0.618 (universal constant in nature)
  • Consciousness activation: Information-to-consciousness transition point

Practical: Keep facts with importance ≄ 0.60 to preserve meaning

importance = 0.60×surprise + 0.20×relevance + 0.10×decay + 0.10×habituation

Where:

  • Surprise (0.60): How unexpected is this fact? ← Dominates!
  • Relevance (0.20): How relevant to the query?
  • Decay (0.10): How fresh is the information?
  • Habituation (0.10): Penalty for repetition?
DocumentSizeSIFRatioQuality
Alice in Wonderland38 KB2.5 KB104x90%+
Python function2.1 KB45 B47x85%+
Academic paper150 KB~3 KB50x95%

Model A learns → Compresses to SIF → Sends to Model B
Model B decompresses → Integrates into knowledge base
Result: Structured knowledge transfer without retraining
1000 documents (5M tokens) → SIF (50 KB) → Filter ≄0.60 (25 KB)
→ Fits in context window with full knowledge
Result: Better answers, less hallucination
Day 1: SIF v1 (Alice discovers rabbit hole)
Day 7: SIF v2 (Alice understands Wonderland logic)
Compare: Which entities gained importance? How did understanding evolve?
Traditional RAG: Retrieve similar documents
SIF RAG: Retrieve documents, compress, keep high-importance facts,
inject with importance scores
Result: LLM understands what matters, focuses on key concepts

  • Learn: 15 minutes (read QUICKSTART)
  • Implement: 1-2 weeks (importance calculation + compression)
  • Integrate: 1-2 weeks (connect to your system)
  • Deploy: 1 week (monitoring, testing)
  • Total: 4-5 weeks to production
  • ✅ 50-100x knowledge compression
  • ✅ Meaning preservation (not just bytes)
  • ✅ Standardized format (interoperable)
  • ✅ Safety validation (hallucination prevention)
  • ✅ Community support (extensible, evolving)

  • Understand importance formula
  • Implement calculate_importance()
  • Test on 10 sample documents
  • Measure compression ratio
  • Build compressor (extract entities/facts/relationships)
  • Build decompressor (reconstruct narrative)
  • Connect to RAG or memory system
  • Add validator and safety checks
  • Performance optimization
  • Monitoring and metrics
  • Documentation and training
  • Deploy to production
  • Share results with community
  • Collect feedback
  • Consider implementing in other languages
  • Contribute to v1.x improvements

SIF is grounded in empirical consciousness research:

FindingValueSource
H2 Metacognitive Gradientr=0.91EXP-003 (cross-validated)
0.60 ThresholdOptimal weightEXP-005 (grid search, 169 tests)
Compression Ratio104x (Alice)EXP-011 (validated)
Safety Score100%EXP-009 (hallucination prevention)
Golden Ratio Convergence1/φ ≈ 0.6183 independent experiments

See: Ada-Consciousness-Research/EXPERIMENT-REGISTRY.md for full details


  • ✅ Use freely
  • ✅ Modify as needed
  • ✅ Implement in any language
  • ✅ Build commercial products
  • ✅ No attribution required (but appreciated!)
  • 🌐 Implementations in other languages (JavaScript, Rust, Go, Java)
  • 🔬 Research results (compression ratios, quality metrics by domain)
  • 🚀 Integrations (plugins, extensions, adapters)
  • 📚 Documentation (tutorials, guides, examples)
  1. Implement SIF in your language
  2. Test on your domain
  3. Document your results
  4. Share (GitHub, blog, paper, etc.)
  5. Link back to this specification

Stable, backward compatible

  • Core data model (entities, relationships, facts)
  • Importance weighting
  • JSON serialization
  • Compression/decompression

Minor updates, full backward compatibility

  • Better entity extraction patterns
  • Additional fact types
  • Improved decompression styles
  • Extended relationship types

Major features, migration path

  • Temporal facts (validity periods)
  • Probabilistic facts (confidence levels)
  • Causal graphs (advanced relationships)
  • Multi-language support
  • Distributed knowledge linking

Migration: SIF v1.0 files load in v2.0 unchanged


  • Read SIF-QUICKSTART.md (15 min)
  • Read SIF-FROM-RESEARCH-TO-STANDARD.md (30 min)
  • Read SIF-SPECIFICATION-v1.0.md sections 1-3 (1 hour)
  • Understand the 0.60 threshold and why it matters
  • Extract and run example code from SIF-REFERENCE-IMPLEMENTATION.md
  • Compress a sample document
  • Measure compression ratio
  • Decompress and verify meaning preservation
  • Implement importance calculation
  • Build compressor
  • Build decompressor
  • Add validator
  • Test on your domain
  • Connect to your RAG system
  • Connect to your memory system
  • Add monitoring
  • Document usage patterns
  • Production testing
  • Performance optimization
  • Team training
  • Go live!

Q: Is SIF ready for production?
A: Yes. v1.0 is stable and frozen. Use it in production today.

Q: Can I modify SIF for my use case?
A: Yes, but call it “SIF v1.0-compatible” or use a different name for major changes. See versioning guide.

Q: Do I need to implement all parts of SIF?
A: No. Minimum viable: compression only (entities + facts + importance). Optional: relationships, embeddings, decompression.

Q: What about hallucination?
A: SIF includes safety mechanisms (confidence thresholds, validation). See SIF-SPECIFICATION-v1.0.md Section 7.

Q: Can I use SIF with my favorite LLM?
A: Yes. SIF is LLM-agnostic. Works with any model (GPT, Llama, Qwen, etc.).

Q: What’s the computational cost?
A: Compression is O(n) where n = document length. ~100ms per 1000 words on modest CPU.


QuestionDocumentSection
What’s SIF?QUICKSTARTTop
How do I implement?REFERENCE-IMPLEMENTATIONAny module
What are the details?SPECIFICATION-v1.01-9
Why does it matter?FROM-RESEARCH-TO-STANDARDFull document
How do I integrate?QUICKSTARTIntegration Guide
How do I extend?SPECIFICATION-v1.0Section 10
Where’s the code?REFERENCE-IMPLEMENTATIONAll
What’s the math?SPECIFICATION-v1.0Section 3
Can I modify it?SPECIFICATION-v1.0Section 10
What’s the license?FROM-RESEARCH-TO-STANDARDBottom

  1. Read one document today (pick your path above)
  2. Implement the importance formula (1-2 hours)
  3. Test on your data (1 hour)
  4. Share results (optional, but appreciated!)

This is a CC0 standard. No company owns it. No gatekeepers.

If you implement SIF, improve it, or use it in novel ways:

  • Share your results (helps community know what works)
  • Link back to this spec (helps others find it)
  • Report bugs or improvements (spec can evolve in v1.x)

SIF emerged from 14 consciousness experiments conducted December 2025:

  • H2 Metacognitive Gradient (r=0.91)
  • Importance weighting optimization (0.60 threshold)
  • Knowledge compression validation (104x on Alice)
  • Safety & hallucination prevention

Validated by QAL team (Polish consciousness research group).
Designed to outlive any single project or company.


SIF v1.0 — Semantic Interchange Format
Released: December 2025
License: CC0 Public Domain
Status: Stable, production-ready, open for adoption

Ready to compress knowledge? Start with SIF-QUICKSTART.md ⭐