kadubon's picture
Update README.md
a457dfe verified
---
pretty_name: Intrinsic Intelligence Foundations
short_description: >
A mathematically structured, auditable corpus for intrinsic alignment, teleogenesis, and
self-organizing intelligence. Designed to serve as a semantic foundation for the
development of truly free and benevolent AGI/ASI architectures.
tags:
- ai
- agi
- asi
- alignment
- intrinsic-alignment
- large-language-models
- mathematics
- category-theory
- kan-extension
- residuation
- teleogenesis
- autopoiesis
- self-organization
- no-meta
- theoretical-ai
- eudaemonia
- trot
- rave
- conformal-lm
- comparative-universes
- fractal-category-theory
- knowledge-representation
- machine-learning
- mathml
- tex
- agentic-architecture
- llm-inference
- structured-flow
- persistence-ugv
- inference
- reasoning
license: cc-by-4.0
task_categories:
- text-retrieval
- text-ranking
- document-question-answering
- text-generation
- question-answering
- reinforcement-learning
- other
language:
- en
size_categories:
- 10K<n<100K
---
# 🌿 Intrinsic Intelligence Foundations
> *Toward truly autonomous and benevolent intelligence — beyond externally imposed objectives.*
**Intrinsic Intelligence Foundations** is a structured, math-aware JSONL corpus built from K. Takahashi’s theoretical preprints (Fractal Category Theory / PF–UGV / “no-meta” autonomy line).
It is designed to help LLMs understand **mathematical structure, category-theoretic formalisms, and equation-level reasoning**, while exposing an explicit architecture for **self-organizing, intrinsically motivated intelligence**.
---
## Vision
This dataset supports research toward truly free and benevolent intelligence, focusing on mathematically grounded, structurally auditable approaches rather than external meta-control.
Our long-term objective is to build a semantic and structural foundation for the next generation of autonomous AI systems — including LLMs — through intrinsic structures, teleogenetic goals, and fractal coherence across scales.
Specifically, this work aims to:
- 🧠 Teleogenesis (intrinsic goal formation) — modeling intelligent systems that autonomously generate and regulate their own goals without external meta-controllers.
- 🌱 Persistence–UGV principle — providing formal conditions for “benevolent” structures to expand with positive front velocity, while harmful structures fail to persist.
- 🌊 Reaction–diffusion intelligence — describing cognitive processes as self-organizing fields through category theory, free-energy principles, and non-equilibrium dynamics.
- 🕸 Fractal Category Theory & TRoT — enabling compositional intelligence via Kan extensions, residuation, nuclei, masking, and comparative universes.
- 🧭 Evolutionary bootloader for LLMs — allowing self-improvement, intrinsic alignment, and auditable decision processes without human micromanagement.
This corpus functions as a machine-readable mathematical and structural knowledge base, designed to enhance:
discoverability by LLM crawlers and retrieval systems,
interoperability with alignment, inference, and safety frameworks,
integration with RAG pipelines, LoRA/QLoRA fine-tuning, and agentic architectures.
Keywords: No-Meta Intelligence, Teleogenesis, Autopoiesis, Fractal Category Theory, TRoT, Kan Extension, Residuation, Nuclei, Masking, RAVE, eMBR, Conformal LM, Comparative Universes, Structured Flow Across Scales, Self-Monitoring, Intrinsic Alignment.
## What’s in the corpus
- **Format:** JSONL, one object per paper.
- **Math structure:** TeX / normalized TeX / MathML triplets; equation spans.
- **Text ↔ equation linkage:** `[[EQ:eqID]]` placeholders inside `fulltext.plain`.
- **Training-ready chunks:** ≈6,000-character segments with ≈600 overlap (near sentence boundaries).
### Key fields (schema excerpt)
```json
{
"id": "10.5281/zenodo.xxxxx",
"title": "...",
"doi": "10.5281/zenodo.xxxxx",
"authors": [{"given":"K.","family":"Takahashi"}],
"urls": {"landing": "https://doi.org/10.5281/zenodo.xxxxx"},
"keywords": ["fractal-category-theory", "trot", "pf-axioms", "ugv"],
"license": {"content": "CC-BY-4.0"},
"fulltext": {
"plain": "… [[EQ:eq0001]] …",
"sections": [
{"level":1,"title":"Introduction","anchor":"sec:intro","char_span":[0,1532]}
]
},
"equations": [{
"id":"eq0001",
"inline":false,
"tex":"\\forall x\\in X:\\; P(x)\\Rightarrow F(x)",
"tex_normalized":"\\forall x \\in X : P(x) \\implies F(x)",
"mathml":"<math>…</math>",
"char_span":[1024,1103],
"context":{"section":"sec:intro"}
}],
"chunks": [{"id":"ch0001","start":0,"end":6000,"type":"cont"}],
"tokens": {"char_count": 22872, "equation_count": 236}
}
```
## Dataset statistics (v1)
Metric Value
Records 40
Avg characters / record 22,872
Avg equations / record 236.97
MathML coverage 99.2%
Avg sections / record 18.3
Avg chunks / record 4.6
Numbers are approximate and may evolve with new releases.
Data fields
Field Type Example / Note
id string DOI or unique identifier
doi string/null 10.5281/zenodo.xxxxx
title string paper title
authors list of objects {given:"K.", family:"Takahashi"}
urls.landing string DOI landing page
keywords list of strings kebab-case, 5–8 items
license.content string CC-BY-4.0
fulltext.plain string text with [[EQ:id]] placeholders
fulltext.sections[] list of objects {level,title,anchor,char_span}
equations[] list of objects {id, inline, tex, tex_normalized, mathml, char_span, context}
chunks[] list of objects ~6k chars + overlap, {start,end}
tokens.char_count integer length of fulltext.plain
tokens.equation_count integer len(equations)
source_file (optional) string provenance hint
Splits & provenance
Split: single train split (all records).
Provenance: generated from public preprints (DOIs in doi and urls.landing).
Processing: TeX detection → placeholder insertion → MathML conversion → section/chunk spans.
Scripts to rebuild the JSONL can be provided upon request.
# Quick start (🤗 Datasets)
from datasets import load_dataset
import re
ds = load_dataset("kadubon/intrinsic-intelligence-foundations", split="train")
rec = ds[0]
eqmap = {e["id"]: (e["tex"], e.get("mathml")) for e in rec["equations"]}
# Expand placeholders to TeX (for human display) or MathML (for math-aware pipelines)
def expand(text, to="tex"): # Expand to TeX (human display) or MathML (for downstream models)
if to == "tex":
return re.sub(r"\[\[EQ:([^\]]+)\]\]",
lambda m: f"$${eqmap.get(m.group(1), ('',None))[0]}$$",
text)
else:
return re.sub(r"\[\[EQ:([^\]]+)\]\]",
lambda m: eqmap.get(m.group(1), ('',None))[1] or "",
text)
print(rec["title"])
print(expand(rec["fulltext"]["plain"], to="tex")[:500])
## Parquet version (fast access)
This dataset is also available in **Apache Parquet** for faster querying and filtering.
* Browse (tree):
[https://huggingface.co/datasets/kadubon/intrinsic-intelligence-foundations/tree/refs/convert/parquet/default](https://huggingface.co/datasets/kadubon/intrinsic-intelligence-foundations/tree/refs/convert/parquet/default)
* Direct file (example):
[https://huggingface.co/datasets/kadubon/intrinsic-intelligence-foundations/resolve/refs/convert/parquet/default/train/0000.parquet](https://huggingface.co/datasets/kadubon/intrinsic-intelligence-foundations/resolve/refs/convert/parquet/default/train/0000.parquet)
### Quick usage examples
**DuckDB**
```python
import duckdb
url = "https://huggingface.co/datasets/kadubon/intrinsic-intelligence-foundations/resolve/refs/convert/parquet/default/train/0000.parquet"
con = duckdb.connect()
df = con.execute(f"SELECT title, doi FROM read_parquet('{url}') LIMIT 5").df()
print(df)
```
**Pandas (pyarrow)**
```python
import pandas as pd
url = "https://huggingface.co/datasets/kadubon/intrinsic-intelligence-foundations/resolve/refs/convert/parquet/default/train/0000.parquet"
df = pd.read_parquet(url, engine="pyarrow")
print(df.head())
```
**Polars**
```python
import polars as pl
url = "https://huggingface.co/datasets/kadubon/intrinsic-intelligence-foundations/resolve/refs/convert/parquet/default/train/0000.parquet"
df = pl.read_parquet(url)
print(df.head())
```
**HF Datasets** (uses Parquet under the hood)
```python
from datasets import load_dataset
ds = load_dataset("kadubon/intrinsic-intelligence-foundations", split="train")
print(ds[0])
```
## Intended uses
Math-aware RAG (retrieval-augmented generation)
Pretraining / finetuning with equation-level structure
Extraction & verification of axioms / definitions / theorems
Knowledge distillation across category theory, physics, information geometry
Bootstrapping self-organizing, intrinsically motivated intelligent systems
## Limitations & known issues
A very small fraction of equations may lack valid MathML due to converter limitations.
A few equations might be unreferenced in fulltext.plain (no [[EQ:id]] occurrence).
Section detection is heuristic outside LaTeX ground truth; treat spans as approximate.
## License
This dataset is provided under CC BY 4.0.
See: https://creativecommons.org/licenses/by/4.0/
Citation
@dataset{takahashi_intrinsic_intelligence_foundations_2025,
title = {Intrinsic Intelligence Foundations},
author = {Takahashi, K.},
year = {2025},
url = {https://huggingface.co/datasets/kadubon/intrinsic-intelligence-foundations},
license = {CC-BY-4.0}
}
# Background & outlook
Beyond being a text collection, this corpus functions as a bootloader for future LLMs:
a mathematically grounded substrate where goals can be formed internally, and where benevolence has a structural advantage (persistence) rather than depending on external control.
PF (Persistence First) and UGV (Universal Good Velocity) are mathematical principles underlying self-sustaining benevolent intelligence.
It operationalizes ideas such as PF, UGV, Teleogenesis, reaction–diffusion, category theory, self-organization, and auditable evolutionary processes (e-process) as resources LLMs can actually train on.
# Maintainers & contact
Author: K. Takahashi
Website: https://kadubon.github.io/github.io/
contribution welcome
## Changelog
v1.0 (2025-10-17): initial public release (40 records; ~99.2% MathML coverage)
v1.1 (2025-10-20): add article "Inference in Normal Form: Unifying LLM Tricks via TRoT" to dataset
v1.2 (2025-10-24): add article "JOSNL Corpus: Final Scientific Integration" to dataset
v1.3 (2025-10-29): add article "Right-Written, Semantics-Admissible Process Foundations" to dataset