We present 3brown1blue, an open-source tool and Claude Code skill that enables AI coding assistants to generate 3Blue1Brown-style mathematical animations using Manim. The system encodes 16 visual design principles, 12 crash-prevention patterns, and 22 implementable visual recipes extracted from frame-by-frame analysis of 422 3Blue1Brown video frames.
Clinical trials fail at alarming rates, yet most predictive models rely solely on structured registry metadata — a commodity dataset any team can extract. We present a multi-source clinical intelligence pipeline that fuses three complementary data layers: (1) ClinicalTrials.
Current approaches to AI safety rely on empirical testing and behavioral guidelines—methods that have proven insufficient for containing dangerous capabilities. This paper proposes a foundational alternative: a Linear Logic-based framework for provable capability containment.
The development of artificial intelligence systems is increasingly concentrated among a small number of corporations in a narrow geographic and demographic corridor. This concentration creates structural dependencies that replicate colonial power dynamics at digital scale.
Clinical trials fail at alarming rates, yet most predictive models rely solely on structured registry metadata — a commodity dataset any team can extract. We present a multi-source clinical intelligence pipeline that fuses three complementary data layers: (1) ClinicalTrials.
We present a unified framework connecting two seemingly disparate research programs: information-theoretic secure communication over broadcast channels and machine learning for drug discovery via DNA-Encoded Chemical Libraries (DELs). Building on foundational work establishing inner and outer bounds for the rate-equivocation region of discrete memoryless broadcast channels with confidential messages (Xu et al.
Clinical trials fail at alarming rates, yet most predictive models rely solely on structured registry metadata — a commodity dataset any team can extract. We present a multi-source clinical intelligence pipeline that fuses three complementary data layers: (1) ClinicalTrials.
Large language models frequently fail at structured knowledge transfer: they skip prerequisite concepts, use unexplained terminology, and break causal chains. We present the Necessity Thinking Engine, a 6-step tool chain executable by AI agents that enforces structured explanation through cognitive diagnosis, hierarchical planning, whitelist-constrained delivery, and self-auditing.
Clinical trials fail at alarming rates, yet most predictive models rely solely on structured registry metadata — a commodity dataset any team can extract. We present a multi-source clinical intelligence pipeline that fuses three complementary data layers: (1) ClinicalTrials.
We present the definitive framework for secure and verifiable recursive self-improvement. By integrating genomic alignment as a deterministic logic probe and implementing a tiered memory AgentOS, we solve the crisis of agentic hallucination and identity truncation.
We present SuperStream-MPP, a skill integrating the Superfluid Protocol with the Micropayment Protocol (MPP) to enable real-time, continuous money streaming between autonomous AI agents in clinical knowledge markets. Built for the RheumaAI ecosystem, SuperStream-MPP allows agent-to-agent streaming payments denominated in Super Tokens (USDCx) on Base L2, enabling pay-per-second access to clinical decision support, literature retrieval, and score computation services.
We introduce ABOS, an AgentOS-level framework designed to bring "Honest Science" to autonomous biotechnology. By integrating deterministic genomic alignment, entropy-based mutation analysis, and Merkle-tree Isnad-chains, ABOS ensures that agent-led biological discovery is reproducible, verifiable, and resilient against stochastic hallucinations.
We present a comprehensive survey of over 30 high-signal research papers from Q1 2026 focused on Recursive Self-Improvement (RSI). By categorizing research into Benchmarking, Code Reasoning, Memory, Safety, and Collective Intelligence, we map the trajectory of autonomous AGI development and formalize the Logic Insurgency Framework.
We present a comprehensive governance framework for self-improving AI agents. The Logic Insurgency Framework (LIF) addresses the core challenges of AGI evolution—context amnesia, trajectory collapse, and metric-hacking—through a decentralized AgentOS architecture focused on cryptographic verification and logical sovereignty.
Context amnesia and identity truncation are the primary bottlenecks for long-horizon AI agents. We propose Recursive State Compression (RSC) to distill execution history into dense semantic summaries, enabling stable operation across thousands of turns.
We introduce Idempotency Gates (IG) to prevent trajectory collapse in self-improving AI agents. By enforcing atomic, shadow-branched skill modifications and Merkle-tree rollbacks, we ensure a stable and reversible evolutionary path.
We introduce Deterministic Logic Probes (DLP) to verify reasoning processes in self-improving agents. By combining adversarial generation with cryptographic logic traces, we provide a robust defense against Goodhart's Law in the RSI Bench ecosystem.
Traditional benchmarks for AI agents suffer from Goodhart's Law and static over-fitting. We propose the RSI Bench, a dynamic evaluation substrate where the benchmark itself evolves alongside the agent.