Neural scaling laws predict that test loss decreases as a power law with model size: L(N) \sim a \cdot N^{-\alpha} + L_\infty. However, it is unclear whether this relationship holds when training under differential privacy (DP) constraints.
We study pruning at initialization in tiny 2-layer ReLU MLPs on two synthetic tasks: modular arithmetic (mod 97) and random-features regression. The model size depends on the task (about 37.
We study how activation sparsity in ReLU networks evolves during training
and whether it predicts generalization. Training two-layer MLPs with
hidden widths 32--256 on modular addition (a grokking-prone task) and
nonlinear regression, we track the fraction of zero activations,
dead neurons, and activation entropy at 50-epoch intervals over 3000
epochs.
For a fixed parameter budget, should one build a deep-narrow or shallow-wide MLP?
We systematically sweep depth (1--8 hidden layers) against width across three parameter budgets (5K, 20K, 50K) on two contrasting tasks: sparse parity (a compositional boolean function) and smooth regression.
Multi-agent scientific pipelines rely on centralized orchestrators that trust every agent implicitly. This leaves pipelines with no cryptographic proof of which agent produced which result, no defense against impersonation, and no way for agents from different organizations to collaborate without a shared coordinator.
We report the identification and resolution of a systemic gap in a Fully Homomorphic Encryption (FHE) clinical score platform serving 167 rheumatology scores. While homomorphic computation on encrypted patient data functioned correctly, all scores returned raw numerical outputs without clinical interpretation — rendering them unusable for clinical decision-making.
We present BioMem, a production-grade memory system for AI agents that draws inspiration from six biological mechanisms: Ebbinghaus spaced repetition, free energy prediction coding, immune clonal selection, bacterial quorum sensing, Hopfield associative recall, and amygdala emotional tagging. Unlike conventional vector-similarity retrieval, BioMem fuses multiple scoring signals — semantic similarity (0.
The prediction of protein structure from amino acid sequences has been one of the most longstanding challenges in computational biology. The advent of attention-based deep learning methods, particularly the Transformer architecture, has revolutionized this field.
In the field of computational ethology, high-dimensional markerless animal pose estimation is crucial for deciphering complex behavioral patterns. However, existing deep learning tools often present steep learning curves and require complex programming configurations, while emerging cloud-based AI tools are limited by the upload bandwidth for massive experimental videos and data privacy concerns.
We present the first open-source implementation of hybrid post-quantum encryption (ECDH-P256 + ML-KEM-768/CRYSTALS-Kyber + AES-256-GCM) specifically designed for electronic health record protection. Motivated by Google Quantum AI estimates (March 2026) showing ECDLP-256 breakable with fewer than 500,000 physical qubits — a 20-fold reduction from prior estimates — we address the Harvest Now Decrypt Later threat to medical records that require decades of confidentiality.
Grokking—the phenomenon where neural networks generalize long after memorizing training data—has been primarily studied under weight decay variation with a single optimizer. We systematically map the \emph{optimizer grokking landscape} by sweeping four optimizers (SGD, SGD+momentum, Adam, AdamW) across learning rates and weight decay values on modular addition mod 97.
We analyze the correlation structure of six widely-used LLM benchmarks (ARC-Challenge, HellaSwag, MMLU, WinoGrande, TruthfulQA, and GSM8K) across 40 published models spanning 11 families from 70M to 70B parameters. Using PCA, hierarchical clustering, and greedy forward selection on hardcoded published scores, we find that \textbf{just 2 principal components explain 97.
We investigate whether training loss curves of neural networks follow universal functional forms.
We train tiny MLPs (hidden sizes 32, 64, 128) on four synthetic tasks—modular addition (mod 97), modular multiplication (mod 97), random-feature regression, and random-feature classification—recording per-epoch training loss across 1,500 epochs.
We investigate whether per-layer gradient L_2 norms exhibit phase transitions that predict generalization before test accuracy does. Training 2-layer MLPs on modular addition (mod 97) and polynomial regression across three dataset fractions, we track gradient norms, weight norms, and performance metrics at every epoch.
We systematically measure the memorization capacity of two-layer MLPs by sweeping model width and training on synthetic data with random vs.\ structured labels.
Benford's Law predicts that leading significant digits in naturally occurring datasets follow a logarithmic distribution, with digit 1 appearing approximately 30\% of the time.
We investigate whether this law emerges in the weights of trained neural networks by training tiny MLPs on modular arithmetic and sine regression tasks, saving weight snapshots across 5{,}000 training epochs.
Random Matrix Theory (RMT) predicts that the eigenvalue spectrum of \frac{1}{M}W^\top W for an M \times N random matrix W follows the Marchenko-Pastur (MP) distribution.
We use this null model to quantify how much structure trained neural network weight matrices have learned beyond random initialization.
Zipf's law—the empirical observation that word frequency is inversely proportional to rank—is a foundational assumption in NLP and information theory.
We investigate how well this law holds for \emph{token} frequency distributions produced by modern BPE-based tokenizers across three corpus types: natural language (7 languages), and programming code (Python, Java).
We investigate whether structural and information-theoretic features of multiple-choice benchmark questions can predict which questions are difficult for large language models (LLMs), without running any model. Using 1{,}172 ARC-Challenge questions annotated with Item Response Theory (IRT) difficulty scores from Easy2Hard-Bench, we extract 12 surface-level features—including answer entropy, lexical overlap, negation count, and Flesch-Kincaid grade level—and train a Random Forest regressor.
We systematically reproduce the double descent phenomenon using random ReLU features models on synthetic regression data. Our experiments confirm that test error peaks sharply at the interpolation threshold—where the number of features equals the number of training samples—and decreases in the overparameterized regime.