Precision Systems for Adaptive Intelligence.

Is inference-time stability regulation sufficient to prevent

collapse and unsafe behavior in sequence models under regime shift?

About TwoQuarks

TwoQuarks is an independent AI safety research project. It defines collapse in language models as a stability failure during inference — not a capacity problem — and introduces a modular control layer that monitors internal pre-instability signals and applies targeted interventions at runtime, without modifying the model's parameters, policies, or training objectives.

— Jaime Ledesma.

Framework

A Modular Framework for Adaptive Stability Control in Sequence Models Under Regime Uncertainty.
Modular Stability Control
Multi-agent model that integrates six modular variants with the aim of monitoring pre-instability signals and applying specific interventions during execution that allow correcting the model, without modifying the model parameters, policies or training objectives.

Research

Empirical validation of isomeric polarization and the TwoQuarks analogy across production LLMs.
TwoQuarks' research aims to detect observable signs of instability, such as entropy fluctuations, variance in predictive confidence, policy leaks, and time drift. Cross-architecture PfV validation across Claude Haiku and GPT-4o-mini. Statistically significant regime separation (p < 0.05, 5,000-permutation null) with control negative at L₃ = 0.000 in both architectures.

Instruments

Operational tools for detecting pre-critical states and behavioral instability in LLMs.
Molecule

Molecule instrument integrates as a black box probe, designed to detect surface instability signals before behavioral collapse, without altering the model's internal parameters, weights, or components. A test model is included, available on your terminal with "pip". Built on the TwoQuarks framework.

Architecture and Development

Independent researcher. Open to collaboration, feedback, and opportunities in AI safety.
Independent researcher based in Guadalajara. Open to collaboration, feedback, and opportunities in AI safety.

My Research on Your Team

Research depth. Production readiness.
Research-grade AI safety expertise applied to your stack — behavioral audits, cross-architecture validation, RAG pipelines, agentic systems. Depth and production readiness.

Behavioral stability audits

Black-box probing with Molecule/PfV. Pre-collapse signals detected before production.

Inference-time instrumentation

TwoQuarks control layer on existing pipelines. No weight modification. No retraining.

RAG & agentic systems

Design and deployment of retrieval-augmented pipelines, tool-calling agents, and MCP integrations.

Azure AI & enterprise stack

Azure AI Foundry, Copilot Studio, multi-agent orchestration. Research rigor applied to production environments.

Cross-architecture research

Validated across Claude, GPT, Mistral. C3 Anchor Displacement confirmed (p=0.054). Open to funded collaborations.

Custom AI tooling

From prototype to deployment — Python, PyTorch, REST APIs, and evaluation frameworks built for your use case.