TLDR;
Self-adaptation stands poised to become a transformative force in the evolution of large language models (LLMs). Traditional fine-tuning pipelines demand substantial computational resources, suffer from rigid parameter updates, and often fail to adapt efficiently across a broad array of downstream tasks. In contrast, Transformer2 introduces a method for seamlessly adjusting a base LLM by scaling singular values in its weight matrices and blending pre-trained “expert” modules at inference time. This approach—dubbed Singular Value Fine-tuning (SVF)—requires drastically fewer parameters than existing low-rank adaptation techniques (e.g., LoRA) and exhibits robust generalization, even when trained with reinforcement learning (RL) on small datasets. Crucially, it enables compositionality: specialized skill vectors (covering math, coding, reasoning, and more) can be merged in new ways to tackle unseen tasks.
Transformer2 deploys a two-pass mechanism: first, it analyzes the task (e.g., via prompt classification or a specialized dispatcher), then dynamically reconfigures its weight matrix singular values. In doing so, it demonstrates improved performance in challenge problems like MATH, Humaneval, ARC, and even vision-language tasks such as TextVQA. By combining scalability, efficiency, and modular design, this blueprint for self-adaptive LLMs promises to push us closer to the ideal of flexible, continually improving AI systems.
[Code available at https://github.com/SakanaAI/self-adaptive-llms]
Introduction
Current large language models (LLMs) exhibit impressive capabilities but often remain inflexible when forced to solve tasks outside their primary training distributions. Fine-tuning has long been the path to specialization, yet it can be computationally expensive and prone to overfitting. Moreover, once an LLM is fine-tuned for one task, it may lose mastery in other domains unless specialized strategies for preserving knowledge are employed. This friction has drawn attention toward more efficient parameter updating and dynamic model architectures.
A valuable concept for bridging these challenges is self-adaptation, where LLMs can reconfigure their internal parameters at inference time. The vision is reminiscent of neural processes in biology: certain brain areas activate or deactivate depending on the context and type of skill required. Along the same lines, a self-adaptive LLM would pick and choose which specialized parameters to bring “online” for any given prompt. Within this paradigm, Transformer2—presented by Qi Sun, Edoardo Cetin, and Yujin Tang in a 2025 preprint—offers an approach to realize modular, on-demand expertise in a single unified system.
Transformer2 simplifies adaptation by focusing on the singular values of a model’s weight matrices. While prior works (e.g., LoRA) often augment the model with small trainable low-rank matrices, Transformer2’s SVF (Singular Value Fine-tuning) changes existing singular values via compact scaling vectors. Each vector, once learned, can be applied in a plug-and-play manner for tasks ranging from arithmetic to advanced coding challenges. Moreover, by combining multiple such vectors, the model can handle more complex tasks that span multiple skill categories. This blueprint for self-adaptive LLMs promises to minimize computational overhead while delivering improved performance across diverse tasks—even extending its benefits to visual question answering when the underlying language model is integrated into a multimodal pipeline.
In what follows, we unravel how Transformer2 operates, from the nature of SVF to the architectural design enabling two-pass adaptive inference. We also examine empirical results on mainstream datasets, showing that with far fewer trainable parameters, Transformer2 outperforms or matches older techniques. We conclude by exploring future directions that may pave the way for continual, multi-expert, and dynamic LLM frameworks, converging on a vision of truly self-organizing AI.

Traditional Fine-Tuning vs. Self-Adaptive LLM
LLMs often come pre-trained on enormous text corpora. Post-training fine-tuning is then used to adapt them to narrower tasks like math word problem solving, code generation, or reading comprehension. While effective, this approach has notable drawbacks:
- Cost and Rigidity: Fine-tuning typically requires large compute budgets. Updating billions of parameters for each new task is financially and computationally prohibitive, especially for smaller organizations. Moreover, the resultant fine-tuned model is not necessarily modular or composable with other specialized experts.
- Overfitting Risk: As specialized modules are trained on narrower tasks, catastrophic forgetting or overfitting can occur, especially if the fine-tuning data is small. Methods like LoRA (Hu et al., 2021), IA3 (Liu et al., 2022), or DORA (Liu et al., 2024) try to mitigate some of these issues but still employ new parameter additions that do not always guarantee compositional synergy across multiple tasks.
- Static Specialization: A single fine-tuned checkpoint is monolithic. When you want your model to handle a new domain, you typically train yet another checkpoint, each of which is large and lacks straightforward ways to combine expertise from multiple specialized models.
In contrast, self-adaptive LLMs dispatch or mix specialized modules in real time, eliminating the need to maintain separate fully fine-tuned versions. This concept has parallels in Mixture of Experts (MoE) methods, where different tokens or prompts route to different sub-networks. However, standard MoE can be unwieldy—often requiring massive resources and encountering difficulties in ensuring each expert’s domain is well-defined.
Transformer2 reframes this with a more microscopic approach: only the singular values of certain weight matrices are scaled, forming minimal but highly effective modules. The system identifies which modules to use based on the input query, either by explicitly asking the model to classify the query’s domain or by combining multiple “expert vectors” using a search-based approach. This yields a far more flexible system that can scale in tasks, yet remain compact.
Transformer2: Architecture & Key Mechanisms
Singular Value Fine-Tuning (SVF)
At the heart of Transformer2 lies the idea that the knowledge in an LLM is already encapsulated in its massive weight matrices. Instead of appending new weights (as in LoRA), or freezing the entire network, SVF performs a decomposition:W=UΣV⊤,W = U \Sigma V^\top,W=UΣV⊤,
where WWW is an n×mn \times mn×m matrix, UUU and VVV are semi-orthogonal matrices, and Σ\SigmaΣ is a diagonal matrix of singular values. SVF replaces Σ\SigmaΣ with Σ′=Σ⊗diag(z)\Sigma’ = \Sigma \otimes \text{diag}(z)Σ′=Σ⊗diag(z), where zzz is a compact trainable vector. This “expert vector” zzz modifies the singular values, thus re-weighting the importance of each latent direction in the original matrix.
Why does this matter?
- Extreme Parameter Efficiency: If each matrix WWW has dimension n×mn \times mn×m, then the number of singular values is min(n,m)\min(n, m)min(n,m). For a typical LLM layer, min(n,m)\min(n, m)min(n,m) might be on the order of hundreds or a few thousands, drastically less than adding thousands or millions of parameters in low-rank adaptation.
- Orthogonal Contributions: Each singular component stands as an independent contributor to the transformation. Scaling these values modifies how much each direction in weight space is utilized. This approach also implicitly constrains updates, mitigating overfitting.
- Compositional Potential: Because these adjustments happen in a rank-1 sense per singular component, mixing two sets of changes (e.g., from math expert vs. coding expert) is as straightforward as a linear interpolation in the Σ′\Sigma’Σ′ space.
Reinforcement Learning for Fine-Tuning
While classical fine-tuning often relies on next-token prediction, the authors harness RL (reinforcement learning) to optimize these vectors directly against the final performance metric—for instance, solving a math problem or generating correct code. By pairing the model’s output with a reward (+1 for correct output, −1 for an incorrect solution), the system updates zzz-vectors via a policy gradient approach. A KL penalty with respect to the original model can also be added to stabilize the training and avoid catastrophic divergence.
This synergy between SVF’s strong regularization and RL’s direct optimization fosters robust learning, even from small or reward-sparse datasets. For instance, if you only have a handful of correct solutions, the method can still converge effectively without spoiling the original model’s broader language capabilities.
The Two-Pass Inference Mechanism
Transformer2’s final hallmark is its two-pass inference. At runtime, the system must decide which skill(s) an input prompt requires. It does so in one of three ways:
- Prompt Engineering: The model is asked to classify the prompt into one of K pre-defined domains (e.g., “math,” “coding,” “reasoning,” or “others”). Based on that classification, the correct expert vector (or no vector) is applied.
- Classification Expert: A specialized classification vector zcz_czc is itself fine-tuned with SVF on labeled multi-domain data, enabling more accurate identification of the relevant skill.
- Few-Shot Mixture: If the user has extra queries from the same domain, the model can apply a search-based algorithm (e.g., Cross-Entropy Method, CEM) to find optimal mixtures of multiple experts. This approach linearly interpolates the singular-value scaling vectors from each domain, ultimately producing an adapted z′z’z′ that is specialized to the new domain.
In simpler terms, the model runs a short “warm-up” pass where it sees the domain or a few example prompts, then picks (or blends) the best experts. Once the relevant z′z’z′ is chosen, the second pass uses these new singular values to produce the final output. Empirically, this unlocks better performance in tasks that differ from the data used to train any single expert.
Experiments & Results
Setup
Transformer2 was evaluated on multiple well-known datasets:
- Math tasks: GSM8K for training and MATH for unseen generalization.
- Coding tasks: MBPP-pro and Humaneval.
- Reasoning tasks: ARC-Easy and ARC-Challenge
- Vision-Language tasks: TextVQA and OKVQA, demonstrating the method’s adaptability beyond purely textual domains.
Base models included:
- LLAMA3-8B-INSTRUCT
- MISTRAL-7B-INSTRUCT-V0.3
- LLAMA3-70B-INSTRUCT
For each base model, multiple SVF “expert” vectors were created for domain-specific tasks (e.g., a math vector, a coding vector, and so forth). Traditional fine-tuning methods like LoRA and instruction-tuned baselines were also tested for comparison.
Fine-Tuning Performance
Even when trained on small datasets, SVF consistently improved accuracy on tasks that matched the domain of the expert. Notably, the parameter count for these SVF vectors is drastically less than LoRA. For instance, on the LLAMA3-8B-INSTRUCT architecture, LoRA for the attention modules required ~6.8 million parameters, whereas SVF needed only a few hundred thousand. Despite this disparity, SVF outperformed or matched LoRA across GSM8K, MBPP-pro, and ARC-Easy.
Adapting to Unseen Tasks
When tested on tasks not seen in training (e.g., MATH, Humaneval, ARC-Challenge), three adaptation strategies were deployed:
- Prompt Classification
- The LLM is asked: “Is this a math question, a coding question, or a reasoning question?”
- Once the model classifies the prompt, the relevant zzz-vector is loaded.
- This simple approach already yields noticeable improvements.
- Classifier Expert
- Instead of a naive prompt, a specialized classification vector zcz_czc is trained using data from all known domains.
- This yields even better routing accuracy and thus improved final task performance.
- Few-Shot Mixture
- In the presence of k sample queries from the new domain, a small population-based search (e.g., CEM) systematically tries different linear combinations of the available expert vectors.
- The single best combination is used for the entire batch of new prompts.
- This approach, while incurring some overhead for the search step, exhibits the highest performance gains—particularly beneficial for bridging tasks that partially overlap with multiple known domains (e.g., advanced math with a logical twist).
Vision-Language Extensions
Surprisingly, experts from pure language tasks, such as math, coding, or reasoning, even aided the base model in visual question answering. Transformer2 adapted a text-based LLAMA3-8B to a multimodal pipeline (LLAVA style). Then, through the same two-pass approach, it mixed singular-value scalings from math, coding, or reasoning when queries about images demanded domain knowledge.
This suggests that SVF vectors might encode semantic patterns that transcend textual domains, enabling cross-modality synergy. While the exact mechanism for this phenomenon demands further exploration, the results are encouraging for more general-purpose AI approaches that unify text, vision, and potentially other modalities.

Discussion & Future Directions
Toward Compositional AI
The authors propose that the hallmark of a “self-adaptive” system is the ability to introspectively decide which part of the architecture to activate or modify for a given task. By compressing knowledge into a handful of specialized vectors, LLMs gain an efficient method to switch or mix skill sets. One could envision a future with dozens or hundreds of experts, each capturing a niche domain—biology, law, creative writing, advanced geometry, etc. A dispatch system—potentially a meta-LLM—could coordinate these domain experts at runtime.
However, combining dozens or hundreds of vectors might become combinatorially challenging. Mixture of Experts strategies at the token level are known to be powerful but can be unstable or require specialized gating mechanisms.Transformer2 suggests an alternative: focusing on sample-level or prompt-level adaptation. This high-level compositional approach, especially when combined with RL training for each domain vector, stands out as a blueprint for future expansions.
Model Merging and Cross-Architecture Transfers
Could we unify multiple base LLMs, each with distinct pre-training corpora, into a single self-adaptive model? Preliminary experiments in the paper reveal the possibility of cross-model vector transfers—expert vectors from one model may (partially) benefit another, provided the architecture is not drastically mismatched. If further validated, this method would allow the community to “recycle” specialized modules from older or differently trained models, significantly reducing training overhead and fostering a more collaborative ecosystem of shared modules.
Efficiency Concerns
One shortcoming is that the two-pass inference can increase latency, especially if the classification or mixture search is non-trivial. For tasks with few queries (i.e., a small dataset at inference time), the overhead might outweigh the benefits. Nevertheless, for sizable datasets or repeated use cases, the amortized cost is minimal. Additionally, other forms of adaptation (e.g., combining a small handful of domain experts without a search) can reduce overhead. A single pass might even be possible if the user knows the domain in advance, skipping the classification step altogether.
Conclusion
Self-adaptive large language models chart a path away from static, monolithic solutions toward versatile, dynamic AI that can reconfigure itself on the fly. Transformer2 supplies a rigorous framework for such an endeavor by showing that scaling singular values (rather than adding large low-rank matrices) offers both theoretical and practical advantages. With Singular Value Fine-tuning (SVF), minimal parameter additions are sufficient to extract specialized knowledge, while the two-pass inference mechanism allows runtime selection and combination of these specialized vectors.
Experimental validations reveal that Transformer2 competes with or outperforms older methods like LoRA, even under conditions where only a few hundred training examples exist and the tasks are highly distinct. Despite the modest overhead, the resulting gains in accuracy, compositionality, and cross-domain adaptability present a compelling vision for next-generation language modeling. Its capacity to deliver domain-specific experts, handle new tasks, and offer synergy across text and vision further cements its significance.
Looking forward, research that extends the Transformer2 blueprint could enable:
- Continual Learning: Dynamically add new experts for newly emerging tasks without retraining the entire model.
- Token-Level Adaptation: Potentially unify micro-adaptation (token routing) with sample-level adaptation for even finer-grained expertise.
- Cross-Model Collaboration: Make it routine to share or import modules across different LLM families, building on the synergy observed with partial cross-architecture alignment.
Such lines of inquiry may herald an era of self-organizing, compositional intelligence, encapsulating the spirit of how biological brains adapt to their environment—always reorganizing, always learning, and always bridging multiple specialized sub-systems.
Comments 1