Tech

Self-adaptive LLM dynamically adjusts its weights to learn new tasks

Share
Share
Self-adaptive AI LLM—Transformer²—dynamically adjusts its weights to learn new tasks
Method overview. Left) At training time, we employ SVF and RL to learn the “expert” vectors z’s that scale the singular values of the weight matrices. Right) At inference time, we propose three distinct methods to adaptively select/combine the learned expert vectors. Credit: arXiv (2025). DOI: 10.48550/arxiv.2501.06252

A trio of AI researchers at Sakana AI, a Japanese startup, has announced the development of a self-adaptive AI LLM called Transformer2. Qi Sun, Edoardo Cetin, and Yujin Tang, have posted their paper on the arXiv preprint server.

As LLMs mature, AI researchers continue to refine them to be more efficient and less energy demanding. In this new study, the research trio has found a way to reduce one of the major inefficiencies in traditional LLMs—the need for fine-tuning if they are asked to do something they have not been trained to do.

Under current scenarios, an LLM’s parameters are adjusted and it is then trained with new samples—afterward, the new parameters remain frozen in place. The research team has introduced a model that makes adjustments to a system of weights when it is introduced to something new, to allow it to adjust dynamically to new types of tasks.

To allow the LLM to carry out dynamic adjustments, the researchers have split the task response into a two-step approach; the first involves analyzing the request and figuring out what will be required to provide a good response. The second involves making adjustments to a system of weights to help it focus its efforts on things that will lead to an answer.

The system of weights uses a math process called Singular Value Decomposition to determine which parts of its own AI system are the most important for providing the best possible answer. Reinforcement learning is applied to create the steps needed to guide the AI’s behavior.

During inference, (which is the part of the system involved in generating responses to the initial query), the system employs three main strategies to achieve its goals—one that is based on the prompt, another that serves as a classifier and the third that applies a few-shot adaptation process (where an AI model learns from a limited training set). Once the weights have been applied, the LLM carries on in similar fashion to other LLMs.

The overall result of using the new approach is that it allows an LLM to adjust itself on the fly when it finds itself faced with an unfamiliar task. Testing of the system showed it capable of performing as well as other LLMs on traditional queries but much more flexible when it came to answering queries that confused other models.

More information:
Qi Sun et al, Transformer2: Self-adaptive LLMs, arXiv (2025). DOI: 10.48550/arxiv.2501.06252

Journal information:
arXiv


© 2025 Science X Network

Citation:
Self-adaptive LLM dynamically adjusts its weights to learn new tasks (2025, January 24)
retrieved 24 January 2025
from

This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.

Share

Leave a comment

Leave a Reply

Your email address will not be published. Required fields are marked *

Related Articles
AI took a huge leap in IQ, and now a quarter of Gen Z thinks AI is conscious
Tech

AI took a huge leap in IQ, and now a quarter of Gen Z thinks AI is conscious

ChatGPT’s o3 model scored a 136 on the Mensa IQ test and...

DeepSeek sees surge in developer use as 3 in 10 businesses adopt the controversial LLM provider
Tech

DeepSeek sees surge in developer use as 3 in 10 businesses adopt the controversial LLM provider

Developers shift from loyalty to flexibility as OpenAI leads, but DeepSeek gains...

China’s CATL launches new EV sodium battery
Tech

China’s CATL launches new EV sodium battery

Chinese battery giant CATL has launched a new sodium-ion battery it says...