New MAP framework enhances parameter-efficient fine-tuning by decoupling direction and magnitude

Researchers unveil the MAP framework, a novel method that boosts Artificial Intelligence fine-tuning efficiency by separating directional and magnitude updates in model parameters.

Researchers have introduced MAP (Matrix Adaptation via Projection), a fresh framework designed to improve the parameter-efficient fine-tuning of large language models by decoupling weight adaptation into distinct directional and magnitude updates. Unlike traditional approaches that modify model weights collectively, MAP normalises pre-trained weights and then allows directional and scalar adjustments during fine-tuning. This geometric strategy enables greater flexibility and precision, resulting in more efficient adaptation with fewer parameter changes.

Developed by a team from leading institutions including Shanghai Jiao Tong University, Harvard, Alibaba Group, and the Singapore Institute of Technology, MAP was detailed in the study ´MAP: Revisiting Weight Decomposition for Low-Rank Adaptation.´ The researchers focused their experiments on foundation models such as LLaMA-7B and LLaMA-3-8B, leveraging the General Language Understanding Evaluation (GLUE) benchmark to assess performance. They carefully tuned hyperparameters such as the LoRA rank, learning rates, and batch sizes. Training employed AdamW optimization and a warmup period to ensure stable convergence.

The MAP approach not only demonstrated improvements in a range of tasks—including question answering, textual entailment, and sentiment analysis—but also reduced the risk of catastrophic forgetting by controlling the magnitude of updates. Its straightforward design allows for easy integration with existing parameter-efficient fine-tuning methods, potentially streamlining the deployment of large language models for domain-specific applications. The research also outlines plans to further explore the mathematical landscape of MAP, examine alternative normalisation methods, and develop intelligent strategies for automatically determining optimal scaling factors as training progresses.

Ultimately, the introduction of MAP signals a step forward for the practical and theoretical development of parameter-efficient fine-tuning strategies, offering the field a scalable and elegant mechanism for adapting large language models to diverse tasks without sacrificing previously acquired knowledge.

73

Impact Score

Saudi Artificial Intelligence startup launches Arabic LLM

Misraj Artificial Intelligence unveiled Kawn, an Arabic large language model, at AWS re:Invent and launched Workforces, a platform for creating and managing Artificial Intelligence agents for enterprises and public institutions.

Introducing Mistral 3: open artificial intelligence models

Mistral 3 is a family of open, multimodal and multilingual Artificial Intelligence models that includes three Ministral edge models and a sparse Mistral Large 3 trained with 41B active and 675B total parameters, released under the Apache 2.0 license.

NVIDIA and Mistral Artificial Intelligence partner to accelerate new family of open models

NVIDIA and Mistral Artificial Intelligence announced a partnership to optimize the Mistral 3 family of open-source multilingual, multimodal models across NVIDIA supercomputing and edge platforms. The collaboration highlights Mistral Large 3, a mixture-of-experts model designed to improve efficiency and accuracy for enterprise artificial intelligence deployments starting Tuesday, Dec. 2.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.