Promote Your Research… Share it Worldwide
Have a story or a research paper to share? Become a contributor and publish your work on AcademicJobs.com.
Submit your Research - Make it Global NewsMohamed bin Zayed University of Artificial Intelligence (MBZUAI) researchers have unveiled a groundbreaking advancement in AI model adaptation known as LoFT, or Low-Rank Adaptation that Behaves Like Full Fine-Tuning. This innovative method dramatically narrows the longstanding performance gap between efficient parameter techniques like LoRA and resource-intensive full fine-tuning, promising to revolutionize how large language models are customized for specific tasks.
In the fast-evolving world of artificial intelligence, fine-tuning pre-trained models has become essential for tailoring powerful systems like LLaMA to specialized applications. Traditional full fine-tuning updates every parameter in a model, delivering top performance but demanding massive computational resources—often infeasible for all but the largest organizations. Enter Low-Rank Adaptation (LoRA), a parameter-efficient fine-tuning (PEFT) approach that injects small, trainable low-rank matrices into the model while freezing the bulk of its weights. This slashes memory and compute needs by orders of magnitude, making fine-tuning accessible on consumer hardware. However, LoRA has consistently lagged behind full fine-tuning in accuracy, particularly at lower ranks where efficiency peaks.
LoFT changes that equation. Developed by a team led by third-year PhD student Nurbek Tastan at MBZUAI, in collaboration with experts from Amazon Science and Michigan State University, LoFT aligns the optimizer's internal dynamics—specifically the first and second moments in algorithms like AdamW—with those of full fine-tuning. By meticulously calibrating gradients, momentum, and variance within the low-rank subspace, LoFT ensures that the adaptation process mirrors the comprehensive updates of full methods, without the overhead.
Understanding the Fine-Tuning Landscape
Before diving into LoFT, it's crucial to grasp the core challenge. Large language models (LLMs) like LLaMA-7B or 70B contain billions of parameters. Full fine-tuning requires duplicating the model and optimizer states, leading to memory footprints exceeding hundreds of gigabytes and training times spanning days on multi-GPU clusters. For context, fine-tuning a 7B model fully might consume 100x more VRAM than LoRA, which trains only 0.1-1% of parameters.
LoRA approximates weight updates as ΔW = U V^T, where U and V are low-rank matrices (rank r << model dimensions). This reduces trainable parameters dramatically—e.g., from 7 billion to mere millions—cutting costs by 90-99%. Yet, empirical gaps persist: on commonsense reasoning tasks, LoRA at rank 16 might score 73% accuracy versus full fine-tuning's 76%.
MBZUAI's insight? The discrepancy stems not just from gradient mismatches (addressed by prior scaling techniques) but from optimizer state misalignment. LoRA's compressed updates distort momentum (first moment) and variance (second moment), causing slower convergence and suboptimal solutions.
LoFT: A Five-Pronged Approach to Mimic Full Fine-Tuning
LoFT tackles this holistically through five synergistic components, implemented seamlessly atop standard LoRA frameworks.
- Gradient Scaling and Projection: Scales gradients to match full-model magnitudes and projects them onto the low-rank subspace, ensuring directionality aligns perfectly.
- Alternating Updates: Updates U and V matrices sequentially rather than jointly, eliminating cross-term distortions in second-order statistics.
- Optimizer State Calibration: Recalibrates first and second moments using projection matrices, bridging the subspace gap. This is LoFT's core innovation—previous methods ignored momentum drift.
- Projected Full Update Construction: Reconstructs a full-model-like update before low-rank projection, preserving global dynamics.
- Aware Gradient Clipping: Applies clipping in the full space equivalent, preventing explosion in low-rank approximations.
Notably, LoFT eliminates the LoRA scaling hyperparameter α, as alignments render it redundant. When rank equals full dimensions, LoFT provably recovers exact AdamW behavior—the first PEFT method to do so.

Benchmark Results: LoFT Shines on Commonsense Reasoning
Tested rigorously on LLaMA-7B, LLaMA2-7B, and LLaMA3-8B across benchmarks like PIQA, HellaSwag, WinoGrande, ARC-Easy, ARC-Challenge, and OpenBookQA, LoFT dominated.
For LLaMA-7B at rank 16, LoFT hit 76.08% average accuracy, beating LoRA's 73.57% and DoRA's 71.11%. At ultra-low rank 4, LoFT's 74.95% surpassed LoRA-rank-16. Even at rank 1, it scored 72.17%, viable where others faltered below 70%.
| Method | Rank 1 | Rank 4 | Rank 16 |
|---|---|---|---|
| LoFT (LLaMA-7B Avg) | 72.17% | 74.95% | 76.08% |
| LoRA | 68.42% | 70.23% | 73.57% |
| DoRA | 65.89% | 67.45% | 71.11% |
Training curves showed LoFT tracking full fine-tuning loss from epoch 1, converging 20-30% faster than LoRA.
Photo by Hardeep Singh on Unsplash
Scaling to Giants: LoFT on LLaMA-70B
LoFT's prowess extends to massive scales. On LLaMA-70B—a model too large for routine full fine-tuning—LoFT at rank 1 achieved competitive results, a literature first. This democratizes adaptation for trillion-parameter behemoths, critical as models grow.
Memory overhead? LoFT adds ~25% over vanilla LoRA but matches or undercuts advanced baselines like DoRA, while delivering superior accuracy.
Versatility Across Modalities: Vision Tasks
Beyond language, LoFT excelled on ViT-Base for image classification (ISIC2019 skin lesions, HAM10000, Diabetic Retinopathy, DomainNet). At rank 16, 76.12% average accuracy edged full fine-tuning's 75.86%, with stability at rank 4 (74.27%).

Implications for AI Efficiency and Sustainability
LoFT slashes fine-tuning costs: 70B models now tunable on fewer GPUs, reducing energy by 90%+ versus full methods. Ideal for edge devices, federated learning, and resource-poor regions.
In UAE's AI Strategy 2031, which invests billions in compute and talent, LoFT amplifies impact. MBZUAI, ranked top-10 globally for AI research, exemplifies this—fostering homegrown experts via fellowships and supercomputers.
UAE's AI blueprint positions the nation as a hub, with MBZUAI central to talent-building and open models for Global South languages.
MBZUAI: UAE's AI Vanguard
Founded to spearhead UAE's AI ambitions, MBZUAI boasts partnerships with Google.org ($1M grants), HPE supercomputers, and global unis. Its PhD programs produce leaders; LoFT, presented at ICLR 2026 in Rio, underscores research prowess.
Quotes from Tastan: "We calibrate momentums beyond gradient scaling." On 70B: "Nobody has done that before." This positions UAE as PEFT innovator.
For students eyeing AI careers, explore research positions or UAE opportunities at AcademicJobs UAE.
Photo by hayato togashi on Unsplash
Future Horizons: LoFT's Ripple Effects
LoFT paves for quantized/privacy-preserving variants, vital for mobile AI and regulations. Experts hail it for low-rank viability on giants, accelerating deployment in healthcare, finance.
In UAE, it bolsters sectors like energy (robotics for oil), agriculture (precision farming). Globally, PEFT like LoFT counters compute walls, enabling broader innovation.
Read the LoFT paper or MBZUAI's announcement.
Why This Matters for UAE Higher Education
MBZUAI's feat highlights UAE universities' ascent: top QS AI rankings, attracting global talent. Amid 57k new enrollments, such research draws students, positions UAE as AI exporter.
For aspiring researchers, MBZUAI fellowships build UAE's AI pipeline, aligning with national goals for economic diversification.

Be the first to comment on this article!
Please keep comments respectful and on-topic.