Company
Date Published
Author
Subash SN, Akshay Nambiar, Patrik Lambert, Milan Gritta, and Amril Nurman
Word count
4604
Language
-
Hacker News points
None

Summary

Tether Data's AI research division has introduced QVAC-fabric-llm, a framework designed to enable cross-platform Low-Rank Adaptation (LoRA) fine-tuning of Large Language Models (LLMs) on a diverse range of consumer hardware, including mobile and desktop GPUs. This framework integrates with the llama.cpp ecosystem to democratize AI by making fine-tuning vendor-independent, thereby facilitating on-device personalization and instruction-tuning. The solution employs the Vulkan API for broad compatibility, allowing modern models like Qwen3 and Gemma3 to be fine-tuned on various devices, from smartphones to servers. Tether Data has also released multi-platform binaries, fine-tuned model adapters, and source code under the Apache 2.0 license to empower the AI community. The project demonstrates successful on-device fine-tuning for applications such as email style transfer and biomedical question answering, offering a scalable, privacy-preserving AI platform that extends fine-tuning capabilities beyond traditional data centers.