Tether has announced a new AI training framework that allows fine-tuning large language models directly on consumer devices such as smartphones and non-Nvidia GPUs. This system, part of the QVAC platform, leverages Microsoft’s BitNet architecture combined with LoRA techniques to significantly reduce memory requirements and computational costs.
According to Tether, the framework supports multiple platforms and is compatible with chips from AMD, Intel, Apple Silicon, and Qualcomm mobile GPUs. Engineers can fine-tune models with up to 1 billion parameters on smartphones in under two hours, and even scale up to 13 billion parameters on mobile devices.
BitNet technology reduces VRAM usage by up to 77.8% compared to 16-bit models and accelerates inference on mobile GPUs. Tether also emphasizes potential applications such as federated learning, reducing reliance on the cloud.
This move reflects a trend among crypto companies expanding into AI and computing infrastructure, alongside the growth of AI agents in the industry.