ThirdAI Unveils BOLT2.5B: The World’s First Generative LLM Trained Only on CPUs
Introducing the World’s First Generative LLM Pre-Trained Only on CPUs. The model is fine-tunable on commodity CPUs in a matter of minutes.
Key Insights into BOLT2.5B:
Rapid Pre-training: In just 20 days, ThirdAI achieved pre-training of the model on 10 Sapphire Rapids CPUs, under the Intel Disruptor program, processing a staggering 2 billion tokens each day, totaling around 40 billion tokens.
Remarkable Performance: Preliminary evaluations show BOLT2.5B's capabilities rivaling OpenAI’s GPT-2 XL model, which was trained using 128 V100s over 10 days, processing a phenomenal 170 billion tokens.
Unlocking the Power of CPUs for GenAI:
Train Your Own: With ThirdAI’s innovative approach, pre-training a GenAI from scratch is now astonishingly simple. Users can harness cloud CPUs or data centers to craft their custom BOLT GenAI models with ease.
Fine-tuning Redefined: BOLT2.5B's 'dynamic sparse' architecture allows even older desktops to fine-tune the model, with rates reaching up to 50 tokens per second. An illustration provided by ThirdAI showcases the model being fine-tuned on a Shakespearean corpus, achieving one epoch of fine-tuning in just 20 minutes on a 2014 dual socket Intel(R) Xeon(R) CPU E5–2680 v3 server
Swift Inference: Designed with CPUs in mind, BOLT2.5B delivers rapid inference, producing tokens at a rate of 20 per second without the need for specialized treatments like quantization or pruning.
"We are at the dawn of a new GenAI era, where the capabilities of commodity computing rise to meet the demands of AI. BOLT2.5B is not just a model; it's a testament to the untapped power of CPUs in the AI realm. With this launch, we invite everyone to experience this transformative shift," says Anshumali Shrivastava, Founder & CEO of ThirdAI.
ThirdAI has hosted BOLT2.5B on Hugging Face and provided a Github notebook for any user wishing to experiment. With only a requirement of 10GB of RAM for inference and 20GB for fine-tuning, there's no longer a need to wait for GPUs to develop and fine-tune a custom LLM.
About ThirdAI: ThirdAI is a Houston-based AI company that aims to make deep learning technology more affordable and attainable for all. It builds and deploys large language models on CPUs without the need for configurations or expensive GPUs. ThirdAI’s models reduce latency, minimize training time, and increase accuracy by retraining only specific parameters when needed, rather than updating the entire AI model. ThirdAI enables companies to train LLMs on their own data with existing hardware, lowering the cost and energy consumption associated with AI training.
For more information visit our website at https://www.thirdai.com/bolt-25b/
Vinod Iyengar
ThirdAI
+ 19087648639
email us here
Visit us on social media:
Twitter
LinkedIn
YouTube
Introducing BOLT2.5B
Legal Disclaimer:
EIN Presswire provides this news content "as is" without warranty of any kind. We do not accept any responsibility or liability for the accuracy, content, images, videos, licenses, completeness, legality, or reliability of the information contained in this article. If you have any complaints or copyright issues related to this article, kindly contact the author above.
