Hugging Face Blog Posts
23 articles on transformer models and the HF ecosystem
July 3, 2024
Accelerating Protein Language Model ProtST on Intel Gaudi 2 (opens in new tab)
Protein Language Models (PLMs) have emerged as potent tools for predicting and designing protein structure and function. At the International...
May 9, 2024
Building Cost-Efficient Enterprise RAG applications with Intel Gaudi 2 and Intel Xeon (opens in new tab)
Retrieval-augmented generation (RAG) enhances text generation with a large language model by incorporating fresh domain knowledge stored in an...
March 20, 2024
A Chatbot on your Laptop: Phi-2 on Intel Meteor Lake (opens in new tab)
Retrieval-augmented generation (RAG) enhances text generation with a large language model by incorporating fresh domain knowledge stored in an...
September 11, 2023
SafeCoder vs. Closed-source Code Assistants (opens in new tab)
Because of their impressive abilities, large language models (LLMs) require significant computing power, which is seldom available on personal...
July 14, 2023
Fine-tuning Stable Diffusion models on Intel CPUs (opens in new tab)
For decades, software developers have designed methodologies, processes, and tools that help them improve code quality and increase productivity. For...
June 13, 2023
Hugging Face and AMD partner on accelerating state-of-the-art models for CPU and GPU platforms (opens in new tab)
Whether language models, large language models, or foundation models, transformers require significant computation for pre-training, fine-tuning, and...
May 23, 2023
Hugging Face Endpoints on Azure (opens in new tab)
Whether language models, large language models, or foundation models, transformers require significant computation for pre-training, fine-tuning, and...
May 23, 2023
Hugging Face and IBM partner on watsonx.ai, the next-generation enterprise studio for AI builders (opens in new tab)
All hype aside, it’s hard to deny the profound impact that AI is having on society and businesses. From startups to enterprises to the public sector,...
May 16, 2023
Smaller is better: Q8-Chat, an efficient generative AI experience on Xeon (opens in new tab)
Large language models (LLMs) are taking the machine learning world by storm. Thanks to their Transformer architecture, LLMs have an uncanny ability...
April 17, 2023
Accelerating Hugging Face Transformers with AWS Inferentia2 (opens in new tab)
Large language models (LLMs) are taking the machine learning world by storm. Thanks to their Transformer architecture, LLMs have an uncanny ability...
March 28, 2023
Accelerating Stable Diffusion Inference on Intel CPUs (opens in new tab)
In the last five years, Transformer models [ 1 ] have become the de facto standard for many machine learning (ML) tasks, such as natural language...
March 1, 2023
How Hugging Face Accelerated Development of Witty Works Writing Assistant (opens in new tab)
If you’re interested in building ML solutions faster, visit the Expert Acceleration Program landing page and contact us here !
February 21, 2023
Hugging Face and AWS partner to make AI more accessible (opens in new tab)
If you’re interested in building ML solutions faster, visit the Expert Acceleration Program landing page and contact us here !
February 6, 2023
Accelerating PyTorch Transformers with Intel Sapphire Rapids, part 2 (opens in new tab)
It’s time to make AI open and accessible to all. That’s the goal of this expanded long-term strategic partnership between Hugging Face and Amazon Web...
January 2, 2023
Accelerating PyTorch Transformers with Intel Sapphire Rapids, part 1 (opens in new tab)
In a recent post , we introduced you to the fourth generation of Intel Xeon CPUs, code-named Sapphire Rapids , and its new Advanced Matrix Extensions...
November 21, 2022
An Overview of Inference Solutions on Hugging Face (opens in new tab)
About a year ago, we showed you how to distribute the training of Hugging Face transformers on a cluster or third-generation Intel Xeon Scalable CPUs...
November 2, 2022
Accelerate your models with Optimum Intel and OpenVINO (opens in new tab)
Every day, developers and organizations are adopting models hosted on Hugging Face to turn ideas into proof-of-concept demos, and demos into...
October 14, 2022
Getting started with Hugging Face Inference Endpoints (opens in new tab)
Last July, we announced that Intel and Hugging Face would collaborate on building state-of-the-art yet simple hardware acceleration tools for...
August 18, 2022
Deep Dive: Vision Transformers On Hugging Face Optimum Graphcore (opens in new tab)
Training machine learning models has become quite simple, especially with the rise of pre-trained models and transfer learning. OK, sometimes it’s...
June 15, 2022
Intel and Hugging Face Partner to Democratize Machine Learning Hardware Acceleration (opens in new tab)
The mission of Hugging Face is to democratize good machine learning and maximize its positive impact across industries and society. Not only do we...
April 26, 2022
Getting Started with Transformers on Habana Gaudi (opens in new tab)
A couple of weeks ago, we’ve had the pleasure to announce that Habana Labs and Hugging Face would partner to accelerate Transformer model training.
November 30, 2021
Getting Started with Hugging Face Transformers for IPUs with Optimum (opens in new tab)
A couple of weeks ago, we’ve had the pleasure to announce that Habana Labs and Hugging Face would partner to accelerate Transformer model training.
November 19, 2021
Accelerating PyTorch distributed fine-tuning with Intel technologies (opens in new tab)
Transformer models have proven to be extremely efficient on a wide range of machine learning tasks, such as natural language processing, audio...