Skip to main content

Hugging Face Blog Posts

23 articles on transformer models and the HF ecosystem

July 3, 2024

Accelerating Protein Language Model ProtST on Intel Gaudi 2 (opens in new tab)

Protein Language Models (PLMs) have emerged as potent tools for predicting and designing protein structure and function. At the International...

local copy

May 9, 2024

Building Cost-Efficient Enterprise RAG applications with Intel Gaudi 2 and Intel Xeon (opens in new tab)

Retrieval-augmented generation (RAG) enhances text generation with a large language model by incorporating fresh domain knowledge stored in an...

local copy

March 20, 2024

A Chatbot on your Laptop: Phi-2 on Intel Meteor Lake (opens in new tab)

Retrieval-augmented generation (RAG) enhances text generation with a large language model by incorporating fresh domain knowledge stored in an...

local copy

September 11, 2023

SafeCoder vs. Closed-source Code Assistants (opens in new tab)

Because of their impressive abilities, large language models (LLMs) require significant computing power, which is seldom available on personal...

local copy

July 14, 2023

Fine-tuning Stable Diffusion models on Intel CPUs (opens in new tab)

For decades, software developers have designed methodologies, processes, and tools that help them improve code quality and increase productivity. For...

local copy

June 13, 2023

Hugging Face and AMD partner on accelerating state-of-the-art models for CPU and GPU platforms (opens in new tab)

Whether language models, large language models, or foundation models, transformers require significant computation for pre-training, fine-tuning, and...

local copy

May 23, 2023

Hugging Face Endpoints on Azure (opens in new tab)

Whether language models, large language models, or foundation models, transformers require significant computation for pre-training, fine-tuning, and...

local copy

May 23, 2023

Hugging Face and IBM partner on watsonx.ai, the next-generation enterprise studio for AI builders (opens in new tab)

All hype aside, it’s hard to deny the profound impact that AI is having on society and businesses. From startups to enterprises to the public sector,...

local copy

May 16, 2023

Smaller is better: Q8-Chat, an efficient generative AI experience on Xeon (opens in new tab)

Large language models (LLMs) are taking the machine learning world by storm. Thanks to their Transformer architecture, LLMs have an uncanny ability...

local copy

April 17, 2023

Accelerating Hugging Face Transformers with AWS Inferentia2 (opens in new tab)

Large language models (LLMs) are taking the machine learning world by storm. Thanks to their Transformer architecture, LLMs have an uncanny ability...

local copy

March 28, 2023

Accelerating Stable Diffusion Inference on Intel CPUs (opens in new tab)

In the last five years, Transformer models [ 1 ] have become the de facto standard for many machine learning (ML) tasks, such as natural language...

local copy

March 1, 2023

How Hugging Face Accelerated Development of Witty Works Writing Assistant (opens in new tab)

If you’re interested in building ML solutions faster, visit the Expert Acceleration Program landing page and contact us here !

local copy

February 21, 2023

Hugging Face and AWS partner to make AI more accessible (opens in new tab)

If you’re interested in building ML solutions faster, visit the Expert Acceleration Program landing page and contact us here !

local copy

February 6, 2023

Accelerating PyTorch Transformers with Intel Sapphire Rapids, part 2 (opens in new tab)

It’s time to make AI open and accessible to all. That’s the goal of this expanded long-term strategic partnership between Hugging Face and Amazon Web...

local copy

January 2, 2023

Accelerating PyTorch Transformers with Intel Sapphire Rapids, part 1 (opens in new tab)

In a recent post , we introduced you to the fourth generation of Intel Xeon CPUs, code-named Sapphire Rapids , and its new Advanced Matrix Extensions...

local copy

November 21, 2022

An Overview of Inference Solutions on Hugging Face (opens in new tab)

About a year ago, we showed you how to distribute the training of Hugging Face transformers on a cluster or third-generation Intel Xeon Scalable CPUs...

local copy

November 2, 2022

Accelerate your models with Optimum Intel and OpenVINO (opens in new tab)

Every day, developers and organizations are adopting models hosted on Hugging Face to turn ideas into proof-of-concept demos, and demos into...

local copy

October 14, 2022

Getting started with Hugging Face Inference Endpoints (opens in new tab)

Last July, we announced that Intel and Hugging Face would collaborate on building state-of-the-art yet simple hardware acceleration tools for...

local copy

August 18, 2022

Deep Dive: Vision Transformers On Hugging Face Optimum Graphcore (opens in new tab)

Training machine learning models has become quite simple, especially with the rise of pre-trained models and transfer learning. OK, sometimes it’s...

local copy

June 15, 2022

Intel and Hugging Face Partner to Democratize Machine Learning Hardware Acceleration (opens in new tab)

The mission of Hugging Face is to democratize good machine learning and maximize its positive impact across industries and society. Not only do we...

local copy

April 26, 2022

Getting Started with Transformers on Habana Gaudi (opens in new tab)

A couple of weeks ago, we’ve had the pleasure to announce that Habana Labs and Hugging Face would partner to accelerate Transformer model training.

local copy

November 30, 2021

Getting Started with Hugging Face Transformers for IPUs with Optimum (opens in new tab)

A couple of weeks ago, we’ve had the pleasure to announce that Habana Labs and Hugging Face would partner to accelerate Transformer model training.

local copy

November 19, 2021

Accelerating PyTorch distributed fine-tuning with Intel technologies (opens in new tab)

Transformer models have proven to be extremely efficient on a wide range of machine learning tasks, such as natural language processing, audio...

local copy