News
Hugging Face Raises Series B!
π£ We are so excited to announce our $40M series B led by Lee Fixel at Addition with participation from Lux Capital, A.Capital Ventures, and betaworks!
Thank you to all our open source contributors, pull requesters, issue openers, notebook creators, model architects, tweeting supporters & community members all over the world π!
We couldn't do what we do & be where we are - in a field dominated by big tech - without you! ππ»
Check us out on TechCrunch and VentureBeat!
π§ Train Transformers faster with Hugging Face in Amazon SageMaker π
We partnered with Amazon SageMaker to enable faster training of Transformers in your AWS cloud! π₯
Head to our blog for walkthroughs, documentation and sample notebooks showing you how to use the new Hugging Face Deep Learning Containers (DLCs) with the SageMaker Python SDK to train models with PyTorch and TensorFlow, and
πData Parallelism
πModel Parallelism
πΈSpot Instances
πCustom Metrics
π€ Transformers v4.4 gets 5 new models!
1οΈβ£ π Multilingual w/ M2M100 and 2οΈβ£ mBART-50
3οΈβ£ π€ Speech w/ Wav2Vec2-XLSR
4οΈβ£ Quantization w/ I-BERT
5οΈβ£ π₯ SOTA NLU w/ DeBERTa-v2
Not to mention:
βοΈ TF models support XLA & AMP
β‘οΈ Trainer supports SageMaker model Parallelism
New v1.5 of π€ Datasets is out π₯
πΎ Tokenized datasets now 4x (!) smaller
π₯ Simpler from_csv/json/text
π New datasets: Common Voice, SST, e.g.:
π€ Common Voice: speech data in 60 languages!
π©π»βπ€ fashionMNIST for CV.
π Shout out to over 800+ users who are already sharing and hosting their datasets on the Hub!
π Any dataset can be loaded with one line of python.
ππ» Check out the full list here
π Learn how to add yours here
Dark Mode is Here!
π π π Get your equipment because it's getting very dark in here... The long-awaited dark mode is now available on Hugging Face π
To try it out, activate in your user settings/theme (you have to be a registered user).
Facebook AI's Wav2Vec2 is now available in Transformers!
Not only for English but for 53 Languages π€―
Check out the tutorials:
ππ» Train Wav2Vec2 on TIMIT
ππ» Train XLSR-Wav2Vec2 on Common Voice
For more interaction, take a look at these colab tutorials:
ππ» Train Wav2Vec2 on TIMIT
ππ» Train XLSR-Wav2Vec2 on Common Voice
AutoNLP is Here!
π Looking for a sneak peek of AutoNLP in action?
Check out this exclusive preview video by Abhishek Thakur that shows just how easy it is to train models using AutoNLP!
FairScale Support Release
FairScale just released support for ZeRO-DP3 and ZeRO-offload (to make large model fine-tuning easier), and you can already start playing with it in π€ Transformers!
This is still highly experimental so expect a few (maybe a lot) of rough edges. The PR gives a few examples, refer to the master documentation for more information.
For more information on what ZeRO-DP and ZeRO-offload are, you're in luck! Sylvain Gugger gave a talk about the topic at PyDataMTL.
The new SOTA is in Transformers!
DeBERTa-v2 beats the human baseline on SuperGLUE and up to a crazy 91.7% dev accuracy on MNLI task. It even beats T5 while 10x smaller!
DeBERTa-v2 was contributed by Pengcheng He from Microsoft Research
Try it directly on the hub or in π€ Transformers by installing from source!
DeBERTa will be available from pypi/anaconda as early as v4.4.0 is out!
π£ Announcing a new Ray + Hugging Face integration!
RAG is a new NLP model that uses external documents to augment its knowledge. The RAG model by Aleksandra Piktus, Patrick Lewis, and more Facebook AI colleagues leverages external knowledge sources like Wikipedia to have direct and dynamic access to information at inference time.
π This integration with RAG:
- Speeds up retrieval calls by 2x
- Improves the scalability of fine-tuning
π Check out our newest guest post by Amog Kamsetty and the Ray team on training a Retrieval Augmented Generation Model.
π§ββοΈTrain a Text Classifier with Unlabeled Data
Weβve added a script to π€ Transformers that allows you to train a text classifier with nothing but a set of specified class names and some unlabeled data!
The script generates proxy-labels on your data from our zero-shot classification pipeline and performs knowledge distillation by training a smaller student model πͺ
The result is an efficient classifier that speeds up inference by 100x or more compared to zero-shot classification π
π Script
π Model Hub Highlights π
Translate text to or between 50 languages with mBART-50 from Facebook AI!
πΊπ³ One-to-Many model: translate from English to 49 other languages
βοΈ Many-to-Many model: translation between any pair of 50 languages
Check out all the mBART-50 models
I-BERT
π₯Brought to you by UC Berkeley, I-BERT is the first quantized model in π€Model Hub! Everything is integer in I-BERT. It brings you 4x speed-up with TensorRT!!
Community
π Hugging Face Reading Group
The Hugging Face Reading Group is back!
We frequently need to manipulate extremely long sequences for application such as document summarization and also in modalities outside of NLP. But how do you efficiently process sequences of over 64K tokens with Transformers?
We took a deep dive into long-range transformers to understand the different approaches, their assumptions, as well as their strengths and weaknesses.
To learn how you can use it for your use-cases, check out the paper summaries and our discussion!
Collaboration between Hugging Face and Tensorflow's implementation of RAG
An incredible collaboration between Hugging Face and Tensorflow's implementation of RAG by (atthachat Chatpatanasiri (Jung)!
ICYMI: RAG is an AI prototype that can read articles to give answers to any questions!
With appropriate training data like ELI5, it can even give free-form answers with reasonable arguments!
πΊ Languages at Hugging Face
Join our new "Languages at Hugging Face" initiative! You can discuss language-specific tools and resources, connect with local groups, and share initiatives to make #NLP easier in your own language!
Tutorials
Fine-tuning Turkish BERT
A Turkish tutorial on fine-tuning Turkish BERT for sentiment analysis on product reviews dataset with native #PyTorch and Hugging Face by Merve Noyan!
Finetuning and Accelerating Transformers
If you want to finetune Transformers and accelerate them by >2X at the same time, the Hugging Face nn_pruning library is now available thanks to François Lagunas!
Check the results on SQuAD and MNLI here
Speed up your Training on TPUs
Do you want to MASSIVELY speed up your trainings on TPU? π
Transformers has TPU support for all Facebook AI PyTorch training scripts thanks to PyTorch/XLA.
Check out our joint blog post with Google which showcases the integration and examples.