Skip to content

Commit

Permalink
No more magic comments (huggingface#1554)
Browse files Browse the repository at this point in the history
* no more magic comments

* Also replace h1 by actual markdown

* nit: remove extra space

* Fix remaining <h1>s

* handling complex h1 heading

* Update README.md

cc @mishig25

---------

Co-authored-by: Mishig Davaadorj <[email protected]>
  • Loading branch information
julien-c and mishig25 authored Oct 3, 2023
1 parent ef9aed0 commit 2d74c93
Show file tree
Hide file tree
Showing 384 changed files with 115 additions and 879 deletions.
2 changes: 0 additions & 2 deletions 1b-sentence-embeddings.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,8 +7,6 @@ authors:

# Train a Sentence Embedding Model with 1 Billion Training Pairs

<!-- {blog_metadata} -->
<!-- {authors} -->

**Sentence embedding** is a method that maps sentences to vectors of real numbers. Ideally, these vectors would capture the semantic of a sentence and be highly generic. Such representations could then be used for many downstream applications such as clustering, text mining, or question answering.

Expand Down
2 changes: 0 additions & 2 deletions 3d-assets.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,8 +7,6 @@ authors:

# Practical 3D Asset Generation: A Step-by-Step Guide

<!-- {blog_metadata} -->
<!-- {authors} -->

## Introduction

Expand Down
2 changes: 0 additions & 2 deletions 4bit-transformers-bitsandbytes.md
Original file line number Diff line number Diff line change
Expand Up @@ -13,8 +13,6 @@ authors:

# Making LLMs even more accessible with bitsandbytes, 4-bit quantization and QLoRA

<!-- {blog_metadata} -->
<!-- {authors} -->

LLMs are known to be large, and running or training them in consumer hardware is a huge challenge for users and accessibility.
Our [LLM.int8 blogpost](https://huggingface.co/blog/hf-bitsandbytes-integration) showed how the techniques in the [LLM.int8 paper](https://arxiv.org/abs/2208.07339) were integrated in transformers using the `bitsandbytes` library.
Expand Down
2 changes: 0 additions & 2 deletions Llama2-for-non-engineers.md
Original file line number Diff line number Diff line change
Expand Up @@ -6,8 +6,6 @@ authors:
- user: abhishek
---

<!-- {blog_metadata} -->
<!-- {authors} -->

# Non-engineers guide: Train a LLaMA 2 chatbot

Expand Down
8 changes: 2 additions & 6 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -29,19 +29,15 @@ authors:
# Train your first Decision Transformer
<!-- {blog_metadata} -->
<!-- {authors} -->
Your content here [...]
```

The blog_metadata and authors HTML comments are meant to mark where in the file will be inserted the following UI elements:
When published, the Hub will insert the following UI elements right after the blogpost's main header (i.e. the line that starts with a single `#`, aka. the `<h1>`):

- "Published on [date]"
- "Update on GitHub" button
- avatars of the authors that were listed in authors.

⚠️ Please keep the blog_metadata and authors comments exactly equal to those strings otherwise they won't be replaced.

5️⃣ Then, you can add your content. It's markdown system so if you wrote your text on notion just control shift v to copy/paste as markdown.

6️⃣ Modify `_blog.yml` to add your blogpost.
Expand Down
4 changes: 1 addition & 3 deletions accelerate-deepspeed.md
Original file line number Diff line number Diff line change
Expand Up @@ -6,10 +6,8 @@ authors:
- user: sgugger
---

<h1>Accelerate Large Model Training using DeepSpeed</h1>
# Accelerate Large Model Training using DeepSpeed

<!-- {blog_metadata} -->
<!-- {authors} -->

In this post we will look at how we can leverage the **[Accelerate](https://github.com/huggingface/accelerate)** library for training large models which enables users to leverage the ZeRO features of **[DeeSpeed](https://www.deepspeed.ai)**.

Expand Down
2 changes: 0 additions & 2 deletions accelerate-large-models.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,8 +7,6 @@ authors:

# How 🤗 Accelerate runs very large models thanks to PyTorch

<!-- {blog_metadata} -->
<!-- {authors} -->

## Load and run large models

Expand Down
2 changes: 0 additions & 2 deletions accelerate-library.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,8 +7,6 @@ authors:

# Introducing 🤗 Accelerate

<!-- {blog_metadata} -->
<!-- {authors} -->

## 🤗 Accelerate

Expand Down
2 changes: 0 additions & 2 deletions accelerate-transformers-with-inferentia2.md
Original file line number Diff line number Diff line change
Expand Up @@ -8,8 +8,6 @@ authors:

# Accelerating Hugging Face Transformers with AWS Inferentia2

<!-- {blog_metadata} -->
<!-- {authors} -->

<script async defer src="https://unpkg.com/medium-zoom-element@0/dist/medium-zoom-element.min.js"></script>

Expand Down
3 changes: 1 addition & 2 deletions accelerated-inference.md
Original file line number Diff line number Diff line change
Expand Up @@ -3,9 +3,8 @@ title: "How we sped up transformer inference 100x for 🤗 API customers"
thumbnail: /blog/assets/09_accelerated_inference/thumbnail.png
---

<h1>How we sped up transformer inference 100x for 🤗 API customers</h1>
# How we sped up transformer inference 100x for 🤗 API customers

<!-- {blog_metadata} -->

🤗 Transformers has become the default library for data scientists all around the world to explore state of the art NLP models and build new NLP features. With over 5,000 pre-trained and fine-tuned models available, in over 250 languages, it is a rich playground, easily accessible whichever framework you are working in.

Expand Down
2 changes: 0 additions & 2 deletions accelerating-pytorch.md
Original file line number Diff line number Diff line change
Expand Up @@ -8,8 +8,6 @@ authors:
# Accelerating PyTorch distributed fine-tuning with Intel technologies


<!-- {blog_metadata} -->
<!-- {authors} -->

For all their amazing performance, state of the art deep learning models often take a long time to train. In order to speed up training jobs, engineering teams rely on distributed training, a divide-and-conquer technique where clustered servers each keep a copy of the model, train it on a subset of the training set, and exchange results to converge to a final model.

Expand Down
2 changes: 0 additions & 2 deletions agents-js.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,8 +7,6 @@ authors:

# Introducing Agents.js: Give tools to your LLMs using JavaScript

<!-- {blog_metadata} -->
<!-- {authors} -->

We have recently been working on Agents.js at [huggingface.js](https://github.com/huggingface/huggingface.js/blob/main/packages/agents/README.md). It's a new library for giving tool access to LLMs from JavaScript in either the browser or the server. It ships with a few multi-modal tools out of the box and can easily be extended with your own tools and language models.

Expand Down
2 changes: 0 additions & 2 deletions ai-comic-factory.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,8 +7,6 @@ authors:

# Deploying the AI Comic Factory using the Inference API

<!-- {blog_metadata} -->
<!-- {authors} -->

We recently announced [Inference for PROs](https://huggingface.co/blog/inference-pro), our new offering that makes larger models accessible to a broader audience. This opportunity opens up new possibilities for running end-user applications using Hugging Face as a platform.

Expand Down
2 changes: 0 additions & 2 deletions ai-residency.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,8 +7,6 @@ authors:

# Announcing the 🤗 AI Research Residency Program 🎉 🎉 🎉

<!-- {blog_metadata} -->
<!-- {authors} -->


The 🤗 Research Residency Program is a 9-month opportunity to launch or advance your career in machine learning research 🚀. The goal of the residency is to help you grow into an impactful AI researcher. Residents will work alongside Researchers from our Science Team. Together, you will pick a research problem and then develop new machine learning techniques to solve it in an open & collaborative way, with the hope of ultimately publishing your work and making it visible to a wide audience.
Expand Down
2 changes: 0 additions & 2 deletions ai-webtv.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,8 +7,6 @@ authors:

# Building an AI WebTV

<!-- {blog_metadata} -->
<!-- {authors} -->

The AI WebTV is an experimental demo to showcase the latest advancements in automatic video and music synthesis.

Expand Down
2 changes: 0 additions & 2 deletions aivsai.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,8 +7,6 @@ authors:
---

# Introducing ⚔️ AI vs. AI ⚔️ a deep reinforcement learning multi-agents competition system
<!-- {blog_metadata} -->
<!-- {authors} -->

<div align="center">
<img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/128_aivsai/thumbnail.png" alt="Thumbnail">
Expand Down
2 changes: 0 additions & 2 deletions ambassadors.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,8 +7,6 @@ authors:

# Student Ambassador Program’s call for applications is open!

<!-- {blog_metadata} -->
<!-- {authors} -->

As an open-source company democratizing machine learning, Hugging Face believes it is essential to **[teach](https://huggingface.co/blog/education)** open-source ML to people from all backgrounds worldwide. **We aim to teach machine learning to 5 million people by 2023**.

Expand Down
2 changes: 0 additions & 2 deletions annotated-diffusion.md
Original file line number Diff line number Diff line change
Expand Up @@ -8,8 +8,6 @@ authors:

# The Annotated Diffusion Model

<!-- {blog_metadata} -->
<!-- {authors} -->

<script async defer src="https://unpkg.com/medium-zoom-element@0/dist/medium-zoom-element.min.js"></script>

Expand Down
2 changes: 0 additions & 2 deletions arxiv.md
Original file line number Diff line number Diff line change
Expand Up @@ -9,8 +9,6 @@ authors:

# Hugging Face Machine Learning Demos on arXiv

<!-- {blog_metadata} -->
<!-- {authors} -->

We’re very excited to announce that Hugging Face has collaborated with arXiv to make papers more accessible, discoverable, and fun! Starting today, [Hugging Face Spaces](https://huggingface.co/spaces) is integrated with arXivLabs through a Demo tab that includes links to demos created by the community or the authors themselves. By going to the Demos tab of your favorite paper, you can find links to open-source demos and try them out immediately 🔥

Expand Down
2 changes: 0 additions & 2 deletions asr-chunking.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,8 +7,6 @@ authors:

# Making automatic speech recognition work on large files with Wav2Vec2 in 🤗 Transformers

<!-- {blog_metadata} -->
<!-- {authors} -->

```
Tl;dr: This post explains how to use the specificities of the Connectionist
Expand Down
2 changes: 0 additions & 2 deletions assisted-generation.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,8 +7,6 @@ authors:

# Assisted Generation: a new direction toward low-latency text generation

<!-- {blog_metadata} -->
<!-- {authors} -->

Large language models are all the rage these days, with many companies investing significant resources to scale them up and unlock new capabilities. However, as humans with ever-decreasing attention spans, we also dislike their slow response times. Latency is critical for a good user experience, and smaller models are often used despite their lower quality (e.g. in [code completion](https://ai.googleblog.com/2022/07/ml-enhanced-code-completion-improves.html)).

Expand Down
2 changes: 0 additions & 2 deletions audio-datasets.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,8 +7,6 @@ authors:

# A Complete Guide to Audio Datasets

<!-- {blog_metadata} -->
<!-- {authors} -->

<!--- Note to reviewer: comments and TODOs are included in this format. --->

Expand Down
2 changes: 0 additions & 2 deletions audioldm2.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,8 +7,6 @@ authors:

# AudioLDM 2, but faster ⚡️

<!-- {blog_metadata} -->
<!-- {authors} -->

<a target="_blank" href="https://colab.research.google.com/github/sanchit-gandhi/notebooks/blob/main/AudioLDM-2.ipynb">
<img src="https://colab.research.google.com/assets/colab-badge.svg" alt="Open In Colab"/>
Expand Down
2 changes: 0 additions & 2 deletions autoformer.md
Original file line number Diff line number Diff line change
Expand Up @@ -10,8 +10,6 @@ authors:

# Yes, Transformers are Effective for Time Series Forecasting (+ Autoformer)

<!-- {blog_metadata} -->
<!-- {authors} -->

<script async defer src="https://unpkg.com/medium-zoom-element@0/dist/medium-zoom-element.min.js"></script>

Expand Down
4 changes: 1 addition & 3 deletions autonlp-prodigy.md
Original file line number Diff line number Diff line change
Expand Up @@ -5,10 +5,8 @@ authors:
- user: abhishek
---

<h1>Active Learning with AutoNLP and Prodigy</h1>
# Active Learning with AutoNLP and Prodigy

<!-- {blog_metadata} -->
<!-- {authors} -->

Active learning in the context of Machine Learning is a process in which you iteratively add labeled data, retrain a model and serve it to the end user. It is an endless process and requires human interaction for labeling/creating the data. In this article, we will discuss how to use [AutoNLP](https://huggingface.co/autonlp) and [Prodigy](https://prodi.gy/) to build an active learning pipeline.

Expand Down
2 changes: 0 additions & 2 deletions autotrain-image-classification.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,8 +7,6 @@ authors:

# Image Classification with AutoTrain

<!-- {blog_metadata} -->
<!-- {authors} -->

<script async defer src="https://unpkg.com/medium-zoom-element@0/dist/medium-zoom-element.min.js"></script>

Expand Down
2 changes: 0 additions & 2 deletions aws-marketplace.md
Original file line number Diff line number Diff line change
Expand Up @@ -9,8 +9,6 @@ authors:

# Hugging Face Platform on the AWS Marketplace: Pay with your AWS Account

<!-- {blog_metadata} -->
<!-- {authors} -->

The [Hugging Face Platform](https://aws.amazon.com/marketplace/pp/prodview-n6vsyhdjkfng2) has landed on the AWS Marketplace. Starting today, you can subscribe to the Hugging Face Platform through AWS Marketplace to pay for your Hugging Face usage directly with your AWS account. This new integrated billing method makes it easy to manage payment for usage of all our managed services by all members of your organization, including Inference Endpoints, Spaces Hardware Upgrades, and AutoTrain to easily train, test and deploy the most popular machine learning models like Llama 2, StarCoder, or BERT.

Expand Down
2 changes: 0 additions & 2 deletions aws-partnership.md
Original file line number Diff line number Diff line change
Expand Up @@ -9,8 +9,6 @@ authors:

# Hugging Face and AWS partner to make AI more accessible

<!-- {blog_metadata} -->
<!-- {authors} -->

It’s time to make AI open and accessible to all. That’s the goal of this expanded long-term strategic partnership between Hugging Face and Amazon Web Services (AWS). Together, the two leaders aim to accelerate the availability of next-generation machine learning models by making them more accessible to the machine learning community and helping developers achieve the highest performance at the lowest cost.

Expand Down
4 changes: 1 addition & 3 deletions bert-101.md
Original file line number Diff line number Diff line change
Expand Up @@ -5,10 +5,8 @@ authors:
- user: britneymuller
---
<html itemscope itemtype="https://schema.org/FAQPage">
<h1>BERT 101 🤗 State Of The Art NLP Model Explained</h1>
# BERT 101 🤗 State Of The Art NLP Model Explained

<!-- {blog_metadata} -->
<!-- {authors} -->

<script async defer src="https://unpkg.com/medium-zoom-element@0/dist/medium-zoom-element.min.js"></script>

Expand Down
2 changes: 0 additions & 2 deletions bert-cpu-scaling-part-1.md
Original file line number Diff line number Diff line change
Expand Up @@ -19,8 +19,6 @@ authors:
}
</style>

<!-- {blog_metadata} -->
<!-- {authors} -->

# Scaling up BERT-like model Inference on modern CPU - Part 1

Expand Down
2 changes: 0 additions & 2 deletions bert-cpu-scaling-part-2.md
Original file line number Diff line number Diff line change
Expand Up @@ -9,8 +9,6 @@ authors:

# Scaling up BERT-like model Inference on modern CPU - Part 2

<!-- {blog_metadata} -->
<!-- {authors} -->

<script async defer src="https://unpkg.com/medium-zoom-element@0/dist/medium-zoom-element.min.js"></script>

Expand Down
4 changes: 1 addition & 3 deletions bert-inferentia-sagemaker.md
Original file line number Diff line number Diff line change
Expand Up @@ -5,10 +5,8 @@ authors:
- user: philschmid
---

<h1>Accelerate BERT inference with Hugging Face Transformers and AWS Inferentia</h1>
# Accelerate BERT inference with Hugging Face Transformers and AWS Inferentia

<!-- {blog_metadata} -->
<!-- {authors} -->

<script async defer src="https://unpkg.com/medium-zoom-element@0/dist/medium-zoom-element.min.js"></script>

Expand Down
4 changes: 1 addition & 3 deletions bertopic.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,10 +7,8 @@ authors:
- user: davanstrien
---

<h1> Introducing BERTopic Integration with the Hugging Face Hub</h1>
# Introducing BERTopic Integration with the Hugging Face Hub

<!-- {blog_metadata} -->
<!-- {authors} -->

[![Open in Colab](https://colab.research.google.com/assets/colab-badge.svg 'open in colab')](https://colab.research.google.com/#fileId=https://huggingface.co/spaces/davanstrien/blog_notebooks/blob/main/BERTopic_hub_starter.ipynb)

Expand Down
2 changes: 0 additions & 2 deletions big-bird.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,8 +7,6 @@ authors:

# Understanding BigBird's Block Sparse Attention

<!-- {blog_metadata} -->
<!-- {authors} -->

## Introduction

Expand Down
2 changes: 0 additions & 2 deletions blip-2.md
Original file line number Diff line number Diff line change
Expand Up @@ -8,8 +8,6 @@ authors:

# Zero-shot image-to-text generation with BLIP-2

<!-- {blog_metadata} -->
<!-- {authors} -->

This guide introduces [BLIP-2](https://huggingface.co/docs/transformers/main/en/model_doc/blip-2) from Salesforce Research
that enables a suite of state-of-the-art visual-language models that are now available in [🤗 Transformers](https://huggingface.co/transformers).
Expand Down
4 changes: 1 addition & 3 deletions bloom-inference-optimization.md
Original file line number Diff line number Diff line change
Expand Up @@ -5,9 +5,7 @@ authors:
- user: Narsil
---

<h1>Optimization story: Bloom inference</h1>
<!-- {blog_metadata} -->
<!-- {authors} -->
# Optimization story: Bloom inference

This article gives you the behind-the-scenes of how we made an efficient inference server that powers bloom.
inference server that powers [https://huggingface.co/bigscience/bloom]().
Expand Down
4 changes: 1 addition & 3 deletions bloom-inference-pytorch-scripts.md
Original file line number Diff line number Diff line change
Expand Up @@ -6,10 +6,8 @@ authors:
- user: sgugger
---

<h1>Incredibly Fast BLOOM Inference with DeepSpeed and Accelerate</h1>
# Incredibly Fast BLOOM Inference with DeepSpeed and Accelerate

<!-- {blog_metadata} -->
<!-- {authors} -->

This article shows how to get an incredibly fast per token throughput when generating with the 176B parameter [BLOOM model](https://huggingface.co/bigscience/bloom).

Expand Down
4 changes: 1 addition & 3 deletions bloom-megatron-deepspeed.md
Original file line number Diff line number Diff line change
Expand Up @@ -5,10 +5,8 @@ authors:
- user: stas
---

<h1>The Technology Behind BLOOM Training</h1>
# The Technology Behind BLOOM Training

<!-- {blog_metadata} -->
<!-- {authors} -->



Expand Down
4 changes: 1 addition & 3 deletions bloom.md
Original file line number Diff line number Diff line change
Expand Up @@ -6,10 +6,8 @@ authors:
---


<h1>🌸 Introducing The World's Largest Open Multilingual Language Model: BLOOM 🌸</h1>
# 🌸 Introducing The World's Largest Open Multilingual Language Model: BLOOM 🌸

<!-- {blog_metadata} -->
<!-- {authors} -->

<a href="https://huggingface.co/bigscience/bloom"><img style="middle" width="950" src="/blog/assets/86_bloom/thumbnail-2.png"></a>

Expand Down
Loading

0 comments on commit 2d74c93

Please sign in to comment.