Skip to content

Commit c988db5

Browse files
committed
Release v4.4.0
1 parent 5c02b97 commit c988db5

File tree

16 files changed

+22
-22
lines changed

16 files changed

+22
-22
lines changed

README.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -201,7 +201,7 @@ Current number of checkpoints: ![](https://img.shields.io/endpoint?url=https://h
201201
1. **[ConvBERT](https://huggingface.co/transformers/model_doc/convbert.html)** (from YituTech) released with the paper [ConvBERT: Improving BERT with Span-based Dynamic Convolution](https://arxiv.org/abs/2008.02496) by Zihang Jiang, Weihao Yu, Daquan Zhou, Yunpeng Chen, Jiashi Feng, Shuicheng Yan.
202202
1. **[CTRL](https://huggingface.co/transformers/model_doc/ctrl.html)** (from Salesforce) released with the paper [CTRL: A Conditional Transformer Language Model for Controllable Generation](https://arxiv.org/abs/1909.05858) by Nitish Shirish Keskar*, Bryan McCann*, Lav R. Varshney, Caiming Xiong and Richard Socher.
203203
1. **[DeBERTa](https://huggingface.co/transformers/model_doc/deberta.html)** (from Microsoft) released with the paper [DeBERTa: Decoding-enhanced BERT with Disentangled Attention](https://arxiv.org/abs/2006.03654) by Pengcheng He, Xiaodong Liu, Jianfeng Gao, Weizhu Chen.
204-
1. **[DeBERTa-v2](https://huggingface.co/transformers/master/model_doc/deberta_v2.html)** (from Microsoft) released with the paper [DeBERTa: Decoding-enhanced BERT with Disentangled Attention](https://arxiv.org/abs/2006.03654) by Pengcheng He, Xiaodong Liu, Jianfeng Gao, Weizhu Chen.
204+
1. **[DeBERTa-v2](https://huggingface.co/transformers/model_doc/deberta_v2.html)** (from Microsoft) released with the paper [DeBERTa: Decoding-enhanced BERT with Disentangled Attention](https://arxiv.org/abs/2006.03654) by Pengcheng He, Xiaodong Liu, Jianfeng Gao, Weizhu Chen.
205205
1. **[DialoGPT](https://huggingface.co/transformers/model_doc/dialogpt.html)** (from Microsoft Research) released with the paper [DialoGPT: Large-Scale Generative Pre-training for Conversational Response Generation](https://arxiv.org/abs/1911.00536) by Yizhe Zhang, Siqi Sun, Michel Galley, Yen-Chun Chen, Chris Brockett, Xiang Gao, Jianfeng Gao, Jingjing Liu, Bill Dolan.
206206
1. **[DistilBERT](https://huggingface.co/transformers/model_doc/distilbert.html)** (from HuggingFace), released together with the paper [DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter](https://arxiv.org/abs/1910.01108) by Victor Sanh, Lysandre Debut and Thomas Wolf. The same method has been applied to compress GPT2 into [DistilGPT2](https://github.com/huggingface/transformers/tree/master/examples/distillation), RoBERTa into [DistilRoBERTa](https://github.com/huggingface/transformers/tree/master/examples/distillation), Multilingual BERT into [DistilmBERT](https://github.com/huggingface/transformers/tree/master/examples/distillation) and a German version of DistilBERT.
207207
1. **[DPR](https://huggingface.co/transformers/model_doc/dpr.html)** (from Facebook) released with the paper [Dense Passage Retrieval
@@ -212,7 +212,7 @@ Min, Patrick Lewis, Ledell Wu, Sergey Edunov, Danqi Chen, and Wen-tau Yih.
212212
1. **[Funnel Transformer](https://huggingface.co/transformers/model_doc/funnel.html)** (from CMU/Google Brain) released with the paper [Funnel-Transformer: Filtering out Sequential Redundancy for Efficient Language Processing](https://arxiv.org/abs/2006.03236) by Zihang Dai, Guokun Lai, Yiming Yang, Quoc V. Le.
213213
1. **[GPT](https://huggingface.co/transformers/model_doc/gpt.html)** (from OpenAI) released with the paper [Improving Language Understanding by Generative Pre-Training](https://blog.openai.com/language-unsupervised/) by Alec Radford, Karthik Narasimhan, Tim Salimans and Ilya Sutskever.
214214
1. **[GPT-2](https://huggingface.co/transformers/model_doc/gpt2.html)** (from OpenAI) released with the paper [Language Models are Unsupervised Multitask Learners](https://blog.openai.com/better-language-models/) by Alec Radford*, Jeffrey Wu*, Rewon Child, David Luan, Dario Amodei** and Ilya Sutskever**.
215-
1. **[I-BERT](https://huggingface.co/transformers/master/model_doc/ibert.html)** (from Berkeley) released with the paper [I-BERT: Integer-only BERT Quantization](https://arxiv.org/abs/2101.01321) by Sehoon Kim, Amir Gholami, Zhewei Yao, Michael W. Mahoney, Kurt Keutzer
215+
1. **[I-BERT](https://huggingface.co/transformers/model_doc/ibert.html)** (from Berkeley) released with the paper [I-BERT: Integer-only BERT Quantization](https://arxiv.org/abs/2101.01321) by Sehoon Kim, Amir Gholami, Zhewei Yao, Michael W. Mahoney, Kurt Keutzer
216216
1. **[LayoutLM](https://huggingface.co/transformers/model_doc/layoutlm.html)** (from Microsoft Research Asia) released with the paper [LayoutLM: Pre-training of Text and Layout for Document Image Understanding](https://arxiv.org/abs/1912.13318) by Yiheng Xu, Minghao Li, Lei Cui, Shaohan Huang, Furu Wei, Ming Zhou.
217217
1. **[LED](https://huggingface.co/transformers/model_doc/led.html)** (from AllenAI) released with the paper [Longformer: The Long-Document Transformer](https://arxiv.org/abs/2004.05150) by Iz Beltagy, Matthew E. Peters, Arman Cohan.
218218
1. **[Longformer](https://huggingface.co/transformers/model_doc/longformer.html)** (from AllenAI) released with the paper [Longformer: The Long-Document Transformer](https://arxiv.org/abs/2004.05150) by Iz Beltagy, Matthew E. Peters, Arman Cohan.

docs/source/conf.py

+2-1
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,8 @@
2626
# The short X.Y version
2727
version = u''
2828
# The full version, including alpha/beta/rc tags
29-
release = u'4.2.0'
29+
release = u'4.4.0'
30+
3031
# Prefix link to point to master, comment this during version release and uncomment below line
3132
extlinks = {'prefix_link': ('https://github.com/huggingface/transformers/blob/master/%s', '')}
3233
# Prefix link to always point to corresponding version, uncomment this during version release

docs/source/index.rst

+5-6
Original file line numberDiff line numberDiff line change
@@ -117,9 +117,9 @@ and conversion utilities for the following models:
117117
12. :doc:`DeBERTa <model_doc/deberta>` (from Microsoft) released with the paper `DeBERTa: Decoding-enhanced BERT with
118118
Disentangled Attention <https://arxiv.org/abs/2006.03654>`__ by Pengcheng He, Xiaodong Liu, Jianfeng Gao, Weizhu
119119
Chen.
120-
13. `DeBERTa-v2 <https://huggingface.co/transformers/master/model_doc/deberta_v2.html>`__ (from Microsoft) released
121-
with the paper `DeBERTa: Decoding-enhanced BERT with Disentangled Attention <https://arxiv.org/abs/2006.03654>`__
122-
by Pengcheng He, Xiaodong Liu, Jianfeng Gao, Weizhu Chen.
120+
13. :doc:`DeBERTa-v2 <model_doc/deberta_v2>` (from Microsoft) released with the paper `DeBERTa: Decoding-enhanced BERT
121+
with Disentangled Attention <https://arxiv.org/abs/2006.03654>`__ by Pengcheng He, Xiaodong Liu, Jianfeng Gao,
122+
Weizhu Chen.
123123
14. :doc:`DialoGPT <model_doc/dialogpt>` (from Microsoft Research) released with the paper `DialoGPT: Large-Scale
124124
Generative Pre-training for Conversational Response Generation <https://arxiv.org/abs/1911.00536>`__ by Yizhe
125125
Zhang, Siqi Sun, Michel Galley, Yen-Chun Chen, Chris Brockett, Xiang Gao, Jianfeng Gao, Jingjing Liu, Bill Dolan.
@@ -148,9 +148,8 @@ and conversion utilities for the following models:
148148
21. :doc:`GPT-2 <model_doc/gpt2>` (from OpenAI) released with the paper `Language Models are Unsupervised Multitask
149149
Learners <https://blog.openai.com/better-language-models/>`__ by Alec Radford*, Jeffrey Wu*, Rewon Child, David
150150
Luan, Dario Amodei** and Ilya Sutskever**.
151-
22. `I-BERT <https://huggingface.co/transformers/master/model_doc/ibert.html>`__ (from Berkeley) released with the
152-
paper `I-BERT: Integer-only BERT Quantization <https://arxiv.org/abs/2101.01321>`__ by Sehoon Kim, Amir Gholami,
153-
Zhewei Yao, Michael W. Mahoney, Kurt Keutzer
151+
22. :doc:`I-BERT <model_doc/ibert>` (from Berkeley) released with the paper `I-BERT: Integer-only BERT Quantization
152+
<https://arxiv.org/abs/2101.01321>`__ by Sehoon Kim, Amir Gholami, Zhewei Yao, Michael W. Mahoney, Kurt Keutzer
154153
23. :doc:`LayoutLM <model_doc/layoutlm>` (from Microsoft Research Asia) released with the paper `LayoutLM: Pre-training
155154
of Text and Layout for Document Image Understanding <https://arxiv.org/abs/1912.13318>`__ by Yiheng Xu, Minghao Li,
156155
Lei Cui, Shaohan Huang, Furu Wei, Ming Zhou.

examples/language-modeling/run_clm.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -48,7 +48,7 @@
4848

4949

5050
# Will error if the minimal version of Transformers is not installed. Remove at your own risks.
51-
check_min_version("4.4.0.dev0")
51+
check_min_version("4.4.0")
5252

5353
logger = logging.getLogger(__name__)
5454

examples/language-modeling/run_mlm.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -48,7 +48,7 @@
4848

4949

5050
# Will error if the minimal version of Transformers is not installed. Remove at your own risks.
51-
check_min_version("4.4.0.dev0")
51+
check_min_version("4.4.0")
5252

5353
logger = logging.getLogger(__name__)
5454
MODEL_CONFIG_CLASSES = list(MODEL_FOR_MASKED_LM_MAPPING.keys())

examples/language-modeling/run_plm.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -44,7 +44,7 @@
4444

4545

4646
# Will error if the minimal version of Transformers is not installed. Remove at your own risks.
47-
check_min_version("4.4.0.dev0")
47+
check_min_version("4.4.0")
4848

4949
logger = logging.getLogger(__name__)
5050

examples/multiple-choice/run_swag.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -46,7 +46,7 @@
4646

4747

4848
# Will error if the minimal version of Transformers is not installed. Remove at your own risks.
49-
check_min_version("4.4.0.dev0")
49+
check_min_version("4.4.0")
5050

5151
logger = logging.getLogger(__name__)
5252

examples/question-answering/run_qa.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -46,7 +46,7 @@
4646

4747

4848
# Will error if the minimal version of Transformers is not installed. Remove at your own risks.
49-
check_min_version("4.4.0.dev0")
49+
check_min_version("4.4.0")
5050

5151
logger = logging.getLogger(__name__)
5252

examples/question-answering/run_qa_beam_search.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -45,7 +45,7 @@
4545

4646

4747
# Will error if the minimal version of Transformers is not installed. Remove at your own risks.
48-
check_min_version("4.4.0.dev0")
48+
check_min_version("4.4.0")
4949

5050
logger = logging.getLogger(__name__)
5151

examples/seq2seq/run_summarization.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -47,7 +47,7 @@
4747

4848

4949
# Will error if the minimal version of Transformers is not installed. Remove at your own risks.
50-
check_min_version("4.4.0.dev0")
50+
check_min_version("4.4.0")
5151

5252
logger = logging.getLogger(__name__)
5353

examples/seq2seq/run_translation.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -46,7 +46,7 @@
4646

4747

4848
# Will error if the minimal version of Transformers is not installed. Remove at your own risks.
49-
check_min_version("4.4.0.dev0")
49+
check_min_version("4.4.0")
5050

5151
logger = logging.getLogger(__name__)
5252

examples/text-classification/run_glue.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -45,7 +45,7 @@
4545

4646

4747
# Will error if the minimal version of Transformers is not installed. Remove at your own risks.
48-
check_min_version("4.4.0.dev0")
48+
check_min_version("4.4.0")
4949

5050
task_to_keys = {
5151
"cola": ("sentence", None),

examples/text-classification/run_xnli.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -45,7 +45,7 @@
4545

4646

4747
# Will error if the minimal version of Transformers is not installed. Remove at your own risks.
48-
check_min_version("4.4.0.dev0")
48+
check_min_version("4.4.0")
4949

5050
logger = logging.getLogger(__name__)
5151

examples/token-classification/run_ner.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -45,7 +45,7 @@
4545

4646

4747
# Will error if the minimal version of Transformers is not installed. Remove at your own risks.
48-
check_min_version("4.4.0.dev0")
48+
check_min_version("4.4.0")
4949

5050
logger = logging.getLogger(__name__)
5151

setup.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -278,7 +278,7 @@ def run(self):
278278

279279
setup(
280280
name="transformers",
281-
version="4.4.0.dev0", # expected format is one of x.y.z.dev0, or x.y.z.rc1 or x.y.z (no to dashes, yes to dots)
281+
version="4.4.0", # expected format is one of x.y.z.dev0, or x.y.z.rc1 or x.y.z (no to dashes, yes to dots)
282282
author="Thomas Wolf, Lysandre Debut, Victor Sanh, Julien Chaumond, Sam Shleifer, Patrick von Platen, Sylvain Gugger, Google AI Language Team Authors, Open AI team Authors, Facebook AI Authors, Carnegie Mellon University Authors",
283283
author_email="thomas@huggingface.co",
284284
description="State-of-the-art Natural Language Processing for TensorFlow 2.0 and PyTorch",

src/transformers/__init__.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@
2222
# to defer the actual importing for when the objects are requested. This way `import transformers` provides the names
2323
# in the namespace without actually importing anything (and especially none of the backends).
2424

25-
__version__ = "4.4.0.dev0"
25+
__version__ = "4.4.0"
2626

2727
# Work around to update TensorFlow's absl.logging threshold which alters the
2828
# default Python logging output behavior when present.

0 commit comments

Comments
 (0)