Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Time-Series] Autoformer model #21891

Merged
merged 134 commits into from
May 30, 2023
Merged
Show file tree
Hide file tree
Changes from 84 commits
Commits
Show all changes
134 commits
Select commit Hold shift + click to select a range
00bb9a5
ran `transformers-cli add-new-model-like`
elisim Mar 1, 2023
33de4c3
Merge remote-tracking branch 'upstream/main' into add-autoformer2021-…
elisim Mar 2, 2023
d3465c8
added `AutoformerLayernorm` and `AutoformerSeriesDecomposition`
elisim Mar 2, 2023
10f18a8
added `decomposition_layer` in `init` and `moving_avg` to config
elisim Mar 2, 2023
d091bab
added `AutoformerAutoCorrelation` to encoder & decoder
elisim Mar 2, 2023
30ff575
removed caninical self attention `AutoformerAttention`
elisim Mar 2, 2023
36cc0eb
added arguments in config and model tester. Init works! 😁
elisim Mar 2, 2023
1439803
Merge remote-tracking branch 'upstream/main' into add-autoformer2021-…
elisim Mar 3, 2023
53bd981
WIP autoformer attention with autocorrlation
elisim Mar 3, 2023
b944a99
fixed `attn_weights` size
elisim Mar 3, 2023
a24b9c8
wip time_delay_agg_training
elisim Mar 3, 2023
59d0509
Merge remote-tracking branch 'upstream/main' into add-autoformer2021-…
elisim Mar 4, 2023
d43e488
fixing sizes and debug time_delay_agg_training
elisim Mar 4, 2023
a2cab3e
aggregation in training works! 😁
elisim Mar 4, 2023
5864a04
`top_k_delays` -> `top_k_delays_index` and added `contiguous()`
elisim Mar 4, 2023
f4b9751
wip time_delay_agg_inference
elisim Mar 4, 2023
56e37e2
finish time_delay_agg_inference 😎
elisim Mar 5, 2023
e377a6a
added resize to autocorrelation
elisim Mar 5, 2023
e653c1b
Merge remote-tracking branch 'upstream/main' into add-autoformer2021-…
elisim Mar 6, 2023
b15fa09
bug fix: added the length of the output signal to `irfft`
elisim Mar 6, 2023
d300534
`attention_mask = None` in the decoder
elisim Mar 6, 2023
d5e0409
fixed test: changed attention expected size, `test_attention_outputs`…
elisim Mar 6, 2023
6c361f3
removed unnecessary code
elisim Mar 6, 2023
4eb7b19
apply AutoformerLayernorm in final norm in enc & dec
elisim Mar 6, 2023
34344ea
added series decomposition to the encoder
elisim Mar 6, 2023
157313c
Merge remote-tracking branch 'upstream/main' into add-autoformer2021-…
elisim Mar 7, 2023
468f2ba
added series decomp to decoder, with inputs
elisim Mar 7, 2023
e3301e6
added trend todos
elisim Mar 7, 2023
4d6575d
added autoformer to README
elisim Mar 7, 2023
a373aa2
added to index
elisim Mar 7, 2023
0dda9ff
added autoformer.mdx
elisim Mar 7, 2023
20b4b2f
remove scaling and init attention_mask in the decoder
elisim Mar 7, 2023
59c4818
make style
elisim Mar 7, 2023
93c2bcb
fix copies
elisim Mar 7, 2023
72c615a
make fix-copies
elisim Mar 7, 2023
0156019
Merge remote-tracking branch 'upstream/main' into add-autoformer2021-…
kashif Mar 7, 2023
7658856
inital fix-copies
kashif Mar 7, 2023
187f982
Merge remote-tracking branch 'upstream/main' into add-autoformer2021-…
elisim Mar 10, 2023
3a87689
Merge branch 'add-autoformer2021-model' of https://github.com/elisim/…
elisim Mar 10, 2023
980c266
fix from https://github.com/huggingface/transformers/pull/22076
elisim Mar 10, 2023
a8f0364
Merge remote-tracking branch 'upstream/main' into add-autoformer2021-…
kashif Mar 14, 2023
648d98b
make style
kashif Mar 14, 2023
156f1cf
fix class names
kashif Mar 14, 2023
0f55bd3
Merge remote-tracking branch 'elisim/main' into add-autoformer2021-model
kashif Mar 21, 2023
b401118
added trend
kashif Mar 21, 2023
b509304
added d_model and projection layers
kashif Mar 21, 2023
e61dac5
added `trend_projection` source, and decomp layer init
elisim Apr 3, 2023
66ae325
Merge branch 'main' into add-autoformer2021-model
elisim Apr 3, 2023
bbb722c
added trend & seasonal init for decoder input
elisim Apr 3, 2023
35ab45d
AutoformerModel cannot be copied as it has the decomp layer too
kashif Apr 4, 2023
9488b63
encoder can be copied from time series transformer
kashif Apr 4, 2023
1d297d8
fixed generation and made distrb. out more robust
kashif Apr 9, 2023
c263e04
use context window to calculate decomposition
kashif Apr 10, 2023
f34750b
use the context_window for decomposition
kashif Apr 11, 2023
4a7981f
use output_params helper
kashif Apr 17, 2023
41fb792
clean up AutoformerAttention
kashif Apr 17, 2023
6250db0
Merge branch 'huggingface:main' into add-autoformer2021-model
kashif Apr 17, 2023
2165b62
subsequences_length off by 1
kashif Apr 18, 2023
5f405ea
make fix copies
kashif Apr 18, 2023
68c5015
fix test
kashif Apr 19, 2023
af1452c
added init for nn.Conv1d
elisim Apr 20, 2023
3c07329
fix IGNORE_NON_TESTED
kashif Apr 21, 2023
b9d1903
added model_doc
kashif Apr 21, 2023
c9566c5
fix ruff
kashif Apr 21, 2023
3616eed
ignore tests
kashif Apr 21, 2023
6a4c437
remove dup
kashif Apr 21, 2023
0f2960b
fix SPECIAL_CASES_TO_ALLOW
kashif Apr 21, 2023
a4c1396
do not copy due to conv1d weight init
kashif Apr 21, 2023
740f889
remove unused imports
kashif Apr 21, 2023
6eaaa57
added short summary
kashif Apr 21, 2023
31103bc
added label_length and made the model non-autoregressive
kashif May 4, 2023
2752fba
added params docs
elisim May 5, 2023
b9f0b20
better doc for `factor`
elisim May 5, 2023
843c616
fix tests
kashif May 5, 2023
301f2aa
Merge remote-tracking branch 'upstream/main' into add-autoformer2021-…
elisim May 6, 2023
cf1c88a
renamed `moving_avg` to `moving_average`
elisim May 6, 2023
3e03295
renamed `factor` to `autocorrelation_factor`
elisim May 6, 2023
d00a541
make style
elisim May 6, 2023
daf151b
Update src/transformers/models/autoformer/configuration_autoformer.py
elisim May 6, 2023
3becbd7
Update src/transformers/models/autoformer/configuration_autoformer.py
elisim May 6, 2023
fa7e042
fix configurations
kashif May 8, 2023
54dde3e
fix integration tests
kashif May 8, 2023
a40a02a
Merge remote-tracking branch 'upstream/main' into add-autoformer2021-…
elisim May 9, 2023
c8d298d
Merge remote-tracking branch 'upstream/main' into add-autoformer2021-…
elisim May 9, 2023
2c64f15
Merge remote-tracking branch 'upstream/main' into add-autoformer2021-…
elisim May 15, 2023
3322728
Update src/transformers/models/autoformer/configuration_autoformer.py
kashif May 16, 2023
21e3491
fixing `lags_sequence` doc
elisim May 21, 2023
6971fdc
Revert "fixing `lags_sequence` doc"
elisim May 21, 2023
34a3a87
Update src/transformers/models/autoformer/modeling_autoformer.py
elisim May 21, 2023
4d09a83
Update src/transformers/models/autoformer/modeling_autoformer.py
elisim May 21, 2023
a3adaf8
Update src/transformers/models/autoformer/modeling_autoformer.py
elisim May 21, 2023
827841b
Apply suggestions from code review
elisim May 21, 2023
5163caa
Update src/transformers/models/autoformer/configuration_autoformer.py
elisim May 21, 2023
6bfe3d2
model layers now take the config
elisim May 21, 2023
a93498c
added `layer_norm_eps` to the config
elisim May 21, 2023
a1fffed
Update src/transformers/models/autoformer/modeling_autoformer.py
elisim May 21, 2023
f162198
added `config.layer_norm_eps` to AutoformerLayernorm
elisim May 21, 2023
b13b6a2
added `config.layer_norm_eps` to all layernorm layers
elisim May 21, 2023
0b88155
Update src/transformers/models/autoformer/configuration_autoformer.py
elisim May 21, 2023
baca5cf
Update src/transformers/models/autoformer/configuration_autoformer.py
elisim May 21, 2023
c9d909f
Update src/transformers/models/autoformer/configuration_autoformer.py
elisim May 21, 2023
678429d
Update src/transformers/models/autoformer/configuration_autoformer.py
elisim May 21, 2023
b6591fa
fix variable names
elisim May 21, 2023
60c75d0
added inital pretrained model
kashif May 22, 2023
0aecf6b
added use_cache docstring
kashif May 22, 2023
c42a1c0
doc strings for trend and use_cache
kashif May 22, 2023
112343f
fix order of args
kashif May 22, 2023
bd9e13b
imports on one line
kashif May 22, 2023
4a14739
fixed get_lagged_subsequences docs
kashif May 22, 2023
75b185f
add docstring for create_network_inputs
kashif May 22, 2023
252e128
get rid of layer_norm_eps config
kashif May 22, 2023
76044e4
add back layernorm
kashif May 22, 2023
30f553b
update fixture location
kashif May 22, 2023
0e1cc40
fix signature
kashif May 22, 2023
24b1e62
use AutoformerModelOutput dataclass
kashif May 22, 2023
091b0c0
fix pretrain config
kashif May 23, 2023
821ec9f
no need as default exists
kashif May 23, 2023
01a6eef
subclass ModelOutput
kashif May 24, 2023
fd89f11
remove layer_norm_eps config
kashif May 24, 2023
ff92998
fix test_model_outputs_equivalence test
kashif May 25, 2023
f891be6
test hidden_states_output
kashif May 25, 2023
39955a7
make fix-copies
kashif May 25, 2023
cf97b82
Update src/transformers/models/autoformer/configuration_autoformer.py
elisim May 29, 2023
9bf042f
removed unused attr
kashif May 29, 2023
dca7474
Update tests/models/autoformer/test_modeling_autoformer.py
kashif May 29, 2023
e754864
Update src/transformers/models/autoformer/modeling_autoformer.py
kashif May 29, 2023
e3e3c3a
Update src/transformers/models/autoformer/modeling_autoformer.py
kashif May 29, 2023
00e4b80
Update src/transformers/models/autoformer/modeling_autoformer.py
kashif May 29, 2023
d057569
Update src/transformers/models/autoformer/modeling_autoformer.py
kashif May 29, 2023
76bb282
Update src/transformers/models/autoformer/modeling_autoformer.py
kashif May 29, 2023
8344329
Update src/transformers/models/autoformer/modeling_autoformer.py
kashif May 29, 2023
d6e247e
use AutoFormerDecoderOutput
kashif May 29, 2023
bb75c44
fix formatting
kashif May 29, 2023
555b6dd
fix formatting
kashif May 29, 2023
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Jump to
Jump to file
Failed to load files.
Diff view
Diff view
1 change: 1 addition & 0 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -279,6 +279,7 @@ Current number of checkpoints: ![](https://img.shields.io/endpoint?url=https://h
1. **[ALIGN](https://huggingface.co/docs/transformers/model_doc/align)** (from Google Research) released with the paper [Scaling Up Visual and Vision-Language Representation Learning With Noisy Text Supervision](https://arxiv.org/abs/2102.05918) by Chao Jia, Yinfei Yang, Ye Xia, Yi-Ting Chen, Zarana Parekh, Hieu Pham, Quoc V. Le, Yunhsuan Sung, Zhen Li, Tom Duerig.
1. **[AltCLIP](https://huggingface.co/docs/transformers/model_doc/altclip)** (from BAAI) released with the paper [AltCLIP: Altering the Language Encoder in CLIP for Extended Language Capabilities](https://arxiv.org/abs/2211.06679) by Chen, Zhongzhi and Liu, Guang and Zhang, Bo-Wen and Ye, Fulong and Yang, Qinghong and Wu, Ledell.
1. **[Audio Spectrogram Transformer](https://huggingface.co/docs/transformers/model_doc/audio-spectrogram-transformer)** (from MIT) released with the paper [AST: Audio Spectrogram Transformer](https://arxiv.org/abs/2104.01778) by Yuan Gong, Yu-An Chung, James Glass.
1. **[Autoformer](https://huggingface.co/docs/transformers/main/model_doc/autoformer)** (from Tsinghua University) released with the paper [Autoformer: Decomposition Transformers with Auto-Correlation for Long-Term Series Forecasting](https://arxiv.org/abs/2106.13008) by Haixu Wu, Jiehui Xu, Jianmin Wang, Mingsheng Long.
1. **[BART](https://huggingface.co/docs/transformers/model_doc/bart)** (from Facebook) released with the paper [BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension](https://arxiv.org/abs/1910.13461) by Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer Levy, Ves Stoyanov and Luke Zettlemoyer.
1. **[BARThez](https://huggingface.co/docs/transformers/model_doc/barthez)** (from École polytechnique) released with the paper [BARThez: a Skilled Pretrained French Sequence-to-Sequence Model](https://arxiv.org/abs/2010.12321) by Moussa Kamal Eddine, Antoine J.-P. Tixier, Michalis Vazirgiannis.
1. **[BARTpho](https://huggingface.co/docs/transformers/model_doc/bartpho)** (from VinAI Research) released with the paper [BARTpho: Pre-trained Sequence-to-Sequence Models for Vietnamese](https://arxiv.org/abs/2109.09701) by Nguyen Luong Tran, Duong Minh Le and Dat Quoc Nguyen.
Expand Down
1 change: 1 addition & 0 deletions README_es.md
Original file line number Diff line number Diff line change
Expand Up @@ -267,6 +267,7 @@ Número actual de puntos de control: ![](https://img.shields.io/endpoint?url=htt
1. **[ALIGN](https://huggingface.co/docs/transformers/model_doc/align)** (from Google Research) released with the paper [Scaling Up Visual and Vision-Language Representation Learning With Noisy Text Supervision](https://arxiv.org/abs/2102.05918) by Chao Jia, Yinfei Yang, Ye Xia, Yi-Ting Chen, Zarana Parekh, Hieu Pham, Quoc V. Le, Yunhsuan Sung, Zhen Li, Tom Duerig.
1. **[AltCLIP](https://huggingface.co/docs/transformers/model_doc/altclip)** (from BAAI) released with the paper [AltCLIP: Altering the Language Encoder in CLIP for Extended Language Capabilities](https://arxiv.org/abs/2211.06679) by Chen, Zhongzhi and Liu, Guang and Zhang, Bo-Wen and Ye, Fulong and Yang, Qinghong and Wu, Ledell.
1. **[Audio Spectrogram Transformer](https://huggingface.co/docs/transformers/model_doc/audio-spectrogram-transformer)** (from MIT) released with the paper [AST: Audio Spectrogram Transformer](https://arxiv.org/abs/2104.01778) by Yuan Gong, Yu-An Chung, James Glass.
1. **[Autoformer](https://huggingface.co/docs/transformers/main/model_doc/autoformer)** (from Tsinghua University) released with the paper [Autoformer: Decomposition Transformers with Auto-Correlation for Long-Term Series Forecasting](https://arxiv.org/abs/2106.13008) by Haixu Wu, Jiehui Xu, Jianmin Wang, Mingsheng Long.
1. **[BART](https://huggingface.co/docs/transformers/model_doc/bart)** (from Facebook) released with the paper [BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension](https://arxiv.org/abs/1910.13461) by Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer Levy, Ves Stoyanov and Luke Zettlemoyer.
1. **[BARThez](https://huggingface.co/docs/transformers/model_doc/barthez)** (from École polytechnique) released with the paper [BARThez: a Skilled Pretrained French Sequence-to-Sequence Model](https://arxiv.org/abs/2010.12321) by Moussa Kamal Eddine, Antoine J.-P. Tixier, Michalis Vazirgiannis.
1. **[BARTpho](https://huggingface.co/docs/transformers/model_doc/bartpho)** (from VinAI Research) released with the paper [BARTpho: Pre-trained Sequence-to-Sequence Models for Vietnamese](https://arxiv.org/abs/2109.09701) by Nguyen Luong Tran, Duong Minh Le and Dat Quoc Nguyen.
Expand Down
1 change: 1 addition & 0 deletions README_hd.md
Original file line number Diff line number Diff line change
Expand Up @@ -239,6 +239,7 @@ conda install -c huggingface transformers
1. **[ALIGN](https://huggingface.co/docs/transformers/model_doc/align)** (Google Research से) Chao Jia, Yinfei Yang, Ye Xia, Yi-Ting Chen, Zarana Parekh, Hieu Pham, Quoc V. Le, Yunhsuan Sung, Zhen Li, Tom Duerig. द्वाराअनुसंधान पत्र [Scaling Up Visual and Vision-Language Representation Learning With Noisy Text Supervision](https://arxiv.org/abs/2102.05918) के साथ जारी किया गया
1. **[AltCLIP](https://huggingface.co/docs/transformers/model_doc/altclip)** (from BAAI) released with the paper [AltCLIP: Altering the Language Encoder in CLIP for Extended Language Capabilities](https://arxiv.org/abs/2211.06679) by Chen, Zhongzhi and Liu, Guang and Zhang, Bo-Wen and Ye, Fulong and Yang, Qinghong and Wu, Ledell.
1. **[Audio Spectrogram Transformer](https://huggingface.co/docs/transformers/model_doc/audio-spectrogram-transformer)** (from MIT) released with the paper [AST: Audio Spectrogram Transformer](https://arxiv.org/abs/2104.01778) by Yuan Gong, Yu-An Chung, James Glass.
1. **[Autoformer](https://huggingface.co/docs/transformers/main/model_doc/autoformer)** (from Tsinghua University) released with the paper [Autoformer: Decomposition Transformers with Auto-Correlation for Long-Term Series Forecasting](https://arxiv.org/abs/2106.13008) by Haixu Wu, Jiehui Xu, Jianmin Wang, Mingsheng Long.
1. **[BART](https://huggingface.co/docs/transformers/model_doc/bart)** (फेसबुक) साथ थीसिस [बार्ट: प्राकृतिक भाषा निर्माण, अनुवाद के लिए अनुक्रम-से-अनुक्रम पूर्व प्रशिक्षण , और समझ] (https://arxiv.org/pdf/1910.13461.pdf) पर निर्भर माइक लुईस, यिनहान लियू, नमन गोयल, मार्जन ग़ज़विनिनेजाद, अब्देलरहमान मोहम्मद, ओमर लेवी, वेस स्टोयानोव और ल्यूक ज़ेटलमॉयर
1. **[BARThez](https://huggingface.co/docs/transformers/model_doc/barthez)** (से École polytechnique) साथ थीसिस [BARThez: a Skilled Pretrained French Sequence-to-Sequence Model](https://arxiv.org/abs/2010.12321) पर निर्भर Moussa Kamal Eddine, Antoine J.-P. Tixier, Michalis Vazirgiannis रिहाई।
1. **[BARTpho](https://huggingface.co/docs/transformers/model_doc/bartpho)** (VinAI Research से) साथ में पेपर [BARTpho: Pre-trained Sequence-to-Sequence Models for Vietnamese](https://arxiv.org/abs/2109.09701)गुयेन लुओंग ट्रान, डुओंग मिन्ह ले और डाट क्वोक गुयेन द्वारा पोस्ट किया गया।
Expand Down
1 change: 1 addition & 0 deletions README_ja.md
Original file line number Diff line number Diff line change
Expand Up @@ -301,6 +301,7 @@ Flax、PyTorch、TensorFlowをcondaでインストールする方法は、それ
1. **[ALIGN](https://huggingface.co/docs/transformers/model_doc/align)** (Google Research から) Chao Jia, Yinfei Yang, Ye Xia, Yi-Ting Chen, Zarana Parekh, Hieu Pham, Quoc V. Le, Yunhsuan Sung, Zhen Li, Tom Duerig. から公開された研究論文 [Scaling Up Visual and Vision-Language Representation Learning With Noisy Text Supervision](https://arxiv.org/abs/2102.05918)
1. **[AltCLIP](https://huggingface.co/docs/transformers/model_doc/altclip)** (BAAI から) Chen, Zhongzhi and Liu, Guang and Zhang, Bo-Wen and Ye, Fulong and Yang, Qinghong and Wu, Ledell から公開された研究論文: [AltCLIP: Altering the Language Encoder in CLIP for Extended Language Capabilities](https://arxiv.org/abs/2211.06679)
1. **[Audio Spectrogram Transformer](https://huggingface.co/docs/transformers/model_doc/audio-spectrogram-transformer)** (MIT から) Yuan Gong, Yu-An Chung, James Glass から公開された研究論文: [AST: Audio Spectrogram Transformer](https://arxiv.org/abs/2104.01778)
1. **[Autoformer](https://huggingface.co/docs/transformers/main/model_doc/autoformer)** (from Tsinghua University) released with the paper [Autoformer: Decomposition Transformers with Auto-Correlation for Long-Term Series Forecasting](https://arxiv.org/abs/2106.13008) by Haixu Wu, Jiehui Xu, Jianmin Wang, Mingsheng Long.
1. **[BART](https://huggingface.co/docs/transformers/model_doc/bart)** (Facebook から) Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer Levy, Ves Stoyanov and Luke Zettlemoyer から公開された研究論文: [BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension](https://arxiv.org/abs/1910.13461)
1. **[BARThez](https://huggingface.co/docs/transformers/model_doc/barthez)** (École polytechnique から) Moussa Kamal Eddine, Antoine J.-P. Tixier, Michalis Vazirgiannis から公開された研究論文: [BARThez: a Skilled Pretrained French Sequence-to-Sequence Model](https://arxiv.org/abs/2010.12321)
1. **[BARTpho](https://huggingface.co/docs/transformers/model_doc/bartpho)** (VinAI Research から) Nguyen Luong Tran, Duong Minh Le and Dat Quoc Nguyen から公開された研究論文: [BARTpho: Pre-trained Sequence-to-Sequence Models for Vietnamese](https://arxiv.org/abs/2109.09701)
Expand Down
1 change: 1 addition & 0 deletions README_ko.md
Original file line number Diff line number Diff line change
Expand Up @@ -216,6 +216,7 @@ Flax, PyTorch, TensorFlow 설치 페이지에서 이들을 conda로 설치하는
1. **[ALIGN](https://huggingface.co/docs/transformers/model_doc/align)** (Google Research 에서 제공)은 Chao Jia, Yinfei Yang, Ye Xia, Yi-Ting Chen, Zarana Parekh, Hieu Pham, Quoc V. Le, Yunhsuan Sung, Zhen Li, Tom Duerig.의 [Scaling Up Visual and Vision-Language Representation Learning With Noisy Text Supervision](https://arxiv.org/abs/2102.05918)논문과 함께 발표했습니다.
1. **[AltCLIP](https://huggingface.co/docs/transformers/model_doc/altclip)** (from BAAI) released with the paper [AltCLIP: Altering the Language Encoder in CLIP for Extended Language Capabilities](https://arxiv.org/abs/2211.06679) by Chen, Zhongzhi and Liu, Guang and Zhang, Bo-Wen and Ye, Fulong and Yang, Qinghong and Wu, Ledell.
1. **[Audio Spectrogram Transformer](https://huggingface.co/docs/transformers/model_doc/audio-spectrogram-transformer)** (from MIT) released with the paper [AST: Audio Spectrogram Transformer](https://arxiv.org/abs/2104.01778) by Yuan Gong, Yu-An Chung, James Glass.
1. **[Autoformer](https://huggingface.co/docs/transformers/main/model_doc/autoformer)** (from Tsinghua University) released with the paper [Autoformer: Decomposition Transformers with Auto-Correlation for Long-Term Series Forecasting](https://arxiv.org/abs/2106.13008) by Haixu Wu, Jiehui Xu, Jianmin Wang, Mingsheng Long.
1. **[BART](https://huggingface.co/docs/transformers/model_doc/bart)** (from Facebook) released with the paper [BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension](https://arxiv.org/pdf/1910.13461.pdf) by Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer Levy, Ves Stoyanov and Luke Zettlemoyer.
1. **[BARThez](https://huggingface.co/docs/transformers/model_doc/barthez)** (from École polytechnique) released with the paper [BARThez: a Skilled Pretrained French Sequence-to-Sequence Model](https://arxiv.org/abs/2010.12321) by Moussa Kamal Eddine, Antoine J.-P. Tixier, Michalis Vazirgiannis.
1. **[BARTpho](https://huggingface.co/docs/transformers/model_doc/bartpho)** (from VinAI Research) released with the paper [BARTpho: Pre-trained Sequence-to-Sequence Models for Vietnamese](https://arxiv.org/abs/2109.09701) by Nguyen Luong Tran, Duong Minh Le and Dat Quoc Nguyen.
Expand Down
1 change: 1 addition & 0 deletions README_zh-hans.md
Original file line number Diff line number Diff line change
Expand Up @@ -240,6 +240,7 @@ conda install -c huggingface transformers
1. **[ALIGN](https://huggingface.co/docs/transformers/model_doc/align)** (来自 Google Research) 伴随论文 [Scaling Up Visual and Vision-Language Representation Learning With Noisy Text Supervision](https://arxiv.org/abs/2102.05918) 由 Chao Jia, Yinfei Yang, Ye Xia, Yi-Ting Chen, Zarana Parekh, Hieu Pham, Quoc V. Le, Yunhsuan Sung, Zhen Li, Tom Duerig 发布。
1. **[AltCLIP](https://huggingface.co/docs/transformers/model_doc/altclip)** (来自 BAAI) 伴随论文 [AltCLIP: Altering the Language Encoder in CLIP for Extended Language Capabilities](https://arxiv.org/abs/2211.06679) 由 Chen, Zhongzhi and Liu, Guang and Zhang, Bo-Wen and Ye, Fulong and Yang, Qinghong and Wu, Ledell 发布。
1. **[Audio Spectrogram Transformer](https://huggingface.co/docs/transformers/model_doc/audio-spectrogram-transformer)** (来自 MIT) 伴随论文 [AST: Audio Spectrogram Transformer](https://arxiv.org/abs/2104.01778) 由 Yuan Gong, Yu-An Chung, James Glass 发布。
1. **[Autoformer](https://huggingface.co/docs/transformers/main/model_doc/autoformer)** (from Tsinghua University) released with the paper [Autoformer: Decomposition Transformers with Auto-Correlation for Long-Term Series Forecasting](https://arxiv.org/abs/2106.13008) by Haixu Wu, Jiehui Xu, Jianmin Wang, Mingsheng Long.
1. **[BART](https://huggingface.co/docs/transformers/model_doc/bart)** (来自 Facebook) 伴随论文 [BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension](https://arxiv.org/pdf/1910.13461.pdf) 由 Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer Levy, Ves Stoyanov and Luke Zettlemoyer 发布。
1. **[BARThez](https://huggingface.co/docs/transformers/model_doc/barthez)** (来自 École polytechnique) 伴随论文 [BARThez: a Skilled Pretrained French Sequence-to-Sequence Model](https://arxiv.org/abs/2010.12321) 由 Moussa Kamal Eddine, Antoine J.-P. Tixier, Michalis Vazirgiannis 发布。
1. **[BARTpho](https://huggingface.co/docs/transformers/model_doc/bartpho)** (来自 VinAI Research) 伴随论文 [BARTpho: Pre-trained Sequence-to-Sequence Models for Vietnamese](https://arxiv.org/abs/2109.09701) 由 Nguyen Luong Tran, Duong Minh Le and Dat Quoc Nguyen 发布。
Expand Down
1 change: 1 addition & 0 deletions README_zh-hant.md
Original file line number Diff line number Diff line change
Expand Up @@ -252,6 +252,7 @@ conda install -c huggingface transformers
1. **[ALIGN](https://huggingface.co/docs/transformers/model_doc/align)** (from Google Research) released with the paper [Scaling Up Visual and Vision-Language Representation Learning With Noisy Text Supervision](https://arxiv.org/abs/2102.05918) by Chao Jia, Yinfei Yang, Ye Xia, Yi-Ting Chen, Zarana Parekh, Hieu Pham, Quoc V. Le, Yunhsuan Sung, Zhen Li, Tom Duerig.
1. **[AltCLIP](https://huggingface.co/docs/transformers/model_doc/altclip)** (from BAAI) released with the paper [AltCLIP: Altering the Language Encoder in CLIP for Extended Language Capabilities](https://arxiv.org/abs/2211.06679) by Chen, Zhongzhi and Liu, Guang and Zhang, Bo-Wen and Ye, Fulong and Yang, Qinghong and Wu, Ledell.
1. **[Audio Spectrogram Transformer](https://huggingface.co/docs/transformers/model_doc/audio-spectrogram-transformer)** (from MIT) released with the paper [AST: Audio Spectrogram Transformer](https://arxiv.org/abs/2104.01778) by Yuan Gong, Yu-An Chung, James Glass.
1. **[Autoformer](https://huggingface.co/docs/transformers/main/model_doc/autoformer)** (from Tsinghua University) released with the paper [Autoformer: Decomposition Transformers with Auto-Correlation for Long-Term Series Forecasting](https://arxiv.org/abs/2106.13008) by Haixu Wu, Jiehui Xu, Jianmin Wang, Mingsheng Long.
1. **[BART](https://huggingface.co/docs/transformers/model_doc/bart)** (from Facebook) released with the paper [BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension](https://arxiv.org/pdf/1910.13461.pdf) by Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer Levy, Ves Stoyanov and Luke Zettlemoyer.
1. **[BARThez](https://huggingface.co/docs/transformers/model_doc/barthez)** (from École polytechnique) released with the paper [BARThez: a Skilled Pretrained French Sequence-to-Sequence Model](https://arxiv.org/abs/2010.12321) by Moussa Kamal Eddine, Antoine J.-P. Tixier, Michalis Vazirgiannis.
1. **[BARTpho](https://huggingface.co/docs/transformers/model_doc/bartpho)** (from VinAI Research) released with the paper [BARTpho: Pre-trained Sequence-to-Sequence Models for Vietnamese](https://arxiv.org/abs/2109.09701) by Nguyen Luong Tran, Duong Minh Le and Dat Quoc Nguyen.
Expand Down
2 changes: 2 additions & 0 deletions docs/source/en/_toctree.yml
Original file line number Diff line number Diff line change
Expand Up @@ -644,6 +644,8 @@
title: Reinforcement learning models
- isExpanded: false
sections:
- local: model_doc/autoformer
title: Autoformer
- local: model_doc/informer
title: Informer
- local: model_doc/time_series_transformer
Expand Down
2 changes: 2 additions & 0 deletions docs/source/en/index.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -53,6 +53,7 @@ The documentation is organized into five sections:
1. **[ALIGN](model_doc/align)** (from Google Research) released with the paper [Scaling Up Visual and Vision-Language Representation Learning With Noisy Text Supervision](https://arxiv.org/abs/2102.05918) by Chao Jia, Yinfei Yang, Ye Xia, Yi-Ting Chen, Zarana Parekh, Hieu Pham, Quoc V. Le, Yunhsuan Sung, Zhen Li, Tom Duerig.
1. **[AltCLIP](model_doc/altclip)** (from BAAI) released with the paper [AltCLIP: Altering the Language Encoder in CLIP for Extended Language Capabilities](https://arxiv.org/abs/2211.06679) by Chen, Zhongzhi and Liu, Guang and Zhang, Bo-Wen and Ye, Fulong and Yang, Qinghong and Wu, Ledell.
1. **[Audio Spectrogram Transformer](model_doc/audio-spectrogram-transformer)** (from MIT) released with the paper [AST: Audio Spectrogram Transformer](https://arxiv.org/abs/2104.01778) by Yuan Gong, Yu-An Chung, James Glass.
1. **[Autoformer](model_doc/autoformer)** (from Tsinghua University) released with the paper [Autoformer: Decomposition Transformers with Auto-Correlation for Long-Term Series Forecasting](https://arxiv.org/abs/2106.13008) by Haixu Wu, Jiehui Xu, Jianmin Wang, Mingsheng Long.
1. **[BART](model_doc/bart)** (from Facebook) released with the paper [BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension](https://arxiv.org/abs/1910.13461) by Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer Levy, Ves Stoyanov and Luke Zettlemoyer.
1. **[BARThez](model_doc/barthez)** (from École polytechnique) released with the paper [BARThez: a Skilled Pretrained French Sequence-to-Sequence Model](https://arxiv.org/abs/2010.12321) by Moussa Kamal Eddine, Antoine J.-P. Tixier, Michalis Vazirgiannis.
1. **[BARTpho](model_doc/bartpho)** (from VinAI Research) released with the paper [BARTpho: Pre-trained Sequence-to-Sequence Models for Vietnamese](https://arxiv.org/abs/2109.09701) by Nguyen Luong Tran, Duong Minh Le and Dat Quoc Nguyen.
Expand Down Expand Up @@ -266,6 +267,7 @@ Flax), PyTorch, and/or TensorFlow.
| ALIGN | ❌ | ❌ | ✅ | ❌ | ❌ |
| AltCLIP | ❌ | ❌ | ✅ | ❌ | ❌ |
| Audio Spectrogram Transformer | ❌ | ❌ | ✅ | ❌ | ❌ |
| Autoformer | ❌ | ❌ | ✅ | ❌ | ❌ |
| BART | ✅ | ✅ | ✅ | ✅ | ✅ |
| BEiT | ❌ | ❌ | ✅ | ❌ | ✅ |
| BERT | ✅ | ✅ | ✅ | ✅ | ✅ |
Expand Down