Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add Chinese-CLIP implementation #20368

Merged
merged 123 commits into from Nov 30, 2022
Merged
Show file tree
Hide file tree
Changes from 114 commits
Commits
Show all changes
123 commits
Select commit Hold shift + click to select a range
70a9f72
init chinese-clip model from clip
yangapku Nov 12, 2022
b094a73
init model tests and docs
yangapku Nov 12, 2022
7f015cb
implement chinese-clip into hf
yangapku Nov 12, 2022
38dd850
implement chinese-clip into hf
yangapku Nov 12, 2022
b731e2d
implement chinese-clip into hf
yangapku Nov 12, 2022
c89a9ae
implement chinese-clip into hf
yangapku Nov 12, 2022
de05b3a
implement chinese-clip into hf
yangapku Nov 12, 2022
ef09771
update usecase example in model implementation
yangapku Nov 14, 2022
4aa01f0
Merge remote-tracking branch 'origin/main' into main_refactor
yangapku Nov 14, 2022
4522004
fix codestyle
yangapku Nov 14, 2022
f91b4f6
fix model_type typo in readme
yangapku Nov 14, 2022
80ea6ff
add placeholder in doc
yangapku Nov 14, 2022
b491bfa
add placeholder in doc
yangapku Nov 14, 2022
131cc0a
update the init script
yangapku Nov 14, 2022
44ab584
update usecase
yangapku Nov 14, 2022
933face
fix codestyle
yangapku Nov 14, 2022
09b4a6a
update testcase
yangapku Nov 15, 2022
90af626
Merge branch 'main' of github.com:yangapku/transformers
yangapku Nov 15, 2022
00951b1
Merge remote-tracking branch 'origin/main'
yangapku Nov 15, 2022
91b74da
update testcase
yangapku Nov 15, 2022
4e46944
update testcase
yangapku Nov 15, 2022
047ef5a
update testcase
yangapku Nov 15, 2022
d74a91e
update testcase
yangapku Nov 15, 2022
75a127b
update testcase
yangapku Nov 15, 2022
32a08cc
update testcase
yangapku Nov 15, 2022
c618cd1
update testcase
yangapku Nov 15, 2022
1778e71
update testcase
yangapku Nov 15, 2022
770b559
Merge remote-tracking branch 'origin/main'
yangapku Nov 17, 2022
e78c979
update testcase
yangapku Nov 17, 2022
40cc6b2
update testcase
yangapku Nov 17, 2022
f189356
update testcase
yangapku Nov 17, 2022
131aa83
forward the convert_rgb
yangapku Nov 17, 2022
e8f3dae
Merge remote-tracking branch 'origin/main'
yangapku Nov 21, 2022
64609f4
update testcase
yangapku Nov 21, 2022
0dc255c
update testcase
yangapku Nov 21, 2022
a35a23e
update testcase
yangapku Nov 21, 2022
017d834
merge the recent update from clip about model_input_name property
yangapku Nov 21, 2022
992f692
update the doc
yangapku Nov 21, 2022
d368c89
update the doc
yangapku Nov 21, 2022
8447d8d
update the doc
yangapku Nov 21, 2022
80f81a2
Merge remote-tracking branch 'origin/main'
yangapku Nov 21, 2022
f97787a
update the doc
yangapku Nov 22, 2022
2d9232d
update the doc
yangapku Nov 22, 2022
94edb63
remove unused imports
yangapku Nov 22, 2022
8ad55f2
reformat code style
yangapku Nov 22, 2022
efb38d2
update the doc
yangapku Nov 22, 2022
f405cb4
fix isort style
yangapku Nov 22, 2022
03ea389
bypass a weird failed unit test which is unrelated with my PR
yangapku Nov 22, 2022
2ac03ed
Merge branch 'main' of github.com:yangapku/transformers
yangapku Nov 22, 2022
222be18
update the doc
yangapku Nov 22, 2022
bdfc90c
update the testcase
yangapku Nov 23, 2022
68325a8
implement independent vision config class
yangapku Nov 23, 2022
5c43dc1
implement independent vision model class
yangapku Nov 23, 2022
79d2887
fix refactor bug
yangapku Nov 23, 2022
4525c24
fix refactor bug
yangapku Nov 23, 2022
b94b68d
fix refactor bug
yangapku Nov 23, 2022
38c9c49
make style
yangapku Nov 23, 2022
1f5816f
fix refactor bug
yangapku Nov 23, 2022
d8104dd
make style
yangapku Nov 23, 2022
c597697
fix refactor bug
yangapku Nov 23, 2022
f41ba8b
fix refactor bug
yangapku Nov 23, 2022
c957754
make style
yangapku Nov 23, 2022
666e7f2
fix refactor bug
yangapku Nov 23, 2022
231e244
fix refactor bug
yangapku Nov 23, 2022
3a92c30
fix refactor bug
yangapku Nov 23, 2022
21229c5
doc-build restyle
yangapku Nov 23, 2022
e91e776
implement independent text config class
yangapku Nov 23, 2022
1b74b79
implement independent text model class
yangapku Nov 23, 2022
0f8f220
Merge branch 'main' of github.com:yangapku/transformers
yangapku Nov 23, 2022
8b01339
implement independent text model class
yangapku Nov 23, 2022
57cb624
make style
yangapku Nov 23, 2022
cb27eaf
make fix-copies
yangapku Nov 23, 2022
3ebd25b
fix refactor bug
yangapku Nov 23, 2022
64db8c2
Merge branch 'main' of github.com:yangapku/transformers
yangapku Nov 23, 2022
8f65ff7
fix refactor bug
yangapku Nov 23, 2022
b376a3f
fix refactor bug
yangapku Nov 23, 2022
dee5404
fix refactor bug
yangapku Nov 23, 2022
930c5ea
fix refactor bug
yangapku Nov 23, 2022
dc842f6
fix refactor bug
yangapku Nov 24, 2022
7ffd61a
fix refactor bug
yangapku Nov 24, 2022
f5aad2e
fix refactor bug
yangapku Nov 24, 2022
6a83676
fix refactor bug
yangapku Nov 24, 2022
ae4c9d4
fix refactor bug
yangapku Nov 24, 2022
5624508
Merge remote-tracking branch 'origin/main'
yangapku Nov 24, 2022
e4032d3
make style
yangapku Nov 24, 2022
a9cbc8b
update doc
yangapku Nov 24, 2022
ad0741e
Merge branch 'main' of github.com:yangapku/transformers
yangapku Nov 24, 2022
2429a28
black and isort
yangapku Nov 24, 2022
97b3187
update doc
yangapku Nov 24, 2022
20d508f
Merge branch 'main' of github.com:yangapku/transformers
yangapku Nov 24, 2022
de1b1d6
Update src/transformers/models/chinese_clip/configuration_chinese_cli…
yangapku Nov 28, 2022
18fe004
Update src/transformers/models/auto/tokenization_auto.py
yangapku Nov 28, 2022
23b04dd
modify the model type from chinese-clip to chinese_clip
yangapku Nov 29, 2022
9440c4e
format the example comment of ChineseCLIPVisionConfig
yangapku Nov 29, 2022
ad06030
correct the copyright comment
yangapku Nov 29, 2022
1dd4767
fix the tokenizer specification
yangapku Nov 29, 2022
277402a
add copied from for loss function
yangapku Nov 29, 2022
050afea
remove unused class
yangapku Nov 29, 2022
6addcba
update CHINESE_CLIP_TEXT_INPUTS_DOCSTRING
yangapku Nov 29, 2022
32ccac1
update CHINESE_CLIP_INPUTS_DOCSTRING
yangapku Nov 29, 2022
7d438f8
update doc
yangapku Nov 29, 2022
857b0cc
update doc
yangapku Nov 29, 2022
e9f6fb1
update code comment in config
yangapku Nov 29, 2022
909bb58
update copied from statement
yangapku Nov 29, 2022
6a24dac
make style
yangapku Nov 29, 2022
de30b69
rename the doc file
yangapku Nov 29, 2022
5d63009
add copied statement
yangapku Nov 29, 2022
2bef9ab
remove unused attention_mask, causal_attention_mask in ChineseCLIPVis…
yangapku Nov 29, 2022
3a8b5a8
remove ChineseCLIPTextPreTrainedModel
yangapku Nov 29, 2022
0e343c8
fix bug
yangapku Nov 29, 2022
9bec7ca
fix bug
yangapku Nov 29, 2022
3f2b97b
fix bug
yangapku Nov 29, 2022
5a8c4c1
update doc
yangapku Nov 29, 2022
a0a3817
make style
yangapku Nov 29, 2022
34d25a3
Update src/transformers/models/chinese_clip/configuration_chinese_cli…
yangapku Nov 29, 2022
dc51647
Update src/transformers/models/chinese_clip/configuration_chinese_cli…
yangapku Nov 29, 2022
cec76c2
update ChineseCLIPImageProcessor in image_processing_auto
yangapku Nov 29, 2022
0d5d9ea
fix config_class of chinesecliptextmodel
yangapku Nov 30, 2022
22d676f
fix the test case
yangapku Nov 30, 2022
a7cfb0a
update the docs
yangapku Nov 30, 2022
b8de02e
remove the copied from comment for ChineseCLIPTextModel, since it has…
yangapku Nov 30, 2022
5d04b82
update the testcase
yangapku Nov 30, 2022
c42aff6
final fix
ydshieh Nov 30, 2022
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Jump to
Jump to file
Failed to load files.
Diff view
Diff view
1 change: 1 addition & 0 deletions README.md
Expand Up @@ -279,6 +279,7 @@ Current number of checkpoints: ![](https://img.shields.io/endpoint?url=https://h
1. **[ByT5](https://huggingface.co/docs/transformers/model_doc/byt5)** (from Google Research) released with the paper [ByT5: Towards a token-free future with pre-trained byte-to-byte models](https://arxiv.org/abs/2105.13626) by Linting Xue, Aditya Barua, Noah Constant, Rami Al-Rfou, Sharan Narang, Mihir Kale, Adam Roberts, Colin Raffel.
1. **[CamemBERT](https://huggingface.co/docs/transformers/model_doc/camembert)** (from Inria/Facebook/Sorbonne) released with the paper [CamemBERT: a Tasty French Language Model](https://arxiv.org/abs/1911.03894) by Louis Martin*, Benjamin Muller*, Pedro Javier Ortiz Suárez*, Yoann Dupont, Laurent Romary, Éric Villemonte de la Clergerie, Djamé Seddah and Benoît Sagot.
1. **[CANINE](https://huggingface.co/docs/transformers/model_doc/canine)** (from Google Research) released with the paper [CANINE: Pre-training an Efficient Tokenization-Free Encoder for Language Representation](https://arxiv.org/abs/2103.06874) by Jonathan H. Clark, Dan Garrette, Iulia Turc, John Wieting.
1. **[Chinese-CLIP](https://huggingface.co/docs/transformers/main/model_doc/chinese_clip)** (from OFA-Sys) released with the paper [Chinese CLIP: Contrastive Vision-Language Pretraining in Chinese](https://arxiv.org/abs/2211.01335) by An Yang, Junshu Pan, Junyang Lin, Rui Men, Yichang Zhang, Jingren Zhou, Chang Zhou.
1. **[CLIP](https://huggingface.co/docs/transformers/model_doc/clip)** (from OpenAI) released with the paper [Learning Transferable Visual Models From Natural Language Supervision](https://arxiv.org/abs/2103.00020) by Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, Gretchen Krueger, Ilya Sutskever.
1. **[CLIPSeg](https://huggingface.co/docs/transformers/main/model_doc/clipseg)** (from University of Göttingen) released with the paper [Image Segmentation Using Text and Image Prompts](https://arxiv.org/abs/2112.10003) by Timo Lüddecke and Alexander Ecker.
1. **[CodeGen](https://huggingface.co/docs/transformers/model_doc/codegen)** (from Salesforce) released with the paper [A Conversational Paradigm for Program Synthesis](https://arxiv.org/abs/2203.13474) by Erik Nijkamp, Bo Pang, Hiroaki Hayashi, Lifu Tu, Huan Wang, Yingbo Zhou, Silvio Savarese, Caiming Xiong.
Expand Down
1 change: 1 addition & 0 deletions README_es.md
Expand Up @@ -279,6 +279,7 @@ Número actual de puntos de control: ![](https://img.shields.io/endpoint?url=htt
1. **[ByT5](https://huggingface.co/docs/transformers/model_doc/byt5)** (from Google Research) released with the paper [ByT5: Towards a token-free future with pre-trained byte-to-byte models](https://arxiv.org/abs/2105.13626) by Linting Xue, Aditya Barua, Noah Constant, Rami Al-Rfou, Sharan Narang, Mihir Kale, Adam Roberts, Colin Raffel.
1. **[CamemBERT](https://huggingface.co/docs/transformers/model_doc/camembert)** (from Inria/Facebook/Sorbonne) released with the paper [CamemBERT: a Tasty French Language Model](https://arxiv.org/abs/1911.03894) by Louis Martin*, Benjamin Muller*, Pedro Javier Ortiz Suárez*, Yoann Dupont, Laurent Romary, Éric Villemonte de la Clergerie, Djamé Seddah and Benoît Sagot.
1. **[CANINE](https://huggingface.co/docs/transformers/model_doc/canine)** (from Google Research) released with the paper [CANINE: Pre-training an Efficient Tokenization-Free Encoder for Language Representation](https://arxiv.org/abs/2103.06874) by Jonathan H. Clark, Dan Garrette, Iulia Turc, John Wieting.
1. **[Chinese-CLIP](https://huggingface.co/docs/transformers/main/model_doc/chinese_clip)** (from OFA-Sys) released with the paper [Chinese CLIP: Contrastive Vision-Language Pretraining in Chinese](https://arxiv.org/abs/2211.01335) by An Yang, Junshu Pan, Junyang Lin, Rui Men, Yichang Zhang, Jingren Zhou, Chang Zhou.
1. **[CLIP](https://huggingface.co/docs/transformers/model_doc/clip)** (from OpenAI) released with the paper [Learning Transferable Visual Models From Natural Language Supervision](https://arxiv.org/abs/2103.00020) by Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, Gretchen Krueger, Ilya Sutskever.
1. **[CLIPSeg](https://huggingface.co/docs/transformers/main/model_doc/clipseg)** (from University of Göttingen) released with the paper [Image Segmentation Using Text and Image Prompts](https://arxiv.org/abs/2112.10003) by Timo Lüddecke and Alexander Ecker.
1. **[CodeGen](https://huggingface.co/docs/transformers/model_doc/codegen)** (from Salesforce) released with the paper [A Conversational Paradigm for Program Synthesis](https://arxiv.org/abs/2203.13474) by Erik Nijkamp, Bo Pang, Hiroaki Hayashi, Lifu Tu, Huan Wang, Yingbo Zhou, Silvio Savarese, Caiming Xiong.
Expand Down
1 change: 1 addition & 0 deletions README_ja.md
Expand Up @@ -314,6 +314,7 @@ Flax、PyTorch、TensorFlowをcondaでインストールする方法は、それ
1. **[ByT5](https://huggingface.co/docs/transformers/model_doc/byt5)** (from Google Research) released with the paper [ByT5: Towards a token-free future with pre-trained byte-to-byte models](https://arxiv.org/abs/2105.13626) by Linting Xue, Aditya Barua, Noah Constant, Rami Al-Rfou, Sharan Narang, Mihir Kale, Adam Roberts, Colin Raffel.
1. **[CamemBERT](https://huggingface.co/docs/transformers/model_doc/camembert)** (from Inria/Facebook/Sorbonne) released with the paper [CamemBERT: a Tasty French Language Model](https://arxiv.org/abs/1911.03894) by Louis Martin*, Benjamin Muller*, Pedro Javier Ortiz Suárez*, Yoann Dupont, Laurent Romary, Éric Villemonte de la Clergerie, Djamé Seddah and Benoît Sagot.
1. **[CANINE](https://huggingface.co/docs/transformers/model_doc/canine)** (from Google Research) released with the paper [CANINE: Pre-training an Efficient Tokenization-Free Encoder for Language Representation](https://arxiv.org/abs/2103.06874) by Jonathan H. Clark, Dan Garrette, Iulia Turc, John Wieting.
1. **[Chinese-CLIP](https://huggingface.co/docs/transformers/main/model_doc/chinese_clip)** (from OFA-Sys) released with the paper [Chinese CLIP: Contrastive Vision-Language Pretraining in Chinese](https://arxiv.org/abs/2211.01335) by An Yang, Junshu Pan, Junyang Lin, Rui Men, Yichang Zhang, Jingren Zhou, Chang Zhou.
1. **[CLIP](https://huggingface.co/docs/transformers/model_doc/clip)** (from OpenAI) released with the paper [Learning Transferable Visual Models From Natural Language Supervision](https://arxiv.org/abs/2103.00020) by Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, Gretchen Krueger, Ilya Sutskever.
1. **[CLIPSeg](https://huggingface.co/docs/transformers/main/model_doc/clipseg)** (from University of Göttingen) released with the paper [Image Segmentation Using Text and Image Prompts](https://arxiv.org/abs/2112.10003) by Timo Lüddecke and Alexander Ecker.
1. **[CodeGen](https://huggingface.co/docs/transformers/model_doc/codegen)** (from Salesforce) released with the paper [A Conversational Paradigm for Program Synthesis](https://arxiv.org/abs/2203.13474) by Erik Nijkamp, Bo Pang, Hiroaki Hayashi, Lifu Tu, Huan Wang, Yingbo Zhou, Silvio Savarese, Caiming Xiong.
Expand Down
1 change: 1 addition & 0 deletions README_ko.md
Expand Up @@ -229,6 +229,7 @@ Flax, PyTorch, TensorFlow 설치 페이지에서 이들을 conda로 설치하는
1. **[ByT5](https://huggingface.co/docs/transformers/model_doc/byt5)** (from Google Research) released with the paper [ByT5: Towards a token-free future with pre-trained byte-to-byte models](https://arxiv.org/abs/2105.13626) by Linting Xue, Aditya Barua, Noah Constant, Rami Al-Rfou, Sharan Narang, Mihir Kale, Adam Roberts, Colin Raffel.
1. **[CamemBERT](https://huggingface.co/docs/transformers/model_doc/camembert)** (from Inria/Facebook/Sorbonne) released with the paper [CamemBERT: a Tasty French Language Model](https://arxiv.org/abs/1911.03894) by Louis Martin*, Benjamin Muller*, Pedro Javier Ortiz Suárez*, Yoann Dupont, Laurent Romary, Éric Villemonte de la Clergerie, Djamé Seddah and Benoît Sagot.
1. **[CANINE](https://huggingface.co/docs/transformers/model_doc/canine)** (from Google Research) released with the paper [CANINE: Pre-training an Efficient Tokenization-Free Encoder for Language Representation](https://arxiv.org/abs/2103.06874) by Jonathan H. Clark, Dan Garrette, Iulia Turc, John Wieting.
1. **[Chinese-CLIP](https://huggingface.co/docs/transformers/main/model_doc/chinese_clip)** (from OFA-Sys) released with the paper [Chinese CLIP: Contrastive Vision-Language Pretraining in Chinese](https://arxiv.org/abs/2211.01335) by An Yang, Junshu Pan, Junyang Lin, Rui Men, Yichang Zhang, Jingren Zhou, Chang Zhou.
1. **[CLIP](https://huggingface.co/docs/transformers/model_doc/clip)** (from OpenAI) released with the paper [Learning Transferable Visual Models From Natural Language Supervision](https://arxiv.org/abs/2103.00020) by Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, Gretchen Krueger, Ilya Sutskever.
1. **[CLIPSeg](https://huggingface.co/docs/transformers/main/model_doc/clipseg)** (from University of Göttingen) released with the paper [Image Segmentation Using Text and Image Prompts](https://arxiv.org/abs/2112.10003) by Timo Lüddecke and Alexander Ecker.
1. **[CodeGen](https://huggingface.co/docs/transformers/model_doc/codegen)** (from Salesforce) released with the paper [A Conversational Paradigm for Program Synthesis](https://arxiv.org/abs/2203.13474) by Erik Nijkamp, Bo Pang, Hiroaki Hayashi, Lifu Tu, Huan Wang, Yingbo Zhou, Silvio Savarese, Caiming Xiong.
Expand Down
1 change: 1 addition & 0 deletions README_zh-hans.md
Expand Up @@ -253,6 +253,7 @@ conda install -c huggingface transformers
1. **[ByT5](https://huggingface.co/docs/transformers/model_doc/byt5)** (来自 Google Research) 伴随论文 [ByT5: Towards a token-free future with pre-trained byte-to-byte models](https://arxiv.org/abs/2105.13626) 由 Linting Xue, Aditya Barua, Noah Constant, Rami Al-Rfou, Sharan Narang, Mihir Kale, Adam Roberts, Colin Raffel 发布。
1. **[CamemBERT](https://huggingface.co/docs/transformers/model_doc/camembert)** (来自 Inria/Facebook/Sorbonne) 伴随论文 [CamemBERT: a Tasty French Language Model](https://arxiv.org/abs/1911.03894) 由 Louis Martin*, Benjamin Muller*, Pedro Javier Ortiz Suárez*, Yoann Dupont, Laurent Romary, Éric Villemonte de la Clergerie, Djamé Seddah and Benoît Sagot 发布。
1. **[CANINE](https://huggingface.co/docs/transformers/model_doc/canine)** (来自 Google Research) 伴随论文 [CANINE: Pre-training an Efficient Tokenization-Free Encoder for Language Representation](https://arxiv.org/abs/2103.06874) 由 Jonathan H. Clark, Dan Garrette, Iulia Turc, John Wieting 发布。
1. **[Chinese-CLIP](https://huggingface.co/docs/transformers/main/model_doc/chinese_clip)** (来自 OFA-Sys) 伴随论文 [Chinese CLIP: Contrastive Vision-Language Pretraining in Chinese](https://arxiv.org/abs/2211.01335) 由 An Yang, Junshu Pan, Junyang Lin, Rui Men, Yichang Zhang, Jingren Zhou, Chang Zhou 发布。
1. **[CLIP](https://huggingface.co/docs/transformers/model_doc/clip)** (来自 OpenAI) 伴随论文 [Learning Transferable Visual Models From Natural Language Supervision](https://arxiv.org/abs/2103.00020) 由 Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, Gretchen Krueger, Ilya Sutskever 发布。
1. **[CLIPSeg](https://huggingface.co/docs/transformers/main/model_doc/clipseg)** (来自 University of Göttingen) 伴随论文 [Image Segmentation Using Text and Image Prompts](https://arxiv.org/abs/2112.10003) 由 Timo Lüddecke and Alexander Ecker 发布。
1. **[CodeGen](https://huggingface.co/docs/transformers/model_doc/codegen)** (来自 Salesforce) 伴随论文 [A Conversational Paradigm for Program Synthesis](https://arxiv.org/abs/2203.13474) 由 Erik Nijkamp, Bo Pang, Hiroaki Hayashi, Lifu Tu, Huan Wang, Yingbo Zhou, Silvio Savarese, Caiming Xiong 发布。
Expand Down
1 change: 1 addition & 0 deletions README_zh-hant.md
Expand Up @@ -265,6 +265,7 @@ conda install -c huggingface transformers
1. **[ByT5](https://huggingface.co/docs/transformers/model_doc/byt5)** (from Google Research) released with the paper [ByT5: Towards a token-free future with pre-trained byte-to-byte models](https://arxiv.org/abs/2105.13626) by Linting Xue, Aditya Barua, Noah Constant, Rami Al-Rfou, Sharan Narang, Mihir Kale, Adam Roberts, Colin Raffel.
1. **[CamemBERT](https://huggingface.co/docs/transformers/model_doc/camembert)** (from Inria/Facebook/Sorbonne) released with the paper [CamemBERT: a Tasty French Language Model](https://arxiv.org/abs/1911.03894) by Louis Martin*, Benjamin Muller*, Pedro Javier Ortiz Suárez*, Yoann Dupont, Laurent Romary, Éric Villemonte de la Clergerie, Djamé Seddah and Benoît Sagot.
1. **[CANINE](https://huggingface.co/docs/transformers/model_doc/canine)** (from Google Research) released with the paper [CANINE: Pre-training an Efficient Tokenization-Free Encoder for Language Representation](https://arxiv.org/abs/2103.06874) by Jonathan H. Clark, Dan Garrette, Iulia Turc, John Wieting.
1. **[Chinese-CLIP](https://huggingface.co/docs/transformers/main/model_doc/chinese_clip)** (from OFA-Sys) released with the paper [Chinese CLIP: Contrastive Vision-Language Pretraining in Chinese](https://arxiv.org/abs/2211.01335) by An Yang, Junshu Pan, Junyang Lin, Rui Men, Yichang Zhang, Jingren Zhou, Chang Zhou.
1. **[CLIP](https://huggingface.co/docs/transformers/model_doc/clip)** (from OpenAI) released with the paper [Learning Transferable Visual Models From Natural Language Supervision](https://arxiv.org/abs/2103.00020) by Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, Gretchen Krueger, Ilya Sutskever.
1. **[CLIPSeg](https://huggingface.co/docs/transformers/main/model_doc/clipseg)** (from University of Göttingen) released with the paper [Image Segmentation Using Text and Image Prompts](https://arxiv.org/abs/2112.10003) by Timo Lüddecke and Alexander Ecker.
1. **[CodeGen](https://huggingface.co/docs/transformers/model_doc/codegen)** (from Salesforce) released with the paper [A Conversational Paradigm for Program Synthesis](https://arxiv.org/abs/2203.13474) by Erik Nijkamp, Bo Pang, Hiroaki Hayashi, Lifu Tu, Huan Wang, Yingbo Zhou, Silvio Savarese, Caiming Xiong.
Expand Down
2 changes: 2 additions & 0 deletions docs/source/en/_toctree.yml
Expand Up @@ -482,6 +482,8 @@
title: Audio models
- isExpanded: false
sections:
- local: model_doc/chinese_clip
title: Chinese-CLIP
- local: model_doc/clip
title: CLIP
- local: model_doc/clipseg
Expand Down
2 changes: 2 additions & 0 deletions docs/source/en/index.mdx
Expand Up @@ -67,6 +67,7 @@ The documentation is organized into five sections:
1. **[ByT5](model_doc/byt5)** (from Google Research) released with the paper [ByT5: Towards a token-free future with pre-trained byte-to-byte models](https://arxiv.org/abs/2105.13626) by Linting Xue, Aditya Barua, Noah Constant, Rami Al-Rfou, Sharan Narang, Mihir Kale, Adam Roberts, Colin Raffel.
1. **[CamemBERT](model_doc/camembert)** (from Inria/Facebook/Sorbonne) released with the paper [CamemBERT: a Tasty French Language Model](https://arxiv.org/abs/1911.03894) by Louis Martin*, Benjamin Muller*, Pedro Javier Ortiz Suárez*, Yoann Dupont, Laurent Romary, Éric Villemonte de la Clergerie, Djamé Seddah and Benoît Sagot.
1. **[CANINE](model_doc/canine)** (from Google Research) released with the paper [CANINE: Pre-training an Efficient Tokenization-Free Encoder for Language Representation](https://arxiv.org/abs/2103.06874) by Jonathan H. Clark, Dan Garrette, Iulia Turc, John Wieting.
1. **[Chinese-CLIP](model_doc/chinese_clip)** (from OFA-Sys) released with the paper [Chinese CLIP: Contrastive Vision-Language Pretraining in Chinese](https://arxiv.org/abs/2211.01335) by An Yang, Junshu Pan, Junyang Lin, Rui Men, Yichang Zhang, Jingren Zhou, Chang Zhou.
1. **[CLIP](model_doc/clip)** (from OpenAI) released with the paper [Learning Transferable Visual Models From Natural Language Supervision](https://arxiv.org/abs/2103.00020) by Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, Gretchen Krueger, Ilya Sutskever.
1. **[CLIPSeg](model_doc/clipseg)** (from University of Göttingen) released with the paper [Image Segmentation Using Text and Image Prompts](https://arxiv.org/abs/2112.10003) by Timo Lüddecke and Alexander Ecker.
1. **[CodeGen](model_doc/codegen)** (from Salesforce) released with the paper [A Conversational Paradigm for Program Synthesis](https://arxiv.org/abs/2203.13474) by Erik Nijkamp, Bo Pang, Hiroaki Hayashi, Lifu Tu, Huan Wang, Yingbo Zhou, Silvio Savarese, Caiming Xiong.
Expand Down Expand Up @@ -232,6 +233,7 @@ Flax), PyTorch, and/or TensorFlow.
| BLOOM | ❌ | ✅ | ✅ | ❌ | ❌ |
| CamemBERT | ✅ | ✅ | ✅ | ✅ | ❌ |
| CANINE | ✅ | ❌ | ✅ | ❌ | ❌ |
| Chinese-CLIP | ❌ | ❌ | ✅ | ❌ | ❌ |
| CLIP | ✅ | ✅ | ✅ | ✅ | ✅ |
| CLIPSeg | ❌ | ❌ | ✅ | ❌ | ❌ |
| CodeGen | ✅ | ✅ | ✅ | ❌ | ❌ |
Expand Down