Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Community Event] Doc Tests Sprint #16292

Open
35 of 192 tasks
patrickvonplaten opened this issue Mar 21, 2022 · 40 comments
Open
35 of 192 tasks

[Community Event] Doc Tests Sprint #16292

patrickvonplaten opened this issue Mar 21, 2022 · 40 comments

Comments

@patrickvonplaten
Copy link
Member

@patrickvonplaten patrickvonplaten commented Mar 21, 2022

This issue is part of our Doc Test Sprint. If you're interested in helping out come join us on Discord and talk with other contributors!

Docstring examples are often the first point of contact when trying out a new library! So far we haven't done a very good job at ensuring that all docstring examples work correctly in 🤗 Transformers - but we're now very dedicated to ensure that all documentation examples work correctly by testing each documentation example via Python's doctest (https://docs.python.org/3/library/doctest.html) on a daily basis.

In short we should do the following for all models for both PyTorch and Tensorflow:

    • Check the current doc examples will run without failure
    • Add an expected output to the doc example and test it via Python's doc test (see Guide to contributing below)

Adding a documentation test for a model is a great way to better understand how the model works, a simple (possibly first) contribution to Transformers and most importantly a very important contribution to the Transformers community 🔥

If you're interested in adding a documentation test, please read through the Guide to contributing below.

This issue is a call for contributors, to make sure docstring exmaples of existing model architectures work correctly. If you wish to contribute, reply in this thread which architectures you'd like to take :)

Guide to contributing:

  1. Ensure you've read our contributing guidelines 📜

  2. Claim your architecture(s) in this thread (confirm no one is working on it) 🎯

  3. Implement the changes as in #15987 (see the diff on the model architectures for a few examples) 💪

    In addition, there are a few things we can also improve, for example :

    • Fix some style issues: for example, change ``decoder_input_ids``` to `decoder_input_ids`.
    • Using a small model checkpoint instead of a large one: for example, change "facebook/bart-large" to "facebook/bart-base" (and adjust the expected outputs if any)
  4. Open the PR and tag me @patrickvonplaten @ydshieh or @patil-suraj (don't forget to run make fixup before your final commit) 🎊

    • Note that some code is copied across our codebase. If you see a line like # Copied from transformers.models.bert..., this means that the code is copied from that source, and our scripts will automatically keep that in sync. If you see that, you should not edit the copied method! Instead, edit the original method it's copied from, and run make fixup to synchronize that across all the copies. Be sure you installed the development dependencies with pip install -e ".[dev]", as described in the contributor guidelines above, to ensure that the code quality tools in make fixup can run.

PyTorch Model Examples added to tests:

Tensorflow Model Examples added to tests:

  • ALBERT (@vumichien)
  • BART
  • BEiT
  • BERT (@vumichien)
  • Bert
  • BigBird (@vumichien)
  • BigBirdPegasus
  • Blenderbot
  • BlenderbotSmall
  • CamemBERT
  • Canine
  • CLIP (@Aanisha)
  • ConvBERT (@simonzli)
  • ConvNext
  • CTRL
  • Data2VecAudio
  • Data2VecText
  • DeBERTa
  • DeBERTa-v2
  • DeiT
  • DETR
  • DistilBERT (@jmwoloso)
  • DPR
  • ELECTRA (@bhadreshpsavani)
  • Encoder
  • FairSeq
  • FlauBERT
  • FNet
  • Funnel
  • GPT2 (@cakiki)
  • GPT-J (@cakiki)
  • Hubert
  • I-BERT
  • ImageGPT
  • LayoutLM
  • LayoutLMv2
  • LED
  • Longformer (@KMFODA)
  • LUKE
  • LXMERT
  • M2M100
  • Marian
  • MaskFormer (@reichenbch)
  • mBART
  • MegatronBert
  • MobileBERT (@vumichien)
  • MPNet
  • mT5
  • Nystromformer
  • OpenAI
  • OpenAI
  • Pegasus
  • Perceiver
  • PLBart
  • PoolFormer
  • ProphetNet
  • QDQBert
  • RAG
  • Realm
  • Reformer
  • ResNet
  • RemBERT
  • RetriBERT
  • RoBERTa (@patrickvonplaten)
  • RoFormer
  • SegFormer
  • SEW
  • SEW-D
  • SpeechEncoderDecoder
  • Speech2Text
  • Speech2Text2
  • Splinter
  • SqueezeBERT
  • Swin (@johko)
  • T5 (@MarkusSagen)
  • TAPAS
  • Transformer-XL (@simonzli)
  • TrOCR (@arnaudstiegler)
  • UniSpeech
  • UniSpeechSat
  • Van
  • ViLT
  • VisionEncoderDecoder
  • VisionTextDualEncoder
  • VisualBert
  • ViT (@johko)
  • ViTMAE
  • Wav2Vec2
  • WavLM
  • XGLM
  • XLM
  • XLM-RoBERTa (@AbinayaM02)
  • XLM-RoBERTa-XL
  • XLMProphetNet
  • XLNet
  • YOSO
@patrickvonplaten patrickvonplaten changed the title Doc tests sprint [Community Event] Doc Tests Sprint Mar 21, 2022
@patrickvonplaten patrickvonplaten pinned this issue Mar 21, 2022
@reichenbch
Copy link
Contributor

@reichenbch reichenbch commented Mar 21, 2022

@patrickvonplaten I would like to start with Maskformer for Tensorflow/Pytorch. Catch up with how the event goes.

@patrickvonplaten
Copy link
Member Author

@patrickvonplaten patrickvonplaten commented Mar 21, 2022

Awesome! Let me know if you have any questions :-)

@KMFODA
Copy link

@KMFODA KMFODA commented Mar 21, 2022

Hello! I'd like to take on Longformer for Tensorflow/Pytorch please.

@MarkusSagen
Copy link
Contributor

@MarkusSagen MarkusSagen commented Mar 21, 2022

@patrickvonplaten I would like to start with T5 for pytorch and tensorflow

@patrickvonplaten
Copy link
Member Author

@patrickvonplaten patrickvonplaten commented Mar 22, 2022

Sounds great!

@patrickvonplaten
Copy link
Member Author

@patrickvonplaten patrickvonplaten commented Mar 22, 2022

LayoutLM is also taken as mentioned by a contributor on Discord!

@cakiki
Copy link
Contributor

@cakiki cakiki commented Mar 22, 2022

@patrickvonplaten I would take GPT and GPT-J (TensorFlow editions) if those are still available.

I'm guessing GPT is GPT2?

@vumichien
Copy link
Contributor

@vumichien vumichien commented Mar 22, 2022

I will take Bert, Albert, and Bigbird for both Tensorflow/Pytorch

@johko
Copy link
Contributor

@johko johko commented Mar 22, 2022

I'll take Swin and ViT for Tensorflow

@jmwoloso
Copy link
Contributor

@jmwoloso jmwoloso commented Mar 22, 2022

I'd like DistilBERT for both TF and PT please

@ydshieh
Copy link
Contributor

@ydshieh ydshieh commented Mar 22, 2022

@patrickvonplaten I would take GPT and GPT-J (TensorFlow editions) if those are still available.

I'm guessing GPT is GPT2?

@cakiki You can go for GPT2 (I updated the name in the test)

@ArEnSc
Copy link

@ArEnSc ArEnSc commented Mar 23, 2022

Can I try GPT2 and GPTJ for Pytorch? if @ydshieh you are not doing so?

@Aanisha
Copy link

@Aanisha Aanisha commented Mar 23, 2022

I would like to try CLIP for Tensorflow and PyTorch.

@NielsRogge
Copy link
Contributor

@NielsRogge NielsRogge commented Mar 23, 2022

I'll take CANINE and TAPAS.

@ydshieh
Copy link
Contributor

@ydshieh ydshieh commented Mar 23, 2022

Can I try GPT2 and GPTJ for Pytorch? if @ydshieh you are not doing so?

@ArEnSc
No, you can work on these 2 models :-) Thank you!

@vumichien
Copy link
Contributor

@vumichien vumichien commented Mar 23, 2022

@ydshieh Since the MobileBertForSequenceClassification is the copy of BertForSequenceClassification, so I think I will do check doc-test of MobileBert as well to overcome the error from make fixup

@abdouaziz
Copy link

@abdouaziz abdouaziz commented Mar 23, 2022

I'll take FlauBERT and CamemBERT.

@ydshieh
Copy link
Contributor

@ydshieh ydshieh commented Mar 23, 2022

@abdouaziz Awesome! Do you plan to work on both PyTorch and TensorFlow versions, or only one of them?

@Tegzes
Copy link
Contributor

@Tegzes Tegzes commented Mar 23, 2022

I would like to work on LUKE model for both TF and PT

@NielsRogge
Copy link
Contributor

@NielsRogge NielsRogge commented Mar 23, 2022

@Tegzes you're lucky because there's no LUKE in TF ;) the list above actually just duplicates all models, but many models aren't available yet in TF.

@Tegzes
Copy link
Contributor

@Tegzes Tegzes commented Mar 23, 2022

In this case, I will also take DeBERTa and DeBERTa-v2 for PyTorch

@abdouaziz
Copy link

@abdouaziz abdouaziz commented Mar 23, 2022

@ydshieh

I plan to work only with PyTorch

@patrickvonplaten
Copy link
Member Author

@patrickvonplaten patrickvonplaten commented Mar 23, 2022

@Tegzes you're lucky because there's no LUKE in TF ;) the list above actually just duplicates all models, but many models aren't available yet in TF.

True - sorry I've been lazy at creating this list!

@arnaudstiegler
Copy link

@arnaudstiegler arnaudstiegler commented Mar 23, 2022

Happy to work on TrOCR (pytorch and TF)

@patrickvonplaten
Copy link
Member Author

@patrickvonplaten patrickvonplaten commented Mar 23, 2022

I take RoBERTa in PT and TF

@AbinayaM02
Copy link

@AbinayaM02 AbinayaM02 commented Mar 23, 2022

I would like to pick up XLM-RoBERTa in PT and TF.

@bhadreshpsavani
Copy link
Contributor

@bhadreshpsavani bhadreshpsavani commented Mar 23, 2022

I can work on ELECTRA for PT and TF

@patrickvonplaten
Copy link
Member Author

@patrickvonplaten patrickvonplaten commented Mar 23, 2022

Hey guys,

We've just merged the first template for Roberta-like model doc tests: #16363 :-)
Lots of models like ELETRA, XLM-RoBERTa, DeBERTa, BERT are very similar in spirit, it would be great if you could try to rebase your PR to the change done in #16363 . Usually all you need to do is to add the correct {expected_outputs}, {expected_loss} and {checkpoint} to the docstring of each model (ideally giving sensible results :-)) until it passes locally and then the file can be added to the tester :-)

@patrickvonplaten
Copy link
Member Author

@patrickvonplaten patrickvonplaten commented Mar 23, 2022

Also if you have open PRs and need help, feel free to ping me or @ydshieh and link the PR here so that we can nicely gather everything :-)

@patrickvonplaten
Copy link
Member Author

@patrickvonplaten patrickvonplaten commented Mar 23, 2022

One of the most difficult tasks here might be to actually find a well-working model. As a tip what you can do:

  1. Find all models of your architecture as it's always stated in the modeling files here:
    # See all RoBERTa models at https://huggingface.co/models?filter=roberta
    e.g. for ELECTRA: https://huggingface.co/models?filter=electra
  2. Now click on the task (in left sidebar) your working on, e.g. say you work on ForSequenceClassification of a text model go under this task metric: https://huggingface.co/models?other=electra&pipeline_tag=text-classification&sort=downloads
  3. Finally, click on the framework metric (in left sidebar) you're working with: e.g. for TF: https://huggingface.co/models?library=tf&other=electra&pipeline_tag=text-classification&sort=downloads . If you see too few or too many not well performing models in TF you might also want to think about converting a good PT model to TF under your Hub name and to use this one instead :-)

@jeremyadamsfisher
Copy link

@jeremyadamsfisher jeremyadamsfisher commented Mar 24, 2022

I'll take a shot with the PyTorch implementation of CTRL

@patrickvonplaten
Copy link
Member Author

@patrickvonplaten patrickvonplaten commented Mar 24, 2022

Here the mirror of RoBERTa for Tensorflow: #16370

@ydshieh
Copy link
Contributor

@ydshieh ydshieh commented Mar 24, 2022

Hi, contributors, thank you very much for participating this sprint ❤️.

Here is one tip that might reduce some issues:

Considering the following 2 facts:

  • A previous file file_utils.py contains some code regarding documentation. It was recently refactorized to different files. It might be a good idea (necessary in some case) to update your working branch in your local clone.
  • The file transformers/utils/documentation_tests.txt will be updated frequently by different contributors during this event.

Some testing issues could be resolved as:

git checkout main  # or `master`, depends on your local clone
git fetch upstream
git pull upstream main  # Hugging Face `transformers` renamed the default branch to `main` recently
git checkout your_working_branch_for_this_sprint
git rebase main  # or `master`

Don't hesitate if you encounter any problem. Enjoy~

@abdouaziz
Copy link

@abdouaziz abdouaziz commented Mar 24, 2022

I take BART and IBERT for PT

@simonzli
Copy link
Contributor

@simonzli simonzli commented Mar 24, 2022

I'd like to take a crack on Transformer-XL and ConvBert

@ydshieh
Copy link
Contributor

@ydshieh ydshieh commented Mar 24, 2022

Hi, contributors!

For the model(s) you work with for this sprint, if you could not find any checkpoint for a downstream task, say XXXModelForTokenClassification model, but there is a checkpoint for the base model, what you could do is:

model = XXXModelForTokenClassification.from_pretrained(base_model_checkpoint_name)
model.save_pretrained(local_path)

Then you can upload this new saved checkpoint to Hugging Face Hub, and you can use this uploaded model for the docstring example.

The head part of the model will have randomly initialized weights, and the result is likely to be imperfect, but it is fine for this sprint :-)

@ydshieh
Copy link
Contributor

@ydshieh ydshieh commented Mar 24, 2022

I'd like to take a crack on Transformer-XL and ConvBert

@simonzli, great :-). Do you plan to work with the PyTorch or TensorFlow version, or both?

@simonzli
Copy link
Contributor

@simonzli simonzli commented Mar 24, 2022

I'd like to take a crack on Transformer-XL and ConvBert

@simonzli, great :-). Do you plan to work with the PyTorch or TensorFlow version, or both?

I'll work on both PyTorch and TensorFlow😊

@AbinayaM02
Copy link

@AbinayaM02 AbinayaM02 commented Mar 25, 2022

@patrickvonplaten: I chose XLM-RoBERTa and it's a sub-class of RoBERTa. The comments in the file for both PyTorch and TF suggests that the superclass should be referred for the appropriate documentation alongside usage examples (XLM-RoBERTa documentations shows RoBERTa examples). Should I still be adding examples for XLM-RoBERTa or should I pick some other model?

@ydshieh
Copy link
Contributor

@ydshieh ydshieh commented Mar 25, 2022

@AbinayaM02 :

Could you show me which line you see suggests that the superclass should be referred for the appropriate documentation in the XLM-RoBERTa model file, please? Thank you :-)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Linked pull requests

Successfully merging a pull request may close this issue.

None yet