fb2b45e562
* add minimal working gpt2 tokenizer * graph mode and output equivalence tests working * not today tensorflow. serialization test passing! * fix style, documentation, docstrings and all that jazz * passing consistency checks * move keras nlp to tf dependencies * fix tf modeling utils and gpt2 attention to enable compiling * fix (I hope) keras nlp dependencies * rever changes on generation * remove debug prints * remove redundant tf dummy objects * add from config, get config and max length settings to address review * let flake ignore the error on distillation you are welcome * test from config * add padding test * address sgugger review |
||
---|---|---|
.. | ||
adversarial | ||
bert-loses-patience | ||
bertabs | ||
bertology | ||
codeparrot | ||
decision_transformer | ||
deebert | ||
distillation | ||
fsner | ||
information-gain-filtration | ||
jax-projects | ||
layoutlmv3 | ||
longform-qa | ||
luke | ||
lxmert | ||
mlm_wwm | ||
mm-imdb | ||
movement-pruning | ||
onnx/summarization | ||
performer | ||
pplm | ||
quantization-qdqbert | ||
rag | ||
rag-end2end-retriever | ||
robust-speech-event | ||
self-training-text-classification | ||
seq2seq-distillation | ||
tapex | ||
visual_bert | ||
wav2vec2 | ||
xtreme-s | ||
zero-shot-distillation | ||
README.md |
README.md
Research projects
This folder contains various research projects using 🤗 Transformers. They are not maintained and require a specific version of 🤗 Transformers that is indicated in the requirements file of each folder. Updating them to the most recent version of the library will require some work.
To use any of them, just run the command
pip install -r requirements.txt
inside the folder of your choice.
If you need help with any of those, contact the author(s), indicated at the top of the README
of each folder.