![]() * added exllama kernels support for awq models * doc * style * Update src/transformers/modeling_utils.py Co-authored-by: Marc Sun <57196510+SunMarc@users.noreply.github.com> * refactor * moved exllama post init to after device dispatching * bump autoawq version * added exllama test * style * configurable exllama kernels * copy exllama_config from gptq * moved exllama version check to post init * moved to quantization dockerfile --------- Co-authored-by: Marc Sun <57196510+SunMarc@users.noreply.github.com> |
||
---|---|---|
.. | ||
transformers-all-latest-gpu | ||
transformers-doc-builder | ||
transformers-gpu | ||
transformers-past-gpu | ||
transformers-pytorch-amd-gpu | ||
transformers-pytorch-deepspeed-amd-gpu | ||
transformers-pytorch-deepspeed-latest-gpu | ||
transformers-pytorch-deepspeed-nightly-gpu | ||
transformers-pytorch-gpu | ||
transformers-pytorch-tpu | ||
transformers-quantization-latest-gpu | ||
transformers-tensorflow-gpu |