check llama
This commit is contained in:
parent
f1a37d9a3c
commit
4f0791f365
|
@ -66,7 +66,7 @@ jobs:
|
|||
working-directory: /transformers
|
||||
run: |
|
||||
cat doc_tests.txt
|
||||
python3 -m pytest -v --make-reports doc_tests_gpu_${{ env.split_keys }} --doctest-modules $(cat doc_tests.txt) -sv --doctest-continue-on-failure --doctest-glob="*.md"
|
||||
python3 temp_llama.py
|
||||
|
||||
- name: Failure short reports
|
||||
if: ${{ failure() }}
|
||||
|
|
|
@ -0,0 +1,4 @@
|
|||
from transformers import LlamaForCausalLM, CodeLlamaTokenizer
|
||||
|
||||
# tokenizer = CodeLlamaTokenizer.from_pretrained("codellama/CodeLlama-7b-hf")
|
||||
model = LlamaForCausalLM.from_pretrained("codellama/CodeLlama-7b-hf")
|
|
@ -71,6 +71,8 @@ if __name__ == "__main__":
|
|||
|
||||
sorted_file_dirs = sorted(refined_test_collection_map.keys())
|
||||
|
||||
test_collection_map = ["src/transformers/models/siglip"]
|
||||
|
||||
test_collection_map = {}
|
||||
for file_dir in sorted_file_dirs:
|
||||
test_collection_map[file_dir] = refined_test_collection_map[file_dir]
|
||||
|
|
Loading…
Reference in New Issue