diff --git a/model_cards/gaochangkuan/model_dir/README.md b/model_cards/gaochangkuan/model_dir/README.md new file mode 100644 index 0000000000..41d3e81ebf --- /dev/null +++ b/model_cards/gaochangkuan/model_dir/README.md @@ -0,0 +1,66 @@ +## Generating Chinese poetry by topic. + +```python +from transformers import * + +tokenizer = BertTokenizer.from_pretrained("gaochangkuan/model_dir") + +model = AutoModelWithLMHead.from_pretrained("gaochangkuan/model_dir") + + +prompt= '''田园躬耕''' + +length= 84 +stop_token='' + +temperature = 1.2 + +repetition_penalty=1.3 + +k= 30 +p= 0.95 + +device ='cuda' +seed=2020 +no_cuda=False + +prompt_text = prompt if prompt else input("Model prompt >>> ") + +encoded_prompt = tokenizer.encode( + ''+prompt_text+'', + add_special_tokens=False, + return_tensors="pt" + ) + +encoded_prompt = encoded_prompt.to(device) + +output_sequences = model.generate( + input_ids=encoded_prompt, + max_length=length, + min_length=10, + do_sample=True, + early_stopping=True, + num_beams=10, + temperature=temperature, + top_k=k, + top_p=p, + repetition_penalty=repetition_penalty, + bad_words_ids=None, + bos_token_id=tokenizer.bos_token_id, + pad_token_id=tokenizer.pad_token_id, + eos_token_id=tokenizer.eos_token_id, + length_penalty=1.2, + no_repeat_ngram_size=2, + num_return_sequences=1, + attention_mask=None, + decoder_start_token_id=tokenizer.bos_token_id,) + + + generated_sequence = output_sequences[0].tolist() +text = tokenizer.decode(generated_sequence) + + +text = text[: text.find(stop_token) if stop_token else None] + +print(''.join(text).replace(' ','').replace('','').replace('','')) +```