Spaces:
Runtime error
Runtime error
Final app for version 2 with Flagging added
Browse files
app.py
CHANGED
|
@@ -7,16 +7,16 @@ model_name = 'anugrahap/gpt2-indo-textgen'
|
|
| 7 |
HF_TOKEN = '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'
|
| 8 |
hf_writer = gr.HuggingFaceDatasetSaver(HF_TOKEN, "output-gpt2-indo-textgen")
|
| 9 |
|
| 10 |
-
#
|
| 11 |
-
|
| 12 |
-
|
| 13 |
-
|
| 14 |
-
|
| 15 |
|
| 16 |
-
#
|
| 17 |
-
|
| 18 |
|
| 19 |
-
generator = pipeline('text-generation', model=
|
| 20 |
|
| 21 |
# create the decoder parameter to generate the text
|
| 22 |
def single_generation(text,min_length,max_length,temperature,top_k,top_p,num_beams,repetition_penalty,do_sample):
|
|
|
|
| 7 |
HF_TOKEN = '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'
|
| 8 |
hf_writer = gr.HuggingFaceDatasetSaver(HF_TOKEN, "output-gpt2-indo-textgen")
|
| 9 |
|
| 10 |
+
# define the tokenization method
|
| 11 |
+
tokenizer = AutoTokenizer.from_pretrained(model_name,
|
| 12 |
+
model_max_length=1e30,
|
| 13 |
+
padding_side='right',
|
| 14 |
+
return_tensors='pt')
|
| 15 |
|
| 16 |
+
# add the EOS token as PAD token to avoid warnings
|
| 17 |
+
model = AutoModelForCausalLM.from_pretrained(model_name, pad_token_id=tokenizer.eos_token_id)
|
| 18 |
|
| 19 |
+
generator = pipeline('text-generation', model=model, tokenizer=tokenizer)
|
| 20 |
|
| 21 |
# create the decoder parameter to generate the text
|
| 22 |
def single_generation(text,min_length,max_length,temperature,top_k,top_p,num_beams,repetition_penalty,do_sample):
|