bezzam HF Staff commited on
Commit
bbceaa8
·
verified ·
1 Parent(s): 0a09522

Increase `max_length` of generated output.

Browse files

Otherwise "Voice chat interface" example had following error:
```
ValueError: Input length of input_ids is 599, but `max_length` is set to 256. This can lead to unexpected behavior. You should consider increasing `max_length` or, better yet, setting `max_new_tokens`.
```

And "Audio analysis inference" example had following error:
```
ValueError: Input length of input_ids is 351, but `max_length` is set to 256. This can lead to unexpected behavior. You should consider increasing `max_length` or, better yet, setting `max_new_tokens`.
```

Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -71,7 +71,7 @@ for message in conversation:
71
  inputs = processor(text=text, audios=audios, return_tensors="pt", padding=True)
72
  inputs.input_ids = inputs.input_ids.to("cuda")
73
 
74
- generate_ids = model.generate(**inputs, max_length=256)
75
  generate_ids = generate_ids[:, inputs.input_ids.size(1):]
76
 
77
  response = processor.batch_decode(generate_ids, skip_special_tokens=True, clean_up_tokenization_spaces=False)[0]
@@ -119,7 +119,7 @@ for message in conversation:
119
  inputs = processor(text=text, audios=audios, return_tensors="pt", padding=True)
120
  inputs.input_ids = inputs.input_ids.to("cuda")
121
 
122
- generate_ids = model.generate(**inputs, max_length=256)
123
  generate_ids = generate_ids[:, inputs.input_ids.size(1):]
124
 
125
  response = processor.batch_decode(generate_ids, skip_special_tokens=True, clean_up_tokenization_spaces=False)[0]
 
71
  inputs = processor(text=text, audios=audios, return_tensors="pt", padding=True)
72
  inputs.input_ids = inputs.input_ids.to("cuda")
73
 
74
+ generate_ids = model.generate(**inputs, max_length=1024)
75
  generate_ids = generate_ids[:, inputs.input_ids.size(1):]
76
 
77
  response = processor.batch_decode(generate_ids, skip_special_tokens=True, clean_up_tokenization_spaces=False)[0]
 
119
  inputs = processor(text=text, audios=audios, return_tensors="pt", padding=True)
120
  inputs.input_ids = inputs.input_ids.to("cuda")
121
 
122
+ generate_ids = model.generate(**inputs, max_length=512)
123
  generate_ids = generate_ids[:, inputs.input_ids.size(1):]
124
 
125
  response = processor.batch_decode(generate_ids, skip_special_tokens=True, clean_up_tokenization_spaces=False)[0]