Spaces:
Sleeping
Sleeping
Ctrl+K
Reduce max_new_tokens in model generation from 2048 to 512 in app.py to optimize response length and improve performance. This change aims to enhance the efficiency of the conversation prediction function.
02386cb