Question about Context Trimming and Token Limits #235
Unanswered
e-razdumina
asked this question in
Q&A
Replies: 1 comment
-
|
Hi @e-razdumina thanks for the question and the kind words! I don't have any particularly novel insights to share on this topic, I know there are a lot of guides and discussions online and ultimately I think it depends on the particular application. In terms of how to handle it in the project:
Hope it helps! |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi! Thank you for the amazing work - it’s been incredibly helpful and inspiring. 🙌
I’ve been exploring a question and would love to get your thoughts:
What’s your take on trimming context or bounding models with max_tokens to manage performance and response quality?
Also, I wasn’t quite sure - is this already handled somewhere in the project, or is it something that might need tuning on the integration side?
Looking forward to your insights!
Beta Was this translation helpful? Give feedback.
All reactions