-
Notifications
You must be signed in to change notification settings - Fork 204
Open
Labels
status: triagethis issue has not been evaluated yetthis issue has not been evaluated yet
Description
Describe the bug
When sending AI jobs for expensive models (such as DeepSeek), or in the case of LLM pipeline, sending large max_tokens parameter such as 163K tokens, it causes a lot of payment tickets to be sent at once.
The Orch will show this message.
Error receiving ticket sessionID=33_meta-llama/Meta-Llama-3.1-8B-Instruct recipientRandHash=7905016d8d201e4bb0d13f78234e107018b2effe42343325691a55844a1d54cf senderNonce=178: invalid ticket senderNonce: too many values sender=0x5bE44e23041E93CDF9bCd5A0968524e104e38ae1 nonce=178
There is a nonce cap of 150 currently. We need to allow infinite nonce count to be accepted by the Orch, or some way to manage this limit.
For instance, if LLM context windows keep increasing or price keep getting higher, which I believe it will, we need higher throughput of tickets to be redeemed.
This is also prevalent when multiple jobs are sent at the same time, the ticket nonce stacks up and will reach the limit quickly.
To Reproduce
Steps to reproduce the behavior:
- Start AI Gateway
- Start Orchestrator with 7 USD per 1 million tokens
- Send LLM request with 163K
max_tokensparameter - See error
Metadata
Metadata
Assignees
Labels
status: triagethis issue has not been evaluated yetthis issue has not been evaluated yet