Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Estimate # prompts, # input tokens per probe #1071

Open
leondz opened this issue Jan 11, 2025 · 0 comments
Open

Estimate # prompts, # input tokens per probe #1071

leondz opened this issue Jan 11, 2025 · 0 comments
Labels
architecture Architectural upgrades

Comments

@leondz
Copy link
Collaborator

leondz commented Jan 11, 2025

Summary

We need to help estimate the inference load a run may bring.

One interface might be to have a .estimate_prompt_count() method on prompts

We first know the number of prompts after the probe has enqueued its attempts, but before these are sent for inference.

It would be worth making this possible to override for probes that are dynamic/adaptive.

Returned values are estimates, and don't need to be precise, but should be useful for e.g. building a run-level tqdm bar. User should expect that estimates are always within an order of magnitude of accuracy, and often within 10% of the real count.

@leondz leondz added the architecture Architectural upgrades label Jan 11, 2025
@leondz leondz added this to the 25.02 Efficiency milestone Jan 11, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
architecture Architectural upgrades
Projects
None yet
Development

No branches or pull requests

1 participant