Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Feature] Monitor token consumption #806

Merged
merged 30 commits into from
May 9, 2024

Conversation

riedgar-ms
Copy link
Collaborator

@riedgar-ms riedgar-ms commented May 7, 2024

Count tokens fed into & retrieved from the underlying LLM. This is just for Tranformer and LLamaCpp models at the current time.

This hooks into get_logits() which is the method which concrete Engines must implement. We don't attempt to distinguish between 'forced' and 'prompt' tokens. Between non-unique tokenisations, a variety of tokenisers and token healing, trying to count tokens exactly is decidedly non-trivial.

guidance/models/_model.py Outdated Show resolved Hide resolved
tests/library/test_gen.py Outdated Show resolved Hide resolved
@codecov-commenter
Copy link

codecov-commenter commented May 7, 2024

Codecov Report

Attention: Patch coverage is 81.25000% with 3 lines in your changes are missing coverage. Please review.

Project coverage is 62.14%. Comparing base (7b4d85f) to head (d860cb2).

Files Patch % Lines
guidance/models/transformers/_transformers.py 60.00% 2 Missing ⚠️
guidance/models/_model.py 80.00% 1 Missing ⚠️

❗ Your organization needs to install the Codecov GitHub app to enable full functionality.

Additional details and impacted files
@@            Coverage Diff             @@
##             main     #806      +/-   ##
==========================================
+ Coverage   56.72%   62.14%   +5.42%     
==========================================
  Files          56       57       +1     
  Lines        4159     4174      +15     
==========================================
+ Hits         2359     2594     +235     
+ Misses       1800     1580     -220     

☔ View full report in Codecov by Sentry.
📢 Have feedback on the report? Share it here.

@riedgar-ms riedgar-ms changed the title [WIP] [Feature] Monitor token consumption [Feature] Monitor token consumption May 9, 2024
@paulbkoch
Copy link
Collaborator

paulbkoch commented May 9, 2024

LGTM

@paulbkoch paulbkoch merged commit 5ad2304 into guidance-ai:main May 9, 2024
124 checks passed
@riedgar-ms riedgar-ms deleted the riedgar-ms/model-metrics-01 branch August 26, 2024 16:20
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants