Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Resolve confusion about "batching" support #174

Open
markmc opened this issue Jul 19, 2024 · 2 comments
Open

Resolve confusion about "batching" support #174

markmc opened this issue Jul 19, 2024 · 2 comments
Labels
question Further information is requested

Comments

@markmc
Copy link
Contributor

markmc commented Jul 19, 2024

Capturing some observations from #157

  1. Pipeline has "batching" support - it can shard the dataset and spawn an instance of the pipeline for each shard - batch_num_workers and batch_size
  2. LLMBlock has "batching" support - it can request multiple chat completions from the OpenAI server using the n argument - num_instructions_to_generate

In ilab we disable (1) with llama-cpp by passing batch_size=None - see instructlab/instructlab#346

In LLMBlock we disable (2) with llama-cpp with the server_supports_batched which checks whether the n argument works

Resolve:

  • Do we want to call both of these "batching"
  • Do we want to different ways of handling backend-specific capabilities?
  • Should the library be trying to probe the backend for its capabilities, or should the library user give it information about the backend?
  • server_supports_batched should be a property on PipelineContext, not something we set on the OpenAI client object
@nathan-weinberg nathan-weinberg added the question Further information is requested label Aug 20, 2024
Copy link

This issue has been automatically marked as stale because it has not had activity within 90 days. It will be automatically closed if no further activity occurs within 30 days.

@github-actions github-actions bot added the stale label Nov 20, 2024
@bbrowning
Copy link
Contributor

This has confused me more than once - some tech debt to tackle here for sure.

@github-actions github-actions bot removed the stale label Nov 21, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

3 participants