You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This issue has been automatically marked as stale because it has not had activity within 90 days. It will be automatically closed if no further activity occurs within 30 days.
Capturing some observations from #157
Pipeline
has "batching" support - it can shard the dataset and spawn an instance of the pipeline for each shard -batch_num_workers
andbatch_size
LLMBlock
has "batching" support - it can request multiple chat completions from the OpenAI server using then
argument -num_instructions_to_generate
In
ilab
we disable (1) with llama-cpp by passingbatch_size=None
- see instructlab/instructlab#346In
LLMBlock
we disable (2) with llama-cpp with theserver_supports_batched
which checks whether then
argument worksResolve:
server_supports_batched
should be a property onPipelineContext
, not something we set on the OpenAI client objectThe text was updated successfully, but these errors were encountered: