Skip to content

Commit 9a7b4b3

Browse files
authored
[Doc] Fix README and change default for accelerator_type (#86)
1 parent bf833f2 commit 9a7b4b3

File tree

2 files changed

+9
-2
lines changed

2 files changed

+9
-2
lines changed

skythought/evals/README.md

Lines changed: 8 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -43,6 +43,13 @@ skythought evaluate --model Qwen/QwQ-32B-Preview --task aime --backend ray --bac
4343

4444
By default, we make use of the configuration in [ray_configs/ray_config.yaml](./ray_configs/ray_config.yaml). You can also customize the following parameters for ray:
4545

46+
- `tensor_parallel_size`: Tensor parallel size per replica. Defaults to 4.
47+
- `accelerator_type`: GPU accelerator type. See [the list of available types](https://docs.ray.io/en/latest/ray-core/accelerator-types.html) for more information. Defaults to None, which means any available GPUs in the Ray cluster will be used.
48+
- `num_replicas`: Number of model replicas to use for inference. Defaults to 2.
49+
- `batch_size`: Batch size per model replica for inference.
50+
- `gpu_memory_utilization`: Fraction of GPU memory allocated to the model executor in vLLM. Defaults to 0.9.
51+
- `dtype`: Data type used for inference. Defaults to "auto".
52+
4653

4754
### Optimized settings for 32B and 7B models
4855

@@ -54,7 +61,7 @@ For 32B models, we recommend using the default backend configuration for best pe
5461
skythought evaluate --model Qwen/QwQ-32B-Preview --task aime24 --backend ray --result-dir ./
5562
```
5663

57-
For 7B models, we recommend using `tensor_parallel_size=1` and `num_replicas=8` for best performance. FOr example, the previous command will change to:
64+
For 7B models, we recommend using `tensor_parallel_size=1` and `num_replicas=8` for best performance. For example, the previous command will change to:
5865

5966
```shell
6067
skythought evaluate --model Qwen/Qwen2-7B-Instruct --task math500 --backend ray --backend-args tensor_parallel_size=1,num_replicas=8 --result-dir ./

skythought/evals/ray_configs/ray_config.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
llm_engine: vllm # currently only vllm supported
2-
accelerator_type: H100 # accelerator name as specified here: https://docs.ray.io/en/master/ray-core/accelerator-types.html#accelerator-types
2+
accelerator_type: null # accelerator name as specified here: https://docs.ray.io/en/master/ray-core/accelerator-types.html#accelerator-types
33
engine_kwargs: # vllm engine kwargs
44
tensor_parallel_size: 4
55
gpu_memory_utilization: 0.9

0 commit comments

Comments
 (0)