Skip to content

Commit b2d1466

Browse files
committed
fix server config
1 parent fe6e866 commit b2d1466

File tree

2 files changed

+3
-3
lines changed

2 files changed

+3
-3
lines changed

cookbook/client/server/megatron/server_config.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -42,7 +42,7 @@ applications:
4242
import_path: sampler
4343
args:
4444
model_id: "ms://Qwen/Qwen3.5-27B" # ModelScope model identifier
45-
nproc_per_node: 8 # Number of GPU processes per node
45+
nproc_per_node: 4 # Number of GPU processes per node
4646
sampler_type: vllm # Inference engine: 'vllm' (fast) or 'torch' (TorchSampler)
4747
engine_args: # vLLM engine-specific settings
4848
max_model_len: 32000 # Maximum sequence length the engine supports
@@ -84,7 +84,7 @@ applications:
8484
model_id: "ms://Qwen/Qwen3.5-27B" # ModelScope model identifier
8585
max_length: 32000 # model max length
8686
max_loras: 5 # model max loras
87-
nproc_per_node: 8 # Number of GPU processes per node
87+
nproc_per_node: 4 # Number of GPU processes per node
8888
device_group:
8989
name: model
9090
ranks: 4 # GPU rank indices

pyproject.toml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ dependencies = [
1010
"datasets>=3.0,<4.0",
1111
"omegaconf>=2.3.0,<3.0.0",
1212
"fastapi",
13-
"modelscope[framework]>=1.34.0",
13+
"modelscope[framework]>=1.35.0",
1414
"safetensors",
1515
"peft>=0.11.0,<=0.19.0",
1616
"transformers",

0 commit comments

Comments
 (0)