File tree Expand file tree Collapse file tree 2 files changed +3
-3
lines changed
cookbook/client/server/megatron Expand file tree Collapse file tree 2 files changed +3
-3
lines changed Original file line number Diff line number Diff line change @@ -42,7 +42,7 @@ applications:
4242 import_path : sampler
4343 args :
4444 model_id : " ms://Qwen/Qwen3.5-27B" # ModelScope model identifier
45- nproc_per_node : 8 # Number of GPU processes per node
45+ nproc_per_node : 4 # Number of GPU processes per node
4646 sampler_type : vllm # Inference engine: 'vllm' (fast) or 'torch' (TorchSampler)
4747 engine_args : # vLLM engine-specific settings
4848 max_model_len : 32000 # Maximum sequence length the engine supports
@@ -84,7 +84,7 @@ applications:
8484 model_id : " ms://Qwen/Qwen3.5-27B" # ModelScope model identifier
8585 max_length : 32000 # model max length
8686 max_loras : 5 # model max loras
87- nproc_per_node : 8 # Number of GPU processes per node
87+ nproc_per_node : 4 # Number of GPU processes per node
8888 device_group :
8989 name : model
9090 ranks : 4 # GPU rank indices
Original file line number Diff line number Diff line change @@ -10,7 +10,7 @@ dependencies = [
1010 " datasets>=3.0,<4.0" ,
1111 " omegaconf>=2.3.0,<3.0.0" ,
1212 " fastapi" ,
13- " modelscope[framework]>=1.34 .0" ,
13+ " modelscope[framework]>=1.35 .0" ,
1414 " safetensors" ,
1515 " peft>=0.11.0,<=0.19.0" ,
1616 " transformers" ,
You can’t perform that action at this time.
0 commit comments