Skip to content

Commit d1c2d8e

Browse files
committed
fix docs
1 parent 9342d74 commit d1c2d8e

File tree

39 files changed

+142
-109
lines changed

39 files changed

+142
-109
lines changed

README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -184,7 +184,7 @@ def train():
184184
# 1000 samples
185185
dataset = Dataset(dataset_meta=DatasetMeta('ms://swift/self-cognition', data_slice=range(1000)))
186186
# Set template to prepare encoding
187-
dataset.set_template('Template', model_id=base_model)
187+
dataset.set_template('Qwen3_5Template', model_id=base_model)
188188
# Preprocess the dataset to standard format
189189
dataset.map(SelfCognitionProcessor('twinkle LLM', 'ModelScope Community'))
190190
# Encode dataset
@@ -242,7 +242,7 @@ api_key='your-api-key'
242242

243243
# Use twinkle dataset to load the data
244244
dataset = Dataset(dataset_meta=DatasetMeta('ms://swift/self-cognition', data_slice=range(500)))
245-
dataset.set_template('Template', model_id=base_model, max_length=256)
245+
dataset.set_template('Qwen3_5Template', model_id=base_model, max_length=256)
246246
dataset.map(SelfCognitionProcessor('twinkle Model', 'ModelScope Team'), load_from_cache_file=False)
247247
dataset.encode(batched=True, load_from_cache_file=False)
248248
dataloader = DataLoader(dataset=dataset, batch_size=8)

README_ZH.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -166,7 +166,7 @@ def train():
166166
# 1000 samples
167167
dataset = Dataset(dataset_meta=DatasetMeta('ms://swift/self-cognition', data_slice=range(1000)))
168168
# Set template to prepare encoding
169-
dataset.set_template('Template', model_id=base_model)
169+
dataset.set_template('Qwen3_5Template', model_id=base_model)
170170
# Preprocess the dataset to standard format
171171
dataset.map(SelfCognitionProcessor('twinkle LLM', 'ModelScope Community'))
172172
# Encode dataset
@@ -224,7 +224,7 @@ api_key='your-api-key'
224224

225225
# Use twinkle dataset to load the data
226226
dataset = Dataset(dataset_meta=DatasetMeta('ms://swift/self-cognition', data_slice=range(500)))
227-
dataset.set_template('Template', model_id=base_model, max_length=256)
227+
dataset.set_template('Qwen3_5Template', model_id=base_model, max_length=256)
228228
dataset.map(SelfCognitionProcessor('twinkle Model', 'ModelScope Team'), load_from_cache_file=False)
229229
dataset.encode(batched=True, load_from_cache_file=False)
230230
dataloader = DataLoader(dataset=dataset, batch_size=8)

cookbook/client/tinker/modelscope/self_cognition.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -34,7 +34,7 @@ def train():
3434
dataset = Dataset(dataset_meta=DatasetMeta('ms://swift/self-cognition', data_slice=range(500)))
3535

3636
# Apply the chat template matching the base model (max 256 tokens per sample)
37-
dataset.set_template('Template', model_id=f'ms://{base_model}', max_length=256)
37+
dataset.set_template('Qwen3_5Template', model_id=f'ms://{base_model}', max_length=256)
3838

3939
# Replace placeholder names with custom model/author identity
4040
dataset.map(SelfCognitionProcessor('twinkle模型', 'twinkle团队'), load_from_cache_file=False)

cookbook/client/tinker/modelscope/short_math_grpo.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -182,7 +182,7 @@ def create_math_dataset():
182182
data_slice=range(DATA_NUM),
183183
)
184184
dataset = Dataset(meta)
185-
dataset.set_template('Template', model_id=BASE_MODEL, max_length=4096, truncation_strategy='delete')
185+
dataset.set_template('Qwen3_5Template', model_id=BASE_MODEL, max_length=4096, truncation_strategy='delete')
186186
dataset.map(MathPreprocessor())
187187
dataset.filter(lambda row: bool(row['messages']))
188188
dataset.encode(add_generation_prompt=True)

cookbook/client/tinker/self_host/self_cognition.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -36,7 +36,7 @@ def train():
3636
dataset = Dataset(dataset_meta=DatasetMeta('ms://swift/self-cognition', data_slice=range(500)))
3737

3838
# Apply the chat template matching the base model (max 256 tokens per sample)
39-
dataset.set_template('Template', model_id=f'ms://{base_model}', max_length=256)
39+
dataset.set_template('Qwen3_5Template', model_id=f'ms://{base_model}', max_length=256)
4040

4141
# Replace placeholder names with custom model/author identity
4242
dataset.map(SelfCognitionProcessor('twinkle模型', 'twinkle团队'), load_from_cache_file=False)

cookbook/client/tinker/self_host/short_math_grpo.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -182,7 +182,7 @@ def create_math_dataset():
182182
data_slice=range(DATA_NUM),
183183
)
184184
dataset = Dataset(meta)
185-
dataset.set_template('Template', model_id=BASE_MODEL, max_length=4096, truncation_strategy='delete')
185+
dataset.set_template('Qwen3_5Template', model_id=BASE_MODEL, max_length=4096, truncation_strategy='delete')
186186
dataset.map(MathPreprocessor())
187187
dataset.filter(lambda row: bool(row['messages']))
188188
dataset.encode(add_generation_prompt=True)

cookbook/client/twinkle/modelscope/self_congnition.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -52,7 +52,7 @@ def train():
5252
dataset = Dataset(dataset_meta=DatasetMeta('ms://swift/self-cognition', data_slice=range(500)))
5353

5454
# Apply a chat template so the data matches the model's expected input format
55-
dataset.set_template('Template', model_id=f'ms://{base_model}', max_length=512)
55+
dataset.set_template('Qwen3_5Template', model_id=f'ms://{base_model}', max_length=512)
5656

5757
# Replace placeholder names in the dataset with custom model/author names
5858
dataset.map('SelfCognitionProcessor', init_args={'model_name': 'twinkle模型', 'model_author': 'ModelScope社区'})
@@ -77,7 +77,7 @@ def train():
7777
model.add_adapter_to_model('default', lora_config, gradient_accumulation_steps=2)
7878

7979
# Set the same chat template used during data preprocessing
80-
model.set_template('Template')
80+
model.set_template('Qwen3_5Template')
8181

8282
# Set the input processor (pads sequences on the right side)
8383
model.set_processor('InputProcessor', padding_side='right')

cookbook/client/twinkle/self_host/grpo.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -55,7 +55,7 @@
5555

5656
def create_gsm8k_dataset():
5757
dataset = Dataset(DatasetMeta('ms://modelscope/gsm8k', subset_name='main', split='train'))
58-
dataset.set_template('Template', model_id=MODEL_ID, max_length=2048)
58+
dataset.set_template('Qwen3_5Template', model_id=MODEL_ID, max_length=2048)
5959
dataset.map('GSM8KProcessor')
6060
dataset.encode(add_generation_prompt=True)
6161
return dataset
@@ -112,11 +112,11 @@ def train():
112112

113113
# Set processor and template for encoding inputs
114114
model.set_processor('InputProcessor')
115-
model.set_template('Template', model_id=MODEL_ID)
115+
model.set_template('Qwen3_5Template', model_id=MODEL_ID)
116116

117117
# Step 4: Configure the sampler
118118
sampler = vLLMSampler(model_id=MODEL_ID)
119-
sampler.set_template('Template', model_id=MODEL_ID)
119+
sampler.set_template('Qwen3_5Template', model_id=MODEL_ID)
120120

121121
# Step 5: Setup metrics and advantage function
122122
advantage_fn = GRPOAdvantage()

cookbook/client/twinkle/self_host/sample.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -42,7 +42,7 @@ def sample():
4242
sampler = vLLMSampler(model_id=MODEL_ID)
4343

4444
# Step 4: Set the chat template so the sampler can encode Trajectory inputs
45-
sampler.set_template('Template', model_id=MODEL_ID)
45+
sampler.set_template('Qwen3_5Template', model_id=MODEL_ID)
4646

4747
# Step 5: Prepare inputs as Trajectory dicts (messages format)
4848
# Each trajectory is a conversation with system and user messages

cookbook/client/twinkle/self_host/self_congnition.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -59,7 +59,7 @@ def train():
5959
dataset = Dataset(dataset_meta=DatasetMeta('ms://swift/self-cognition', data_slice=range(500)))
6060

6161
# Apply a chat template so the data matches the model's expected input format
62-
dataset.set_template('Template', model_id=f'ms://{base_model}', max_length=512)
62+
dataset.set_template('Qwen3_5Template', model_id=f'ms://{base_model}', max_length=512)
6363

6464
# Replace placeholder names in the dataset with custom model/author names
6565
dataset.map('SelfCognitionProcessor', init_args={'model_name': 'twinkle模型', 'model_author': 'ModelScope社区'})
@@ -84,7 +84,7 @@ def train():
8484
model.add_adapter_to_model('default', lora_config, gradient_accumulation_steps=2)
8585

8686
# Set the same chat template used during data preprocessing
87-
model.set_template('Template')
87+
model.set_template('Qwen3_5Template')
8888

8989
# Set the input processor (pads sequences on the right side)
9090
model.set_processor('InputProcessor', padding_side='right')

0 commit comments

Comments
 (0)