-
Notifications
You must be signed in to change notification settings - Fork 14
Expand file tree
/
Copy pathDream_test_flops_script.py
More file actions
281 lines (258 loc) · 12.2 KB
/
Dream_test_flops_script.py
File metadata and controls
281 lines (258 loc) · 12.2 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
import os
import argparse
import torch
from transformers import AutoModel, AutoTokenizer
from calflops import calculate_flops
from dllm_cache.cache import dLLMCacheConfig,dLLMCache
from dllm_cache.hooks import register_cache_Dream
import threading
from queue import Queue
import pandas as pd
import gc
from utils import generate
from dataclasses import asdict
def parse_args():
parser = argparse.ArgumentParser(description="Calculate FLOPs and speed.")
parser.add_argument("--model_name", type=str, default="Dream-org/Dream-v0-Instruct-7B",
help="Name of the pretrained model to load from Hugging Face.")
parser.add_argument("--batch_size", type=int, default=1,
help="Batch size for FLOPs calculation.默认是 1")
parser.add_argument("--prompt_interval_steps", type=int, nargs='+', default=[100],
help="List of steps intervals for caching prompts.")
parser.add_argument("--gen_interval_steps", type=int, nargs='+', default=[7],
help="List of steps intervals for caching generation.")
parser.add_argument("--transfer_ratio", type=float, default=0.25,
help="Transfer ratio for FLOPs calculation in Generate Gap.")
parser.add_argument("--gpu_ids", type=int, nargs='+', default=[0],
help="List of GPU IDs to run the model on (e.g., 0, 1, 2).")
parser.add_argument("--gen_length", type=int, default=256,
help="Generation length")
parser.add_argument("--steps", type=int, default=256,
help="Number of times to repeat the FLOPs calculation with cache for averaging.")
parser.add_argument("--avg_prompt_length", type=int, default=893,
help="Maximum sequence length for FLOPs calculation.")
parser.add_argument("--cache_order", type=int, default=0,
help="Maximum order for caching.")
parser.add_argument("--prompt", type=str, default="*", help="Input prompt")
parser.add_argument("--block_length", type=int, default=8, help="Block length")
parser.add_argument("--temperature", type=float, default=0.0, help="Temperature for generation")
parser.add_argument("--cfg_scale", type=float, default=0.0, help="CFG scale for generation")
parser.add_argument("--remasking", type=str, default="low_confidence", help="Remasking strategy")
parser.add_argument("--num_runs", type=int, default=3, help="Number of runs to average generation time")
parser.add_argument("--is_feature_cache", action="store_true", help="Enable feature cache")
parser.add_argument("--is_cfg_cache", action="store_true", help="Enable cfg cache")
return parser.parse_args()
def convert_to_float(value):
"""Convert calflops string (e.g., '500 GFLOPS' or '7.68 TMACs') to float (in FLOPS or MACs)."""
if isinstance(value, str):
value = value.strip()
if "GFLOPS" in value:
return float(value.replace("GFLOPS", "")) * 1e9
elif "TFLOPS" in value:
return float(value.replace("TFLOPS", "")) * 1e12
elif "MFLOPS" in value:
return float(value.replace("MFLOPS", "")) * 1e6
elif "FLOPS" in value:
return float(value.replace("FLOPS", ""))
elif "GMACs" in value:
return float(value.replace("GMACs", "")) * 1e9
elif "TMACs" in value:
return float(value.replace("TMACs", "")) * 1e12
elif "MMACs" in value:
return float(value.replace("MMACs", "")) * 1e6
elif "MACs" in value:
return float(value.replace("MACs", ""))
else:
return float(value)
return float(value)
def run_experiment(gpu_id, args, prompt_interval_steps, gen_interval_steps, result_lock):
device = torch.device(f"cuda:{gpu_id}" if torch.cuda.is_available() else "cpu")
print(f"Starting experiment on GPU {gpu_id} (device: {device})")
model = AutoModel.from_pretrained(args.model_name, trust_remote_code=True, torch_dtype=torch.bfloat16).to(device).eval()
tokenizer = AutoTokenizer.from_pretrained(args.model_name, trust_remote_code=True,max_length=2048)
# 使用cache
if args.is_feature_cache:
FeatureCache.new_instance(**asdict(FeatureCacheConfig(
prompt_interval_steps=prompt_interval_steps,
gen_interval_steps=gen_interval_steps,
transfer_ratio=args.transfer_ratio,
cfg_interval_steps= 1,
)))
register_cache_Dream(model,"model.layers")
print(f"Cache is enabled, prompt_interval_steps={prompt_interval_steps}, gen_interval_steps={gen_interval_steps},transfer_ratio={args.transfer_ratio}")
else:
prompt_interval_steps = -1
gen_interval_steps = -1
args.transfer_ratio = -1
print("Cache is not enabled")
cache = dLLMCache()
cache.reset_cache(int(args.avg_prompt_length))
flops_cached_total = 0.0
macs_cached_total = 0.0
print("开始计算Flops")
# Repeat FLOPs calculation with cache
for i in range(args.steps):
flops_cached, macs_cached, _ = calculate_flops(
model=model,
input_shape=(1, args.avg_prompt_length + args.gen_length),
print_detailed=False,
print_results=False,
transformer_tokenizer=tokenizer,
output_precision=4
)
flops_cached_val = convert_to_float(flops_cached)
macs_cached_val = convert_to_float(macs_cached)
flops_cached_total += flops_cached_val
macs_cached_total += macs_cached_val
# Calculate averages
avg_flops_cached = flops_cached_total / args.gen_length
avg_macs_cached = macs_cached_total /args.gen_length
total_flops_cached = flops_cached_total
total_macs_cached = macs_cached_total
# 显示结果
result = (
f"\n| Prompt Interval Steps: {prompt_interval_steps} | Gen Interval Steps: {gen_interval_steps}|Transfer_Ratio {args.transfer_ratio}=================\n"
f"With Cache - FLOPs/tokens: {avg_flops_cached / 1e12:.8f} TFLOPS MACs/tokens: {avg_macs_cached / 1e12:.8f} TMACs \n"
f"With Cache - FLOPs: {total_flops_cached / 1e12:.4f} TFLOPS MACs: {total_macs_cached / 1e12:.4f} TMACs \n"
)
print(result)
input_ids_attn_mask = tokenizer(args.prompt, max_length=2048)
input_ids = torch.tensor(input_ids_attn_mask["input_ids"]).to(device).repeat(args.batch_size, args.avg_prompt_length)
attention_mask = torch.tensor(input_ids_attn_mask["attention_mask"]).to(device).repeat(args.batch_size, args.avg_prompt_length).to(torch.bool)
# 这里是把原本一个 token 的 prompt 变成 avg_prompt_length 长度
total_length = input_ids.shape[1] + args.gen_length
print(f"平均提示词的长度:{input_ids.shape[1]},生成部分的的长度{args.gen_length},平均总长度计算长度{total_length}")
if args.is_feature_cache:
FeatureCache.new_instance(**asdict(FeatureCacheConfig(
prompt_interval_steps=prompt_interval_steps,
gen_interval_steps=gen_interval_steps,
transfer_ratio=args.transfer_ratio,
cfg_interval_steps= 1,
)))
register_cache_Dream(model,"model.layers")
print(f"Cache is enabled, prompt_interval_steps={prompt_interval_steps}, gen_interval_steps={gen_interval_steps},transfer_ratio={args.transfer_ratio}")
else:
prompt_interval_steps = -1
gen_interval_steps = -1
args.transfer_ratio = -1
print("Cache is not enabled")
# 预热(避免首次运行的初始化开销)
cache.reset_cache(int(args.avg_prompt_length))
with torch.no_grad():
_ = model.diffusion_generate(
input_ids,
attention_mask=attention_mask,
max_new_tokens= args.gen_length,
output_history=False,
return_dict_in_generate=True,
steps=args.steps,
cfg_scale=args.cfg_scale,
temperature=0.2,
top_p=0.95,
alg="entropy",
alg_temp=0.0,
)
cache.reset_cache(int(args.avg_prompt_length))
times = []
for i in range(args.num_runs):
start_event = torch.cuda.Event(enable_timing=True)
end_event = torch.cuda.Event(enable_timing=True)
with torch.no_grad():
start_event.record()
_ = model.diffusion_generate(
input_ids,
attention_mask=attention_mask,
max_new_tokens= args.gen_length,
output_history=False,
return_dict_in_generate=True,
steps=args.steps,
cfg_scale=args.cfg_scale,
temperature=0.2,
top_p=0.95,
alg="entropy",
alg_temp=0.0,
)
end_event.record()
torch.cuda.synchronize() # CPU与GPU进行同步
elapsed_time_ms = start_event.elapsed_time(end_event) # 求出间隔时间
times.append(elapsed_time_ms / 1000.0) # 转换为秒
print(f"Run {i+1}: {elapsed_time_ms/1000.0:.3f} seconds")
arg_run_time = sum(times) / len(times) / args.batch_size/args.gen_length
total_run_time = sum(times) / len(times) / args.batch_size
token_per_second = args.gen_length / total_run_time
print(f"Average generation time per token over {args.num_runs} runs: {arg_run_time:.8f} seconds")
# Prepare data for CSV
log_entry = {
"avg_run_time_per_token": round(arg_run_time, 8),
"avg_token_per_second":round(token_per_second,8),
"total_run_time":total_run_time,
"avg_flops_per_token": round(avg_flops_cached / 1e12, 8),
"avg_macs_per_token": round(avg_macs_cached / 1e12, 8),
"total_flops":total_flops_cached/ 1e12,
"total_macs":total_macs_cached/ 1e12,
"prompt_interval_steps": prompt_interval_steps,
"gen_interval_steps": gen_interval_steps,
"transfer_ratio": args.transfer_ratio,
"steps": args.steps,
"avg_prompt_length":args.avg_prompt_length
}
# Ensure the ./speed_flops directory exists
output_dir = './speed_flops'
os.makedirs(output_dir, exist_ok=True)
csv_file = os.path.join(output_dir, 'flops_results.csv')
print("GPU")
with torch.profiler.profile(
activities=[torch.profiler.ProfilerActivity.CUDA],
record_shapes=True,
profile_memory=True
) as prof:
with torch.no_grad():
_ = model.diffusion_generate(
input_ids,
attention_mask=attention_mask,
max_new_tokens= args.gen_length,
output_history=False,
return_dict_in_generate=True,
steps=args.steps,
cfg_scale=args.cfg_scale,
temperature=0.2,
top_p=0.95,
alg="entropy",
alg_temp=0.0,
)
print(prof.key_averages().table(sort_by="self_cuda_memory_usage", row_limit=10))
# Write to CSV with thread-safe locking
with result_lock:
df = pd.DataFrame([log_entry])
if os.path.exists(csv_file):
df.to_csv(csv_file, mode='a', header=False, index=False)
else:
df.to_csv(csv_file, mode='w', header=True, index=False)
del model
torch.cuda.empty_cache() # Clear GPU memory after experiment
gc.collect()
def thread_worker(gpu_queue, args, prompt_interval_steps, gen_interval_steps, result_lock):
gpu_id = gpu_queue.get() # 获取空闲 GPU
try:
run_experiment(gpu_id, args, prompt_interval_steps, gen_interval_steps, result_lock)
finally:
gpu_queue.put(gpu_id) # 归还 GPU
def main():
args = parse_args()
available_gpus = torch.cuda.device_count()
for gpu_id in args.gpu_ids:
if gpu_id >= available_gpus:
raise ValueError(f"GPU ID {gpu_id} is not available. Only {available_gpus} GPUs are detected.")
gpu_queue = Queue()
for gpu_id in args.gpu_ids:
gpu_queue.put(gpu_id)
result_lock = threading.Lock()
for prompt_steps in args.prompt_interval_steps:
for gen_steps in args.gen_interval_steps:
thread = threading.Thread(
target=thread_worker,
args=(gpu_queue, args, prompt_steps, gen_steps, result_lock)
)
thread.start()
if __name__ == "__main__":
main()