|
{ |
|
"config": { |
|
"name": "pytorch_generate", |
|
"backend": { |
|
"name": "pytorch", |
|
"version": "2.3.0+cu121", |
|
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", |
|
"task": "text-generation", |
|
"library": "transformers", |
|
"model": "google/gemma-2b", |
|
"processor": "google/gemma-2b", |
|
"device": "cuda", |
|
"device_ids": "0", |
|
"seed": 42, |
|
"inter_op_num_threads": null, |
|
"intra_op_num_threads": null, |
|
"model_kwargs": {}, |
|
"processor_kwargs": {}, |
|
"hub_kwargs": {}, |
|
"no_weights": true, |
|
"device_map": null, |
|
"torch_dtype": "float16", |
|
"eval_mode": true, |
|
"to_bettertransformer": false, |
|
"low_cpu_mem_usage": null, |
|
"attn_implementation": null, |
|
"cache_implementation": null, |
|
"autocast_enabled": false, |
|
"autocast_dtype": null, |
|
"torch_compile": false, |
|
"torch_compile_target": "forward", |
|
"torch_compile_config": { |
|
"backend": "inductor", |
|
"mode": "reduce-overhead", |
|
"fullgraph": true |
|
}, |
|
"quantization_scheme": null, |
|
"quantization_config": {}, |
|
"deepspeed_inference": false, |
|
"deepspeed_inference_config": {}, |
|
"peft_type": null, |
|
"peft_config": {} |
|
}, |
|
"scenario": { |
|
"name": "inference", |
|
"_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario", |
|
"iterations": 2, |
|
"duration": 0, |
|
"warmup_runs": 10, |
|
"input_shapes": { |
|
"batch_size": 1, |
|
"num_choices": 2, |
|
"sequence_length": 7 |
|
}, |
|
"new_tokens": null, |
|
"latency": true, |
|
"memory": true, |
|
"energy": false, |
|
"forward_kwargs": {}, |
|
"generate_kwargs": { |
|
"max_new_tokens": 128, |
|
"min_new_tokens": 128, |
|
"do_sample": false |
|
}, |
|
"call_kwargs": {} |
|
}, |
|
"launcher": { |
|
"name": "process", |
|
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", |
|
"device_isolation": true, |
|
"device_isolation_action": "warn", |
|
"start_method": "spawn" |
|
}, |
|
"environment": { |
|
"cpu": " AMD EPYC 7R32", |
|
"cpu_count": 16, |
|
"cpu_ram_mb": 66697.29792, |
|
"system": "Linux", |
|
"machine": "x86_64", |
|
"platform": "Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.29", |
|
"processor": "x86_64", |
|
"python_version": "3.8.10", |
|
"gpu": [ |
|
"NVIDIA A10G" |
|
], |
|
"gpu_count": 1, |
|
"gpu_vram_mb": 24146608128, |
|
"optimum_benchmark_version": "0.2.1", |
|
"optimum_benchmark_commit": null, |
|
"transformers_version": "4.42.0.dev0", |
|
"transformers_commit": "25245ec26dc29bcf6102e1b4ddd0dfd02e720cf5", |
|
"accelerate_version": "0.32.0.dev0", |
|
"accelerate_commit": null, |
|
"diffusers_version": null, |
|
"diffusers_commit": null, |
|
"optimum_version": "1.21.0.dev0", |
|
"optimum_commit": null, |
|
"timm_version": "0.9.16", |
|
"timm_commit": null, |
|
"peft_version": "0.11.2.dev0", |
|
"peft_commit": null |
|
} |
|
}, |
|
"report": { |
|
"prefill": { |
|
"memory": { |
|
"unit": "MB", |
|
"max_ram": 1723.301888, |
|
"max_global_vram": 6793.199616, |
|
"max_process_vram": 0.0, |
|
"max_reserved": 6146.752512, |
|
"max_allocated": 5034.576896 |
|
}, |
|
"latency": { |
|
"unit": "s", |
|
"count": 2, |
|
"total": 0.04076508712768555, |
|
"mean": 0.020382543563842774, |
|
"stdev": 0.0007646236419677731, |
|
"p50": 0.020382543563842774, |
|
"p90": 0.02099424247741699, |
|
"p95": 0.02107070484161377, |
|
"p99": 0.021131874732971193, |
|
"values": [ |
|
0.021147167205810547, |
|
0.019617919921875 |
|
] |
|
}, |
|
"throughput": { |
|
"unit": "tokens/s", |
|
"value": 343.43113155011315 |
|
}, |
|
"energy": null, |
|
"efficiency": null |
|
}, |
|
"decode": { |
|
"memory": { |
|
"unit": "MB", |
|
"max_ram": 1723.379712, |
|
"max_global_vram": 6795.296768, |
|
"max_process_vram": 0.0, |
|
"max_reserved": 6148.849664, |
|
"max_allocated": 5034.577408 |
|
}, |
|
"latency": { |
|
"unit": "s", |
|
"count": 2, |
|
"total": 4.85029150390625, |
|
"mean": 2.425145751953125, |
|
"stdev": 0.008099121093749995, |
|
"p50": 2.425145751953125, |
|
"p90": 2.431625048828125, |
|
"p95": 2.4324349609375, |
|
"p99": 2.433082890625, |
|
"values": [ |
|
2.417046630859375, |
|
2.433244873046875 |
|
] |
|
}, |
|
"throughput": { |
|
"unit": "tokens/s", |
|
"value": 52.3679865004892 |
|
}, |
|
"energy": null, |
|
"efficiency": null |
|
}, |
|
"per_token": { |
|
"memory": null, |
|
"latency": { |
|
"unit": "s", |
|
"count": 253, |
|
"total": 4.849364982604982, |
|
"mean": 0.019167450524130356, |
|
"stdev": 0.001259601118010216, |
|
"p50": 0.01899622344970703, |
|
"p90": 0.019580928802490235, |
|
"p95": 0.019634585952758788, |
|
"p99": 0.019947479553222656, |
|
"values": [ |
|
0.021135360717773437, |
|
0.01902899169921875, |
|
0.0188538875579834, |
|
0.01884569549560547, |
|
0.018912256240844725, |
|
0.018852863311767578, |
|
0.018897920608520507, |
|
0.018911231994628908, |
|
0.01886617660522461, |
|
0.018890752792358398, |
|
0.018881536483764647, |
|
0.01887129592895508, |
|
0.018873344421386717, |
|
0.01886720085144043, |
|
0.018913280487060546, |
|
0.019179519653320314, |
|
0.019162111282348633, |
|
0.018896896362304686, |
|
0.018990079879760743, |
|
0.019382272720336914, |
|
0.019591167449951173, |
|
0.018919424057006837, |
|
0.018905088424682616, |
|
0.018877439498901367, |
|
0.01887129592895508, |
|
0.018928640365600585, |
|
0.018893823623657227, |
|
0.01887129592895508, |
|
0.01886412811279297, |
|
0.018936832427978514, |
|
0.018891775131225585, |
|
0.018900991439819336, |
|
0.018923519134521484, |
|
0.018921472549438476, |
|
0.018926591873168946, |
|
0.01887948799133301, |
|
0.018966527938842775, |
|
0.018900991439819336, |
|
0.01889587211608887, |
|
0.01883033561706543, |
|
0.018861055374145508, |
|
0.018873344421386717, |
|
0.01883545684814453, |
|
0.01885491180419922, |
|
0.018888704299926756, |
|
0.01883443260192871, |
|
0.01883545684814453, |
|
0.01885593605041504, |
|
0.01887129592895508, |
|
0.01887539291381836, |
|
0.019124223709106446, |
|
0.01926246452331543, |
|
0.019975168228149414, |
|
0.01909350395202637, |
|
0.018749439239501953, |
|
0.018844671249389648, |
|
0.019014656066894533, |
|
0.01903718376159668, |
|
0.018991104125976564, |
|
0.018972671508789063, |
|
0.01900441551208496, |
|
0.018980863571166993, |
|
0.018982912063598634, |
|
0.018946048736572265, |
|
0.01902899169921875, |
|
0.018997247695922852, |
|
0.018829376220703124, |
|
0.01893574333190918, |
|
0.018958335876464845, |
|
0.01900851249694824, |
|
0.01898700714111328, |
|
0.01901158332824707, |
|
0.018981887817382814, |
|
0.018974720001220705, |
|
0.01903308868408203, |
|
0.01900032043457031, |
|
0.019086336135864256, |
|
0.0188221435546875, |
|
0.01883647918701172, |
|
0.01902592086791992, |
|
0.019147775650024415, |
|
0.019019775390625, |
|
0.019009536743164062, |
|
0.0190894718170166, |
|
0.01899616050720215, |
|
0.018997247695922852, |
|
0.018938880920410156, |
|
0.018815999984741212, |
|
0.018971647262573242, |
|
0.018940927505493164, |
|
0.018844671249389648, |
|
0.01881088066101074, |
|
0.018975744247436522, |
|
0.01901055908203125, |
|
0.01899622344970703, |
|
0.019359743118286133, |
|
0.019581951141357423, |
|
0.019569664001464843, |
|
0.01898700714111328, |
|
0.018993183135986327, |
|
0.01903919982910156, |
|
0.01903104019165039, |
|
0.018984960556030273, |
|
0.019221567153930665, |
|
0.01953171157836914, |
|
0.019083263397216797, |
|
0.019096576690673828, |
|
0.01901260757446289, |
|
0.018997247695922852, |
|
0.018949119567871094, |
|
0.018952192306518553, |
|
0.01901055908203125, |
|
0.018998271942138673, |
|
0.019021823883056642, |
|
0.01905254364013672, |
|
0.0190382080078125, |
|
0.019519487380981446, |
|
0.019512319564819337, |
|
0.019577856063842772, |
|
0.018991104125976564, |
|
0.018963455200195312, |
|
0.019001344680786132, |
|
0.018985984802246093, |
|
0.018938880920410156, |
|
0.018941951751708985, |
|
0.018969600677490234, |
|
0.038586368560791014, |
|
0.018783231735229493, |
|
0.018768896102905275, |
|
0.018869247436523438, |
|
0.018779136657714843, |
|
0.018868223190307617, |
|
0.0188272647857666, |
|
0.01881292724609375, |
|
0.018770944595336913, |
|
0.01883443260192871, |
|
0.018865152359008788, |
|
0.01881395149230957, |
|
0.018908159255981445, |
|
0.018888736724853517, |
|
0.018698207855224608, |
|
0.018932735443115235, |
|
0.018873344421386717, |
|
0.01884160041809082, |
|
0.018844671249389648, |
|
0.01886617660522461, |
|
0.01884774398803711, |
|
0.01881292724609375, |
|
0.01882111930847168, |
|
0.018884607315063476, |
|
0.01883750343322754, |
|
0.01863680076599121, |
|
0.018740224838256835, |
|
0.01880985641479492, |
|
0.018923519134521484, |
|
0.018865152359008788, |
|
0.018880512237548826, |
|
0.019098623275756836, |
|
0.019495935440063478, |
|
0.019430400848388672, |
|
0.0194201602935791, |
|
0.019352575302124024, |
|
0.018844671249389648, |
|
0.018856960296630858, |
|
0.01886207962036133, |
|
0.018899967193603515, |
|
0.018918399810791017, |
|
0.018880512237548826, |
|
0.018918399810791017, |
|
0.018899967193603515, |
|
0.01883340835571289, |
|
0.018811904907226562, |
|
0.018851839065551757, |
|
0.018886655807495118, |
|
0.018918399810791017, |
|
0.018933759689331055, |
|
0.01885081672668457, |
|
0.018944000244140623, |
|
0.018851839065551757, |
|
0.019364864349365234, |
|
0.019388416290283202, |
|
0.019338239669799806, |
|
0.018917375564575196, |
|
0.019721216201782226, |
|
0.01963520050048828, |
|
0.019580928802490235, |
|
0.019555328369140625, |
|
0.019573759078979493, |
|
0.01942937660217285, |
|
0.019629056930541993, |
|
0.01959219169616699, |
|
0.01963929557800293, |
|
0.019591167449951173, |
|
0.01966592025756836, |
|
0.019554304122924804, |
|
0.019611648559570313, |
|
0.019596288681030274, |
|
0.01970790481567383, |
|
0.019647487640380858, |
|
0.01992192077636719, |
|
0.019148799896240236, |
|
0.01904025650024414, |
|
0.019009536743164062, |
|
0.019083263397216797, |
|
0.019491840362548828, |
|
0.019116031646728517, |
|
0.01904844856262207, |
|
0.019013631820678712, |
|
0.01905356788635254, |
|
0.01943244743347168, |
|
0.019501056671142578, |
|
0.019361791610717775, |
|
0.019388416290283202, |
|
0.01905766487121582, |
|
0.01941606330871582, |
|
0.019543039321899415, |
|
0.019086336135864256, |
|
0.019070976257324217, |
|
0.019615743637084963, |
|
0.019373056411743163, |
|
0.019321855545043946, |
|
0.01900441551208496, |
|
0.019280895233154297, |
|
0.01962188720703125, |
|
0.019689472198486328, |
|
0.019589120864868165, |
|
0.019620864868164063, |
|
0.019539968490600586, |
|
0.01963929557800293, |
|
0.01963417625427246, |
|
0.019696640014648437, |
|
0.01921023941040039, |
|
0.019139583587646485, |
|
0.01903513526916504, |
|
0.019165184020996092, |
|
0.019339263916015623, |
|
0.01927065658569336, |
|
0.019096576690673828, |
|
0.019042303085327148, |
|
0.01904332733154297, |
|
0.01906892776489258, |
|
0.0190248966217041, |
|
0.01901568031311035, |
|
0.01904332733154297, |
|
0.019090431213378906, |
|
0.019173376083374022, |
|
0.019132415771484376, |
|
0.01904844856262207, |
|
0.019076095581054688, |
|
0.019005439758300782, |
|
0.019309568405151366, |
|
0.019573759078979493, |
|
0.019580928802490235 |
|
] |
|
}, |
|
"throughput": { |
|
"unit": "tokens/s", |
|
"value": 52.17177937885252 |
|
}, |
|
"energy": null, |
|
"efficiency": null |
|
} |
|
} |
|
} |