Félix Marty commited on
Commit
c08eef4
1 Parent(s): e85a888

remove old

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. dana/configs/db/Inference/infos/benchmarks.series.json +0 -1
  2. dana/configs/db/Inference/infos/benchmarks.statusSeries.json +0 -1
  3. dana/configs/db/Inference/infos/builds.json +0 -1
  4. dana/configs/db/Inference/series/llama_1gpu/_0_forward/_latency_s/_.json +0 -1
  5. dana/configs/db/Inference/series/llama_1gpu/_0_forward/_throughpu/t_samples_/s_.json +0 -1
  6. dana/configs/db/Inference/series/llama_1gpu/_0_generat/e_latency_/s_.json +0 -1
  7. dana/configs/db/Inference/series/llama_1gpu/_0_generat/e_throughp/ut_tokens_/s_.json +0 -1
  8. dana/configs/db/Inference/series/llama_1gpu/_1_forward/_latency_s/_.json +0 -1
  9. dana/configs/db/Inference/series/llama_1gpu/_1_forward/_throughpu/t_samples_/s_.json +0 -1
  10. dana/configs/db/Inference/series/llama_1gpu/_1_generat/e_latency_/s_.json +0 -1
  11. dana/configs/db/Inference/series/llama_1gpu/_1_generat/e_throughp/ut_tokens_/s_.json +0 -1
  12. dana/configs/db/Inference/series/llama_1gpu/_2_forward/_latency_s/_.json +0 -1
  13. dana/configs/db/Inference/series/llama_1gpu/_2_forward/_throughpu/t_samples_/s_.json +0 -1
  14. dana/configs/db/Inference/series/llama_1gpu/_2_generat/e_latency_/s_.json +0 -1
  15. dana/configs/db/Inference/series/llama_1gpu/_2_generat/e_throughp/ut_tokens_/s_.json +0 -1
  16. dana/configs/db/Inference/series/llama_1gpu/_3_forward/_latency_s/_.json +0 -1
  17. dana/configs/db/Inference/series/llama_1gpu/_3_forward/_throughpu/t_samples_/s_.json +0 -1
  18. dana/configs/db/Inference/series/llama_1gpu/_3_generat/e_latency_/s_.json +0 -1
  19. dana/configs/db/Inference/series/llama_1gpu/_3_generat/e_throughp/ut_tokens_/s_.json +0 -1
  20. dana/configs/db/Training/infos/benchmarks.series.json +0 -1
  21. dana/configs/db/Training/infos/benchmarks.statusSeries.json +0 -1
  22. dana/configs/db/Training/infos/builds.json +0 -1
  23. dana/configs/db/Training/series/bert_1gpu_/0_training/_runtime_s/_.json +0 -1
  24. dana/configs/db/Training/series/bert_1gpu_/0_training/_throughpu/t_samples_/s_.json +0 -1
  25. dana/configs/db/Training/series/bert_1gpu_/1_training/_runtime_s/_.json +0 -1
  26. dana/configs/db/Training/series/bert_1gpu_/1_training/_throughpu/t_samples_/s_.json +0 -1
  27. dana/configs/db/admin/globalStats.json +0 -1
  28. dana/configs/db/admin/projects.json +0 -1
  29. raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/bert_1gpu_training/0/.config/config.yaml +0 -75
  30. raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/bert_1gpu_training/0/.config/hydra.yaml +0 -174
  31. raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/bert_1gpu_training/0/.config/overrides.yaml +0 -2
  32. raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/bert_1gpu_training/0/experiment.log +0 -17
  33. raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/bert_1gpu_training/0/hydra_config.yaml +0 -75
  34. raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/bert_1gpu_training/0/training_results.csv +0 -2
  35. raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/bert_1gpu_training/1/.config/config.yaml +0 -75
  36. raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/bert_1gpu_training/1/.config/hydra.yaml +0 -174
  37. raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/bert_1gpu_training/1/.config/overrides.yaml +0 -2
  38. raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/bert_1gpu_training/1/experiment.log +0 -16
  39. raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/bert_1gpu_training/1/hydra_config.yaml +0 -75
  40. raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/bert_1gpu_training/1/training_results.csv +0 -2
  41. raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/bert_1gpu_training/multirun.yaml +0 -246
  42. raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/llama_1gpu_inference/0/.config/config.yaml +0 -73
  43. raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/llama_1gpu_inference/0/.config/hydra.yaml +0 -174
  44. raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/llama_1gpu_inference/0/.config/overrides.yaml +0 -2
  45. raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/llama_1gpu_inference/0/experiment.log +0 -27
  46. raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/llama_1gpu_inference/0/hydra_config.yaml +0 -79
  47. raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/llama_1gpu_inference/0/inference_results.csv +0 -2
  48. raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/llama_1gpu_inference/1/.config/config.yaml +0 -73
  49. raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/llama_1gpu_inference/1/.config/hydra.yaml +0 -174
  50. raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/llama_1gpu_inference/1/.config/overrides.yaml +0 -2
dana/configs/db/Inference/infos/benchmarks.series.json DELETED
@@ -1 +0,0 @@
1
- {"llama_1gpu_0_forward_latency_s_":{"status":{"error":"Unable to find first average","lastBuildId":14055},"description":"\nbenchmark.input_shapes.batch_size: 1\nbackend.torch_dtype: float16\nbenchmark.input_shapes.sequence_length: 200\nbenchmark.new_tokens: 200","state":"similarNeedstriage"},"llama_1gpu_0_forward_throughput_samples_s_":{"status":{"error":"Unable to find first average","lastBuildId":14055},"description":"\nbenchmark.input_shapes.batch_size: 1\nbackend.torch_dtype: float16\nbenchmark.input_shapes.sequence_length: 200\nbenchmark.new_tokens: 200","state":"similarNeedstriage"},"llama_1gpu_0_generate_latency_s_":{"status":{"error":"Unable to find first average","lastBuildId":14055},"description":"\nbenchmark.input_shapes.batch_size: 1\nbackend.torch_dtype: float16\nbenchmark.input_shapes.sequence_length: 200\nbenchmark.new_tokens: 200","state":"similarNeedstriage"},"llama_1gpu_0_generate_throughput_tokens_s_":{"status":{"error":"Unable to find first average","lastBuildId":14055},"description":"\nbenchmark.input_shapes.batch_size: 1\nbackend.torch_dtype: float16\nbenchmark.input_shapes.sequence_length: 200\nbenchmark.new_tokens: 200","state":"similarNeedstriage"},"llama_1gpu_1_forward_latency_s_":{"status":{"error":"Unable to find first average","lastBuildId":14055},"description":"\nbenchmark.input_shapes.batch_size: 1\nbackend.torch_dtype: float32\nbenchmark.input_shapes.sequence_length: 200\nbenchmark.new_tokens: 200","state":"similarNeedstriage"},"llama_1gpu_1_forward_throughput_samples_s_":{"status":{"error":"Unable to find first average","lastBuildId":14055},"description":"\nbenchmark.input_shapes.batch_size: 1\nbackend.torch_dtype: float32\nbenchmark.input_shapes.sequence_length: 200\nbenchmark.new_tokens: 200","state":"similarNeedstriage"},"llama_1gpu_1_generate_latency_s_":{"status":{"error":"Unable to find first average","lastBuildId":14055},"description":"\nbenchmark.input_shapes.batch_size: 1\nbackend.torch_dtype: float32\nbenchmark.input_shapes.sequence_length: 200\nbenchmark.new_tokens: 200","state":"similarNeedstriage"},"llama_1gpu_1_generate_throughput_tokens_s_":{"status":{"error":"Unable to find first average","lastBuildId":14055},"description":"\nbenchmark.input_shapes.batch_size: 1\nbackend.torch_dtype: float32\nbenchmark.input_shapes.sequence_length: 200\nbenchmark.new_tokens: 200","state":"similarNeedstriage"},"llama_1gpu_2_forward_latency_s_":{"status":{"error":"Unable to find first average","lastBuildId":14055},"description":"\nbenchmark.input_shapes.batch_size: 16\nbackend.torch_dtype: float16\nbenchmark.input_shapes.sequence_length: 200\nbenchmark.new_tokens: 200","state":"similarNeedstriage"},"llama_1gpu_2_forward_throughput_samples_s_":{"status":{"error":"Unable to find first average","lastBuildId":14055},"description":"\nbenchmark.input_shapes.batch_size: 16\nbackend.torch_dtype: float16\nbenchmark.input_shapes.sequence_length: 200\nbenchmark.new_tokens: 200","state":"similarNeedstriage"},"llama_1gpu_2_generate_latency_s_":{"status":{"error":"Unable to find first average","lastBuildId":14055},"description":"\nbenchmark.input_shapes.batch_size: 16\nbackend.torch_dtype: float16\nbenchmark.input_shapes.sequence_length: 200\nbenchmark.new_tokens: 200","state":"similarNeedstriage"},"llama_1gpu_2_generate_throughput_tokens_s_":{"status":{"error":"Unable to find first average","lastBuildId":14055},"description":"\nbenchmark.input_shapes.batch_size: 16\nbackend.torch_dtype: float16\nbenchmark.input_shapes.sequence_length: 200\nbenchmark.new_tokens: 200","state":"similarNeedstriage"},"llama_1gpu_3_forward_latency_s_":{"status":{"error":"Unable to find first average","lastBuildId":14055},"description":"\nbenchmark.input_shapes.batch_size: 16\nbackend.torch_dtype: float32\nbenchmark.input_shapes.sequence_length: 200\nbenchmark.new_tokens: 200","state":"similarNeedstriage"},"llama_1gpu_3_forward_throughput_samples_s_":{"status":{"error":"Unable to find first average","lastBuildId":14055},"description":"\nbenchmark.input_shapes.batch_size: 16\nbackend.torch_dtype: float32\nbenchmark.input_shapes.sequence_length: 200\nbenchmark.new_tokens: 200","state":"similarNeedstriage"},"llama_1gpu_3_generate_latency_s_":{"status":{"error":"Unable to find first average","lastBuildId":14055},"description":"\nbenchmark.input_shapes.batch_size: 16\nbackend.torch_dtype: float32\nbenchmark.input_shapes.sequence_length: 200\nbenchmark.new_tokens: 200","state":"similarNeedstriage"},"llama_1gpu_3_generate_throughput_tokens_s_":{"status":{"error":"Unable to find first average","lastBuildId":14055},"description":"\nbenchmark.input_shapes.batch_size: 16\nbackend.torch_dtype: float32\nbenchmark.input_shapes.sequence_length: 200\nbenchmark.new_tokens: 200","state":"similarNeedstriage"}}
 
 
dana/configs/db/Inference/infos/benchmarks.statusSeries.json DELETED
@@ -1 +0,0 @@
1
- {"0":{"numSeries":16,"numSeriesSimilar":0,"numSeriesImproved":0,"numSeriesRegression":0,"numSeriesUndefined":16,"time":1695816293533}}
 
 
dana/configs/db/Inference/infos/builds.json DELETED
@@ -1 +0,0 @@
1
- {"14054":{"buildId":14054,"infos":{"hash":"153755ee386ac73e04814a94337abcb1208ff5d1","abbrevHash":"153755ee","authorName":"Younes Belkada","authorEmail":"49240599+younesbelkada@users.noreply.github.com","subject":"[`FA` / `tests`] Add use_cache tests for FA models (#26415)","url":null}},"14055":{"buildId":14055,"infos":{"hash":"946bac798caefada3f5f1c9fecdcfd587ed24ac7","abbrevHash":"946bac79","authorName":"statelesshz","authorEmail":"hzji210@gmail.com","subject":"add bf16 mixed precision support for NPU (#26163)","url":null}}}
 
 
dana/configs/db/Inference/series/llama_1gpu/_0_forward/_latency_s/_.json DELETED
@@ -1 +0,0 @@
1
- {"projectId":"Inference","description":"\nbenchmark.input_shapes.batch_size: 1\nbackend.torch_dtype: float16\nbenchmark.input_shapes.sequence_length: 200\nbenchmark.new_tokens: 200","analyse":{"benchmark":{"range":"10%","required":5,"trend":"smaller"}},"assignee":{"compares":{}},"samples":{"14054":0.00239,"14055":0.00312},"state":{"analyse":"similarNeedstriage","compares":{}},"lastBuildId":"14055","analyseResult":{"summary":{"error":"Unable to find first average","lastBuildId":14055}}}
 
 
dana/configs/db/Inference/series/llama_1gpu/_0_forward/_throughpu/t_samples_/s_.json DELETED
@@ -1 +0,0 @@
1
- {"projectId":"Inference","description":"\nbenchmark.input_shapes.batch_size: 1\nbackend.torch_dtype: float16\nbenchmark.input_shapes.sequence_length: 200\nbenchmark.new_tokens: 200","analyse":{"benchmark":{"range":"10%","required":5,"trend":"smaller"}},"assignee":{"compares":{}},"samples":{"14054":418,"14055":321},"state":{"analyse":"similarNeedstriage","compares":{}},"lastBuildId":"14055","analyseResult":{"summary":{"error":"Unable to find first average","lastBuildId":14055}}}
 
 
dana/configs/db/Inference/series/llama_1gpu/_0_generat/e_latency_/s_.json DELETED
@@ -1 +0,0 @@
1
- {"projectId":"Inference","description":"\nbenchmark.input_shapes.batch_size: 1\nbackend.torch_dtype: float16\nbenchmark.input_shapes.sequence_length: 200\nbenchmark.new_tokens: 200","analyse":{"benchmark":{"range":"10%","required":5,"trend":"smaller"}},"assignee":{"compares":{}},"samples":{"14054":0.491,"14055":0.639},"state":{"analyse":"similarNeedstriage","compares":{}},"lastBuildId":"14055","analyseResult":{"summary":{"error":"Unable to find first average","lastBuildId":14055}}}
 
 
dana/configs/db/Inference/series/llama_1gpu/_0_generat/e_throughp/ut_tokens_/s_.json DELETED
@@ -1 +0,0 @@
1
- {"projectId":"Inference","description":"\nbenchmark.input_shapes.batch_size: 1\nbackend.torch_dtype: float16\nbenchmark.input_shapes.sequence_length: 200\nbenchmark.new_tokens: 200","analyse":{"benchmark":{"range":"10%","required":5,"trend":"smaller"}},"assignee":{"compares":{}},"samples":{"14054":407,"14055":313},"state":{"analyse":"similarNeedstriage","compares":{}},"lastBuildId":"14055","analyseResult":{"summary":{"error":"Unable to find first average","lastBuildId":14055}}}
 
 
dana/configs/db/Inference/series/llama_1gpu/_1_forward/_latency_s/_.json DELETED
@@ -1 +0,0 @@
1
- {"projectId":"Inference","description":"\nbenchmark.input_shapes.batch_size: 1\nbackend.torch_dtype: float32\nbenchmark.input_shapes.sequence_length: 200\nbenchmark.new_tokens: 200","analyse":{"benchmark":{"range":"10%","required":5,"trend":"smaller"}},"assignee":{"compares":{}},"samples":{"14054":0.00328,"14055":0.00332},"state":{"analyse":"similarNeedstriage","compares":{}},"lastBuildId":"14055","analyseResult":{"summary":{"error":"Unable to find first average","lastBuildId":14055}}}
 
 
dana/configs/db/Inference/series/llama_1gpu/_1_forward/_throughpu/t_samples_/s_.json DELETED
@@ -1 +0,0 @@
1
- {"projectId":"Inference","description":"\nbenchmark.input_shapes.batch_size: 1\nbackend.torch_dtype: float32\nbenchmark.input_shapes.sequence_length: 200\nbenchmark.new_tokens: 200","analyse":{"benchmark":{"range":"10%","required":5,"trend":"smaller"}},"assignee":{"compares":{}},"samples":{"14054":305,"14055":301},"state":{"analyse":"similarNeedstriage","compares":{}},"lastBuildId":"14055","analyseResult":{"summary":{"error":"Unable to find first average","lastBuildId":14055}}}
 
 
dana/configs/db/Inference/series/llama_1gpu/_1_generat/e_latency_/s_.json DELETED
@@ -1 +0,0 @@
1
- {"projectId":"Inference","description":"\nbenchmark.input_shapes.batch_size: 1\nbackend.torch_dtype: float32\nbenchmark.input_shapes.sequence_length: 200\nbenchmark.new_tokens: 200","analyse":{"benchmark":{"range":"10%","required":5,"trend":"smaller"}},"assignee":{"compares":{}},"samples":{"14054":0.53,"14055":0.537},"state":{"analyse":"similarNeedstriage","compares":{}},"lastBuildId":"14055","analyseResult":{"summary":{"error":"Unable to find first average","lastBuildId":14055}}}
 
 
dana/configs/db/Inference/series/llama_1gpu/_1_generat/e_throughp/ut_tokens_/s_.json DELETED
@@ -1 +0,0 @@
1
- {"projectId":"Inference","description":"\nbenchmark.input_shapes.batch_size: 1\nbackend.torch_dtype: float32\nbenchmark.input_shapes.sequence_length: 200\nbenchmark.new_tokens: 200","analyse":{"benchmark":{"range":"10%","required":5,"trend":"smaller"}},"assignee":{"compares":{}},"samples":{"14054":377,"14055":372},"state":{"analyse":"similarNeedstriage","compares":{}},"lastBuildId":"14055","analyseResult":{"summary":{"error":"Unable to find first average","lastBuildId":14055}}}
 
 
dana/configs/db/Inference/series/llama_1gpu/_2_forward/_latency_s/_.json DELETED
@@ -1 +0,0 @@
1
- {"projectId":"Inference","description":"\nbenchmark.input_shapes.batch_size: 16\nbackend.torch_dtype: float16\nbenchmark.input_shapes.sequence_length: 200\nbenchmark.new_tokens: 200","analyse":{"benchmark":{"range":"10%","required":5,"trend":"smaller"}},"assignee":{"compares":{}},"samples":{"14054":0.0041,"14055":0.00487},"state":{"analyse":"similarNeedstriage","compares":{}},"lastBuildId":"14055","analyseResult":{"summary":{"error":"Unable to find first average","lastBuildId":14055}}}
 
 
dana/configs/db/Inference/series/llama_1gpu/_2_forward/_throughpu/t_samples_/s_.json DELETED
@@ -1 +0,0 @@
1
- {"projectId":"Inference","description":"\nbenchmark.input_shapes.batch_size: 16\nbackend.torch_dtype: float16\nbenchmark.input_shapes.sequence_length: 200\nbenchmark.new_tokens: 200","analyse":{"benchmark":{"range":"10%","required":5,"trend":"smaller"}},"assignee":{"compares":{}},"samples":{"14054":3900,"14055":3290},"state":{"analyse":"similarNeedstriage","compares":{}},"lastBuildId":"14055","analyseResult":{"summary":{"error":"Unable to find first average","lastBuildId":14055}}}
 
 
dana/configs/db/Inference/series/llama_1gpu/_2_generat/e_latency_/s_.json DELETED
@@ -1 +0,0 @@
1
- {"projectId":"Inference","description":"\nbenchmark.input_shapes.batch_size: 16\nbackend.torch_dtype: float16\nbenchmark.input_shapes.sequence_length: 200\nbenchmark.new_tokens: 200","analyse":{"benchmark":{"range":"10%","required":5,"trend":"smaller"}},"assignee":{"compares":{}},"samples":{"14054":0.655,"14055":0.765},"state":{"analyse":"similarNeedstriage","compares":{}},"lastBuildId":"14055","analyseResult":{"summary":{"error":"Unable to find first average","lastBuildId":14055}}}
 
 
dana/configs/db/Inference/series/llama_1gpu/_2_generat/e_throughp/ut_tokens_/s_.json DELETED
@@ -1 +0,0 @@
1
- {"projectId":"Inference","description":"\nbenchmark.input_shapes.batch_size: 16\nbackend.torch_dtype: float16\nbenchmark.input_shapes.sequence_length: 200\nbenchmark.new_tokens: 200","analyse":{"benchmark":{"range":"10%","required":5,"trend":"smaller"}},"assignee":{"compares":{}},"samples":{"14054":4890,"14055":4180},"state":{"analyse":"similarNeedstriage","compares":{}},"lastBuildId":"14055","analyseResult":{"summary":{"error":"Unable to find first average","lastBuildId":14055}}}
 
 
dana/configs/db/Inference/series/llama_1gpu/_3_forward/_latency_s/_.json DELETED
@@ -1 +0,0 @@
1
- {"projectId":"Inference","description":"\nbenchmark.input_shapes.batch_size: 16\nbackend.torch_dtype: float32\nbenchmark.input_shapes.sequence_length: 200\nbenchmark.new_tokens: 200","analyse":{"benchmark":{"range":"10%","required":5,"trend":"smaller"}},"assignee":{"compares":{}},"samples":{"14054":0.00457,"14055":0.00609},"state":{"analyse":"similarNeedstriage","compares":{}},"lastBuildId":"14055","analyseResult":{"summary":{"error":"Unable to find first average","lastBuildId":14055}}}
 
 
dana/configs/db/Inference/series/llama_1gpu/_3_forward/_throughpu/t_samples_/s_.json DELETED
@@ -1 +0,0 @@
1
- {"projectId":"Inference","description":"\nbenchmark.input_shapes.batch_size: 16\nbackend.torch_dtype: float32\nbenchmark.input_shapes.sequence_length: 200\nbenchmark.new_tokens: 200","analyse":{"benchmark":{"range":"10%","required":5,"trend":"smaller"}},"assignee":{"compares":{}},"samples":{"14054":3500,"14055":2630},"state":{"analyse":"similarNeedstriage","compares":{}},"lastBuildId":"14055","analyseResult":{"summary":{"error":"Unable to find first average","lastBuildId":14055}}}
 
 
dana/configs/db/Inference/series/llama_1gpu/_3_generat/e_latency_/s_.json DELETED
@@ -1 +0,0 @@
1
- {"projectId":"Inference","description":"\nbenchmark.input_shapes.batch_size: 16\nbackend.torch_dtype: float32\nbenchmark.input_shapes.sequence_length: 200\nbenchmark.new_tokens: 200","analyse":{"benchmark":{"range":"10%","required":5,"trend":"smaller"}},"assignee":{"compares":{}},"samples":{"14054":0.539,"14055":0.822},"state":{"analyse":"similarNeedstriage","compares":{}},"lastBuildId":"14055","analyseResult":{"summary":{"error":"Unable to find first average","lastBuildId":14055}}}
 
 
dana/configs/db/Inference/series/llama_1gpu/_3_generat/e_throughp/ut_tokens_/s_.json DELETED
@@ -1 +0,0 @@
1
- {"projectId":"Inference","description":"\nbenchmark.input_shapes.batch_size: 16\nbackend.torch_dtype: float32\nbenchmark.input_shapes.sequence_length: 200\nbenchmark.new_tokens: 200","analyse":{"benchmark":{"range":"10%","required":5,"trend":"smaller"}},"assignee":{"compares":{}},"samples":{"14054":5940,"14055":3890},"state":{"analyse":"similarNeedstriage","compares":{}},"lastBuildId":"14055","analyseResult":{"summary":{"error":"Unable to find first average","lastBuildId":14055}}}
 
 
dana/configs/db/Training/infos/benchmarks.series.json DELETED
@@ -1 +0,0 @@
1
- {"bert_1gpu_0_training_runtime_s_":{"status":{"error":"Unable to find first average","lastBuildId":14055},"description":"\n+benchmark.training_arguments.per_device_train_batch_size: None\nbackend.torch_dtype: float16\nbenchmark.dataset_shapes.sequence_length: 256","state":"similarNeedstriage"},"bert_1gpu_0_training_throughput_samples_s_":{"status":{"error":"Unable to find first average","lastBuildId":14055},"description":"\n+benchmark.training_arguments.per_device_train_batch_size: None\nbackend.torch_dtype: float16\nbenchmark.dataset_shapes.sequence_length: 256","state":"similarNeedstriage"},"bert_1gpu_1_training_runtime_s_":{"status":{"error":"Unable to find first average","lastBuildId":14055},"description":"\n+benchmark.training_arguments.per_device_train_batch_size: None\nbackend.torch_dtype: float32\nbenchmark.dataset_shapes.sequence_length: 256","state":"similarNeedstriage"},"bert_1gpu_1_training_throughput_samples_s_":{"status":{"error":"Unable to find first average","lastBuildId":14055},"description":"\n+benchmark.training_arguments.per_device_train_batch_size: None\nbackend.torch_dtype: float32\nbenchmark.dataset_shapes.sequence_length: 256","state":"similarNeedstriage"}}
 
 
dana/configs/db/Training/infos/benchmarks.statusSeries.json DELETED
@@ -1 +0,0 @@
1
- {"0":{"numSeries":4,"numSeriesSimilar":0,"numSeriesImproved":0,"numSeriesRegression":0,"numSeriesUndefined":4,"time":1695816293579}}
 
 
dana/configs/db/Training/infos/builds.json DELETED
@@ -1 +0,0 @@
1
- {"14054":{"buildId":14054,"infos":{"hash":"153755ee386ac73e04814a94337abcb1208ff5d1","abbrevHash":"153755ee","authorName":"Younes Belkada","authorEmail":"49240599+younesbelkada@users.noreply.github.com","subject":"[`FA` / `tests`] Add use_cache tests for FA models (#26415)","url":null}},"14055":{"buildId":14055,"infos":{"hash":"946bac798caefada3f5f1c9fecdcfd587ed24ac7","abbrevHash":"946bac79","authorName":"statelesshz","authorEmail":"hzji210@gmail.com","subject":"add bf16 mixed precision support for NPU (#26163)","url":null}}}
 
 
dana/configs/db/Training/series/bert_1gpu_/0_training/_runtime_s/_.json DELETED
@@ -1 +0,0 @@
1
- {"projectId":"Training","description":"\n+benchmark.training_arguments.per_device_train_batch_size: None\nbackend.torch_dtype: float16\nbenchmark.dataset_shapes.sequence_length: 256","analyse":{"benchmark":{"range":"10%","required":5,"trend":"smaller"}},"assignee":{"compares":{}},"samples":{"14054":37.33880257606506,"14055":32.68557357788086},"state":{"analyse":"similarNeedstriage","compares":{}},"lastBuildId":"14055","analyseResult":{"summary":{"error":"Unable to find first average","lastBuildId":14055}}}
 
 
dana/configs/db/Training/series/bert_1gpu_/0_training/_throughpu/t_samples_/s_.json DELETED
@@ -1 +0,0 @@
1
- {"projectId":"Training","description":"\n+benchmark.training_arguments.per_device_train_batch_size: None\nbackend.torch_dtype: float16\nbenchmark.dataset_shapes.sequence_length: 256","analyse":{"benchmark":{"range":"10%","required":5,"trend":"smaller"}},"assignee":{"compares":{}},"samples":{"14054":86.55874792492082,"14055":98.88154455356337},"state":{"analyse":"similarNeedstriage","compares":{}},"lastBuildId":"14055","analyseResult":{"summary":{"error":"Unable to find first average","lastBuildId":14055}}}
 
 
dana/configs/db/Training/series/bert_1gpu_/1_training/_runtime_s/_.json DELETED
@@ -1 +0,0 @@
1
- {"projectId":"Training","description":"\n+benchmark.training_arguments.per_device_train_batch_size: None\nbackend.torch_dtype: float32\nbenchmark.dataset_shapes.sequence_length: 256","analyse":{"benchmark":{"range":"10%","required":5,"trend":"smaller"}},"assignee":{"compares":{}},"samples":{"14054":13.53417706489563,"14055":13.473761320114136},"state":{"analyse":"similarNeedstriage","compares":{}},"lastBuildId":"14055","analyseResult":{"summary":{"error":"Unable to find first average","lastBuildId":14055}}}
 
 
dana/configs/db/Training/series/bert_1gpu_/1_training/_throughpu/t_samples_/s_.json DELETED
@@ -1 +0,0 @@
1
- {"projectId":"Training","description":"\n+benchmark.training_arguments.per_device_train_batch_size: None\nbackend.torch_dtype: float32\nbenchmark.dataset_shapes.sequence_length: 256","analyse":{"benchmark":{"range":"10%","required":5,"trend":"smaller"}},"assignee":{"compares":{}},"samples":{"14054":238.802845899144,"14055":239.87362720869555},"state":{"analyse":"similarNeedstriage","compares":{}},"lastBuildId":"14055","analyseResult":{"summary":{"error":"Unable to find first average","lastBuildId":14055}}}
 
 
dana/configs/db/admin/globalStats.json DELETED
@@ -1 +0,0 @@
1
- {"numSamples":40,"numSeries":20,"projects":{"Inference":{"numSamples":32,"numSeries":16},"Training":{"numSamples":8,"numSeries":4}}}
 
 
dana/configs/db/admin/projects.json DELETED
@@ -1 +0,0 @@
1
- {"Inference":{"description":"Benchmarks related to inference","users":"","useBugTracker":false},"Training":{"description":"Benchmarks related to training","users":"","useBugTracker":false}}
 
 
raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/bert_1gpu_training/0/.config/config.yaml DELETED
@@ -1,75 +0,0 @@
1
- backend:
2
- name: pytorch
3
- version: ${pytorch_version:}
4
- _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
5
- seed: 42
6
- inter_op_num_threads: null
7
- intra_op_num_threads: null
8
- initial_isolation_check: true
9
- continous_isolation_check: true
10
- delete_cache: false
11
- no_weights: false
12
- device_map: null
13
- torch_dtype: float16
14
- disable_grad: ${is_inference:${benchmark.name}}
15
- eval_mode: ${is_inference:${benchmark.name}}
16
- amp_autocast: false
17
- amp_dtype: null
18
- torch_compile: false
19
- torch_compile_config: {}
20
- bettertransformer: false
21
- quantization_scheme: null
22
- quantization_config: {}
23
- use_ddp: false
24
- ddp_config: {}
25
- peft_strategy: null
26
- peft_config: {}
27
- benchmark:
28
- name: training
29
- _target_: optimum_benchmark.benchmarks.training.benchmark.TrainingBenchmark
30
- warmup_steps: 40
31
- dataset_shapes:
32
- dataset_size: 1500
33
- sequence_length: 256
34
- num_choices: 1
35
- feature_size: 80
36
- nb_max_frames: 3000
37
- audio_sequence_length: 16000
38
- training_arguments:
39
- skip_memory_metrics: true
40
- output_dir: ./trainer_output
41
- use_cpu: ${is_cpu:${device}}
42
- ddp_find_unused_parameters: false
43
- do_train: true
44
- do_eval: false
45
- do_predict: false
46
- report_to: none
47
- per_device_train_batch_size: 32
48
- experiment_name: bert_1gpu_training
49
- model: bert-base-uncased
50
- device: cuda
51
- task: text-classification
52
- hub_kwargs:
53
- revision: main
54
- cache_dir: null
55
- force_download: false
56
- local_files_only: false
57
- environment:
58
- optimum_version: 1.13.1
59
- transformers_version: 4.34.0.dev0
60
- accelerate_version: 0.23.0
61
- diffusers_version: null
62
- python_version: 3.10.12
63
- system: Linux
64
- cpu: ' AMD EPYC 7643 48-Core Processor'
65
- cpu_count: 96
66
- cpu_ram_mb: 1082028
67
- gpus:
68
- - Instinct MI210
69
- - Instinct MI210
70
- - Instinct MI210
71
- - Instinct MI210
72
- - Instinct MI210
73
- - Instinct MI210
74
- - Instinct MI210
75
- - Instinct MI210
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/bert_1gpu_training/0/.config/hydra.yaml DELETED
@@ -1,174 +0,0 @@
1
- hydra:
2
- run:
3
- dir: runs/${oc.env:COMMIT_DATE_GMT}_${oc.env:COMMIT_SHA}/${experiment_name}
4
- sweep:
5
- dir: sweeps/${oc.env:COMMIT_DATE_GMT}_${oc.env:COMMIT_SHA}/${experiment_name}
6
- subdir: ${hydra.job.num}
7
- launcher:
8
- _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
- sweeper:
10
- _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
- max_batch_size: null
12
- params:
13
- +benchmark.training_arguments.per_device_train_batch_size: '32'
14
- backend.torch_dtype: float16,float32
15
- help:
16
- app_name: ${hydra.job.name}
17
- header: '${hydra.help.app_name} is powered by Hydra.
18
-
19
- '
20
- footer: 'Powered by Hydra (https://hydra.cc)
21
-
22
- Use --hydra-help to view Hydra specific help
23
-
24
- '
25
- template: '${hydra.help.header}
26
-
27
- == Configuration groups ==
28
-
29
- Compose your configuration from those groups (group=option)
30
-
31
-
32
- $APP_CONFIG_GROUPS
33
-
34
-
35
- == Config ==
36
-
37
- Override anything in the config (foo.bar=value)
38
-
39
-
40
- $CONFIG
41
-
42
-
43
- ${hydra.help.footer}
44
-
45
- '
46
- hydra_help:
47
- template: 'Hydra (${hydra.runtime.version})
48
-
49
- See https://hydra.cc for more info.
50
-
51
-
52
- == Flags ==
53
-
54
- $FLAGS_HELP
55
-
56
-
57
- == Configuration groups ==
58
-
59
- Compose your configuration from those groups (For example, append hydra/job_logging=disabled
60
- to command line)
61
-
62
-
63
- $HYDRA_CONFIG_GROUPS
64
-
65
-
66
- Use ''--cfg hydra'' to Show the Hydra config.
67
-
68
- '
69
- hydra_help: ???
70
- hydra_logging:
71
- version: 1
72
- formatters:
73
- colorlog:
74
- (): colorlog.ColoredFormatter
75
- format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
76
- handlers:
77
- console:
78
- class: logging.StreamHandler
79
- formatter: colorlog
80
- stream: ext://sys.stdout
81
- root:
82
- level: INFO
83
- handlers:
84
- - console
85
- disable_existing_loggers: false
86
- job_logging:
87
- version: 1
88
- formatters:
89
- simple:
90
- format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
- colorlog:
92
- (): colorlog.ColoredFormatter
93
- format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
94
- - %(message)s'
95
- log_colors:
96
- DEBUG: purple
97
- INFO: green
98
- WARNING: yellow
99
- ERROR: red
100
- CRITICAL: red
101
- handlers:
102
- console:
103
- class: logging.StreamHandler
104
- formatter: colorlog
105
- stream: ext://sys.stdout
106
- file:
107
- class: logging.FileHandler
108
- formatter: simple
109
- filename: ${hydra.job.name}.log
110
- root:
111
- level: INFO
112
- handlers:
113
- - console
114
- - file
115
- disable_existing_loggers: false
116
- env: {}
117
- mode: MULTIRUN
118
- searchpath: []
119
- callbacks: {}
120
- output_subdir: .hydra
121
- overrides:
122
- hydra:
123
- - hydra.mode=MULTIRUN
124
- task:
125
- - +benchmark.training_arguments.per_device_train_batch_size=32
126
- - backend.torch_dtype=float16
127
- job:
128
- name: experiment
129
- chdir: true
130
- override_dirname: +benchmark.training_arguments.per_device_train_batch_size=32,backend.torch_dtype=float16
131
- id: '0'
132
- num: 0
133
- config_name: bert_1gpu_training
134
- env_set: {}
135
- env_copy: []
136
- config:
137
- override_dirname:
138
- kv_sep: '='
139
- item_sep: ','
140
- exclude_keys: []
141
- runtime:
142
- version: 1.3.2
143
- version_base: '1.3'
144
- cwd: /home/user/transformers-regression
145
- config_sources:
146
- - path: hydra.conf
147
- schema: pkg
148
- provider: hydra
149
- - path: optimum_benchmark
150
- schema: pkg
151
- provider: main
152
- - path: hydra_plugins.hydra_colorlog.conf
153
- schema: pkg
154
- provider: hydra-colorlog
155
- - path: /home/user/transformers-regression/configs
156
- schema: file
157
- provider: command-line
158
- - path: ''
159
- schema: structured
160
- provider: schema
161
- output_dir: /home/user/transformers-regression/sweeps/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/bert_1gpu_training/0
162
- choices:
163
- benchmark: training
164
- backend: pytorch
165
- hydra/env: default
166
- hydra/callbacks: null
167
- hydra/job_logging: colorlog
168
- hydra/hydra_logging: colorlog
169
- hydra/hydra_help: default
170
- hydra/help: default
171
- hydra/sweeper: basic
172
- hydra/launcher: basic
173
- hydra/output: default
174
- verbose: false
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/bert_1gpu_training/0/.config/overrides.yaml DELETED
@@ -1,2 +0,0 @@
1
- - +benchmark.training_arguments.per_device_train_batch_size=32
2
- - backend.torch_dtype=float16
 
 
 
raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/bert_1gpu_training/0/experiment.log DELETED
@@ -1,17 +0,0 @@
1
- [2023-09-27 11:57:31,928][experiment][WARNING] - Multiple GPUs detected but CUDA_DEVICE_ORDER is not set. This means that code might allocate resources from the wrong GPUs even if CUDA_VISIBLE_DEVICES is set. Pytorch uses the `FASTEST_FIRST` order by default, which is not guaranteed to be the same as nvidia-smi. `CUDA_DEVICE_ORDER` will be set to `PCI_BUS_ID` to ensure that the GPUs are allocated in the same order as nvidia-smi.
2
- [2023-09-27 11:57:34,386][pytorch][INFO] - + Inferred AutoModel class AutoModelForSequenceClassification for task text-classification and model_type bert
3
- [2023-09-27 11:57:34,386][backend][INFO] - Configuring pytorch backend
4
- [2023-09-27 11:57:34,389][backend][INFO] - + Checking initial device(s) isolation of CUDA device(s): [0]
5
- [2023-09-27 11:57:34,517][backend][INFO] - + Checking continuous device(s) isolation of CUDA device(s): [0]
6
- [2023-09-27 11:57:34,530][pytorch][INFO] - + Loading model on device: cuda
7
- [2023-09-27 11:57:35,285][benchmark][INFO] - Configuring training benchmark
8
- [2023-09-27 11:57:35,286][training][INFO] - Running training benchmark
9
- [2023-09-27 11:57:35,286][dataset_generator][INFO] - Using text-classification task generator
10
- [2023-09-27 11:57:35,335][pytorch][INFO] - + Setting dataset format to `torch`.
11
- [2023-09-27 11:57:35,335][pytorch][INFO] - + Wrapping training arguments with transformers.TrainingArguments
12
- [2023-09-27 11:57:35,337][pytorch][INFO] - + Wrapping model with transformers.Trainer
13
- [2023-09-27 11:57:35,341][pytorch][INFO] - + Starting training
14
- [2023-09-27 11:57:55,096][pytorch][INFO] - + Training finished successfully
15
- [2023-09-27 11:57:55,097][training][INFO] - Saving training results
16
- [2023-09-27 11:57:55,100][backend][INFO] - Cleaning pytorch backend
17
- [2023-09-27 11:57:55,100][backend][INFO] - + Deleting pretrained model
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/bert_1gpu_training/0/hydra_config.yaml DELETED
@@ -1,75 +0,0 @@
1
- backend:
2
- name: pytorch
3
- version: 2.1.0+rocm5.6
4
- _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
5
- seed: 42
6
- inter_op_num_threads: null
7
- intra_op_num_threads: null
8
- initial_isolation_check: true
9
- continous_isolation_check: true
10
- delete_cache: false
11
- no_weights: false
12
- device_map: null
13
- torch_dtype: float16
14
- disable_grad: false
15
- eval_mode: false
16
- amp_autocast: false
17
- amp_dtype: null
18
- torch_compile: false
19
- torch_compile_config: {}
20
- bettertransformer: false
21
- quantization_scheme: null
22
- quantization_config: {}
23
- use_ddp: false
24
- ddp_config: {}
25
- peft_strategy: null
26
- peft_config: {}
27
- benchmark:
28
- name: training
29
- _target_: optimum_benchmark.benchmarks.training.benchmark.TrainingBenchmark
30
- warmup_steps: 40
31
- dataset_shapes:
32
- dataset_size: 1500
33
- sequence_length: 256
34
- num_choices: 1
35
- feature_size: 80
36
- nb_max_frames: 3000
37
- audio_sequence_length: 16000
38
- training_arguments:
39
- skip_memory_metrics: true
40
- output_dir: ./trainer_output
41
- use_cpu: false
42
- ddp_find_unused_parameters: false
43
- do_train: true
44
- do_eval: false
45
- do_predict: false
46
- report_to: none
47
- per_device_train_batch_size: 32
48
- experiment_name: bert_1gpu_training
49
- model: bert-base-uncased
50
- device: cuda
51
- task: text-classification
52
- hub_kwargs:
53
- revision: main
54
- cache_dir: null
55
- force_download: false
56
- local_files_only: false
57
- environment:
58
- optimum_version: 1.13.1
59
- transformers_version: 4.34.0.dev0
60
- accelerate_version: 0.23.0
61
- diffusers_version: null
62
- python_version: 3.10.12
63
- system: Linux
64
- cpu: ' AMD EPYC 7643 48-Core Processor'
65
- cpu_count: 96
66
- cpu_ram_mb: 1082028
67
- gpus:
68
- - Instinct MI210
69
- - Instinct MI210
70
- - Instinct MI210
71
- - Instinct MI210
72
- - Instinct MI210
73
- - Instinct MI210
74
- - Instinct MI210
75
- - Instinct MI210
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/bert_1gpu_training/0/training_results.csv DELETED
@@ -1,2 +0,0 @@
1
- warmup.runtime(s),warmup.throughput(samples/s),training.runtime(s),training.throughput(samples/s),overall_training.runtime(s),overall_training.throughput(samples/s)
2
- 6.112920045852661,209.39256368458834,13.53417706489563,238.802845899144,19.647098064422607,164.50266545228763
 
 
 
raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/bert_1gpu_training/1/.config/config.yaml DELETED
@@ -1,75 +0,0 @@
1
- backend:
2
- name: pytorch
3
- version: ${pytorch_version:}
4
- _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
5
- seed: 42
6
- inter_op_num_threads: null
7
- intra_op_num_threads: null
8
- initial_isolation_check: true
9
- continous_isolation_check: true
10
- delete_cache: false
11
- no_weights: false
12
- device_map: null
13
- torch_dtype: float32
14
- disable_grad: ${is_inference:${benchmark.name}}
15
- eval_mode: ${is_inference:${benchmark.name}}
16
- amp_autocast: false
17
- amp_dtype: null
18
- torch_compile: false
19
- torch_compile_config: {}
20
- bettertransformer: false
21
- quantization_scheme: null
22
- quantization_config: {}
23
- use_ddp: false
24
- ddp_config: {}
25
- peft_strategy: null
26
- peft_config: {}
27
- benchmark:
28
- name: training
29
- _target_: optimum_benchmark.benchmarks.training.benchmark.TrainingBenchmark
30
- warmup_steps: 40
31
- dataset_shapes:
32
- dataset_size: 1500
33
- sequence_length: 256
34
- num_choices: 1
35
- feature_size: 80
36
- nb_max_frames: 3000
37
- audio_sequence_length: 16000
38
- training_arguments:
39
- skip_memory_metrics: true
40
- output_dir: ./trainer_output
41
- use_cpu: ${is_cpu:${device}}
42
- ddp_find_unused_parameters: false
43
- do_train: true
44
- do_eval: false
45
- do_predict: false
46
- report_to: none
47
- per_device_train_batch_size: 32
48
- experiment_name: bert_1gpu_training
49
- model: bert-base-uncased
50
- device: cuda
51
- task: text-classification
52
- hub_kwargs:
53
- revision: main
54
- cache_dir: null
55
- force_download: false
56
- local_files_only: false
57
- environment:
58
- optimum_version: 1.13.1
59
- transformers_version: 4.34.0.dev0
60
- accelerate_version: 0.23.0
61
- diffusers_version: null
62
- python_version: 3.10.12
63
- system: Linux
64
- cpu: ' AMD EPYC 7643 48-Core Processor'
65
- cpu_count: 96
66
- cpu_ram_mb: 1082028
67
- gpus:
68
- - Instinct MI210
69
- - Instinct MI210
70
- - Instinct MI210
71
- - Instinct MI210
72
- - Instinct MI210
73
- - Instinct MI210
74
- - Instinct MI210
75
- - Instinct MI210
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/bert_1gpu_training/1/.config/hydra.yaml DELETED
@@ -1,174 +0,0 @@
1
- hydra:
2
- run:
3
- dir: runs/${oc.env:COMMIT_DATE_GMT}_${oc.env:COMMIT_SHA}/${experiment_name}
4
- sweep:
5
- dir: sweeps/${oc.env:COMMIT_DATE_GMT}_${oc.env:COMMIT_SHA}/${experiment_name}
6
- subdir: ${hydra.job.num}
7
- launcher:
8
- _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
- sweeper:
10
- _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
- max_batch_size: null
12
- params:
13
- +benchmark.training_arguments.per_device_train_batch_size: '32'
14
- backend.torch_dtype: float16,float32
15
- help:
16
- app_name: ${hydra.job.name}
17
- header: '${hydra.help.app_name} is powered by Hydra.
18
-
19
- '
20
- footer: 'Powered by Hydra (https://hydra.cc)
21
-
22
- Use --hydra-help to view Hydra specific help
23
-
24
- '
25
- template: '${hydra.help.header}
26
-
27
- == Configuration groups ==
28
-
29
- Compose your configuration from those groups (group=option)
30
-
31
-
32
- $APP_CONFIG_GROUPS
33
-
34
-
35
- == Config ==
36
-
37
- Override anything in the config (foo.bar=value)
38
-
39
-
40
- $CONFIG
41
-
42
-
43
- ${hydra.help.footer}
44
-
45
- '
46
- hydra_help:
47
- template: 'Hydra (${hydra.runtime.version})
48
-
49
- See https://hydra.cc for more info.
50
-
51
-
52
- == Flags ==
53
-
54
- $FLAGS_HELP
55
-
56
-
57
- == Configuration groups ==
58
-
59
- Compose your configuration from those groups (For example, append hydra/job_logging=disabled
60
- to command line)
61
-
62
-
63
- $HYDRA_CONFIG_GROUPS
64
-
65
-
66
- Use ''--cfg hydra'' to Show the Hydra config.
67
-
68
- '
69
- hydra_help: ???
70
- hydra_logging:
71
- version: 1
72
- formatters:
73
- colorlog:
74
- (): colorlog.ColoredFormatter
75
- format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
76
- handlers:
77
- console:
78
- class: logging.StreamHandler
79
- formatter: colorlog
80
- stream: ext://sys.stdout
81
- root:
82
- level: INFO
83
- handlers:
84
- - console
85
- disable_existing_loggers: false
86
- job_logging:
87
- version: 1
88
- formatters:
89
- simple:
90
- format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
- colorlog:
92
- (): colorlog.ColoredFormatter
93
- format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
94
- - %(message)s'
95
- log_colors:
96
- DEBUG: purple
97
- INFO: green
98
- WARNING: yellow
99
- ERROR: red
100
- CRITICAL: red
101
- handlers:
102
- console:
103
- class: logging.StreamHandler
104
- formatter: colorlog
105
- stream: ext://sys.stdout
106
- file:
107
- class: logging.FileHandler
108
- formatter: simple
109
- filename: ${hydra.job.name}.log
110
- root:
111
- level: INFO
112
- handlers:
113
- - console
114
- - file
115
- disable_existing_loggers: false
116
- env: {}
117
- mode: MULTIRUN
118
- searchpath: []
119
- callbacks: {}
120
- output_subdir: .hydra
121
- overrides:
122
- hydra:
123
- - hydra.mode=MULTIRUN
124
- task:
125
- - +benchmark.training_arguments.per_device_train_batch_size=32
126
- - backend.torch_dtype=float32
127
- job:
128
- name: experiment
129
- chdir: true
130
- override_dirname: +benchmark.training_arguments.per_device_train_batch_size=32,backend.torch_dtype=float32
131
- id: '1'
132
- num: 1
133
- config_name: bert_1gpu_training
134
- env_set: {}
135
- env_copy: []
136
- config:
137
- override_dirname:
138
- kv_sep: '='
139
- item_sep: ','
140
- exclude_keys: []
141
- runtime:
142
- version: 1.3.2
143
- version_base: '1.3'
144
- cwd: /home/user/transformers-regression
145
- config_sources:
146
- - path: hydra.conf
147
- schema: pkg
148
- provider: hydra
149
- - path: optimum_benchmark
150
- schema: pkg
151
- provider: main
152
- - path: hydra_plugins.hydra_colorlog.conf
153
- schema: pkg
154
- provider: hydra-colorlog
155
- - path: /home/user/transformers-regression/configs
156
- schema: file
157
- provider: command-line
158
- - path: ''
159
- schema: structured
160
- provider: schema
161
- output_dir: /home/user/transformers-regression/sweeps/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/bert_1gpu_training/1
162
- choices:
163
- benchmark: training
164
- backend: pytorch
165
- hydra/env: default
166
- hydra/callbacks: null
167
- hydra/job_logging: colorlog
168
- hydra/hydra_logging: colorlog
169
- hydra/hydra_help: default
170
- hydra/help: default
171
- hydra/sweeper: basic
172
- hydra/launcher: basic
173
- hydra/output: default
174
- verbose: false
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/bert_1gpu_training/1/.config/overrides.yaml DELETED
@@ -1,2 +0,0 @@
1
- - +benchmark.training_arguments.per_device_train_batch_size=32
2
- - backend.torch_dtype=float32
 
 
 
raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/bert_1gpu_training/1/experiment.log DELETED
@@ -1,16 +0,0 @@
1
- [2023-09-27 11:57:56,611][pytorch][INFO] - + Inferred AutoModel class AutoModelForSequenceClassification for task text-classification and model_type bert
2
- [2023-09-27 11:57:56,612][backend][INFO] - Configuring pytorch backend
3
- [2023-09-27 11:57:56,612][backend][INFO] - + Checking initial device(s) isolation of CUDA device(s): [0]
4
- [2023-09-27 11:57:56,733][backend][INFO] - + Checking continuous device(s) isolation of CUDA device(s): [0]
5
- [2023-09-27 11:57:56,749][pytorch][INFO] - + Loading model on device: cuda
6
- [2023-09-27 11:57:57,432][benchmark][INFO] - Configuring training benchmark
7
- [2023-09-27 11:57:57,432][training][INFO] - Running training benchmark
8
- [2023-09-27 11:57:57,433][dataset_generator][INFO] - Using text-classification task generator
9
- [2023-09-27 11:57:57,467][pytorch][INFO] - + Setting dataset format to `torch`.
10
- [2023-09-27 11:57:57,468][pytorch][INFO] - + Wrapping training arguments with transformers.TrainingArguments
11
- [2023-09-27 11:57:57,469][pytorch][INFO] - + Wrapping model with transformers.Trainer
12
- [2023-09-27 11:57:57,474][pytorch][INFO] - + Starting training
13
- [2023-09-27 11:58:50,280][pytorch][INFO] - + Training finished successfully
14
- [2023-09-27 11:58:50,281][training][INFO] - Saving training results
15
- [2023-09-27 11:58:50,282][backend][INFO] - Cleaning pytorch backend
16
- [2023-09-27 11:58:50,282][backend][INFO] - + Deleting pretrained model
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/bert_1gpu_training/1/hydra_config.yaml DELETED
@@ -1,75 +0,0 @@
1
- backend:
2
- name: pytorch
3
- version: 2.1.0+rocm5.6
4
- _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
5
- seed: 42
6
- inter_op_num_threads: null
7
- intra_op_num_threads: null
8
- initial_isolation_check: true
9
- continous_isolation_check: true
10
- delete_cache: false
11
- no_weights: false
12
- device_map: null
13
- torch_dtype: float32
14
- disable_grad: false
15
- eval_mode: false
16
- amp_autocast: false
17
- amp_dtype: null
18
- torch_compile: false
19
- torch_compile_config: {}
20
- bettertransformer: false
21
- quantization_scheme: null
22
- quantization_config: {}
23
- use_ddp: false
24
- ddp_config: {}
25
- peft_strategy: null
26
- peft_config: {}
27
- benchmark:
28
- name: training
29
- _target_: optimum_benchmark.benchmarks.training.benchmark.TrainingBenchmark
30
- warmup_steps: 40
31
- dataset_shapes:
32
- dataset_size: 1500
33
- sequence_length: 256
34
- num_choices: 1
35
- feature_size: 80
36
- nb_max_frames: 3000
37
- audio_sequence_length: 16000
38
- training_arguments:
39
- skip_memory_metrics: true
40
- output_dir: ./trainer_output
41
- use_cpu: false
42
- ddp_find_unused_parameters: false
43
- do_train: true
44
- do_eval: false
45
- do_predict: false
46
- report_to: none
47
- per_device_train_batch_size: 32
48
- experiment_name: bert_1gpu_training
49
- model: bert-base-uncased
50
- device: cuda
51
- task: text-classification
52
- hub_kwargs:
53
- revision: main
54
- cache_dir: null
55
- force_download: false
56
- local_files_only: false
57
- environment:
58
- optimum_version: 1.13.1
59
- transformers_version: 4.34.0.dev0
60
- accelerate_version: 0.23.0
61
- diffusers_version: null
62
- python_version: 3.10.12
63
- system: Linux
64
- cpu: ' AMD EPYC 7643 48-Core Processor'
65
- cpu_count: 96
66
- cpu_ram_mb: 1082028
67
- gpus:
68
- - Instinct MI210
69
- - Instinct MI210
70
- - Instinct MI210
71
- - Instinct MI210
72
- - Instinct MI210
73
- - Instinct MI210
74
- - Instinct MI210
75
- - Instinct MI210
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/bert_1gpu_training/1/training_results.csv DELETED
@@ -1,2 +0,0 @@
1
- warmup.runtime(s),warmup.throughput(samples/s),training.runtime(s),training.throughput(samples/s),overall_training.runtime(s),overall_training.throughput(samples/s)
2
- 15.344141006469727,83.41946280735421,37.33880257606506,86.55874792492082,52.682945013046265,61.34812697353263
 
 
 
raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/bert_1gpu_training/multirun.yaml DELETED
@@ -1,246 +0,0 @@
1
- hydra:
2
- run:
3
- dir: runs/${oc.env:COMMIT_DATE_GMT}_${oc.env:COMMIT_SHA}/${experiment_name}
4
- sweep:
5
- dir: sweeps/${oc.env:COMMIT_DATE_GMT}_${oc.env:COMMIT_SHA}/${experiment_name}
6
- subdir: ${hydra.job.num}
7
- launcher:
8
- _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
- sweeper:
10
- _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
- max_batch_size: null
12
- params:
13
- +benchmark.training_arguments.per_device_train_batch_size: '32'
14
- backend.torch_dtype: float16,float32
15
- help:
16
- app_name: ${hydra.job.name}
17
- header: '${hydra.help.app_name} is powered by Hydra.
18
-
19
- '
20
- footer: 'Powered by Hydra (https://hydra.cc)
21
-
22
- Use --hydra-help to view Hydra specific help
23
-
24
- '
25
- template: '${hydra.help.header}
26
-
27
- == Configuration groups ==
28
-
29
- Compose your configuration from those groups (group=option)
30
-
31
-
32
- $APP_CONFIG_GROUPS
33
-
34
-
35
- == Config ==
36
-
37
- Override anything in the config (foo.bar=value)
38
-
39
-
40
- $CONFIG
41
-
42
-
43
- ${hydra.help.footer}
44
-
45
- '
46
- hydra_help:
47
- template: 'Hydra (${hydra.runtime.version})
48
-
49
- See https://hydra.cc for more info.
50
-
51
-
52
- == Flags ==
53
-
54
- $FLAGS_HELP
55
-
56
-
57
- == Configuration groups ==
58
-
59
- Compose your configuration from those groups (For example, append hydra/job_logging=disabled
60
- to command line)
61
-
62
-
63
- $HYDRA_CONFIG_GROUPS
64
-
65
-
66
- Use ''--cfg hydra'' to Show the Hydra config.
67
-
68
- '
69
- hydra_help: ???
70
- hydra_logging:
71
- version: 1
72
- formatters:
73
- colorlog:
74
- (): colorlog.ColoredFormatter
75
- format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
76
- handlers:
77
- console:
78
- class: logging.StreamHandler
79
- formatter: colorlog
80
- stream: ext://sys.stdout
81
- root:
82
- level: INFO
83
- handlers:
84
- - console
85
- disable_existing_loggers: false
86
- job_logging:
87
- version: 1
88
- formatters:
89
- simple:
90
- format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
- colorlog:
92
- (): colorlog.ColoredFormatter
93
- format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
94
- - %(message)s'
95
- log_colors:
96
- DEBUG: purple
97
- INFO: green
98
- WARNING: yellow
99
- ERROR: red
100
- CRITICAL: red
101
- handlers:
102
- console:
103
- class: logging.StreamHandler
104
- formatter: colorlog
105
- stream: ext://sys.stdout
106
- file:
107
- class: logging.FileHandler
108
- formatter: simple
109
- filename: ${hydra.job.name}.log
110
- root:
111
- level: INFO
112
- handlers:
113
- - console
114
- - file
115
- disable_existing_loggers: false
116
- env: {}
117
- mode: MULTIRUN
118
- searchpath: []
119
- callbacks: {}
120
- output_subdir: .hydra
121
- overrides:
122
- hydra:
123
- - hydra.mode=MULTIRUN
124
- task: []
125
- job:
126
- name: experiment
127
- chdir: true
128
- override_dirname: ''
129
- id: ???
130
- num: ???
131
- config_name: bert_1gpu_training
132
- env_set: {}
133
- env_copy: []
134
- config:
135
- override_dirname:
136
- kv_sep: '='
137
- item_sep: ','
138
- exclude_keys: []
139
- runtime:
140
- version: 1.3.2
141
- version_base: '1.3'
142
- cwd: /home/user/transformers-regression
143
- config_sources:
144
- - path: hydra.conf
145
- schema: pkg
146
- provider: hydra
147
- - path: optimum_benchmark
148
- schema: pkg
149
- provider: main
150
- - path: hydra_plugins.hydra_colorlog.conf
151
- schema: pkg
152
- provider: hydra-colorlog
153
- - path: /home/user/transformers-regression/configs
154
- schema: file
155
- provider: command-line
156
- - path: ''
157
- schema: structured
158
- provider: schema
159
- output_dir: ???
160
- choices:
161
- benchmark: training
162
- backend: pytorch
163
- hydra/env: default
164
- hydra/callbacks: null
165
- hydra/job_logging: colorlog
166
- hydra/hydra_logging: colorlog
167
- hydra/hydra_help: default
168
- hydra/help: default
169
- hydra/sweeper: basic
170
- hydra/launcher: basic
171
- hydra/output: default
172
- verbose: false
173
- backend:
174
- name: pytorch
175
- version: ${pytorch_version:}
176
- _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
177
- seed: 42
178
- inter_op_num_threads: null
179
- intra_op_num_threads: null
180
- initial_isolation_check: true
181
- continous_isolation_check: true
182
- delete_cache: false
183
- no_weights: false
184
- device_map: null
185
- torch_dtype: null
186
- disable_grad: ${is_inference:${benchmark.name}}
187
- eval_mode: ${is_inference:${benchmark.name}}
188
- amp_autocast: false
189
- amp_dtype: null
190
- torch_compile: false
191
- torch_compile_config: {}
192
- bettertransformer: false
193
- quantization_scheme: null
194
- quantization_config: {}
195
- use_ddp: false
196
- ddp_config: {}
197
- peft_strategy: null
198
- peft_config: {}
199
- benchmark:
200
- name: training
201
- _target_: optimum_benchmark.benchmarks.training.benchmark.TrainingBenchmark
202
- warmup_steps: 40
203
- dataset_shapes:
204
- dataset_size: 1500
205
- sequence_length: 256
206
- num_choices: 1
207
- feature_size: 80
208
- nb_max_frames: 3000
209
- audio_sequence_length: 16000
210
- training_arguments:
211
- skip_memory_metrics: true
212
- output_dir: ./trainer_output
213
- use_cpu: ${is_cpu:${device}}
214
- ddp_find_unused_parameters: false
215
- do_train: true
216
- do_eval: false
217
- do_predict: false
218
- report_to: none
219
- experiment_name: bert_1gpu_training
220
- model: bert-base-uncased
221
- device: cuda
222
- task: text-classification
223
- hub_kwargs:
224
- revision: main
225
- cache_dir: null
226
- force_download: false
227
- local_files_only: false
228
- environment:
229
- optimum_version: 1.13.1
230
- transformers_version: 4.34.0.dev0
231
- accelerate_version: 0.23.0
232
- diffusers_version: null
233
- python_version: 3.10.12
234
- system: Linux
235
- cpu: ' AMD EPYC 7643 48-Core Processor'
236
- cpu_count: 96
237
- cpu_ram_mb: 1082028
238
- gpus:
239
- - Instinct MI210
240
- - Instinct MI210
241
- - Instinct MI210
242
- - Instinct MI210
243
- - Instinct MI210
244
- - Instinct MI210
245
- - Instinct MI210
246
- - Instinct MI210
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/llama_1gpu_inference/0/.config/config.yaml DELETED
@@ -1,73 +0,0 @@
1
- backend:
2
- name: pytorch
3
- version: ${pytorch_version:}
4
- _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
5
- seed: 42
6
- inter_op_num_threads: null
7
- intra_op_num_threads: null
8
- initial_isolation_check: true
9
- continous_isolation_check: true
10
- delete_cache: false
11
- no_weights: false
12
- device_map: null
13
- torch_dtype: float16
14
- disable_grad: ${is_inference:${benchmark.name}}
15
- eval_mode: ${is_inference:${benchmark.name}}
16
- amp_autocast: false
17
- amp_dtype: null
18
- torch_compile: false
19
- torch_compile_config: {}
20
- bettertransformer: false
21
- quantization_scheme: null
22
- quantization_config: {}
23
- use_ddp: false
24
- ddp_config: {}
25
- peft_strategy: null
26
- peft_config: {}
27
- benchmark:
28
- name: inference
29
- _target_: optimum_benchmark.benchmarks.inference.benchmark.InferenceBenchmark
30
- duration: 10
31
- warmup_runs: 10
32
- memory: false
33
- energy: false
34
- input_shapes:
35
- batch_size: 1
36
- sequence_length: 200
37
- num_choices: 1
38
- feature_size: 80
39
- nb_max_frames: 3000
40
- audio_sequence_length: 16000
41
- new_tokens: 200
42
- can_diffuse: ${can_diffuse:${task}}
43
- can_generate: ${can_generate:${task}}
44
- forward_kwargs: {}
45
- generate_kwargs: {}
46
- experiment_name: llama_1gpu_inference
47
- model: fxmarty/tiny-llama-fast-tokenizer
48
- device: cuda
49
- task: ${infer_task:${model}}
50
- hub_kwargs:
51
- revision: main
52
- cache_dir: null
53
- force_download: false
54
- local_files_only: false
55
- environment:
56
- optimum_version: 1.13.1
57
- transformers_version: 4.34.0.dev0
58
- accelerate_version: 0.23.0
59
- diffusers_version: null
60
- python_version: 3.10.12
61
- system: Linux
62
- cpu: ' AMD EPYC 7643 48-Core Processor'
63
- cpu_count: 96
64
- cpu_ram_mb: 1082028
65
- gpus:
66
- - Instinct MI210
67
- - Instinct MI210
68
- - Instinct MI210
69
- - Instinct MI210
70
- - Instinct MI210
71
- - Instinct MI210
72
- - Instinct MI210
73
- - Instinct MI210
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/llama_1gpu_inference/0/.config/hydra.yaml DELETED
@@ -1,174 +0,0 @@
1
- hydra:
2
- run:
3
- dir: runs/${oc.env:COMMIT_DATE_GMT}_${oc.env:COMMIT_SHA}/${experiment_name}
4
- sweep:
5
- dir: sweeps/${oc.env:COMMIT_DATE_GMT}_${oc.env:COMMIT_SHA}/${experiment_name}
6
- subdir: ${hydra.job.num}
7
- launcher:
8
- _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
- sweeper:
10
- _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
- max_batch_size: null
12
- params:
13
- benchmark.input_shapes.batch_size: 1,16
14
- backend.torch_dtype: float16,float32
15
- help:
16
- app_name: ${hydra.job.name}
17
- header: '${hydra.help.app_name} is powered by Hydra.
18
-
19
- '
20
- footer: 'Powered by Hydra (https://hydra.cc)
21
-
22
- Use --hydra-help to view Hydra specific help
23
-
24
- '
25
- template: '${hydra.help.header}
26
-
27
- == Configuration groups ==
28
-
29
- Compose your configuration from those groups (group=option)
30
-
31
-
32
- $APP_CONFIG_GROUPS
33
-
34
-
35
- == Config ==
36
-
37
- Override anything in the config (foo.bar=value)
38
-
39
-
40
- $CONFIG
41
-
42
-
43
- ${hydra.help.footer}
44
-
45
- '
46
- hydra_help:
47
- template: 'Hydra (${hydra.runtime.version})
48
-
49
- See https://hydra.cc for more info.
50
-
51
-
52
- == Flags ==
53
-
54
- $FLAGS_HELP
55
-
56
-
57
- == Configuration groups ==
58
-
59
- Compose your configuration from those groups (For example, append hydra/job_logging=disabled
60
- to command line)
61
-
62
-
63
- $HYDRA_CONFIG_GROUPS
64
-
65
-
66
- Use ''--cfg hydra'' to Show the Hydra config.
67
-
68
- '
69
- hydra_help: ???
70
- hydra_logging:
71
- version: 1
72
- formatters:
73
- colorlog:
74
- (): colorlog.ColoredFormatter
75
- format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
76
- handlers:
77
- console:
78
- class: logging.StreamHandler
79
- formatter: colorlog
80
- stream: ext://sys.stdout
81
- root:
82
- level: INFO
83
- handlers:
84
- - console
85
- disable_existing_loggers: false
86
- job_logging:
87
- version: 1
88
- formatters:
89
- simple:
90
- format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
- colorlog:
92
- (): colorlog.ColoredFormatter
93
- format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
94
- - %(message)s'
95
- log_colors:
96
- DEBUG: purple
97
- INFO: green
98
- WARNING: yellow
99
- ERROR: red
100
- CRITICAL: red
101
- handlers:
102
- console:
103
- class: logging.StreamHandler
104
- formatter: colorlog
105
- stream: ext://sys.stdout
106
- file:
107
- class: logging.FileHandler
108
- formatter: simple
109
- filename: ${hydra.job.name}.log
110
- root:
111
- level: INFO
112
- handlers:
113
- - console
114
- - file
115
- disable_existing_loggers: false
116
- env: {}
117
- mode: MULTIRUN
118
- searchpath: []
119
- callbacks: {}
120
- output_subdir: .hydra
121
- overrides:
122
- hydra:
123
- - hydra.mode=MULTIRUN
124
- task:
125
- - benchmark.input_shapes.batch_size=1
126
- - backend.torch_dtype=float16
127
- job:
128
- name: experiment
129
- chdir: true
130
- override_dirname: backend.torch_dtype=float16,benchmark.input_shapes.batch_size=1
131
- id: '0'
132
- num: 0
133
- config_name: llama2_1gpu_inference
134
- env_set: {}
135
- env_copy: []
136
- config:
137
- override_dirname:
138
- kv_sep: '='
139
- item_sep: ','
140
- exclude_keys: []
141
- runtime:
142
- version: 1.3.2
143
- version_base: '1.3'
144
- cwd: /home/user/transformers-regression
145
- config_sources:
146
- - path: hydra.conf
147
- schema: pkg
148
- provider: hydra
149
- - path: optimum_benchmark
150
- schema: pkg
151
- provider: main
152
- - path: hydra_plugins.hydra_colorlog.conf
153
- schema: pkg
154
- provider: hydra-colorlog
155
- - path: /home/user/transformers-regression/configs
156
- schema: file
157
- provider: command-line
158
- - path: ''
159
- schema: structured
160
- provider: schema
161
- output_dir: /home/user/transformers-regression/sweeps/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/llama_1gpu_inference/0
162
- choices:
163
- benchmark: inference
164
- backend: pytorch
165
- hydra/env: default
166
- hydra/callbacks: null
167
- hydra/job_logging: colorlog
168
- hydra/hydra_logging: colorlog
169
- hydra/hydra_help: default
170
- hydra/help: default
171
- hydra/sweeper: basic
172
- hydra/launcher: basic
173
- hydra/output: default
174
- verbose: false
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/llama_1gpu_inference/0/.config/overrides.yaml DELETED
@@ -1,2 +0,0 @@
1
- - benchmark.input_shapes.batch_size=1
2
- - backend.torch_dtype=float16
 
 
 
raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/llama_1gpu_inference/0/experiment.log DELETED
@@ -1,27 +0,0 @@
1
- [2023-09-27 11:58:54,429][inference][INFO] - `new_tokens` was set to 200. `max_new_tokens` and `min_new_tokens` will be set to 200.
2
- [2023-09-27 11:58:54,571][experiment][WARNING] - Multiple GPUs detected but CUDA_DEVICE_ORDER is not set. This means that code might allocate resources from the wrong GPUs even if CUDA_VISIBLE_DEVICES is set. Pytorch uses the `FASTEST_FIRST` order by default, which is not guaranteed to be the same as nvidia-smi. `CUDA_DEVICE_ORDER` will be set to `PCI_BUS_ID` to ensure that the GPUs are allocated in the same order as nvidia-smi.
3
- [2023-09-27 11:58:56,806][pytorch][INFO] - + Inferred AutoModel class AutoModelForCausalLM for task text-generation and model_type llama
4
- [2023-09-27 11:58:56,806][backend][INFO] - Configuring pytorch backend
5
- [2023-09-27 11:58:56,807][backend][INFO] - + Checking initial device(s) isolation of CUDA device(s): [0]
6
- [2023-09-27 11:58:56,931][backend][INFO] - + Checking continuous device(s) isolation of CUDA device(s): [0]
7
- [2023-09-27 11:58:56,946][pytorch][INFO] - + Disabling gradients
8
- [2023-09-27 11:58:56,947][pytorch][INFO] - + Loading model on device: cuda
9
- [2023-09-27 11:58:57,576][pytorch][INFO] - + Turning on model's eval mode
10
- [2023-09-27 11:58:57,583][benchmark][INFO] - Configuring inference benchmark
11
- [2023-09-27 11:58:57,584][inference][INFO] - Running inference benchmark
12
- [2023-09-27 11:58:57,584][input_generator][INFO] - Using llama model type generator
13
- [2023-09-27 11:58:57,606][inference][INFO] - + Preparing input for the forward pass
14
- [2023-09-27 11:58:57,606][inference][INFO] - + Warming up the forward pass
15
- [2023-09-27 11:58:57,936][inference][INFO] - + Tracking forward pass latency and throughput
16
- [2023-09-27 11:58:57,937][latency_tracker][INFO] - Tracked Pytorch devices: [0]
17
- [2023-09-27 11:59:08,120][inference][INFO] - + Forward pass latency: 3.28e-03 (s)
18
- [2023-09-27 11:59:08,122][inference][INFO] - + Forward pass throughput: 305.00 (samples/s)
19
- [2023-09-27 11:59:08,122][inference][INFO] - + Preparing input for the generation pass
20
- [2023-09-27 11:59:08,122][inference][INFO] - + Warming up the generation pass
21
- [2023-09-27 11:59:09,154][inference][INFO] - + Tracking generation latency and throughput
22
- [2023-09-27 11:59:09,154][latency_tracker][INFO] - Tracked Pytorch devices: [0]
23
- [2023-09-27 11:59:19,765][inference][INFO] - + Generation pass latency: 5.30e-01 (s)
24
- [2023-09-27 11:59:19,766][inference][INFO] - + Generation pass throughput: 377.00 (tokens/s)
25
- [2023-09-27 11:59:19,766][inference][INFO] - Saving inference results
26
- [2023-09-27 11:59:19,773][backend][INFO] - Cleaning pytorch backend
27
- [2023-09-27 11:59:19,773][backend][INFO] - + Deleting pretrained model
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/llama_1gpu_inference/0/hydra_config.yaml DELETED
@@ -1,79 +0,0 @@
1
- backend:
2
- name: pytorch
3
- version: 2.1.0+rocm5.6
4
- _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
5
- seed: 42
6
- inter_op_num_threads: null
7
- intra_op_num_threads: null
8
- initial_isolation_check: true
9
- continous_isolation_check: true
10
- delete_cache: false
11
- no_weights: false
12
- device_map: null
13
- torch_dtype: float16
14
- disable_grad: true
15
- eval_mode: true
16
- amp_autocast: false
17
- amp_dtype: null
18
- torch_compile: false
19
- torch_compile_config: {}
20
- bettertransformer: false
21
- quantization_scheme: null
22
- quantization_config: {}
23
- use_ddp: false
24
- ddp_config: {}
25
- peft_strategy: null
26
- peft_config: {}
27
- benchmark:
28
- name: inference
29
- _target_: optimum_benchmark.benchmarks.inference.benchmark.InferenceBenchmark
30
- duration: 10
31
- warmup_runs: 10
32
- memory: false
33
- energy: false
34
- input_shapes:
35
- batch_size: 1
36
- sequence_length: 200
37
- num_choices: 1
38
- feature_size: 80
39
- nb_max_frames: 3000
40
- audio_sequence_length: 16000
41
- new_tokens: 200
42
- can_diffuse: false
43
- can_generate: true
44
- forward_kwargs: {}
45
- generate_kwargs:
46
- max_new_tokens: 200
47
- min_new_tokens: 200
48
- do_sample: false
49
- use_cache: true
50
- pad_token_id: 0
51
- num_beams: 1
52
- experiment_name: llama_1gpu_inference
53
- model: fxmarty/tiny-llama-fast-tokenizer
54
- device: cuda
55
- task: text-generation
56
- hub_kwargs:
57
- revision: main
58
- cache_dir: null
59
- force_download: false
60
- local_files_only: false
61
- environment:
62
- optimum_version: 1.13.1
63
- transformers_version: 4.34.0.dev0
64
- accelerate_version: 0.23.0
65
- diffusers_version: null
66
- python_version: 3.10.12
67
- system: Linux
68
- cpu: ' AMD EPYC 7643 48-Core Processor'
69
- cpu_count: 96
70
- cpu_ram_mb: 1082028
71
- gpus:
72
- - Instinct MI210
73
- - Instinct MI210
74
- - Instinct MI210
75
- - Instinct MI210
76
- - Instinct MI210
77
- - Instinct MI210
78
- - Instinct MI210
79
- - Instinct MI210
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/llama_1gpu_inference/0/inference_results.csv DELETED
@@ -1,2 +0,0 @@
1
- forward.latency(s),forward.throughput(samples/s),generate.latency(s),generate.throughput(tokens/s)
2
- 0.00328,305.0,0.53,377.0
 
 
 
raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/llama_1gpu_inference/1/.config/config.yaml DELETED
@@ -1,73 +0,0 @@
1
- backend:
2
- name: pytorch
3
- version: ${pytorch_version:}
4
- _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
5
- seed: 42
6
- inter_op_num_threads: null
7
- intra_op_num_threads: null
8
- initial_isolation_check: true
9
- continous_isolation_check: true
10
- delete_cache: false
11
- no_weights: false
12
- device_map: null
13
- torch_dtype: float32
14
- disable_grad: ${is_inference:${benchmark.name}}
15
- eval_mode: ${is_inference:${benchmark.name}}
16
- amp_autocast: false
17
- amp_dtype: null
18
- torch_compile: false
19
- torch_compile_config: {}
20
- bettertransformer: false
21
- quantization_scheme: null
22
- quantization_config: {}
23
- use_ddp: false
24
- ddp_config: {}
25
- peft_strategy: null
26
- peft_config: {}
27
- benchmark:
28
- name: inference
29
- _target_: optimum_benchmark.benchmarks.inference.benchmark.InferenceBenchmark
30
- duration: 10
31
- warmup_runs: 10
32
- memory: false
33
- energy: false
34
- input_shapes:
35
- batch_size: 1
36
- sequence_length: 200
37
- num_choices: 1
38
- feature_size: 80
39
- nb_max_frames: 3000
40
- audio_sequence_length: 16000
41
- new_tokens: 200
42
- can_diffuse: ${can_diffuse:${task}}
43
- can_generate: ${can_generate:${task}}
44
- forward_kwargs: {}
45
- generate_kwargs: {}
46
- experiment_name: llama_1gpu_inference
47
- model: fxmarty/tiny-llama-fast-tokenizer
48
- device: cuda
49
- task: ${infer_task:${model}}
50
- hub_kwargs:
51
- revision: main
52
- cache_dir: null
53
- force_download: false
54
- local_files_only: false
55
- environment:
56
- optimum_version: 1.13.1
57
- transformers_version: 4.34.0.dev0
58
- accelerate_version: 0.23.0
59
- diffusers_version: null
60
- python_version: 3.10.12
61
- system: Linux
62
- cpu: ' AMD EPYC 7643 48-Core Processor'
63
- cpu_count: 96
64
- cpu_ram_mb: 1082028
65
- gpus:
66
- - Instinct MI210
67
- - Instinct MI210
68
- - Instinct MI210
69
- - Instinct MI210
70
- - Instinct MI210
71
- - Instinct MI210
72
- - Instinct MI210
73
- - Instinct MI210
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/llama_1gpu_inference/1/.config/hydra.yaml DELETED
@@ -1,174 +0,0 @@
1
- hydra:
2
- run:
3
- dir: runs/${oc.env:COMMIT_DATE_GMT}_${oc.env:COMMIT_SHA}/${experiment_name}
4
- sweep:
5
- dir: sweeps/${oc.env:COMMIT_DATE_GMT}_${oc.env:COMMIT_SHA}/${experiment_name}
6
- subdir: ${hydra.job.num}
7
- launcher:
8
- _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
- sweeper:
10
- _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
- max_batch_size: null
12
- params:
13
- benchmark.input_shapes.batch_size: 1,16
14
- backend.torch_dtype: float16,float32
15
- help:
16
- app_name: ${hydra.job.name}
17
- header: '${hydra.help.app_name} is powered by Hydra.
18
-
19
- '
20
- footer: 'Powered by Hydra (https://hydra.cc)
21
-
22
- Use --hydra-help to view Hydra specific help
23
-
24
- '
25
- template: '${hydra.help.header}
26
-
27
- == Configuration groups ==
28
-
29
- Compose your configuration from those groups (group=option)
30
-
31
-
32
- $APP_CONFIG_GROUPS
33
-
34
-
35
- == Config ==
36
-
37
- Override anything in the config (foo.bar=value)
38
-
39
-
40
- $CONFIG
41
-
42
-
43
- ${hydra.help.footer}
44
-
45
- '
46
- hydra_help:
47
- template: 'Hydra (${hydra.runtime.version})
48
-
49
- See https://hydra.cc for more info.
50
-
51
-
52
- == Flags ==
53
-
54
- $FLAGS_HELP
55
-
56
-
57
- == Configuration groups ==
58
-
59
- Compose your configuration from those groups (For example, append hydra/job_logging=disabled
60
- to command line)
61
-
62
-
63
- $HYDRA_CONFIG_GROUPS
64
-
65
-
66
- Use ''--cfg hydra'' to Show the Hydra config.
67
-
68
- '
69
- hydra_help: ???
70
- hydra_logging:
71
- version: 1
72
- formatters:
73
- colorlog:
74
- (): colorlog.ColoredFormatter
75
- format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
76
- handlers:
77
- console:
78
- class: logging.StreamHandler
79
- formatter: colorlog
80
- stream: ext://sys.stdout
81
- root:
82
- level: INFO
83
- handlers:
84
- - console
85
- disable_existing_loggers: false
86
- job_logging:
87
- version: 1
88
- formatters:
89
- simple:
90
- format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
- colorlog:
92
- (): colorlog.ColoredFormatter
93
- format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
94
- - %(message)s'
95
- log_colors:
96
- DEBUG: purple
97
- INFO: green
98
- WARNING: yellow
99
- ERROR: red
100
- CRITICAL: red
101
- handlers:
102
- console:
103
- class: logging.StreamHandler
104
- formatter: colorlog
105
- stream: ext://sys.stdout
106
- file:
107
- class: logging.FileHandler
108
- formatter: simple
109
- filename: ${hydra.job.name}.log
110
- root:
111
- level: INFO
112
- handlers:
113
- - console
114
- - file
115
- disable_existing_loggers: false
116
- env: {}
117
- mode: MULTIRUN
118
- searchpath: []
119
- callbacks: {}
120
- output_subdir: .hydra
121
- overrides:
122
- hydra:
123
- - hydra.mode=MULTIRUN
124
- task:
125
- - benchmark.input_shapes.batch_size=1
126
- - backend.torch_dtype=float32
127
- job:
128
- name: experiment
129
- chdir: true
130
- override_dirname: backend.torch_dtype=float32,benchmark.input_shapes.batch_size=1
131
- id: '1'
132
- num: 1
133
- config_name: llama2_1gpu_inference
134
- env_set: {}
135
- env_copy: []
136
- config:
137
- override_dirname:
138
- kv_sep: '='
139
- item_sep: ','
140
- exclude_keys: []
141
- runtime:
142
- version: 1.3.2
143
- version_base: '1.3'
144
- cwd: /home/user/transformers-regression
145
- config_sources:
146
- - path: hydra.conf
147
- schema: pkg
148
- provider: hydra
149
- - path: optimum_benchmark
150
- schema: pkg
151
- provider: main
152
- - path: hydra_plugins.hydra_colorlog.conf
153
- schema: pkg
154
- provider: hydra-colorlog
155
- - path: /home/user/transformers-regression/configs
156
- schema: file
157
- provider: command-line
158
- - path: ''
159
- schema: structured
160
- provider: schema
161
- output_dir: /home/user/transformers-regression/sweeps/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/llama_1gpu_inference/1
162
- choices:
163
- benchmark: inference
164
- backend: pytorch
165
- hydra/env: default
166
- hydra/callbacks: null
167
- hydra/job_logging: colorlog
168
- hydra/hydra_logging: colorlog
169
- hydra/hydra_help: default
170
- hydra/help: default
171
- hydra/sweeper: basic
172
- hydra/launcher: basic
173
- hydra/output: default
174
- verbose: false
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
raw_results/2023-09-27_10:21:54_153755ee386ac73e04814a94337abcb1208ff5d1/llama_1gpu_inference/1/.config/overrides.yaml DELETED
@@ -1,2 +0,0 @@
1
- - benchmark.input_shapes.batch_size=1
2
- - backend.torch_dtype=float32