Dataset Preview
Full Screen Viewer
Full Screen
The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
The dataset generation failed
Error code: DatasetGenerationError Exception: ArrowNotImplementedError Message: Cannot write struct type 'model_kwargs' with no child field to Parquet. Consider adding a dummy child field. Traceback: Traceback (most recent call last): File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1870, in _prepare_split_single writer.write_table(table) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 620, in write_table self._build_writer(inferred_schema=pa_table.schema) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 441, in _build_writer self.pa_writer = self._WRITER_CLASS(self.stream, schema) File "/src/services/worker/.venv/lib/python3.9/site-packages/pyarrow/parquet/core.py", line 1010, in __init__ self.writer = _parquet.ParquetWriter( File "pyarrow/_parquet.pyx", line 2157, in pyarrow._parquet.ParquetWriter.__cinit__ File "pyarrow/error.pxi", line 154, in pyarrow.lib.pyarrow_internal_check_status File "pyarrow/error.pxi", line 91, in pyarrow.lib.check_status pyarrow.lib.ArrowNotImplementedError: Cannot write struct type 'model_kwargs' with no child field to Parquet. Consider adding a dummy child field. During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1886, in _prepare_split_single num_examples, num_bytes = writer.finalize() File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 639, in finalize self._build_writer(self.schema) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 441, in _build_writer self.pa_writer = self._WRITER_CLASS(self.stream, schema) File "/src/services/worker/.venv/lib/python3.9/site-packages/pyarrow/parquet/core.py", line 1010, in __init__ self.writer = _parquet.ParquetWriter( File "pyarrow/_parquet.pyx", line 2157, in pyarrow._parquet.ParquetWriter.__cinit__ File "pyarrow/error.pxi", line 154, in pyarrow.lib.pyarrow_internal_check_status File "pyarrow/error.pxi", line 91, in pyarrow.lib.check_status pyarrow.lib.ArrowNotImplementedError: Cannot write struct type 'model_kwargs' with no child field to Parquet. Consider adding a dummy child field. The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1417, in compute_config_parquet_and_info_response parquet_operations = convert_to_parquet(builder) File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1049, in convert_to_parquet builder.download_and_prepare( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 924, in download_and_prepare self._download_and_prepare( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1000, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1741, in _prepare_split for job_id, done, content in self._prepare_split_single( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1897, in _prepare_split_single raise DatasetGenerationError("An error occurred while generating the dataset") from e datasets.exceptions.DatasetGenerationError: An error occurred while generating the dataset
Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
config
dict | report
dict | name
string | backend
dict | scenario
dict | launcher
dict | environment
dict | print_report
bool | log_report
bool | overall
dict | warmup
dict | train
dict |
---|---|---|---|---|---|---|---|---|---|---|---|
{
"name": "cuda_training_transformers_fill-mask_google-bert/bert-base-uncased",
"backend": {
"name": "pytorch",
"version": "2.5.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "fill-mask",
"library": "transformers",
"model_type": "bert",
"model": "google-bert/bert-base-uncased",
"processor": "google-bert/bert-base-uncased",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.248768,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.227-219.884.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": null,
"transformers_version": "4.46.3",
"transformers_commit": null,
"accelerate_version": "1.1.1",
"accelerate_commit": null,
"diffusers_version": "0.31.0",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.11",
"timm_commit": null,
"peft_version": "0.13.2",
"peft_commit": null
},
"print_report": true,
"log_report": true
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1307.738112,
"max_global_vram": 3176.660992,
"max_process_vram": 0,
"max_reserved": 2520.776704,
"max_allocated": 2211.86048
},
"latency": {
"unit": "s",
"values": [
0.3425331115722656,
0.04516454315185547,
0.04459622573852539,
0.044355583190917966,
0.04437401580810547
],
"count": 5,
"total": 0.5210234794616699,
"mean": 0.10420469589233398,
"p50": 0.04459622573852539,
"p90": 0.22358568420410158,
"p95": 0.2830593978881836,
"p99": 0.3306383688354492,
"stdev": 0.11916456657974374,
"stdev_": 114.35623467762579
},
"throughput": {
"unit": "samples/s",
"value": 95.96496505620213
},
"energy": {
"unit": "kWh",
"cpu": 0.000008661581222222145,
"ram": 0.000004727325612405731,
"gpu": 0.00001422445582399982,
"total": 0.0000276133626586277
},
"efficiency": {
"unit": "samples/kWh",
"value": 362143.50724414707
}
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1307.738112,
"max_global_vram": 3176.660992,
"max_process_vram": 0,
"max_reserved": 2520.776704,
"max_allocated": 2211.86048
},
"latency": {
"unit": "s",
"values": [
0.3425331115722656,
0.04516454315185547
],
"count": 2,
"total": 0.3876976547241211,
"mean": 0.19384882736206055,
"p50": 0.19384882736206055,
"p90": 0.3127962547302246,
"p95": 0.3276646831512451,
"p99": 0.33955942588806154,
"stdev": 0.14868428421020508,
"stdev_": 76.70115225020189
},
"throughput": {
"unit": "samples/s",
"value": 20.634636043111122
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1307.738112,
"max_global_vram": 3176.660992,
"max_process_vram": 0,
"max_reserved": 2520.776704,
"max_allocated": 2211.86048
},
"latency": {
"unit": "s",
"values": [
0.04459622573852539,
0.044355583190917966,
0.04437401580810547
],
"count": 3,
"total": 0.1333258247375488,
"mean": 0.044441941579182936,
"p50": 0.04437401580810547,
"p90": 0.044551783752441405,
"p95": 0.0445740047454834,
"p99": 0.04459178153991699,
"stdev": 0.00010935459664809051,
"stdev_": 0.24606169929198893
},
"throughput": {
"unit": "samples/s",
"value": 135.00760288138406
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
null | null | cuda_training_transformers_fill-mask_google-bert/bert-base-uncased | {
"name": "pytorch",
"version": "2.5.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "fill-mask",
"library": "transformers",
"model_type": "bert",
"model": "google-bert/bert-base-uncased",
"processor": "google-bert/bert-base-uncased",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.248768,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.227-219.884.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": null,
"transformers_version": "4.46.3",
"transformers_commit": null,
"accelerate_version": "1.1.1",
"accelerate_commit": null,
"diffusers_version": "0.31.0",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.11",
"timm_commit": null,
"peft_version": "0.13.2",
"peft_commit": null
} | true | true | null | null | null |
null | null | null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 1307.738112,
"max_global_vram": 3176.660992,
"max_process_vram": 0,
"max_reserved": 2520.776704,
"max_allocated": 2211.86048
},
"latency": {
"unit": "s",
"values": [
0.3425331115722656,
0.04516454315185547,
0.04459622573852539,
0.044355583190917966,
0.04437401580810547
],
"count": 5,
"total": 0.5210234794616699,
"mean": 0.10420469589233398,
"p50": 0.04459622573852539,
"p90": 0.22358568420410158,
"p95": 0.2830593978881836,
"p99": 0.3306383688354492,
"stdev": 0.11916456657974374,
"stdev_": 114.35623467762579
},
"throughput": {
"unit": "samples/s",
"value": 95.96496505620213
},
"energy": {
"unit": "kWh",
"cpu": 0.000008661581222222145,
"ram": 0.000004727325612405731,
"gpu": 0.00001422445582399982,
"total": 0.0000276133626586277
},
"efficiency": {
"unit": "samples/kWh",
"value": 362143.50724414707
}
} | {
"memory": {
"unit": "MB",
"max_ram": 1307.738112,
"max_global_vram": 3176.660992,
"max_process_vram": 0,
"max_reserved": 2520.776704,
"max_allocated": 2211.86048
},
"latency": {
"unit": "s",
"values": [
0.3425331115722656,
0.04516454315185547
],
"count": 2,
"total": 0.3876976547241211,
"mean": 0.19384882736206055,
"p50": 0.19384882736206055,
"p90": 0.3127962547302246,
"p95": 0.3276646831512451,
"p99": 0.33955942588806154,
"stdev": 0.14868428421020508,
"stdev_": 76.70115225020189
},
"throughput": {
"unit": "samples/s",
"value": 20.634636043111122
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 1307.738112,
"max_global_vram": 3176.660992,
"max_process_vram": 0,
"max_reserved": 2520.776704,
"max_allocated": 2211.86048
},
"latency": {
"unit": "s",
"values": [
0.04459622573852539,
0.044355583190917966,
0.04437401580810547
],
"count": 3,
"total": 0.1333258247375488,
"mean": 0.044441941579182936,
"p50": 0.04437401580810547,
"p90": 0.044551783752441405,
"p95": 0.0445740047454834,
"p99": 0.04459178153991699,
"stdev": 0.00010935459664809051,
"stdev_": 0.24606169929198893
},
"throughput": {
"unit": "samples/s",
"value": 135.00760288138406
},
"energy": null,
"efficiency": null
} |
{
"name": "cuda_training_transformers_fill-mask_google-bert/bert-base-uncased",
"backend": {
"name": "pytorch",
"version": "2.2.2",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "fill-mask",
"model": "google-bert/bert-base-uncased",
"library": "transformers",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.29792,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": null,
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1063.8336,
"max_global_vram": 3169.32096,
"max_process_vram": 0,
"max_reserved": 2520.776704,
"max_allocated": 2211.86048
},
"latency": {
"unit": "s",
"count": 5,
"total": 0.7448790740966797,
"mean": 0.14897581481933594,
"stdev": 0.2054173633207176,
"p50": 0.04632883071899414,
"p90": 0.35471870422363283,
"p95": 0.4572641067504882,
"p99": 0.5393004287719726,
"values": [
0.5598095092773437,
0.04708249664306641,
0.04632883071899414,
0.04576665496826172,
0.04589158248901367
]
},
"throughput": {
"unit": "samples/s",
"value": 67.1249894630687
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1063.8336,
"max_global_vram": 3169.32096,
"max_process_vram": 0,
"max_reserved": 2520.776704,
"max_allocated": 2211.86048
},
"latency": {
"unit": "s",
"count": 2,
"total": 0.6068920059204101,
"mean": 0.3034460029602051,
"stdev": 0.25636350631713867,
"p50": 0.3034460029602051,
"p90": 0.508536808013916,
"p95": 0.5341731586456299,
"p99": 0.554682239151001,
"values": [
0.5598095092773437,
0.04708249664306641
]
},
"throughput": {
"unit": "samples/s",
"value": 13.181916917602548
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1063.8336,
"max_global_vram": 3169.32096,
"max_process_vram": 0,
"max_reserved": 2520.776704,
"max_allocated": 2211.86048
},
"latency": {
"unit": "s",
"count": 3,
"total": 0.13798706817626955,
"mean": 0.045995689392089846,
"stdev": 0.00024102431292157434,
"p50": 0.04589158248901367,
"p90": 0.04624138107299805,
"p95": 0.0462851058959961,
"p99": 0.04632008575439454,
"values": [
0.04632883071899414,
0.04576665496826172,
0.04589158248901367
]
},
"throughput": {
"unit": "samples/s",
"value": 130.4470066499722
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
{
"name": "cuda_training_transformers_image-classification_google/vit-base-patch16-224",
"backend": {
"name": "pytorch",
"version": "2.5.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "image-classification",
"library": "transformers",
"model_type": "vit",
"model": "google/vit-base-patch16-224",
"processor": "google/vit-base-patch16-224",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.248768,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.227-219.884.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": null,
"transformers_version": "4.46.3",
"transformers_commit": null,
"accelerate_version": "1.1.1",
"accelerate_commit": null,
"diffusers_version": "0.31.0",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.11",
"timm_commit": null,
"peft_version": "0.13.2",
"peft_commit": null
},
"print_report": true,
"log_report": true
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1729.032192,
"max_global_vram": 2618.81856,
"max_process_vram": 0,
"max_reserved": 1956.642816,
"max_allocated": 1755.291648
},
"latency": {
"unit": "s",
"values": [
0.33363262939453125,
0.04060774230957031,
0.04029017639160156,
0.03999337768554687,
0.03973427200317383
],
"count": 5,
"total": 0.4942581977844238,
"mean": 0.09885163955688477,
"p50": 0.04029017639160156,
"p90": 0.21642267456054692,
"p95": 0.27502765197753903,
"p99": 0.3219116339111328,
"stdev": 0.11739085812852307,
"stdev_": 118.75458885127526
},
"throughput": {
"unit": "samples/s",
"value": 101.16170095737705
},
"energy": {
"unit": "kWh",
"cpu": 0.000009976675155555153,
"ram": 0.0000054126004284134765,
"gpu": 0.00002172585071399946,
"total": 0.00003711512629796809
},
"efficiency": {
"unit": "samples/kWh",
"value": 269431.927018593
}
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1729.032192,
"max_global_vram": 2618.81856,
"max_process_vram": 0,
"max_reserved": 1956.642816,
"max_allocated": 1755.291648
},
"latency": {
"unit": "s",
"values": [
0.33363262939453125,
0.04060774230957031
],
"count": 2,
"total": 0.37424037170410157,
"mean": 0.18712018585205079,
"p50": 0.18712018585205079,
"p90": 0.30433014068603514,
"p95": 0.31898138504028317,
"p99": 0.3307023805236816,
"stdev": 0.14651244354248047,
"stdev_": 78.29857739577204
},
"throughput": {
"unit": "samples/s",
"value": 21.37663545910892
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1729.032192,
"max_global_vram": 2618.81856,
"max_process_vram": 0,
"max_reserved": 1956.642816,
"max_allocated": 1755.291648
},
"latency": {
"unit": "s",
"values": [
0.04029017639160156,
0.03999337768554687,
0.03973427200317383
],
"count": 3,
"total": 0.12001782608032227,
"mean": 0.04000594202677409,
"p50": 0.03999337768554687,
"p90": 0.040230816650390626,
"p95": 0.04026049652099609,
"p99": 0.040284240417480466,
"stdev": 0.00022712084781918436,
"stdev_": 0.5677177846910418
},
"throughput": {
"unit": "samples/s",
"value": 149.97772070920072
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
null | null | cuda_training_transformers_image-classification_google/vit-base-patch16-224 | {
"name": "pytorch",
"version": "2.5.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "image-classification",
"library": "transformers",
"model_type": "vit",
"model": "google/vit-base-patch16-224",
"processor": "google/vit-base-patch16-224",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.248768,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.227-219.884.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": null,
"transformers_version": "4.46.3",
"transformers_commit": null,
"accelerate_version": "1.1.1",
"accelerate_commit": null,
"diffusers_version": "0.31.0",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.11",
"timm_commit": null,
"peft_version": "0.13.2",
"peft_commit": null
} | true | true | null | null | null |
null | null | null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 1729.032192,
"max_global_vram": 2618.81856,
"max_process_vram": 0,
"max_reserved": 1956.642816,
"max_allocated": 1755.291648
},
"latency": {
"unit": "s",
"values": [
0.33363262939453125,
0.04060774230957031,
0.04029017639160156,
0.03999337768554687,
0.03973427200317383
],
"count": 5,
"total": 0.4942581977844238,
"mean": 0.09885163955688477,
"p50": 0.04029017639160156,
"p90": 0.21642267456054692,
"p95": 0.27502765197753903,
"p99": 0.3219116339111328,
"stdev": 0.11739085812852307,
"stdev_": 118.75458885127526
},
"throughput": {
"unit": "samples/s",
"value": 101.16170095737705
},
"energy": {
"unit": "kWh",
"cpu": 0.000009976675155555153,
"ram": 0.0000054126004284134765,
"gpu": 0.00002172585071399946,
"total": 0.00003711512629796809
},
"efficiency": {
"unit": "samples/kWh",
"value": 269431.927018593
}
} | {
"memory": {
"unit": "MB",
"max_ram": 1729.032192,
"max_global_vram": 2618.81856,
"max_process_vram": 0,
"max_reserved": 1956.642816,
"max_allocated": 1755.291648
},
"latency": {
"unit": "s",
"values": [
0.33363262939453125,
0.04060774230957031
],
"count": 2,
"total": 0.37424037170410157,
"mean": 0.18712018585205079,
"p50": 0.18712018585205079,
"p90": 0.30433014068603514,
"p95": 0.31898138504028317,
"p99": 0.3307023805236816,
"stdev": 0.14651244354248047,
"stdev_": 78.29857739577204
},
"throughput": {
"unit": "samples/s",
"value": 21.37663545910892
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 1729.032192,
"max_global_vram": 2618.81856,
"max_process_vram": 0,
"max_reserved": 1956.642816,
"max_allocated": 1755.291648
},
"latency": {
"unit": "s",
"values": [
0.04029017639160156,
0.03999337768554687,
0.03973427200317383
],
"count": 3,
"total": 0.12001782608032227,
"mean": 0.04000594202677409,
"p50": 0.03999337768554687,
"p90": 0.040230816650390626,
"p95": 0.04026049652099609,
"p99": 0.040284240417480466,
"stdev": 0.00022712084781918436,
"stdev_": 0.5677177846910418
},
"throughput": {
"unit": "samples/s",
"value": 149.97772070920072
},
"energy": null,
"efficiency": null
} |
{
"name": "cuda_training_transformers_image-classification_google/vit-base-patch16-224",
"backend": {
"name": "pytorch",
"version": "2.2.2",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "image-classification",
"model": "google/vit-base-patch16-224",
"library": "transformers",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.29792,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": null,
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1446.354944,
"max_global_vram": 2628.255744,
"max_process_vram": 0,
"max_reserved": 1956.642816,
"max_allocated": 1756.126208
},
"latency": {
"unit": "s",
"count": 5,
"total": 0.48406525039672854,
"mean": 0.09681305007934571,
"stdev": 0.1110534407118009,
"p50": 0.04146995162963867,
"p90": 0.20794796142578126,
"p95": 0.26343380432128904,
"p99": 0.3078224786376953,
"values": [
0.3189196472167969,
0.04103168106079102,
0.041490432739257815,
0.04146995162963867,
0.04115353775024414
]
},
"throughput": {
"unit": "samples/s",
"value": 103.29185984538483
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1446.354944,
"max_global_vram": 2628.255744,
"max_process_vram": 0,
"max_reserved": 1956.642816,
"max_allocated": 1756.126208
},
"latency": {
"unit": "s",
"count": 2,
"total": 0.3599513282775879,
"mean": 0.17997566413879396,
"stdev": 0.13894398307800293,
"p50": 0.17997566413879396,
"p90": 0.2911308506011963,
"p95": 0.30502524890899657,
"p99": 0.31614076755523685,
"values": [
0.3189196472167969,
0.04103168106079102
]
},
"throughput": {
"unit": "samples/s",
"value": 22.225227055782792
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1446.354944,
"max_global_vram": 2628.255744,
"max_process_vram": 0,
"max_reserved": 1956.642816,
"max_allocated": 1756.126208
},
"latency": {
"unit": "s",
"count": 3,
"total": 0.12411392211914063,
"mean": 0.041371307373046874,
"stdev": 0.00015421321911462263,
"p50": 0.04146995162963867,
"p90": 0.04148633651733399,
"p95": 0.0414883846282959,
"p99": 0.041490023117065435,
"values": [
0.041490432739257815,
0.04146995162963867,
0.04115353775024414
]
},
"throughput": {
"unit": "samples/s",
"value": 145.02804917180256
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
{
"name": "cuda_training_transformers_multiple-choice_FacebookAI/roberta-base",
"backend": {
"name": "pytorch",
"version": "2.5.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "multiple-choice",
"library": "transformers",
"model_type": "roberta",
"model": "FacebookAI/roberta-base",
"processor": "FacebookAI/roberta-base",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.248768,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.227-219.884.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": null,
"transformers_version": "4.46.3",
"transformers_commit": null,
"accelerate_version": "1.1.1",
"accelerate_commit": null,
"diffusers_version": "0.31.0",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.11",
"timm_commit": null,
"peft_version": "0.13.2",
"peft_commit": null
},
"print_report": true,
"log_report": true
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1345.72032,
"max_global_vram": 3384.27904,
"max_process_vram": 0,
"max_reserved": 2728.394752,
"max_allocated": 2516.23424
},
"latency": {
"unit": "s",
"values": [
0.35756851196289063,
0.047387649536132816,
0.04659695816040039,
0.04753299331665039,
0.046721023559570314
],
"count": 5,
"total": 0.5458071365356446,
"mean": 0.10916142730712892,
"p50": 0.047387649536132816,
"p90": 0.23355430450439457,
"p95": 0.2955614082336425,
"p99": 0.34516709121704103,
"stdev": 0.12420407402590627,
"stdev_": 113.78018507989496
},
"throughput": {
"unit": "samples/s",
"value": 91.60745005527184
},
"energy": {
"unit": "kWh",
"cpu": 0.000009039191384721572,
"ram": 0.000004909473005839129,
"gpu": 0.000015694734778000288,
"total": 0.000029643399168560992
},
"efficiency": {
"unit": "samples/kWh",
"value": 337343.2292004399
}
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1345.72032,
"max_global_vram": 3384.27904,
"max_process_vram": 0,
"max_reserved": 2728.394752,
"max_allocated": 2516.23424
},
"latency": {
"unit": "s",
"values": [
0.35756851196289063,
0.047387649536132816
],
"count": 2,
"total": 0.40495616149902347,
"mean": 0.20247808074951174,
"p50": 0.20247808074951174,
"p90": 0.32655042572021487,
"p95": 0.34205946884155275,
"p99": 0.35446670333862307,
"stdev": 0.15509043121337893,
"stdev_": 76.59615828008727
},
"throughput": {
"unit": "samples/s",
"value": 19.755224788743686
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1345.72032,
"max_global_vram": 3384.27904,
"max_process_vram": 0,
"max_reserved": 2728.394752,
"max_allocated": 2516.23424
},
"latency": {
"unit": "s",
"values": [
0.04659695816040039,
0.04753299331665039,
0.046721023559570314
],
"count": 3,
"total": 0.14085097503662108,
"mean": 0.04695032501220703,
"p50": 0.046721023559570314,
"p90": 0.04737059936523438,
"p95": 0.04745179634094238,
"p99": 0.04751675392150879,
"stdev": 0.0004151102829931439,
"stdev_": 0.8841478368578188
},
"throughput": {
"unit": "samples/s",
"value": 127.79464249587213
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
null | null | cuda_training_transformers_multiple-choice_FacebookAI/roberta-base | {
"name": "pytorch",
"version": "2.5.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "multiple-choice",
"library": "transformers",
"model_type": "roberta",
"model": "FacebookAI/roberta-base",
"processor": "FacebookAI/roberta-base",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.248768,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.227-219.884.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": null,
"transformers_version": "4.46.3",
"transformers_commit": null,
"accelerate_version": "1.1.1",
"accelerate_commit": null,
"diffusers_version": "0.31.0",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.11",
"timm_commit": null,
"peft_version": "0.13.2",
"peft_commit": null
} | true | true | null | null | null |
null | null | null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 1345.72032,
"max_global_vram": 3384.27904,
"max_process_vram": 0,
"max_reserved": 2728.394752,
"max_allocated": 2516.23424
},
"latency": {
"unit": "s",
"values": [
0.35756851196289063,
0.047387649536132816,
0.04659695816040039,
0.04753299331665039,
0.046721023559570314
],
"count": 5,
"total": 0.5458071365356446,
"mean": 0.10916142730712892,
"p50": 0.047387649536132816,
"p90": 0.23355430450439457,
"p95": 0.2955614082336425,
"p99": 0.34516709121704103,
"stdev": 0.12420407402590627,
"stdev_": 113.78018507989496
},
"throughput": {
"unit": "samples/s",
"value": 91.60745005527184
},
"energy": {
"unit": "kWh",
"cpu": 0.000009039191384721572,
"ram": 0.000004909473005839129,
"gpu": 0.000015694734778000288,
"total": 0.000029643399168560992
},
"efficiency": {
"unit": "samples/kWh",
"value": 337343.2292004399
}
} | {
"memory": {
"unit": "MB",
"max_ram": 1345.72032,
"max_global_vram": 3384.27904,
"max_process_vram": 0,
"max_reserved": 2728.394752,
"max_allocated": 2516.23424
},
"latency": {
"unit": "s",
"values": [
0.35756851196289063,
0.047387649536132816
],
"count": 2,
"total": 0.40495616149902347,
"mean": 0.20247808074951174,
"p50": 0.20247808074951174,
"p90": 0.32655042572021487,
"p95": 0.34205946884155275,
"p99": 0.35446670333862307,
"stdev": 0.15509043121337893,
"stdev_": 76.59615828008727
},
"throughput": {
"unit": "samples/s",
"value": 19.755224788743686
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 1345.72032,
"max_global_vram": 3384.27904,
"max_process_vram": 0,
"max_reserved": 2728.394752,
"max_allocated": 2516.23424
},
"latency": {
"unit": "s",
"values": [
0.04659695816040039,
0.04753299331665039,
0.046721023559570314
],
"count": 3,
"total": 0.14085097503662108,
"mean": 0.04695032501220703,
"p50": 0.046721023559570314,
"p90": 0.04737059936523438,
"p95": 0.04745179634094238,
"p99": 0.04751675392150879,
"stdev": 0.0004151102829931439,
"stdev_": 0.8841478368578188
},
"throughput": {
"unit": "samples/s",
"value": 127.79464249587213
},
"energy": null,
"efficiency": null
} |
{
"name": "cuda_training_transformers_multiple-choice_FacebookAI/roberta-base",
"backend": {
"name": "pytorch",
"version": "2.2.2",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "multiple-choice",
"model": "FacebookAI/roberta-base",
"library": "transformers",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.29792,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": null,
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1093.496832,
"max_global_vram": 3379.03616,
"max_process_vram": 0,
"max_reserved": 2730.491904,
"max_allocated": 2516.23424
},
"latency": {
"unit": "s",
"count": 5,
"total": 0.8026234703063965,
"mean": 0.16052469406127928,
"stdev": 0.22240891148008993,
"p50": 0.04907724761962891,
"p90": 0.38326721343994147,
"p95": 0.49430444412231433,
"p99": 0.5831342286682129,
"values": [
0.6053416748046875,
0.05015552139282226,
0.04897484970092773,
0.04907417678833008,
0.04907724761962891
]
},
"throughput": {
"unit": "samples/s",
"value": 62.295711313940046
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1093.496832,
"max_global_vram": 3379.03616,
"max_process_vram": 0,
"max_reserved": 2730.491904,
"max_allocated": 2516.23424
},
"latency": {
"unit": "s",
"count": 2,
"total": 0.6554971961975098,
"mean": 0.3277485980987549,
"stdev": 0.27759307670593264,
"p50": 0.3277485980987549,
"p90": 0.549823059463501,
"p95": 0.5775823671340942,
"p99": 0.5997898132705688,
"values": [
0.6053416748046875,
0.05015552139282226
]
},
"throughput": {
"unit": "samples/s",
"value": 12.204476306546239
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1093.496832,
"max_global_vram": 3379.03616,
"max_process_vram": 0,
"max_reserved": 2730.491904,
"max_allocated": 2516.23424
},
"latency": {
"unit": "s",
"count": 3,
"total": 0.14712627410888673,
"mean": 0.049042091369628914,
"stdev": 0.000047563564546161045,
"p50": 0.04907417678833008,
"p90": 0.04907663345336914,
"p95": 0.049076940536499025,
"p99": 0.04907718620300293,
"values": [
0.04897484970092773,
0.04907417678833008,
0.04907724761962891
]
},
"throughput": {
"unit": "samples/s",
"value": 122.34388527149387
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
{
"name": "cuda_training_transformers_text-classification_FacebookAI/roberta-base",
"backend": {
"name": "pytorch",
"version": "2.5.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-classification",
"library": "transformers",
"model_type": "roberta",
"model": "FacebookAI/roberta-base",
"processor": "FacebookAI/roberta-base",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.248768,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.227-219.884.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": null,
"transformers_version": "4.46.3",
"transformers_commit": null,
"accelerate_version": "1.1.1",
"accelerate_commit": null,
"diffusers_version": "0.31.0",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.11",
"timm_commit": null,
"peft_version": "0.13.2",
"peft_commit": null
},
"print_report": true,
"log_report": true
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1332.40832,
"max_global_vram": 3384.27904,
"max_process_vram": 0,
"max_reserved": 2728.394752,
"max_allocated": 2516.250112
},
"latency": {
"unit": "s",
"values": [
0.35358001708984377,
0.046873600006103515,
0.048377857208251954,
0.04718991851806641,
0.04760780715942383
],
"count": 5,
"total": 0.5436291999816895,
"mean": 0.10872583999633789,
"p50": 0.04760780715942383,
"p90": 0.23149915313720706,
"p95": 0.29253958511352535,
"p99": 0.3413719306945801,
"stdev": 0.12242812604019773,
"stdev_": 112.6026030650316
},
"throughput": {
"unit": "samples/s",
"value": 91.97445612134908
},
"energy": {
"unit": "kWh",
"cpu": 0.000009134229152777431,
"ram": 0.000004986785702676617,
"gpu": 0.00001595223498400064,
"total": 0.00003007324983945469
},
"efficiency": {
"unit": "samples/kWh",
"value": 332521.4286246002
}
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1332.40832,
"max_global_vram": 3384.27904,
"max_process_vram": 0,
"max_reserved": 2728.394752,
"max_allocated": 2516.250112
},
"latency": {
"unit": "s",
"values": [
0.35358001708984377,
0.046873600006103515
],
"count": 2,
"total": 0.4004536170959473,
"mean": 0.20022680854797364,
"p50": 0.20022680854797364,
"p90": 0.32290937538146974,
"p95": 0.3382446962356567,
"p99": 0.35051295291900636,
"stdev": 0.15335320854187012,
"stdev_": 76.58974822301442
},
"throughput": {
"unit": "samples/s",
"value": 19.977344837125614
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1332.40832,
"max_global_vram": 3384.27904,
"max_process_vram": 0,
"max_reserved": 2728.394752,
"max_allocated": 2516.250112
},
"latency": {
"unit": "s",
"values": [
0.048377857208251954,
0.04718991851806641,
0.04760780715942383
],
"count": 3,
"total": 0.1431755828857422,
"mean": 0.0477251942952474,
"p50": 0.04760780715942383,
"p90": 0.04822384719848633,
"p95": 0.04830085220336914,
"p99": 0.04836245620727539,
"stdev": 0.0004920260072073315,
"stdev_": 1.030956530346339
},
"throughput": {
"unit": "samples/s",
"value": 125.719760570938
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
null | null | cuda_training_transformers_text-classification_FacebookAI/roberta-base | {
"name": "pytorch",
"version": "2.5.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-classification",
"library": "transformers",
"model_type": "roberta",
"model": "FacebookAI/roberta-base",
"processor": "FacebookAI/roberta-base",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.248768,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.227-219.884.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": null,
"transformers_version": "4.46.3",
"transformers_commit": null,
"accelerate_version": "1.1.1",
"accelerate_commit": null,
"diffusers_version": "0.31.0",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.11",
"timm_commit": null,
"peft_version": "0.13.2",
"peft_commit": null
} | true | true | null | null | null |
null | null | null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 1332.40832,
"max_global_vram": 3384.27904,
"max_process_vram": 0,
"max_reserved": 2728.394752,
"max_allocated": 2516.250112
},
"latency": {
"unit": "s",
"values": [
0.35358001708984377,
0.046873600006103515,
0.048377857208251954,
0.04718991851806641,
0.04760780715942383
],
"count": 5,
"total": 0.5436291999816895,
"mean": 0.10872583999633789,
"p50": 0.04760780715942383,
"p90": 0.23149915313720706,
"p95": 0.29253958511352535,
"p99": 0.3413719306945801,
"stdev": 0.12242812604019773,
"stdev_": 112.6026030650316
},
"throughput": {
"unit": "samples/s",
"value": 91.97445612134908
},
"energy": {
"unit": "kWh",
"cpu": 0.000009134229152777431,
"ram": 0.000004986785702676617,
"gpu": 0.00001595223498400064,
"total": 0.00003007324983945469
},
"efficiency": {
"unit": "samples/kWh",
"value": 332521.4286246002
}
} | {
"memory": {
"unit": "MB",
"max_ram": 1332.40832,
"max_global_vram": 3384.27904,
"max_process_vram": 0,
"max_reserved": 2728.394752,
"max_allocated": 2516.250112
},
"latency": {
"unit": "s",
"values": [
0.35358001708984377,
0.046873600006103515
],
"count": 2,
"total": 0.4004536170959473,
"mean": 0.20022680854797364,
"p50": 0.20022680854797364,
"p90": 0.32290937538146974,
"p95": 0.3382446962356567,
"p99": 0.35051295291900636,
"stdev": 0.15335320854187012,
"stdev_": 76.58974822301442
},
"throughput": {
"unit": "samples/s",
"value": 19.977344837125614
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 1332.40832,
"max_global_vram": 3384.27904,
"max_process_vram": 0,
"max_reserved": 2728.394752,
"max_allocated": 2516.250112
},
"latency": {
"unit": "s",
"values": [
0.048377857208251954,
0.04718991851806641,
0.04760780715942383
],
"count": 3,
"total": 0.1431755828857422,
"mean": 0.0477251942952474,
"p50": 0.04760780715942383,
"p90": 0.04822384719848633,
"p95": 0.04830085220336914,
"p99": 0.04836245620727539,
"stdev": 0.0004920260072073315,
"stdev_": 1.030956530346339
},
"throughput": {
"unit": "samples/s",
"value": 125.719760570938
},
"energy": null,
"efficiency": null
} |
{
"name": "cuda_training_transformers_text-classification_FacebookAI/roberta-base",
"backend": {
"name": "pytorch",
"version": "2.2.2",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-classification",
"model": "FacebookAI/roberta-base",
"library": "transformers",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.29792,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": null,
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1080.1152,
"max_global_vram": 3379.03616,
"max_process_vram": 0,
"max_reserved": 2730.491904,
"max_allocated": 2516.250112
},
"latency": {
"unit": "s",
"count": 5,
"total": 0.778464241027832,
"mean": 0.1556928482055664,
"stdev": 0.21129125718980005,
"p50": 0.05020159912109375,
"p90": 0.36730080566406254,
"p95": 0.4727875488281249,
"p99": 0.557176943359375,
"values": [
0.5782742919921875,
0.050840576171875,
0.04969574356079102,
0.049452030181884765,
0.05020159912109375
]
},
"throughput": {
"unit": "samples/s",
"value": 64.22902602948513
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1080.1152,
"max_global_vram": 3379.03616,
"max_process_vram": 0,
"max_reserved": 2730.491904,
"max_allocated": 2516.250112
},
"latency": {
"unit": "s",
"count": 2,
"total": 0.6291148681640625,
"mean": 0.31455743408203124,
"stdev": 0.26371685791015625,
"p50": 0.31455743408203124,
"p90": 0.5255309204101563,
"p95": 0.5519026062011718,
"p99": 0.5729999548339844,
"values": [
0.5782742919921875,
0.050840576171875
]
},
"throughput": {
"unit": "samples/s",
"value": 12.716278703357135
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1080.1152,
"max_global_vram": 3379.03616,
"max_process_vram": 0,
"max_reserved": 2730.491904,
"max_allocated": 2516.250112
},
"latency": {
"unit": "s",
"count": 3,
"total": 0.14934937286376954,
"mean": 0.049783124287923176,
"stdev": 0.0003121857804388603,
"p50": 0.04969574356079102,
"p90": 0.05010042800903321,
"p95": 0.050151013565063476,
"p99": 0.0501914820098877,
"values": [
0.04969574356079102,
0.049452030181884765,
0.05020159912109375
]
},
"throughput": {
"unit": "samples/s",
"value": 120.52276922795568
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
{
"name": "cuda_training_transformers_text-generation_openai-community/gpt2",
"backend": {
"name": "pytorch",
"version": "2.5.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-generation",
"library": "transformers",
"model_type": "gpt2",
"model": "openai-community/gpt2",
"processor": "openai-community/gpt2",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.248768,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.227-219.884.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": null,
"transformers_version": "4.46.3",
"transformers_commit": null,
"accelerate_version": "1.1.1",
"accelerate_commit": null,
"diffusers_version": "0.31.0",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.11",
"timm_commit": null,
"peft_version": "0.13.2",
"peft_commit": null
},
"print_report": true,
"log_report": true
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1359.392768,
"max_global_vram": 3566.731264,
"max_process_vram": 0,
"max_reserved": 2910.846976,
"max_allocated": 2523.776
},
"latency": {
"unit": "s",
"values": [
0.36362753295898437,
0.045039615631103515,
0.04426649475097656,
0.04447641754150391,
0.04424300765991211
],
"count": 5,
"total": 0.5416530685424804,
"mean": 0.10833061370849609,
"p50": 0.04447641754150391,
"p90": 0.23619236602783206,
"p95": 0.29990994949340816,
"p99": 0.3508840162658691,
"stdev": 0.12764878250214318,
"stdev_": 117.83260348328666
},
"throughput": {
"unit": "samples/s",
"value": 92.31000967933893
},
"energy": {
"unit": "kWh",
"cpu": 0.00000903662806805588,
"ram": 0.000004827364832180661,
"gpu": 0.000014324733682000157,
"total": 0.000028188726582236697
},
"efficiency": {
"unit": "samples/kWh",
"value": 354751.74697325856
}
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1359.392768,
"max_global_vram": 3566.731264,
"max_process_vram": 0,
"max_reserved": 2910.846976,
"max_allocated": 2523.776
},
"latency": {
"unit": "s",
"values": [
0.36362753295898437,
0.045039615631103515
],
"count": 2,
"total": 0.4086671485900879,
"mean": 0.20433357429504395,
"p50": 0.20433357429504395,
"p90": 0.33176874122619626,
"p95": 0.3476981370925903,
"p99": 0.36044165378570553,
"stdev": 0.15929395866394044,
"stdev_": 77.95779974657061
},
"throughput": {
"unit": "samples/s",
"value": 19.575833358762026
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1359.392768,
"max_global_vram": 3566.731264,
"max_process_vram": 0,
"max_reserved": 2910.846976,
"max_allocated": 2523.776
},
"latency": {
"unit": "s",
"values": [
0.04426649475097656,
0.04447641754150391,
0.04424300765991211
],
"count": 3,
"total": 0.13298591995239256,
"mean": 0.044328639984130856,
"p50": 0.04426649475097656,
"p90": 0.04443443298339844,
"p95": 0.04445542526245117,
"p99": 0.04447221908569336,
"stdev": 0.00010493352088830884,
"stdev_": 0.23671721245198102
},
"throughput": {
"unit": "samples/s",
"value": 135.35267497825177
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
null | null | cuda_training_transformers_text-generation_openai-community/gpt2 | {
"name": "pytorch",
"version": "2.5.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-generation",
"library": "transformers",
"model_type": "gpt2",
"model": "openai-community/gpt2",
"processor": "openai-community/gpt2",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.248768,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.227-219.884.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": null,
"transformers_version": "4.46.3",
"transformers_commit": null,
"accelerate_version": "1.1.1",
"accelerate_commit": null,
"diffusers_version": "0.31.0",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.11",
"timm_commit": null,
"peft_version": "0.13.2",
"peft_commit": null
} | true | true | null | null | null |
null | null | null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 1359.392768,
"max_global_vram": 3566.731264,
"max_process_vram": 0,
"max_reserved": 2910.846976,
"max_allocated": 2523.776
},
"latency": {
"unit": "s",
"values": [
0.36362753295898437,
0.045039615631103515,
0.04426649475097656,
0.04447641754150391,
0.04424300765991211
],
"count": 5,
"total": 0.5416530685424804,
"mean": 0.10833061370849609,
"p50": 0.04447641754150391,
"p90": 0.23619236602783206,
"p95": 0.29990994949340816,
"p99": 0.3508840162658691,
"stdev": 0.12764878250214318,
"stdev_": 117.83260348328666
},
"throughput": {
"unit": "samples/s",
"value": 92.31000967933893
},
"energy": {
"unit": "kWh",
"cpu": 0.00000903662806805588,
"ram": 0.000004827364832180661,
"gpu": 0.000014324733682000157,
"total": 0.000028188726582236697
},
"efficiency": {
"unit": "samples/kWh",
"value": 354751.74697325856
}
} | {
"memory": {
"unit": "MB",
"max_ram": 1359.392768,
"max_global_vram": 3566.731264,
"max_process_vram": 0,
"max_reserved": 2910.846976,
"max_allocated": 2523.776
},
"latency": {
"unit": "s",
"values": [
0.36362753295898437,
0.045039615631103515
],
"count": 2,
"total": 0.4086671485900879,
"mean": 0.20433357429504395,
"p50": 0.20433357429504395,
"p90": 0.33176874122619626,
"p95": 0.3476981370925903,
"p99": 0.36044165378570553,
"stdev": 0.15929395866394044,
"stdev_": 77.95779974657061
},
"throughput": {
"unit": "samples/s",
"value": 19.575833358762026
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 1359.392768,
"max_global_vram": 3566.731264,
"max_process_vram": 0,
"max_reserved": 2910.846976,
"max_allocated": 2523.776
},
"latency": {
"unit": "s",
"values": [
0.04426649475097656,
0.04447641754150391,
0.04424300765991211
],
"count": 3,
"total": 0.13298591995239256,
"mean": 0.044328639984130856,
"p50": 0.04426649475097656,
"p90": 0.04443443298339844,
"p95": 0.04445542526245117,
"p99": 0.04447221908569336,
"stdev": 0.00010493352088830884,
"stdev_": 0.23671721245198102
},
"throughput": {
"unit": "samples/s",
"value": 135.35267497825177
},
"energy": null,
"efficiency": null
} |
{
"name": "cuda_training_transformers_text-generation_openai-community/gpt2",
"backend": {
"name": "pytorch",
"version": "2.2.2",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-generation",
"model": "openai-community/gpt2",
"library": "transformers",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.29792,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": null,
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1107.173376,
"max_global_vram": 3563.585536,
"max_process_vram": 0,
"max_reserved": 2915.04128,
"max_allocated": 2523.776
},
"latency": {
"unit": "s",
"count": 5,
"total": 0.8139294586181639,
"mean": 0.16278589172363278,
"stdev": 0.2273662362653359,
"p50": 0.04927385711669922,
"p90": 0.3902293930053711,
"p95": 0.5038737297058105,
"p99": 0.594789199066162,
"values": [
0.61751806640625,
0.049296382904052735,
0.04860006332397461,
0.0492410888671875,
0.04927385711669922
]
},
"throughput": {
"unit": "samples/s",
"value": 61.4303849929289
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1107.173376,
"max_global_vram": 3563.585536,
"max_process_vram": 0,
"max_reserved": 2915.04128,
"max_allocated": 2523.776
},
"latency": {
"unit": "s",
"count": 2,
"total": 0.6668144493103026,
"mean": 0.3334072246551513,
"stdev": 0.2841108417510986,
"p50": 0.3334072246551513,
"p90": 0.5606958980560303,
"p95": 0.5891069822311401,
"p99": 0.6118358495712279,
"values": [
0.61751806640625,
0.049296382904052735
]
},
"throughput": {
"unit": "samples/s",
"value": 11.99734050195603
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1107.173376,
"max_global_vram": 3563.585536,
"max_process_vram": 0,
"max_reserved": 2915.04128,
"max_allocated": 2523.776
},
"latency": {
"unit": "s",
"count": 3,
"total": 0.1471150093078613,
"mean": 0.04903833643595377,
"stdev": 0.00031019448743967263,
"p50": 0.0492410888671875,
"p90": 0.04926730346679687,
"p95": 0.049270580291748044,
"p99": 0.04927320175170898,
"values": [
0.04860006332397461,
0.0492410888671875,
0.04927385711669922
]
},
"throughput": {
"unit": "samples/s",
"value": 122.35325331307405
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
{
"name": "cuda_training_transformers_token-classification_microsoft/deberta-v3-base",
"backend": {
"name": "pytorch",
"version": "2.5.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "token-classification",
"library": "transformers",
"model_type": "deberta-v2",
"model": "microsoft/deberta-v3-base",
"processor": "microsoft/deberta-v3-base",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.248768,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.227-219.884.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": null,
"transformers_version": "4.46.3",
"transformers_commit": null,
"accelerate_version": "1.1.1",
"accelerate_commit": null,
"diffusers_version": "0.31.0",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.11",
"timm_commit": null,
"peft_version": "0.13.2",
"peft_commit": null
},
"print_report": true,
"log_report": true
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1395.720192,
"max_global_vram": 4604.821504,
"max_process_vram": 0,
"max_reserved": 3948.937216,
"max_allocated": 3702.95552
},
"latency": {
"unit": "s",
"values": [
0.423625732421875,
0.08791551971435548,
0.08667033386230469,
0.08652799987792968,
0.08687104034423829
],
"count": 5,
"total": 0.7716106262207032,
"mean": 0.15432212524414063,
"p50": 0.08687104034423829,
"p90": 0.28934164733886725,
"p95": 0.35648368988037105,
"p99": 0.4101973239135742,
"stdev": 0.1346526845350235,
"stdev_": 87.25429637649191
},
"throughput": {
"unit": "samples/s",
"value": 64.7995223250056
},
"energy": {
"unit": "kWh",
"cpu": 0.000011613126145138847,
"ram": 0.000006328201977180253,
"gpu": 0.00001950196004599919,
"total": 0.00003744328816831829
},
"efficiency": {
"unit": "samples/kWh",
"value": 267070.5616196724
}
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1395.720192,
"max_global_vram": 4604.821504,
"max_process_vram": 0,
"max_reserved": 3948.937216,
"max_allocated": 3702.95552
},
"latency": {
"unit": "s",
"values": [
0.423625732421875,
0.08791551971435548
],
"count": 2,
"total": 0.5115412521362305,
"mean": 0.25577062606811524,
"p50": 0.25577062606811524,
"p90": 0.39005471115112306,
"p95": 0.40684022178649903,
"p99": 0.4202686302947998,
"stdev": 0.16785510635375978,
"stdev_": 65.62720236258
},
"throughput": {
"unit": "samples/s",
"value": 15.639012428795263
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1395.720192,
"max_global_vram": 4604.821504,
"max_process_vram": 0,
"max_reserved": 3948.937216,
"max_allocated": 3702.95552
},
"latency": {
"unit": "s",
"values": [
0.08667033386230469,
0.08652799987792968,
0.08687104034423829
],
"count": 3,
"total": 0.2600693740844726,
"mean": 0.08668979136149087,
"p50": 0.08667033386230469,
"p90": 0.08683089904785157,
"p95": 0.08685096969604493,
"p99": 0.08686702621459962,
"stdev": 0.00014071990166578441,
"stdev_": 0.1623258049831859
},
"throughput": {
"unit": "samples/s",
"value": 69.2123017689636
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
null | null | cuda_training_transformers_token-classification_microsoft/deberta-v3-base | {
"name": "pytorch",
"version": "2.5.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "token-classification",
"library": "transformers",
"model_type": "deberta-v2",
"model": "microsoft/deberta-v3-base",
"processor": "microsoft/deberta-v3-base",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.248768,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.227-219.884.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": null,
"transformers_version": "4.46.3",
"transformers_commit": null,
"accelerate_version": "1.1.1",
"accelerate_commit": null,
"diffusers_version": "0.31.0",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.11",
"timm_commit": null,
"peft_version": "0.13.2",
"peft_commit": null
} | true | true | null | null | null |
null | null | null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 1395.720192,
"max_global_vram": 4604.821504,
"max_process_vram": 0,
"max_reserved": 3948.937216,
"max_allocated": 3702.95552
},
"latency": {
"unit": "s",
"values": [
0.423625732421875,
0.08791551971435548,
0.08667033386230469,
0.08652799987792968,
0.08687104034423829
],
"count": 5,
"total": 0.7716106262207032,
"mean": 0.15432212524414063,
"p50": 0.08687104034423829,
"p90": 0.28934164733886725,
"p95": 0.35648368988037105,
"p99": 0.4101973239135742,
"stdev": 0.1346526845350235,
"stdev_": 87.25429637649191
},
"throughput": {
"unit": "samples/s",
"value": 64.7995223250056
},
"energy": {
"unit": "kWh",
"cpu": 0.000011613126145138847,
"ram": 0.000006328201977180253,
"gpu": 0.00001950196004599919,
"total": 0.00003744328816831829
},
"efficiency": {
"unit": "samples/kWh",
"value": 267070.5616196724
}
} | {
"memory": {
"unit": "MB",
"max_ram": 1395.720192,
"max_global_vram": 4604.821504,
"max_process_vram": 0,
"max_reserved": 3948.937216,
"max_allocated": 3702.95552
},
"latency": {
"unit": "s",
"values": [
0.423625732421875,
0.08791551971435548
],
"count": 2,
"total": 0.5115412521362305,
"mean": 0.25577062606811524,
"p50": 0.25577062606811524,
"p90": 0.39005471115112306,
"p95": 0.40684022178649903,
"p99": 0.4202686302947998,
"stdev": 0.16785510635375978,
"stdev_": 65.62720236258
},
"throughput": {
"unit": "samples/s",
"value": 15.639012428795263
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 1395.720192,
"max_global_vram": 4604.821504,
"max_process_vram": 0,
"max_reserved": 3948.937216,
"max_allocated": 3702.95552
},
"latency": {
"unit": "s",
"values": [
0.08667033386230469,
0.08652799987792968,
0.08687104034423829
],
"count": 3,
"total": 0.2600693740844726,
"mean": 0.08668979136149087,
"p50": 0.08667033386230469,
"p90": 0.08683089904785157,
"p95": 0.08685096969604493,
"p99": 0.08686702621459962,
"stdev": 0.00014071990166578441,
"stdev_": 0.1623258049831859
},
"throughput": {
"unit": "samples/s",
"value": 69.2123017689636
},
"energy": null,
"efficiency": null
} |
{
"name": "cuda_training_transformers_token-classification_microsoft/deberta-v3-base",
"backend": {
"name": "pytorch",
"version": "2.2.2",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "token-classification",
"model": "microsoft/deberta-v3-base",
"library": "transformers",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.29792,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": null,
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1140.764672,
"max_global_vram": 4597.481472,
"max_process_vram": 0,
"max_reserved": 3948.937216,
"max_allocated": 3702.95552
},
"latency": {
"unit": "s",
"count": 5,
"total": 1.0522736740112304,
"mean": 0.21045473480224608,
"stdev": 0.2521626361507063,
"p50": 0.08427519989013672,
"p90": 0.46303764648437507,
"p95": 0.5889081359863281,
"p99": 0.6896045275878907,
"values": [
0.7147786254882813,
0.08542617797851562,
0.08360550689697266,
0.08418816375732421,
0.08427519989013672
]
},
"throughput": {
"unit": "samples/s",
"value": 47.51615595342393
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1140.764672,
"max_global_vram": 4597.481472,
"max_process_vram": 0,
"max_reserved": 3948.937216,
"max_allocated": 3702.95552
},
"latency": {
"unit": "s",
"count": 2,
"total": 0.8002048034667969,
"mean": 0.40010240173339845,
"stdev": 0.3146762237548828,
"p50": 0.40010240173339845,
"p90": 0.6518433807373047,
"p95": 0.683311003112793,
"p99": 0.7084851010131836,
"values": [
0.7147786254882813,
0.08542617797851562
]
},
"throughput": {
"unit": "samples/s",
"value": 9.997440611879489
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1140.764672,
"max_global_vram": 4597.481472,
"max_process_vram": 0,
"max_reserved": 3948.937216,
"max_allocated": 3702.95552
},
"latency": {
"unit": "s",
"count": 3,
"total": 0.2520688705444336,
"mean": 0.08402295684814454,
"stdev": 0.0002973125946472148,
"p50": 0.08418816375732421,
"p90": 0.08425779266357422,
"p95": 0.08426649627685547,
"p99": 0.08427345916748047,
"values": [
0.08360550689697266,
0.08418816375732421,
0.08427519989013672
]
},
"throughput": {
"unit": "samples/s",
"value": 71.40905563278207
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
No dataset card yet
New: Create and edit this dataset card directly on the website!
Contribute a Dataset Card- Downloads last month
- 277