Dataset Preview
The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
The dataset generation failed
Error code: DatasetGenerationError
Exception: ArrowNotImplementedError
Message: Cannot write struct type 'model_kwargs' with no child field to Parquet. Consider adding a dummy child field.
Traceback: Traceback (most recent call last):
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1869, in _prepare_split_single
writer.write_table(table)
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 578, in write_table
self._build_writer(inferred_schema=pa_table.schema)
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 399, in _build_writer
self.pa_writer = self._WRITER_CLASS(self.stream, schema)
File "/src/services/worker/.venv/lib/python3.9/site-packages/pyarrow/parquet/core.py", line 1010, in __init__
self.writer = _parquet.ParquetWriter(
File "pyarrow/_parquet.pyx", line 2157, in pyarrow._parquet.ParquetWriter.__cinit__
File "pyarrow/error.pxi", line 154, in pyarrow.lib.pyarrow_internal_check_status
File "pyarrow/error.pxi", line 91, in pyarrow.lib.check_status
pyarrow.lib.ArrowNotImplementedError: Cannot write struct type 'model_kwargs' with no child field to Parquet. Consider adding a dummy child field.
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1885, in _prepare_split_single
num_examples, num_bytes = writer.finalize()
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 597, in finalize
self._build_writer(self.schema)
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 399, in _build_writer
self.pa_writer = self._WRITER_CLASS(self.stream, schema)
File "/src/services/worker/.venv/lib/python3.9/site-packages/pyarrow/parquet/core.py", line 1010, in __init__
self.writer = _parquet.ParquetWriter(
File "pyarrow/_parquet.pyx", line 2157, in pyarrow._parquet.ParquetWriter.__cinit__
File "pyarrow/error.pxi", line 154, in pyarrow.lib.pyarrow_internal_check_status
File "pyarrow/error.pxi", line 91, in pyarrow.lib.check_status
pyarrow.lib.ArrowNotImplementedError: Cannot write struct type 'model_kwargs' with no child field to Parquet. Consider adding a dummy child field.
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1392, in compute_config_parquet_and_info_response
parquet_operations = convert_to_parquet(builder)
File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1041, in convert_to_parquet
builder.download_and_prepare(
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 924, in download_and_prepare
self._download_and_prepare(
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 999, in _download_and_prepare
self._prepare_split(split_generator, **prepare_split_kwargs)
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1740, in _prepare_split
for job_id, done, content in self._prepare_split_single(
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1896, in _prepare_split_single
raise DatasetGenerationError("An error occurred while generating the dataset") from e
datasets.exceptions.DatasetGenerationError: An error occurred while generating the datasetNeed help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
config dict | report dict | name string | backend dict | scenario dict | launcher dict | environment dict | print_report bool | log_report bool | load dict | prefill dict | decode dict | per_token dict |
|---|---|---|---|---|---|---|---|---|---|---|---|---|
{
"name": "2024-10-10-13-15-21/openvino",
"backend": {
"name": "openvino",
"version": "2024.4.0",
"_target_": "optimum_benchmark.backends.openvino.backend.OVBackend",
"task": "text-generation",
"library": "transformers",
"model_type": "llama",
"model": "TinyLlama/TinyLlama-1.1B-Chat-v1.0... | {
"load": {
"memory": {
"unit": "MB",
"max_ram": 4841.96352,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
25.399079136550426
],
"count": 1,
"... | null | null | null | null | null | null | null | null | null | null | null |
null | null | 2024-10-10-13-15-21/openvino | {
"name": "openvino",
"version": "2024.4.0",
"_target_": "optimum_benchmark.backends.openvino.backend.OVBackend",
"task": "text-generation",
"library": "transformers",
"model_type": "llama",
"model": "TinyLlama/TinyLlama-1.1B-Chat-v1.0",
"processor": "TinyLlama/TinyLlama-1.1B-Chat-v1.0",
"device": "cp... | {
"name": "inference",
"_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
"iterations": 10,
"duration": 10,
"warmup_runs": 10,
"input_shapes": {
"batch_size": 2,
"num_choices": 2,
"sequence_length": 16
},
"new_tokens": null,
"memory": true,
"latency": true,
... | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": null,
"numactl": true,
"numactl_kwargs": {
"cpunodebind": 0,
"membind": 0
},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7R13 Processor",
"cpu_count": 64,
"cpu_ram_mb": 529717.026816,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.205-195.807.amzn2.x86_64-x86_64-with-glibc2.36",
"processor": "",
"python_version": "3.10.15",
"optimum_benchmark_version": "0.5.0",
"optimum_benchmark_c... | false | true | null | null | null | null |
null | null | null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 4841.96352,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
25.399079136550426
],
"count": 1,
"total": 25.399079136550426,
"mean": 25... | {
"memory": {
"unit": "MB",
"max_ram": 3465.314304,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
0.20681462809443474,
0.14498980715870857,
0.1491953432559967,
0.142987... | {
"memory": {
"unit": "MB",
"max_ram": 3465.314304,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
3.7909555956721306,
1.7467840760946274,
1.8483475632965565,
3.08324761... | {
"memory": null,
"latency": {
"unit": "s",
"values": [
0.1709301434457302,
0.13552312925457954,
0.17639076337218285,
0.12885065376758575,
0.18046993762254715,
0.17862163111567497,
0.11986216530203819,
0.029551289975643158,
0.07815691828727722,
0.0... |
{
"name": "2024-10-10-13-15-21/pytorch",
"backend": {
"name": "pytorch",
"version": "2.4.1",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-generation",
"library": "transformers",
"model_type": "llama",
"model": "TinyLlama/TinyLlama-1.1B-Chat-v1.0"... | {
"load": {
"memory": {
"unit": "MB",
"max_ram": 7341.576192,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
3.724861901253462
],
"count": 1,
"... | null | null | null | null | null | null | null | null | null | null | null |
null | null | 2024-10-10-13-15-21/pytorch | {
"name": "pytorch",
"version": "2.4.1",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-generation",
"library": "transformers",
"model_type": "llama",
"model": "TinyLlama/TinyLlama-1.1B-Chat-v1.0",
"processor": "TinyLlama/TinyLlama-1.1B-Chat-v1.0",
"device": "cp... | {
"name": "inference",
"_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
"iterations": 10,
"duration": 10,
"warmup_runs": 10,
"input_shapes": {
"batch_size": 2,
"num_choices": 2,
"sequence_length": 16
},
"new_tokens": null,
"memory": true,
"latency": true,
... | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": null,
"numactl": true,
"numactl_kwargs": {
"cpunodebind": 0,
"membind": 0
},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7R13 Processor",
"cpu_count": 64,
"cpu_ram_mb": 529717.026816,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.205-195.807.amzn2.x86_64-x86_64-with-glibc2.36",
"processor": "",
"python_version": "3.10.15",
"optimum_benchmark_version": "0.5.0",
"optimum_benchmark_c... | false | true | null | null | null | null |
null | null | null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 7341.576192,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
3.724861901253462
],
"count": 1,
"total": 3.724861901253462,
"mean": 3.7... | {
"memory": {
"unit": "MB",
"max_ram": 5306.855424,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
0.7200786955654621,
0.792355839163065,
2.4664254635572433,
0.694083433... | {
"memory": {
"unit": "MB",
"max_ram": 5306.855424,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
14.600170020014048,
10.227855522185564,
14.517007939517498,
10.3927390... | {
"memory": null,
"latency": {
"unit": "s",
"values": [
0.4060986116528511,
0.3879626765847206,
0.38639020919799805,
0.3389683812856674,
0.3950040340423584,
0.3980884477496147,
0.4014359675347805,
0.408795777708292,
0.30595654249191284,
0.299078408... |
{
"name": "2024-10-10-13-27-38/openvino",
"backend": {
"name": "openvino",
"version": "2024.4.0",
"_target_": "optimum_benchmark.backends.openvino.backend.OVBackend",
"task": "text-generation",
"library": "transformers",
"model_type": "qwen2",
"model": "Qwen/Qwen2.5-7B-Instruct",
"pr... | {
"load": {
"memory": {
"unit": "MB",
"max_ram": 28568.141824,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
85.84128818660975
],
"count": 1,
... | null | null | null | null | null | null | null | null | null | null | null |
null | null | 2024-10-10-13-27-38/openvino | {
"name": "openvino",
"version": "2024.4.0",
"_target_": "optimum_benchmark.backends.openvino.backend.OVBackend",
"task": "text-generation",
"library": "transformers",
"model_type": "qwen2",
"model": "Qwen/Qwen2.5-7B-Instruct",
"processor": "Qwen/Qwen2.5-7B-Instruct",
"device": "cpu",
"device_ids": ... | {
"name": "inference",
"_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
"iterations": 10,
"duration": 10,
"warmup_runs": 10,
"input_shapes": {
"batch_size": 2,
"num_choices": 2,
"sequence_length": 16
},
"new_tokens": null,
"memory": true,
"latency": true,
... | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": null,
"numactl": true,
"numactl_kwargs": {
"cpunodebind": 0,
"membind": 0
},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7R13 Processor",
"cpu_count": 64,
"cpu_ram_mb": 529717.026816,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.205-195.807.amzn2.x86_64-x86_64-with-glibc2.36",
"processor": "",
"python_version": "3.10.15",
"optimum_benchmark_version": "0.5.0",
"optimum_benchmark_c... | false | true | null | null | null | null |
null | null | null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 28568.141824,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
85.84128818660975
],
"count": 1,
"total": 85.84128818660975,
"mean": 85... | {
"memory": {
"unit": "MB",
"max_ram": 17400.188928,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
1.155863493680954,
0.7591285742819309,
0.7738370560109615,
0.75509426... | {
"memory": {
"unit": "MB",
"max_ram": 17400.459264,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
6.200990695506334,
5.851446453481913,
6.216963429003954,
5.7086634449... | {
"memory": null,
"latency": {
"unit": "s",
"values": [
0.4139234870672226,
0.41735056415200233,
0.30081597715616226,
0.1818118579685688,
0.15186486765742302,
0.23906059190630913,
0.15295346081256866,
0.16011932492256165,
0.2003607526421547,
0.1834... |
{
"name": "2024-10-10-13-27-38/pytorch",
"backend": {
"name": "pytorch",
"version": "2.4.1",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-generation",
"library": "transformers",
"model_type": "qwen2",
"model": "Qwen/Qwen2.5-7B-Instruct",
"pro... | {
"load": {
"memory": {
"unit": "MB",
"max_ram": 35904.417792,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
17.653250481933355
],
"count": 1,
... | null | null | null | null | null | null | null | null | null | null | null |
null | null | 2024-10-10-13-27-38/pytorch | {
"name": "pytorch",
"version": "2.4.1",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-generation",
"library": "transformers",
"model_type": "qwen2",
"model": "Qwen/Qwen2.5-7B-Instruct",
"processor": "Qwen/Qwen2.5-7B-Instruct",
"device": "cpu",
"device_ids": ... | {
"name": "inference",
"_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
"iterations": 10,
"duration": 10,
"warmup_runs": 10,
"input_shapes": {
"batch_size": 2,
"num_choices": 2,
"sequence_length": 16
},
"new_tokens": null,
"memory": true,
"latency": true,
... | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": null,
"numactl": true,
"numactl_kwargs": {
"cpunodebind": 0,
"membind": 0
},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7R13 Processor",
"cpu_count": 64,
"cpu_ram_mb": 529717.026816,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.205-195.807.amzn2.x86_64-x86_64-with-glibc2.36",
"processor": "",
"python_version": "3.10.15",
"optimum_benchmark_version": "0.5.0",
"optimum_benchmark_c... | false | true | null | null | null | null |
null | null | null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 35904.417792,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
17.653250481933355
],
"count": 1,
"total": 17.653250481933355,
"mean": ... | {
"memory": {
"unit": "MB",
"max_ram": 32586.477568,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
7.114257633686066,
6.9122952707111835,
7.246219478547573,
7.030736085... | {
"memory": {
"unit": "MB",
"max_ram": 32586.477568,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"values": [
63.17382514476776,
65.35659051686525,
55.37452078983188,
50.915375493... | {
"memory": null,
"latency": {
"unit": "s",
"values": [
1.7681835927069187,
3.4078874364495277,
1.8021652027964592,
1.8068931549787521,
2.261068418622017,
1.726471833884716,
1.9878739155828953,
1.5795339308679104,
1.8304941393435001,
1.563472054898... |
No dataset card yet
- Downloads last month
- 4