machine
stringclasses 4
values | hardware
stringclasses 2
values | subsets
stringclasses 2
values | backends
stringclasses 1
value | model
stringclasses 47
values | success
bool 2
classes | traceback
stringlengths 0
7.23k
| last_updated
stringlengths 26
26
| run_id
stringclasses 12
values | run_start_time
stringclasses 12
values |
---|---|---|---|---|---|---|---|---|---|
debug | cuda | unquantized | pytorch | openai-community/gpt2 | true | 2025-02-05T16:11:02.440342 | 2fdb9d56-2aba-4a3b-ac44-63bdc885f925 | 2025-02-05T16:09:58.183268 |
|
1xT4 | cuda | unquantized | pytorch | openai-community/gpt2 | false | Traceback (most recent call last):
File "/llm-perf-backend/llm_perf/common/benchmark_runner.py", line 297, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 149, in load_transformers_model
self.load_transformers_model_with_no_weights()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 124, in load_transformers_model_with_no_weights
self.pretrained_model = self.automodel_loader.from_pretrained(
File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py", line 564, in from_pretrained
return model_class.from_pretrained(
File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 3826, in from_pretrained
config = cls._autoset_attn_implementation(
File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 1556, in _autoset_attn_implementation
cls._check_and_enable_flash_attn_2(
File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 1658, in _check_and_enable_flash_attn_2
raise ImportError(f"{preface} the package flash_attn seems to be not installed. {install_message}")
ImportError: FlashAttention2 has been toggled on, but it cannot be used due to the following error: the package flash_attn seems to be not installed. Please refer to the documentation of https://huggingface.co/docs/transformers/perf_infer_gpu_one#flashattention-2 to install Flash Attention 2.
| 2025-02-05T16:15:46.371592 | 806946df-514d-4d5f-9b01-0150f2b9ee22 | 2025-02-05T16:14:52.531351 |
1xT4 | cuda | unquantized | pytorch | openai-community/gpt2 | false | Traceback (most recent call last):
File "/llm-perf-backend/llm_perf/common/benchmark_runner.py", line 297, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 644, in _locate
obj = getattr(obj, part)
AttributeError: module 'optimum_benchmark.backends.pytorch' has no attribute 'backend'
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1603, in _get_module
return importlib.import_module("." + module_name, self.__name__)
File "/usr/lib/python3.10/importlib/__init__.py", line 126, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
File "<frozen importlib._bootstrap>", line 1050, in _gcd_import
File "<frozen importlib._bootstrap>", line 1027, in _find_and_load
File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked
File "<frozen importlib._bootstrap>", line 688, in _load_unlocked
File "<frozen importlib._bootstrap_external>", line 883, in exec_module
File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed
File "/usr/local/lib/python3.10/dist-packages/transformers/trainer.py", line 215, in <module>
from peft import PeftModel
File "/usr/local/lib/python3.10/dist-packages/peft/__init__.py", line 22, in <module>
from .auto import (
File "/usr/local/lib/python3.10/dist-packages/peft/auto.py", line 32, in <module>
from .mapping import MODEL_TYPE_TO_PEFT_MODEL_MAPPING
File "/usr/local/lib/python3.10/dist-packages/peft/mapping.py", line 22, in <module>
from peft.tuners.xlora.model import XLoraModel
File "/usr/local/lib/python3.10/dist-packages/peft/tuners/__init__.py", line 21, in <module>
from .lora import LoraConfig, LoraModel, LoftQConfig, LoraRuntimeConfig
File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/__init__.py", line 20, in <module>
from .model import LoraModel
File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/model.py", line 50, in <module>
from .awq import dispatch_awq
File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/awq.py", line 26, in <module>
from awq.modules.linear import WQLinear_GEMM
File "/usr/local/lib/python3.10/dist-packages/awq/__init__.py", line 2, in <module>
from awq.models.auto import AutoAWQForCausalLM
File "/usr/local/lib/python3.10/dist-packages/awq/models/__init__.py", line 1, in <module>
from .mpt import MptAWQForCausalLM
File "/usr/local/lib/python3.10/dist-packages/awq/models/mpt.py", line 1, in <module>
from .base import BaseAWQForCausalLM
File "/usr/local/lib/python3.10/dist-packages/awq/models/base.py", line 52, in <module>
from awq.quantize.quantizer import AwqQuantizer
File "/usr/local/lib/python3.10/dist-packages/awq/quantize/quantizer.py", line 10, in <module>
from awq.quantize.scale import apply_scale, apply_clip
File "/usr/local/lib/python3.10/dist-packages/awq/quantize/scale.py", line 8, in <module>
from transformers.models.llama.modeling_llama import LlamaRMSNorm
File "/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py", line 32, in <module>
from ...modeling_flash_attention_utils import _flash_attention_forward
File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py", line 27, in <module>
from flash_attn.bert_padding import index_first_axis, pad_input, unpad_input # noqa
File "/usr/local/lib/python3.10/dist-packages/flash_attn/__init__.py", line 3, in <module>
from flash_attn.flash_attn_interface import (
File "/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py", line 15, in <module>
import flash_attn_2_cuda as flash_attn_gpu
ImportError: /usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 650, in _locate
obj = import_module(mod)
File "/usr/lib/python3.10/importlib/__init__.py", line 126, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
File "<frozen importlib._bootstrap>", line 1050, in _gcd_import
File "<frozen importlib._bootstrap>", line 1027, in _find_and_load
File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked
File "<frozen importlib._bootstrap>", line 688, in _load_unlocked
File "<frozen importlib._bootstrap_external>", line 883, in exec_module
File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 10, in <module>
from transformers import (
File "<frozen importlib._bootstrap>", line 1075, in _handle_fromlist
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1593, in __getattr__
module = self._get_module(self._class_to_module[name])
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1605, in _get_module
raise RuntimeError(
RuntimeError: Failed to import transformers.trainer because of the following error (look up to see its traceback):
/usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 69, in run
backend_factory: Type[Backend] = get_class(backend_config._target_)
File "/usr/local/lib/python3.10/dist-packages/hydra/utils.py", line 40, in get_class
raise e
File "/usr/local/lib/python3.10/dist-packages/hydra/utils.py", line 31, in get_class
cls = _locate(path)
File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 658, in _locate
raise ImportError(
ImportError: Error loading 'optimum_benchmark.backends.pytorch.backend.PyTorchBackend':
RuntimeError('Failed to import transformers.trainer because of the following error (look up to see its traceback):\n/usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv')
| 2025-02-05T16:17:40.048300 | 03886520-a1ca-4afc-a74c-76bca3bbd085 | 2025-02-05T16:16:54.711736 |
1xT4 | cuda | unquantized | pytorch | openai-community/gpt2 | false | Traceback (most recent call last):
File "/llm-perf-backend/llm_perf/common/benchmark_runner.py", line 297, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 644, in _locate
obj = getattr(obj, part)
AttributeError: module 'optimum_benchmark.backends.pytorch' has no attribute 'backend'
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1603, in _get_module
return importlib.import_module("." + module_name, self.__name__)
File "/usr/lib/python3.10/importlib/__init__.py", line 126, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
File "<frozen importlib._bootstrap>", line 1050, in _gcd_import
File "<frozen importlib._bootstrap>", line 1027, in _find_and_load
File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked
File "<frozen importlib._bootstrap>", line 688, in _load_unlocked
File "<frozen importlib._bootstrap_external>", line 883, in exec_module
File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed
File "/usr/local/lib/python3.10/dist-packages/transformers/trainer.py", line 215, in <module>
from peft import PeftModel
File "/usr/local/lib/python3.10/dist-packages/peft/__init__.py", line 22, in <module>
from .auto import (
File "/usr/local/lib/python3.10/dist-packages/peft/auto.py", line 32, in <module>
from .mapping import MODEL_TYPE_TO_PEFT_MODEL_MAPPING
File "/usr/local/lib/python3.10/dist-packages/peft/mapping.py", line 22, in <module>
from peft.tuners.xlora.model import XLoraModel
File "/usr/local/lib/python3.10/dist-packages/peft/tuners/__init__.py", line 21, in <module>
from .lora import LoraConfig, LoraModel, LoftQConfig, LoraRuntimeConfig
File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/__init__.py", line 20, in <module>
from .model import LoraModel
File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/model.py", line 50, in <module>
from .awq import dispatch_awq
File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/awq.py", line 26, in <module>
from awq.modules.linear import WQLinear_GEMM
File "/usr/local/lib/python3.10/dist-packages/awq/__init__.py", line 2, in <module>
from awq.models.auto import AutoAWQForCausalLM
File "/usr/local/lib/python3.10/dist-packages/awq/models/__init__.py", line 1, in <module>
from .mpt import MptAWQForCausalLM
File "/usr/local/lib/python3.10/dist-packages/awq/models/mpt.py", line 1, in <module>
from .base import BaseAWQForCausalLM
File "/usr/local/lib/python3.10/dist-packages/awq/models/base.py", line 52, in <module>
from awq.quantize.quantizer import AwqQuantizer
File "/usr/local/lib/python3.10/dist-packages/awq/quantize/quantizer.py", line 10, in <module>
from awq.quantize.scale import apply_scale, apply_clip
File "/usr/local/lib/python3.10/dist-packages/awq/quantize/scale.py", line 8, in <module>
from transformers.models.llama.modeling_llama import LlamaRMSNorm
File "/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py", line 32, in <module>
from ...modeling_flash_attention_utils import _flash_attention_forward
File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py", line 27, in <module>
from flash_attn.bert_padding import index_first_axis, pad_input, unpad_input # noqa
File "/usr/local/lib/python3.10/dist-packages/flash_attn/__init__.py", line 3, in <module>
from flash_attn.flash_attn_interface import (
File "/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py", line 15, in <module>
import flash_attn_2_cuda as flash_attn_gpu
ImportError: /usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 650, in _locate
obj = import_module(mod)
File "/usr/lib/python3.10/importlib/__init__.py", line 126, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
File "<frozen importlib._bootstrap>", line 1050, in _gcd_import
File "<frozen importlib._bootstrap>", line 1027, in _find_and_load
File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked
File "<frozen importlib._bootstrap>", line 688, in _load_unlocked
File "<frozen importlib._bootstrap_external>", line 883, in exec_module
File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 10, in <module>
from transformers import (
File "<frozen importlib._bootstrap>", line 1075, in _handle_fromlist
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1593, in __getattr__
module = self._get_module(self._class_to_module[name])
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1605, in _get_module
raise RuntimeError(
RuntimeError: Failed to import transformers.trainer because of the following error (look up to see its traceback):
/usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 69, in run
backend_factory: Type[Backend] = get_class(backend_config._target_)
File "/usr/local/lib/python3.10/dist-packages/hydra/utils.py", line 40, in get_class
raise e
File "/usr/local/lib/python3.10/dist-packages/hydra/utils.py", line 31, in get_class
cls = _locate(path)
File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 658, in _locate
raise ImportError(
ImportError: Error loading 'optimum_benchmark.backends.pytorch.backend.PyTorchBackend':
RuntimeError('Failed to import transformers.trainer because of the following error (look up to see its traceback):\n/usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv')
| 2025-02-05T16:17:54.962479 | 03886520-a1ca-4afc-a74c-76bca3bbd085 | 2025-02-05T16:16:54.711736 |
1xT4 | cuda | unquantized | pytorch | facebook/opt-125m | false | Traceback (most recent call last):
File "/llm-perf-backend/llm_perf/common/benchmark_runner.py", line 297, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 644, in _locate
obj = getattr(obj, part)
AttributeError: module 'optimum_benchmark.backends.pytorch' has no attribute 'backend'
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1603, in _get_module
return importlib.import_module("." + module_name, self.__name__)
File "/usr/lib/python3.10/importlib/__init__.py", line 126, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
File "<frozen importlib._bootstrap>", line 1050, in _gcd_import
File "<frozen importlib._bootstrap>", line 1027, in _find_and_load
File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked
File "<frozen importlib._bootstrap>", line 688, in _load_unlocked
File "<frozen importlib._bootstrap_external>", line 883, in exec_module
File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed
File "/usr/local/lib/python3.10/dist-packages/transformers/trainer.py", line 215, in <module>
from peft import PeftModel
File "/usr/local/lib/python3.10/dist-packages/peft/__init__.py", line 22, in <module>
from .auto import (
File "/usr/local/lib/python3.10/dist-packages/peft/auto.py", line 32, in <module>
from .mapping import MODEL_TYPE_TO_PEFT_MODEL_MAPPING
File "/usr/local/lib/python3.10/dist-packages/peft/mapping.py", line 22, in <module>
from peft.tuners.xlora.model import XLoraModel
File "/usr/local/lib/python3.10/dist-packages/peft/tuners/__init__.py", line 21, in <module>
from .lora import LoraConfig, LoraModel, LoftQConfig, LoraRuntimeConfig
File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/__init__.py", line 20, in <module>
from .model import LoraModel
File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/model.py", line 50, in <module>
from .awq import dispatch_awq
File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/awq.py", line 26, in <module>
from awq.modules.linear import WQLinear_GEMM
File "/usr/local/lib/python3.10/dist-packages/awq/__init__.py", line 2, in <module>
from awq.models.auto import AutoAWQForCausalLM
File "/usr/local/lib/python3.10/dist-packages/awq/models/__init__.py", line 1, in <module>
from .mpt import MptAWQForCausalLM
File "/usr/local/lib/python3.10/dist-packages/awq/models/mpt.py", line 1, in <module>
from .base import BaseAWQForCausalLM
File "/usr/local/lib/python3.10/dist-packages/awq/models/base.py", line 52, in <module>
from awq.quantize.quantizer import AwqQuantizer
File "/usr/local/lib/python3.10/dist-packages/awq/quantize/quantizer.py", line 10, in <module>
from awq.quantize.scale import apply_scale, apply_clip
File "/usr/local/lib/python3.10/dist-packages/awq/quantize/scale.py", line 8, in <module>
from transformers.models.llama.modeling_llama import LlamaRMSNorm
File "/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py", line 32, in <module>
from ...modeling_flash_attention_utils import _flash_attention_forward
File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py", line 27, in <module>
from flash_attn.bert_padding import index_first_axis, pad_input, unpad_input # noqa
File "/usr/local/lib/python3.10/dist-packages/flash_attn/__init__.py", line 3, in <module>
from flash_attn.flash_attn_interface import (
File "/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py", line 15, in <module>
import flash_attn_2_cuda as flash_attn_gpu
ImportError: /usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 650, in _locate
obj = import_module(mod)
File "/usr/lib/python3.10/importlib/__init__.py", line 126, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
File "<frozen importlib._bootstrap>", line 1050, in _gcd_import
File "<frozen importlib._bootstrap>", line 1027, in _find_and_load
File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked
File "<frozen importlib._bootstrap>", line 688, in _load_unlocked
File "<frozen importlib._bootstrap_external>", line 883, in exec_module
File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 10, in <module>
from transformers import (
File "<frozen importlib._bootstrap>", line 1075, in _handle_fromlist
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1593, in __getattr__
module = self._get_module(self._class_to_module[name])
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1605, in _get_module
raise RuntimeError(
RuntimeError: Failed to import transformers.trainer because of the following error (look up to see its traceback):
/usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 69, in run
backend_factory: Type[Backend] = get_class(backend_config._target_)
File "/usr/local/lib/python3.10/dist-packages/hydra/utils.py", line 40, in get_class
raise e
File "/usr/local/lib/python3.10/dist-packages/hydra/utils.py", line 31, in get_class
cls = _locate(path)
File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 658, in _locate
raise ImportError(
ImportError: Error loading 'optimum_benchmark.backends.pytorch.backend.PyTorchBackend':
RuntimeError('Failed to import transformers.trainer because of the following error (look up to see its traceback):\n/usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv')
| 2025-02-05T16:18:45.323625 | 03886520-a1ca-4afc-a74c-76bca3bbd085 | 2025-02-05T16:16:54.711736 |
1xT4 | cuda | unquantized | pytorch | openai-community/gpt2 | false | Traceback (most recent call last):
File "/llm-perf-backend/llm_perf/common/benchmark_runner.py", line 297, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 149, in load_transformers_model
self.load_transformers_model_with_no_weights()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 124, in load_transformers_model_with_no_weights
self.pretrained_model = self.automodel_loader.from_pretrained(
File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py", line 564, in from_pretrained
return model_class.from_pretrained(
File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 3826, in from_pretrained
config = cls._autoset_attn_implementation(
File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 1556, in _autoset_attn_implementation
cls._check_and_enable_flash_attn_2(
File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 1658, in _check_and_enable_flash_attn_2
raise ImportError(f"{preface} the package flash_attn seems to be not installed. {install_message}")
ImportError: FlashAttention2 has been toggled on, but it cannot be used due to the following error: the package flash_attn seems to be not installed. Please refer to the documentation of https://huggingface.co/docs/transformers/perf_infer_gpu_one#flashattention-2 to install Flash Attention 2.
| 2025-02-05T16:27:33.315082 | 8d5167da-766b-4715-902c-37597ace0904 | 2025-02-05T16:26:39.182355 |
1xT4 | cuda | unquantized | pytorch | openai-community/gpt2 | false | Traceback (most recent call last):
File "/llm-perf-backend/llm_perf/common/benchmark_runner.py", line 297, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 149, in load_transformers_model
self.load_transformers_model_with_no_weights()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 124, in load_transformers_model_with_no_weights
self.pretrained_model = self.automodel_loader.from_pretrained(
File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py", line 564, in from_pretrained
return model_class.from_pretrained(
File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 3826, in from_pretrained
config = cls._autoset_attn_implementation(
File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 1556, in _autoset_attn_implementation
cls._check_and_enable_flash_attn_2(
File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 1658, in _check_and_enable_flash_attn_2
raise ImportError(f"{preface} the package flash_attn seems to be not installed. {install_message}")
ImportError: FlashAttention2 has been toggled on, but it cannot be used due to the following error: the package flash_attn seems to be not installed. Please refer to the documentation of https://huggingface.co/docs/transformers/perf_infer_gpu_one#flashattention-2 to install Flash Attention 2.
| 2025-02-05T16:27:57.180163 | 8d5167da-766b-4715-902c-37597ace0904 | 2025-02-05T16:26:39.182355 |
1xT4 | cuda | unquantized | pytorch | facebook/opt-125m | false | Traceback (most recent call last):
File "/llm-perf-backend/llm_perf/common/benchmark_runner.py", line 297, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 149, in load_transformers_model
self.load_transformers_model_with_no_weights()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 124, in load_transformers_model_with_no_weights
self.pretrained_model = self.automodel_loader.from_pretrained(
File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py", line 564, in from_pretrained
return model_class.from_pretrained(
File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 3826, in from_pretrained
config = cls._autoset_attn_implementation(
File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 1556, in _autoset_attn_implementation
cls._check_and_enable_flash_attn_2(
File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 1658, in _check_and_enable_flash_attn_2
raise ImportError(f"{preface} the package flash_attn seems to be not installed. {install_message}")
ImportError: FlashAttention2 has been toggled on, but it cannot be used due to the following error: the package flash_attn seems to be not installed. Please refer to the documentation of https://huggingface.co/docs/transformers/perf_infer_gpu_one#flashattention-2 to install Flash Attention 2.
| 2025-02-05T16:28:54.977003 | 8d5167da-766b-4715-902c-37597ace0904 | 2025-02-05T16:26:39.182355 |
1xT4 | cuda | unquantized | pytorch | facebook/opt-125m | false | Traceback (most recent call last):
File "/llm-perf-backend/llm_perf/common/benchmark_runner.py", line 297, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 149, in load_transformers_model
self.load_transformers_model_with_no_weights()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 124, in load_transformers_model_with_no_weights
self.pretrained_model = self.automodel_loader.from_pretrained(
File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py", line 564, in from_pretrained
return model_class.from_pretrained(
File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 3826, in from_pretrained
config = cls._autoset_attn_implementation(
File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 1556, in _autoset_attn_implementation
cls._check_and_enable_flash_attn_2(
File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 1658, in _check_and_enable_flash_attn_2
raise ImportError(f"{preface} the package flash_attn seems to be not installed. {install_message}")
ImportError: FlashAttention2 has been toggled on, but it cannot be used due to the following error: the package flash_attn seems to be not installed. Please refer to the documentation of https://huggingface.co/docs/transformers/perf_infer_gpu_one#flashattention-2 to install Flash Attention 2.
| 2025-02-05T16:29:17.544956 | 8d5167da-766b-4715-902c-37597ace0904 | 2025-02-05T16:26:39.182355 |
1xT4 | cuda | unquantized | pytorch | openai-community/gpt2 | false | Traceback (most recent call last):
File "/llm-perf-backend/llm_perf/common/benchmark_runner.py", line 297, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 644, in _locate
obj = getattr(obj, part)
AttributeError: module 'optimum_benchmark.backends.pytorch' has no attribute 'backend'
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1603, in _get_module
return importlib.import_module("." + module_name, self.__name__)
File "/usr/lib/python3.10/importlib/__init__.py", line 126, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
File "<frozen importlib._bootstrap>", line 1050, in _gcd_import
File "<frozen importlib._bootstrap>", line 1027, in _find_and_load
File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked
File "<frozen importlib._bootstrap>", line 688, in _load_unlocked
File "<frozen importlib._bootstrap_external>", line 883, in exec_module
File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed
File "/usr/local/lib/python3.10/dist-packages/transformers/trainer.py", line 215, in <module>
from peft import PeftModel
File "/usr/local/lib/python3.10/dist-packages/peft/__init__.py", line 22, in <module>
from .auto import (
File "/usr/local/lib/python3.10/dist-packages/peft/auto.py", line 32, in <module>
from .mapping import MODEL_TYPE_TO_PEFT_MODEL_MAPPING
File "/usr/local/lib/python3.10/dist-packages/peft/mapping.py", line 22, in <module>
from peft.tuners.xlora.model import XLoraModel
File "/usr/local/lib/python3.10/dist-packages/peft/tuners/__init__.py", line 21, in <module>
from .lora import LoraConfig, LoraModel, LoftQConfig, LoraRuntimeConfig
File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/__init__.py", line 20, in <module>
from .model import LoraModel
File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/model.py", line 50, in <module>
from .awq import dispatch_awq
File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/awq.py", line 26, in <module>
from awq.modules.linear import WQLinear_GEMM
File "/usr/local/lib/python3.10/dist-packages/awq/__init__.py", line 2, in <module>
from awq.models.auto import AutoAWQForCausalLM
File "/usr/local/lib/python3.10/dist-packages/awq/models/__init__.py", line 1, in <module>
from .mpt import MptAWQForCausalLM
File "/usr/local/lib/python3.10/dist-packages/awq/models/mpt.py", line 1, in <module>
from .base import BaseAWQForCausalLM
File "/usr/local/lib/python3.10/dist-packages/awq/models/base.py", line 52, in <module>
from awq.quantize.quantizer import AwqQuantizer
File "/usr/local/lib/python3.10/dist-packages/awq/quantize/quantizer.py", line 10, in <module>
from awq.quantize.scale import apply_scale, apply_clip
File "/usr/local/lib/python3.10/dist-packages/awq/quantize/scale.py", line 8, in <module>
from transformers.models.llama.modeling_llama import LlamaRMSNorm
File "/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py", line 32, in <module>
from ...modeling_flash_attention_utils import _flash_attention_forward
File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py", line 27, in <module>
from flash_attn.bert_padding import index_first_axis, pad_input, unpad_input # noqa
File "/usr/local/lib/python3.10/dist-packages/flash_attn/__init__.py", line 3, in <module>
from flash_attn.flash_attn_interface import (
File "/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py", line 15, in <module>
import flash_attn_2_cuda as flash_attn_gpu
ImportError: /usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 650, in _locate
obj = import_module(mod)
File "/usr/lib/python3.10/importlib/__init__.py", line 126, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
File "<frozen importlib._bootstrap>", line 1050, in _gcd_import
File "<frozen importlib._bootstrap>", line 1027, in _find_and_load
File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked
File "<frozen importlib._bootstrap>", line 688, in _load_unlocked
File "<frozen importlib._bootstrap_external>", line 883, in exec_module
File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 10, in <module>
from transformers import (
File "<frozen importlib._bootstrap>", line 1075, in _handle_fromlist
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1593, in __getattr__
module = self._get_module(self._class_to_module[name])
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1605, in _get_module
raise RuntimeError(
RuntimeError: Failed to import transformers.trainer because of the following error (look up to see its traceback):
/usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 69, in run
backend_factory: Type[Backend] = get_class(backend_config._target_)
File "/usr/local/lib/python3.10/dist-packages/hydra/utils.py", line 40, in get_class
raise e
File "/usr/local/lib/python3.10/dist-packages/hydra/utils.py", line 31, in get_class
cls = _locate(path)
File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 658, in _locate
raise ImportError(
ImportError: Error loading 'optimum_benchmark.backends.pytorch.backend.PyTorchBackend':
RuntimeError('Failed to import transformers.trainer because of the following error (look up to see its traceback):\n/usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv')
| 2025-02-05T16:31:42.373752 | 4b0db4ab-2fe1-4e13-a7c6-93344a8b08a2 | 2025-02-05T16:30:57.040646 |
1xT4 | cuda | unquantized | pytorch | openai-community/gpt2 | false | Traceback (most recent call last):
File "/llm-perf-backend/llm_perf/common/benchmark_runner.py", line 297, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 644, in _locate
obj = getattr(obj, part)
AttributeError: module 'optimum_benchmark.backends.pytorch' has no attribute 'backend'
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1603, in _get_module
return importlib.import_module("." + module_name, self.__name__)
File "/usr/lib/python3.10/importlib/__init__.py", line 126, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
File "<frozen importlib._bootstrap>", line 1050, in _gcd_import
File "<frozen importlib._bootstrap>", line 1027, in _find_and_load
File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked
File "<frozen importlib._bootstrap>", line 688, in _load_unlocked
File "<frozen importlib._bootstrap_external>", line 883, in exec_module
File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed
File "/usr/local/lib/python3.10/dist-packages/transformers/trainer.py", line 215, in <module>
from peft import PeftModel
File "/usr/local/lib/python3.10/dist-packages/peft/__init__.py", line 22, in <module>
from .auto import (
File "/usr/local/lib/python3.10/dist-packages/peft/auto.py", line 32, in <module>
from .mapping import MODEL_TYPE_TO_PEFT_MODEL_MAPPING
File "/usr/local/lib/python3.10/dist-packages/peft/mapping.py", line 22, in <module>
from peft.tuners.xlora.model import XLoraModel
File "/usr/local/lib/python3.10/dist-packages/peft/tuners/__init__.py", line 21, in <module>
from .lora import LoraConfig, LoraModel, LoftQConfig, LoraRuntimeConfig
File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/__init__.py", line 20, in <module>
from .model import LoraModel
File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/model.py", line 50, in <module>
from .awq import dispatch_awq
File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/awq.py", line 26, in <module>
from awq.modules.linear import WQLinear_GEMM
File "/usr/local/lib/python3.10/dist-packages/awq/__init__.py", line 2, in <module>
from awq.models.auto import AutoAWQForCausalLM
File "/usr/local/lib/python3.10/dist-packages/awq/models/__init__.py", line 1, in <module>
from .mpt import MptAWQForCausalLM
File "/usr/local/lib/python3.10/dist-packages/awq/models/mpt.py", line 1, in <module>
from .base import BaseAWQForCausalLM
File "/usr/local/lib/python3.10/dist-packages/awq/models/base.py", line 52, in <module>
from awq.quantize.quantizer import AwqQuantizer
File "/usr/local/lib/python3.10/dist-packages/awq/quantize/quantizer.py", line 10, in <module>
from awq.quantize.scale import apply_scale, apply_clip
File "/usr/local/lib/python3.10/dist-packages/awq/quantize/scale.py", line 8, in <module>
from transformers.models.llama.modeling_llama import LlamaRMSNorm
File "/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py", line 32, in <module>
from ...modeling_flash_attention_utils import _flash_attention_forward
File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py", line 27, in <module>
from flash_attn.bert_padding import index_first_axis, pad_input, unpad_input # noqa
File "/usr/local/lib/python3.10/dist-packages/flash_attn/__init__.py", line 3, in <module>
from flash_attn.flash_attn_interface import (
File "/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py", line 15, in <module>
import flash_attn_2_cuda as flash_attn_gpu
ImportError: /usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 650, in _locate
obj = import_module(mod)
File "/usr/lib/python3.10/importlib/__init__.py", line 126, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
File "<frozen importlib._bootstrap>", line 1050, in _gcd_import
File "<frozen importlib._bootstrap>", line 1027, in _find_and_load
File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked
File "<frozen importlib._bootstrap>", line 688, in _load_unlocked
File "<frozen importlib._bootstrap_external>", line 883, in exec_module
File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 10, in <module>
from transformers import (
File "<frozen importlib._bootstrap>", line 1075, in _handle_fromlist
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1593, in __getattr__
module = self._get_module(self._class_to_module[name])
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1605, in _get_module
raise RuntimeError(
RuntimeError: Failed to import transformers.trainer because of the following error (look up to see its traceback):
/usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 69, in run
backend_factory: Type[Backend] = get_class(backend_config._target_)
File "/usr/local/lib/python3.10/dist-packages/hydra/utils.py", line 40, in get_class
raise e
File "/usr/local/lib/python3.10/dist-packages/hydra/utils.py", line 31, in get_class
cls = _locate(path)
File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 658, in _locate
raise ImportError(
ImportError: Error loading 'optimum_benchmark.backends.pytorch.backend.PyTorchBackend':
RuntimeError('Failed to import transformers.trainer because of the following error (look up to see its traceback):\n/usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv')
| 2025-02-06T08:45:39.449306 | 73a95cce-e60e-4115-8874-4aaa173b2aa3 | 2025-02-06T08:44:53.179701 |
1xT4 | cuda | unquantized | pytorch | openai-community/gpt2 | false | Traceback (most recent call last):
File "/llm-perf-backend/llm_perf/common/benchmark_runner.py", line 297, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 644, in _locate
obj = getattr(obj, part)
AttributeError: module 'optimum_benchmark.backends.pytorch' has no attribute 'backend'
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1603, in _get_module
return importlib.import_module("." + module_name, self.__name__)
File "/usr/lib/python3.10/importlib/__init__.py", line 126, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
File "<frozen importlib._bootstrap>", line 1050, in _gcd_import
File "<frozen importlib._bootstrap>", line 1027, in _find_and_load
File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked
File "<frozen importlib._bootstrap>", line 688, in _load_unlocked
File "<frozen importlib._bootstrap_external>", line 883, in exec_module
File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed
File "/usr/local/lib/python3.10/dist-packages/transformers/trainer.py", line 215, in <module>
from peft import PeftModel
File "/usr/local/lib/python3.10/dist-packages/peft/__init__.py", line 22, in <module>
from .auto import (
File "/usr/local/lib/python3.10/dist-packages/peft/auto.py", line 32, in <module>
from .mapping import MODEL_TYPE_TO_PEFT_MODEL_MAPPING
File "/usr/local/lib/python3.10/dist-packages/peft/mapping.py", line 22, in <module>
from peft.tuners.xlora.model import XLoraModel
File "/usr/local/lib/python3.10/dist-packages/peft/tuners/__init__.py", line 21, in <module>
from .lora import LoraConfig, LoraModel, LoftQConfig, LoraRuntimeConfig
File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/__init__.py", line 20, in <module>
from .model import LoraModel
File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/model.py", line 50, in <module>
from .awq import dispatch_awq
File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/awq.py", line 26, in <module>
from awq.modules.linear import WQLinear_GEMM
File "/usr/local/lib/python3.10/dist-packages/awq/__init__.py", line 2, in <module>
from awq.models.auto import AutoAWQForCausalLM
File "/usr/local/lib/python3.10/dist-packages/awq/models/__init__.py", line 1, in <module>
from .mpt import MptAWQForCausalLM
File "/usr/local/lib/python3.10/dist-packages/awq/models/mpt.py", line 1, in <module>
from .base import BaseAWQForCausalLM
File "/usr/local/lib/python3.10/dist-packages/awq/models/base.py", line 52, in <module>
from awq.quantize.quantizer import AwqQuantizer
File "/usr/local/lib/python3.10/dist-packages/awq/quantize/quantizer.py", line 10, in <module>
from awq.quantize.scale import apply_scale, apply_clip
File "/usr/local/lib/python3.10/dist-packages/awq/quantize/scale.py", line 8, in <module>
from transformers.models.llama.modeling_llama import LlamaRMSNorm
File "/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py", line 32, in <module>
from ...modeling_flash_attention_utils import _flash_attention_forward
File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py", line 27, in <module>
from flash_attn.bert_padding import index_first_axis, pad_input, unpad_input # noqa
File "/usr/local/lib/python3.10/dist-packages/flash_attn/__init__.py", line 3, in <module>
from flash_attn.flash_attn_interface import (
File "/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py", line 15, in <module>
import flash_attn_2_cuda as flash_attn_gpu
ImportError: /usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 650, in _locate
obj = import_module(mod)
File "/usr/lib/python3.10/importlib/__init__.py", line 126, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
File "<frozen importlib._bootstrap>", line 1050, in _gcd_import
File "<frozen importlib._bootstrap>", line 1027, in _find_and_load
File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked
File "<frozen importlib._bootstrap>", line 688, in _load_unlocked
File "<frozen importlib._bootstrap_external>", line 883, in exec_module
File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 10, in <module>
from transformers import (
File "<frozen importlib._bootstrap>", line 1075, in _handle_fromlist
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1593, in __getattr__
module = self._get_module(self._class_to_module[name])
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1605, in _get_module
raise RuntimeError(
RuntimeError: Failed to import transformers.trainer because of the following error (look up to see its traceback):
/usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 69, in run
backend_factory: Type[Backend] = get_class(backend_config._target_)
File "/usr/local/lib/python3.10/dist-packages/hydra/utils.py", line 40, in get_class
raise e
File "/usr/local/lib/python3.10/dist-packages/hydra/utils.py", line 31, in get_class
cls = _locate(path)
File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 658, in _locate
raise ImportError(
ImportError: Error loading 'optimum_benchmark.backends.pytorch.backend.PyTorchBackend':
RuntimeError('Failed to import transformers.trainer because of the following error (look up to see its traceback):\n/usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv')
| 2025-02-06T08:45:54.648082 | 73a95cce-e60e-4115-8874-4aaa173b2aa3 | 2025-02-06T08:44:53.179701 |
1xT4 | cuda | unquantized | pytorch | bigscience/bloomz-560m | false | Traceback (most recent call last):
File "/llm-perf-backend/llm_perf/common/benchmark_runner.py", line 297, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 149, in load_transformers_model
self.load_transformers_model_with_no_weights()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 124, in load_transformers_model_with_no_weights
self.pretrained_model = self.automodel_loader.from_pretrained(
File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py", line 564, in from_pretrained
return model_class.from_pretrained(
File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 3826, in from_pretrained
config = cls._autoset_attn_implementation(
File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 1565, in _autoset_attn_implementation
config = cls._check_and_enable_sdpa(
File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 1731, in _check_and_enable_sdpa
raise ValueError(
ValueError: BloomForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation="eager"` meanwhile. Example: `model = AutoModel.from_pretrained("openai/whisper-tiny", attn_implementation="eager")`
| 2025-02-06T09:18:17.866903 | 1c99483e-b56b-4326-aee9-66f38518c30e | 2025-02-06T09:16:35.806219 |
1xT4 | cuda | unquantized | pytorch | openai-community/gpt2 | true | 2025-02-06T09:20:44.677569 | 0e3562c9-963a-49c9-be25-cfa94c03003a | 2025-02-06T09:19:41.190913 |
|
1xT4 | cuda | unquantized | pytorch | openai-community/gpt2 | true | 2025-02-06T09:21:49.942995 | 0e3562c9-963a-49c9-be25-cfa94c03003a | 2025-02-06T09:19:41.190913 |
|
1xT4 | cuda | unquantized | pytorch | openai-community/gpt2 | true | 2025-02-06T09:22:56.522260 | 0e3562c9-963a-49c9-be25-cfa94c03003a | 2025-02-06T09:19:41.190913 |
|
1xT4 | cuda | unquantized | pytorch | openai-community/gpt2 | true | 2025-02-06T09:24:03.915016 | 0e3562c9-963a-49c9-be25-cfa94c03003a | 2025-02-06T09:19:41.190913 |
|
32vCPU-C7i | cpu | unquantized | pytorch | openai-community/gpt2 | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T09:56:50.675962 | 938a1a37-c311-4be1-85f4-dd27b03cce3c | 2025-02-06T09:56:41.233415 |
32vCPU-C7i | cpu | unquantized | pytorch | openai-community/gpt2 | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T09:57:01.606291 | 938a1a37-c311-4be1-85f4-dd27b03cce3c | 2025-02-06T09:56:41.233415 |
32vCPU-C7i | cpu | unquantized | pytorch | openai-community/gpt2 | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T09:57:13.683017 | 938a1a37-c311-4be1-85f4-dd27b03cce3c | 2025-02-06T09:56:41.233415 |
32vCPU-C7i | cpu | unquantized | pytorch | openai-community/gpt2 | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T09:57:24.310483 | 938a1a37-c311-4be1-85f4-dd27b03cce3c | 2025-02-06T09:56:41.233415 |
32vCPU-C7i | cpu | unquantized | pytorch | openai-community/gpt2 | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T09:57:36.126144 | 938a1a37-c311-4be1-85f4-dd27b03cce3c | 2025-02-06T09:56:41.233415 |
32vCPU-C7i | cpu | unquantized | pytorch | openai-community/gpt2 | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T09:57:46.794500 | 938a1a37-c311-4be1-85f4-dd27b03cce3c | 2025-02-06T09:56:41.233415 |
32vCPU-C7i | cpu | unquantized | pytorch | facebook/opt-125m | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T09:57:57.638603 | 938a1a37-c311-4be1-85f4-dd27b03cce3c | 2025-02-06T09:56:41.233415 |
32vCPU-C7i | cpu | unquantized | pytorch | facebook/opt-125m | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T09:58:11.213695 | 938a1a37-c311-4be1-85f4-dd27b03cce3c | 2025-02-06T09:56:41.233415 |
32vCPU-C7i | cpu | unquantized | pytorch | facebook/opt-125m | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T09:58:24.607254 | 938a1a37-c311-4be1-85f4-dd27b03cce3c | 2025-02-06T09:56:41.233415 |
32vCPU-C7i | cpu | unquantized | pytorch | facebook/opt-125m | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T09:58:39.142446 | 938a1a37-c311-4be1-85f4-dd27b03cce3c | 2025-02-06T09:56:41.233415 |
32vCPU-C7i | cpu | unquantized | pytorch | facebook/opt-125m | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T09:58:54.351007 | 938a1a37-c311-4be1-85f4-dd27b03cce3c | 2025-02-06T09:56:41.233415 |
32vCPU-C7i | cpu | unquantized | pytorch | facebook/opt-125m | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T09:59:11.381282 | 938a1a37-c311-4be1-85f4-dd27b03cce3c | 2025-02-06T09:56:41.233415 |
32vCPU-C7i | cpu | unquantized | pytorch | openai-community/gpt2 | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:09:31.650354 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
32vCPU-C7i | cpu | unquantized | pytorch | openai-community/gpt2 | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:09:41.822548 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
32vCPU-C7i | cpu | unquantized | pytorch | openai-community/gpt2 | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:09:50.955328 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
32vCPU-C7i | cpu | unquantized | pytorch | openai-community/gpt2 | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:10:02.091578 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
32vCPU-C7i | cpu | unquantized | pytorch | openai-community/gpt2 | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:10:13.644524 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
1xT4 | cuda | torchao | pytorch | bigscience/bloomz-560m | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model
self.process_quantization_config()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config
self.quantization_config = AutoQuantizationConfig.from_dict(
File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict
return target_cls.from_dict(quantization_config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict
config = cls(**config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__
self.post_init()
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init
raise ValueError(
ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7fab04143250>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
| 2025-02-06T10:10:26.426187 | bd27598f-6f94-4e3d-9747-85c8ce835edd | 2025-02-06T10:09:56.868153 |
32vCPU-C7i | cpu | unquantized | pytorch | openai-community/gpt2 | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:10:27.799185 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
32vCPU-C7i | cpu | unquantized | pytorch | facebook/opt-125m | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:10:44.632379 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
1xT4 | cuda | torchao | pytorch | meta-llama/Llama-3.1-8B-Instruct | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model
self.process_quantization_config()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config
self.quantization_config = AutoQuantizationConfig.from_dict(
File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict
return target_cls.from_dict(quantization_config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict
config = cls(**config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__
self.post_init()
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init
raise ValueError(
ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f885846fd90>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
| 2025-02-06T10:10:54.800330 | bd27598f-6f94-4e3d-9747-85c8ce835edd | 2025-02-06T10:09:56.868153 |
32vCPU-C7i | cpu | unquantized | pytorch | facebook/opt-125m | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:11:03.810607 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
32vCPU-C7i | cpu | unquantized | pytorch | facebook/opt-125m | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:11:20.997163 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
32vCPU-C7i | cpu | unquantized | pytorch | facebook/opt-125m | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:11:40.876875 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
1xT4 | cuda | torchao | pytorch | distilbert/distilgpt2 | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model
self.process_quantization_config()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config
self.quantization_config = AutoQuantizationConfig.from_dict(
File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict
return target_cls.from_dict(quantization_config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict
config = cls(**config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__
self.post_init()
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init
raise ValueError(
ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7fb8da1fa710>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
| 2025-02-06T10:11:48.635299 | bd27598f-6f94-4e3d-9747-85c8ce835edd | 2025-02-06T10:09:56.868153 |
32vCPU-C7i | cpu | unquantized | pytorch | facebook/opt-125m | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:12:00.639319 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
1xT4 | cuda | torchao | pytorch | distilbert/distilgpt2 | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model
self.process_quantization_config()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config
self.quantization_config = AutoQuantizationConfig.from_dict(
File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict
return target_cls.from_dict(quantization_config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict
config = cls(**config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__
self.post_init()
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init
raise ValueError(
ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f4c647c4ca0>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
| 2025-02-06T10:12:14.822852 | bd27598f-6f94-4e3d-9747-85c8ce835edd | 2025-02-06T10:09:56.868153 |
32vCPU-C7i | cpu | unquantized | pytorch | facebook/opt-125m | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:12:18.359729 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
32vCPU-C7i | cpu | unquantized | pytorch | bigscience/bloomz-560m | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:12:38.778996 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
1xT4 | cuda | torchao | pytorch | meta-llama/Llama-3.1-405B | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model
self.process_quantization_config()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config
self.quantization_config = AutoQuantizationConfig.from_dict(
File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict
return target_cls.from_dict(quantization_config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict
config = cls(**config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__
self.post_init()
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init
raise ValueError(
ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f239014fb50>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
| 2025-02-06T10:12:44.565974 | bd27598f-6f94-4e3d-9747-85c8ce835edd | 2025-02-06T10:09:56.868153 |
32vCPU-C7i | cpu | unquantized | pytorch | bigscience/bloomz-560m | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:12:58.240254 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
1xT4 | cuda | torchao | pytorch | meta-llama/Llama-3.1-405B | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model
self.process_quantization_config()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config
self.quantization_config = AutoQuantizationConfig.from_dict(
File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict
return target_cls.from_dict(quantization_config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict
config = cls(**config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__
self.post_init()
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init
raise ValueError(
ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f4838b1e050>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
| 2025-02-06T10:13:11.834624 | bd27598f-6f94-4e3d-9747-85c8ce835edd | 2025-02-06T10:09:56.868153 |
32vCPU-C7i | cpu | unquantized | pytorch | bigscience/bloomz-560m | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:13:17.293850 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
1xA10 | cuda | torchao | pytorch | bigscience/bloomz-560m | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model
self.process_quantization_config()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config
self.quantization_config = AutoQuantizationConfig.from_dict(
File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict
return target_cls.from_dict(quantization_config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict
config = cls(**config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__
self.post_init()
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init
raise ValueError(
ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f5ffc2832e0>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
| 2025-02-06T10:13:19.241876 | fc85a78b-b92a-46d9-b36b-715516040a48 | 2025-02-06T10:12:52.699176 |
32vCPU-C7i | cpu | unquantized | pytorch | bigscience/bloomz-560m | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:13:36.395832 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
1xT4 | cuda | torchao | pytorch | meta-llama/Meta-Llama-3-8B-Instruct | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model
self.process_quantization_config()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config
self.quantization_config = AutoQuantizationConfig.from_dict(
File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict
return target_cls.from_dict(quantization_config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict
config = cls(**config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__
self.post_init()
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init
raise ValueError(
ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f096e25bd00>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
| 2025-02-06T10:13:40.308349 | bd27598f-6f94-4e3d-9747-85c8ce835edd | 2025-02-06T10:09:56.868153 |
1xA10 | cuda | torchao | pytorch | distilbert/distilgpt2 | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model
self.process_quantization_config()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config
self.quantization_config = AutoQuantizationConfig.from_dict(
File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict
return target_cls.from_dict(quantization_config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict
config = cls(**config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__
self.post_init()
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init
raise ValueError(
ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f45002ca710>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
| 2025-02-06T10:13:45.720524 | fc85a78b-b92a-46d9-b36b-715516040a48 | 2025-02-06T10:12:52.699176 |
32vCPU-C7i | cpu | unquantized | pytorch | bigscience/bloomz-560m | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:13:56.268390 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
1xT4 | cuda | torchao | pytorch | meta-llama/Meta-Llama-3-8B-Instruct | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model
self.process_quantization_config()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config
self.quantization_config = AutoQuantizationConfig.from_dict(
File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict
return target_cls.from_dict(quantization_config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict
config = cls(**config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__
self.post_init()
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init
raise ValueError(
ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7efff024a050>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
| 2025-02-06T10:14:07.388246 | bd27598f-6f94-4e3d-9747-85c8ce835edd | 2025-02-06T10:09:56.868153 |
1xA10 | cuda | torchao | pytorch | distilbert/distilgpt2 | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model
self.process_quantization_config()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config
self.quantization_config = AutoQuantizationConfig.from_dict(
File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict
return target_cls.from_dict(quantization_config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict
config = cls(**config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__
self.post_init()
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init
raise ValueError(
ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f4b05ab4ca0>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
| 2025-02-06T10:14:10.815562 | fc85a78b-b92a-46d9-b36b-715516040a48 | 2025-02-06T10:12:52.699176 |
32vCPU-C7i | cpu | unquantized | pytorch | bigscience/bloomz-560m | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:14:14.324658 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
32vCPU-C7i | cpu | unquantized | pytorch | meta-llama/Llama-3.1-8B-Instruct | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:14:32.663277 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
1xT4 | cuda | torchao | pytorch | microsoft/Phi-3-mini-4k-instruct | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model
self.process_quantization_config()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config
self.quantization_config = AutoQuantizationConfig.from_dict(
File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict
return target_cls.from_dict(quantization_config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict
config = cls(**config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__
self.post_init()
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init
raise ValueError(
ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f2335b3c280>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
| 2025-02-06T10:14:35.179953 | bd27598f-6f94-4e3d-9747-85c8ce835edd | 2025-02-06T10:09:56.868153 |
1xA10 | cuda | torchao | pytorch | meta-llama/Llama-3.1-405B | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model
self.process_quantization_config()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config
self.quantization_config = AutoQuantizationConfig.from_dict(
File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict
return target_cls.from_dict(quantization_config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict
config = cls(**config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__
self.post_init()
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init
raise ValueError(
ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f352c733b50>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
| 2025-02-06T10:14:36.775187 | fc85a78b-b92a-46d9-b36b-715516040a48 | 2025-02-06T10:12:52.699176 |
32vCPU-C7i | cpu | unquantized | pytorch | meta-llama/Llama-3.1-8B-Instruct | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:14:52.253129 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
1xT4 | cuda | torchao | pytorch | microsoft/Phi-3-mini-4k-instruct | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model
self.process_quantization_config()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config
self.quantization_config = AutoQuantizationConfig.from_dict(
File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict
return target_cls.from_dict(quantization_config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict
config = cls(**config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__
self.post_init()
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init
raise ValueError(
ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f2e74226b00>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
| 2025-02-06T10:15:02.590667 | bd27598f-6f94-4e3d-9747-85c8ce835edd | 2025-02-06T10:09:56.868153 |
1xA10 | cuda | torchao | pytorch | meta-llama/Llama-3.1-405B | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model
self.process_quantization_config()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config
self.quantization_config = AutoQuantizationConfig.from_dict(
File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict
return target_cls.from_dict(quantization_config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict
config = cls(**config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__
self.post_init()
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init
raise ValueError(
ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f7bd4552050>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
| 2025-02-06T10:15:04.568477 | fc85a78b-b92a-46d9-b36b-715516040a48 | 2025-02-06T10:12:52.699176 |
32vCPU-C7i | cpu | unquantized | pytorch | meta-llama/Llama-3.1-8B-Instruct | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:15:10.538588 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
32vCPU-C7i | cpu | unquantized | pytorch | meta-llama/Llama-3.1-8B-Instruct | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:15:30.458954 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
1xA10 | cuda | torchao | pytorch | meta-llama/Meta-Llama-3-8B-Instruct | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model
self.process_quantization_config()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config
self.quantization_config = AutoQuantizationConfig.from_dict(
File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict
return target_cls.from_dict(quantization_config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict
config = cls(**config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__
self.post_init()
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init
raise ValueError(
ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f38f025fd00>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
| 2025-02-06T10:15:34.638005 | fc85a78b-b92a-46d9-b36b-715516040a48 | 2025-02-06T10:12:52.699176 |
32vCPU-C7i | cpu | unquantized | pytorch | meta-llama/Llama-3.1-8B-Instruct | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:15:46.892472 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
1xT4 | cuda | torchao | pytorch | meta-llama/Meta-Llama-3-8B | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model
self.process_quantization_config()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config
self.quantization_config = AutoQuantizationConfig.from_dict(
File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict
return target_cls.from_dict(quantization_config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict
config = cls(**config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__
self.post_init()
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init
raise ValueError(
ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f51b4142050>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
| 2025-02-06T10:15:57.063359 | bd27598f-6f94-4e3d-9747-85c8ce835edd | 2025-02-06T10:09:56.868153 |
1xA10 | cuda | torchao | pytorch | meta-llama/Meta-Llama-3-8B-Instruct | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model
self.process_quantization_config()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config
self.quantization_config = AutoQuantizationConfig.from_dict(
File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict
return target_cls.from_dict(quantization_config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict
config = cls(**config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__
self.post_init()
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init
raise ValueError(
ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f5158816050>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
| 2025-02-06T10:16:02.322652 | fc85a78b-b92a-46d9-b36b-715516040a48 | 2025-02-06T10:12:52.699176 |
32vCPU-C7i | cpu | unquantized | pytorch | meta-llama/Llama-3.1-405B | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:16:04.758642 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
32vCPU-C7i | cpu | unquantized | pytorch | meta-llama/Llama-3.1-405B | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:16:21.725048 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
1xT4 | cuda | torchao | pytorch | liuhaotian/llava-v1.5-7b | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model
self.process_quantization_config()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config
self.quantization_config = AutoQuantizationConfig.from_dict(
File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict
return target_cls.from_dict(quantization_config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict
config = cls(**config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__
self.post_init()
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init
raise ValueError(
ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f6190389fc0>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
| 2025-02-06T10:16:23.619766 | bd27598f-6f94-4e3d-9747-85c8ce835edd | 2025-02-06T10:09:56.868153 |
1xA10 | cuda | torchao | pytorch | microsoft/Phi-3-mini-4k-instruct | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model
self.process_quantization_config()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config
self.quantization_config = AutoQuantizationConfig.from_dict(
File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict
return target_cls.from_dict(quantization_config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict
config = cls(**config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__
self.post_init()
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init
raise ValueError(
ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7fedf8184280>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
| 2025-02-06T10:16:29.309172 | fc85a78b-b92a-46d9-b36b-715516040a48 | 2025-02-06T10:12:52.699176 |
32vCPU-C7i | cpu | unquantized | pytorch | meta-llama/Llama-3.1-405B | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:16:39.801791 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
1xT4 | cuda | torchao | pytorch | liuhaotian/llava-v1.5-7b | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model
self.process_quantization_config()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config
self.quantization_config = AutoQuantizationConfig.from_dict(
File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict
return target_cls.from_dict(quantization_config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict
config = cls(**config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__
self.post_init()
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init
raise ValueError(
ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7fc4004204c0>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
| 2025-02-06T10:16:50.135365 | bd27598f-6f94-4e3d-9747-85c8ce835edd | 2025-02-06T10:09:56.868153 |
1xA10 | cuda | torchao | pytorch | microsoft/Phi-3-mini-4k-instruct | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model
self.process_quantization_config()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config
self.quantization_config = AutoQuantizationConfig.from_dict(
File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict
return target_cls.from_dict(quantization_config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict
config = cls(**config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__
self.post_init()
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init
raise ValueError(
ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f047c42eb00>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
| 2025-02-06T10:16:54.690294 | fc85a78b-b92a-46d9-b36b-715516040a48 | 2025-02-06T10:12:52.699176 |
32vCPU-C7i | cpu | unquantized | pytorch | meta-llama/Llama-3.1-405B | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:16:58.191029 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
32vCPU-C7i | cpu | unquantized | pytorch | meta-llama/Llama-3.1-405B | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:17:16.353456 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
1xA10 | cuda | torchao | pytorch | meta-llama/Meta-Llama-3-8B | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model
self.process_quantization_config()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config
self.quantization_config = AutoQuantizationConfig.from_dict(
File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict
return target_cls.from_dict(quantization_config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict
config = cls(**config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__
self.post_init()
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init
raise ValueError(
ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7fe6a00a3be0>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
| 2025-02-06T10:17:20.742154 | fc85a78b-b92a-46d9-b36b-715516040a48 | 2025-02-06T10:12:52.699176 |
32vCPU-C7i | cpu | unquantized | pytorch | meta-llama/Llama-3.1-405B | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:17:32.869921 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
1xT4 | cuda | torchao | pytorch | trl-internal-testing/dummy-GPT2-correct-vocab | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model
self.process_quantization_config()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config
self.quantization_config = AutoQuantizationConfig.from_dict(
File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict
return target_cls.from_dict(quantization_config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict
config = cls(**config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__
self.post_init()
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init
raise ValueError(
ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f20ccbd0e50>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
| 2025-02-06T10:17:47.243376 | bd27598f-6f94-4e3d-9747-85c8ce835edd | 2025-02-06T10:09:56.868153 |
32vCPU-C7i | cpu | unquantized | pytorch | microsoft/Phi-3-mini-4k-instruct | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:17:52.456368 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
1xA10 | cuda | torchao | pytorch | liuhaotian/llava-v1.5-7b | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model
self.process_quantization_config()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config
self.quantization_config = AutoQuantizationConfig.from_dict(
File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict
return target_cls.from_dict(quantization_config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict
config = cls(**config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__
self.post_init()
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init
raise ValueError(
ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7ff7e0a25fc0>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
| 2025-02-06T10:18:12.501138 | fc85a78b-b92a-46d9-b36b-715516040a48 | 2025-02-06T10:12:52.699176 |
1xT4 | cuda | torchao | pytorch | datificate/gpt2-small-spanish | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model
self.process_quantization_config()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config
self.quantization_config = AutoQuantizationConfig.from_dict(
File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict
return target_cls.from_dict(quantization_config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict
config = cls(**config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__
self.post_init()
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init
raise ValueError(
ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7fbaa352a9e0>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
| 2025-02-06T10:18:14.516366 | bd27598f-6f94-4e3d-9747-85c8ce835edd | 2025-02-06T10:09:56.868153 |
32vCPU-C7i | cpu | unquantized | pytorch | microsoft/Phi-3-mini-4k-instruct | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:18:30.576978 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
1xA10 | cuda | torchao | pytorch | liuhaotian/llava-v1.5-7b | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model
self.process_quantization_config()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config
self.quantization_config = AutoQuantizationConfig.from_dict(
File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict
return target_cls.from_dict(quantization_config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict
config = cls(**config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__
self.post_init()
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init
raise ValueError(
ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7fd2b035c4c0>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
| 2025-02-06T10:18:38.645285 | fc85a78b-b92a-46d9-b36b-715516040a48 | 2025-02-06T10:12:52.699176 |
1xT4 | cuda | torchao | pytorch | datificate/gpt2-small-spanish | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model
self.process_quantization_config()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config
self.quantization_config = AutoQuantizationConfig.from_dict(
File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict
return target_cls.from_dict(quantization_config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict
config = cls(**config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__
self.post_init()
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init
raise ValueError(
ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f0e567f0dc0>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
| 2025-02-06T10:18:41.917489 | bd27598f-6f94-4e3d-9747-85c8ce835edd | 2025-02-06T10:09:56.868153 |
32vCPU-C7i | cpu | unquantized | pytorch | liuhaotian/llava-v1.5-7b | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:18:49.888665 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
1xA10 | cuda | torchao | pytorch | trl-internal-testing/dummy-GPT2-correct-vocab | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model
self.process_quantization_config()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config
self.quantization_config = AutoQuantizationConfig.from_dict(
File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict
return target_cls.from_dict(quantization_config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict
config = cls(**config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__
self.post_init()
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init
raise ValueError(
ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f49fc1aaa70>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
| 2025-02-06T10:19:04.730530 | fc85a78b-b92a-46d9-b36b-715516040a48 | 2025-02-06T10:12:52.699176 |
32vCPU-C7i | cpu | unquantized | pytorch | liuhaotian/llava-v1.5-7b | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch
raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}")
RuntimeError: Isolated process exited with non-zero code -6
| 2025-02-06T10:19:07.250478 | 9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5 | 2025-02-06T10:09:24.419768 |
1xT4 | cuda | torchao | pytorch | Qwen/Qwen2-0.5B | false | Traceback (most recent call last):
File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark
benchmark_report = Benchmark.launch(benchmark_config)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch
report = launcher.launch(worker=Benchmark.run, worker_args=[config])
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch
raise ChildProcessError(response["traceback"])
ChildProcessError: Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target
report = worker(*worker_args)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run
report = scenario.run(backend)
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run
self.run_model_loading_tracking()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking
self.backend.load()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load
self.load_transformers_model()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model
self.process_quantization_config()
File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config
self.quantization_config = AutoQuantizationConfig.from_dict(
File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict
return target_cls.from_dict(quantization_config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict
config = cls(**config_dict)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__
self.post_init()
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init
raise ValueError(
ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f91153d6b90>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
| 2025-02-06T10:19:09.232979 | bd27598f-6f94-4e3d-9747-85c8ce835edd | 2025-02-06T10:09:56.868153 |
End of preview. Expand
in Dataset Viewer.
README.md exists but content is empty.
- Downloads last month
- 55