machine
stringclasses
4 values
hardware
stringclasses
2 values
subsets
stringclasses
2 values
backends
stringclasses
1 value
model
stringclasses
47 values
success
bool
2 classes
traceback
stringlengths
0
7.23k
last_updated
stringlengths
26
26
run_id
stringclasses
12 values
run_start_time
stringclasses
12 values
debug
cuda
unquantized
pytorch
openai-community/gpt2
true
2025-02-05T16:11:02.440342
2fdb9d56-2aba-4a3b-ac44-63bdc885f925
2025-02-05T16:09:58.183268
1xT4
cuda
unquantized
pytorch
openai-community/gpt2
false
Traceback (most recent call last): File "/llm-perf-backend/llm_perf/common/benchmark_runner.py", line 297, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 149, in load_transformers_model self.load_transformers_model_with_no_weights() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 124, in load_transformers_model_with_no_weights self.pretrained_model = self.automodel_loader.from_pretrained( File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py", line 564, in from_pretrained return model_class.from_pretrained( File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 3826, in from_pretrained config = cls._autoset_attn_implementation( File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 1556, in _autoset_attn_implementation cls._check_and_enable_flash_attn_2( File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 1658, in _check_and_enable_flash_attn_2 raise ImportError(f"{preface} the package flash_attn seems to be not installed. {install_message}") ImportError: FlashAttention2 has been toggled on, but it cannot be used due to the following error: the package flash_attn seems to be not installed. Please refer to the documentation of https://huggingface.co/docs/transformers/perf_infer_gpu_one#flashattention-2 to install Flash Attention 2.
2025-02-05T16:15:46.371592
806946df-514d-4d5f-9b01-0150f2b9ee22
2025-02-05T16:14:52.531351
1xT4
cuda
unquantized
pytorch
openai-community/gpt2
false
Traceback (most recent call last): File "/llm-perf-backend/llm_perf/common/benchmark_runner.py", line 297, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 644, in _locate obj = getattr(obj, part) AttributeError: module 'optimum_benchmark.backends.pytorch' has no attribute 'backend' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1603, in _get_module return importlib.import_module("." + module_name, self.__name__) File "/usr/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "<frozen importlib._bootstrap>", line 1050, in _gcd_import File "<frozen importlib._bootstrap>", line 1027, in _find_and_load File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 688, in _load_unlocked File "<frozen importlib._bootstrap_external>", line 883, in exec_module File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed File "/usr/local/lib/python3.10/dist-packages/transformers/trainer.py", line 215, in <module> from peft import PeftModel File "/usr/local/lib/python3.10/dist-packages/peft/__init__.py", line 22, in <module> from .auto import ( File "/usr/local/lib/python3.10/dist-packages/peft/auto.py", line 32, in <module> from .mapping import MODEL_TYPE_TO_PEFT_MODEL_MAPPING File "/usr/local/lib/python3.10/dist-packages/peft/mapping.py", line 22, in <module> from peft.tuners.xlora.model import XLoraModel File "/usr/local/lib/python3.10/dist-packages/peft/tuners/__init__.py", line 21, in <module> from .lora import LoraConfig, LoraModel, LoftQConfig, LoraRuntimeConfig File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/__init__.py", line 20, in <module> from .model import LoraModel File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/model.py", line 50, in <module> from .awq import dispatch_awq File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/awq.py", line 26, in <module> from awq.modules.linear import WQLinear_GEMM File "/usr/local/lib/python3.10/dist-packages/awq/__init__.py", line 2, in <module> from awq.models.auto import AutoAWQForCausalLM File "/usr/local/lib/python3.10/dist-packages/awq/models/__init__.py", line 1, in <module> from .mpt import MptAWQForCausalLM File "/usr/local/lib/python3.10/dist-packages/awq/models/mpt.py", line 1, in <module> from .base import BaseAWQForCausalLM File "/usr/local/lib/python3.10/dist-packages/awq/models/base.py", line 52, in <module> from awq.quantize.quantizer import AwqQuantizer File "/usr/local/lib/python3.10/dist-packages/awq/quantize/quantizer.py", line 10, in <module> from awq.quantize.scale import apply_scale, apply_clip File "/usr/local/lib/python3.10/dist-packages/awq/quantize/scale.py", line 8, in <module> from transformers.models.llama.modeling_llama import LlamaRMSNorm File "/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py", line 32, in <module> from ...modeling_flash_attention_utils import _flash_attention_forward File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py", line 27, in <module> from flash_attn.bert_padding import index_first_axis, pad_input, unpad_input # noqa File "/usr/local/lib/python3.10/dist-packages/flash_attn/__init__.py", line 3, in <module> from flash_attn.flash_attn_interface import ( File "/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py", line 15, in <module> import flash_attn_2_cuda as flash_attn_gpu ImportError: /usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 650, in _locate obj = import_module(mod) File "/usr/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "<frozen importlib._bootstrap>", line 1050, in _gcd_import File "<frozen importlib._bootstrap>", line 1027, in _find_and_load File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 688, in _load_unlocked File "<frozen importlib._bootstrap_external>", line 883, in exec_module File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 10, in <module> from transformers import ( File "<frozen importlib._bootstrap>", line 1075, in _handle_fromlist File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1593, in __getattr__ module = self._get_module(self._class_to_module[name]) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1605, in _get_module raise RuntimeError( RuntimeError: Failed to import transformers.trainer because of the following error (look up to see its traceback): /usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 69, in run backend_factory: Type[Backend] = get_class(backend_config._target_) File "/usr/local/lib/python3.10/dist-packages/hydra/utils.py", line 40, in get_class raise e File "/usr/local/lib/python3.10/dist-packages/hydra/utils.py", line 31, in get_class cls = _locate(path) File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 658, in _locate raise ImportError( ImportError: Error loading 'optimum_benchmark.backends.pytorch.backend.PyTorchBackend': RuntimeError('Failed to import transformers.trainer because of the following error (look up to see its traceback):\n/usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv')
2025-02-05T16:17:40.048300
03886520-a1ca-4afc-a74c-76bca3bbd085
2025-02-05T16:16:54.711736
1xT4
cuda
unquantized
pytorch
openai-community/gpt2
false
Traceback (most recent call last): File "/llm-perf-backend/llm_perf/common/benchmark_runner.py", line 297, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 644, in _locate obj = getattr(obj, part) AttributeError: module 'optimum_benchmark.backends.pytorch' has no attribute 'backend' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1603, in _get_module return importlib.import_module("." + module_name, self.__name__) File "/usr/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "<frozen importlib._bootstrap>", line 1050, in _gcd_import File "<frozen importlib._bootstrap>", line 1027, in _find_and_load File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 688, in _load_unlocked File "<frozen importlib._bootstrap_external>", line 883, in exec_module File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed File "/usr/local/lib/python3.10/dist-packages/transformers/trainer.py", line 215, in <module> from peft import PeftModel File "/usr/local/lib/python3.10/dist-packages/peft/__init__.py", line 22, in <module> from .auto import ( File "/usr/local/lib/python3.10/dist-packages/peft/auto.py", line 32, in <module> from .mapping import MODEL_TYPE_TO_PEFT_MODEL_MAPPING File "/usr/local/lib/python3.10/dist-packages/peft/mapping.py", line 22, in <module> from peft.tuners.xlora.model import XLoraModel File "/usr/local/lib/python3.10/dist-packages/peft/tuners/__init__.py", line 21, in <module> from .lora import LoraConfig, LoraModel, LoftQConfig, LoraRuntimeConfig File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/__init__.py", line 20, in <module> from .model import LoraModel File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/model.py", line 50, in <module> from .awq import dispatch_awq File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/awq.py", line 26, in <module> from awq.modules.linear import WQLinear_GEMM File "/usr/local/lib/python3.10/dist-packages/awq/__init__.py", line 2, in <module> from awq.models.auto import AutoAWQForCausalLM File "/usr/local/lib/python3.10/dist-packages/awq/models/__init__.py", line 1, in <module> from .mpt import MptAWQForCausalLM File "/usr/local/lib/python3.10/dist-packages/awq/models/mpt.py", line 1, in <module> from .base import BaseAWQForCausalLM File "/usr/local/lib/python3.10/dist-packages/awq/models/base.py", line 52, in <module> from awq.quantize.quantizer import AwqQuantizer File "/usr/local/lib/python3.10/dist-packages/awq/quantize/quantizer.py", line 10, in <module> from awq.quantize.scale import apply_scale, apply_clip File "/usr/local/lib/python3.10/dist-packages/awq/quantize/scale.py", line 8, in <module> from transformers.models.llama.modeling_llama import LlamaRMSNorm File "/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py", line 32, in <module> from ...modeling_flash_attention_utils import _flash_attention_forward File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py", line 27, in <module> from flash_attn.bert_padding import index_first_axis, pad_input, unpad_input # noqa File "/usr/local/lib/python3.10/dist-packages/flash_attn/__init__.py", line 3, in <module> from flash_attn.flash_attn_interface import ( File "/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py", line 15, in <module> import flash_attn_2_cuda as flash_attn_gpu ImportError: /usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 650, in _locate obj = import_module(mod) File "/usr/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "<frozen importlib._bootstrap>", line 1050, in _gcd_import File "<frozen importlib._bootstrap>", line 1027, in _find_and_load File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 688, in _load_unlocked File "<frozen importlib._bootstrap_external>", line 883, in exec_module File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 10, in <module> from transformers import ( File "<frozen importlib._bootstrap>", line 1075, in _handle_fromlist File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1593, in __getattr__ module = self._get_module(self._class_to_module[name]) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1605, in _get_module raise RuntimeError( RuntimeError: Failed to import transformers.trainer because of the following error (look up to see its traceback): /usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 69, in run backend_factory: Type[Backend] = get_class(backend_config._target_) File "/usr/local/lib/python3.10/dist-packages/hydra/utils.py", line 40, in get_class raise e File "/usr/local/lib/python3.10/dist-packages/hydra/utils.py", line 31, in get_class cls = _locate(path) File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 658, in _locate raise ImportError( ImportError: Error loading 'optimum_benchmark.backends.pytorch.backend.PyTorchBackend': RuntimeError('Failed to import transformers.trainer because of the following error (look up to see its traceback):\n/usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv')
2025-02-05T16:17:54.962479
03886520-a1ca-4afc-a74c-76bca3bbd085
2025-02-05T16:16:54.711736
1xT4
cuda
unquantized
pytorch
facebook/opt-125m
false
Traceback (most recent call last): File "/llm-perf-backend/llm_perf/common/benchmark_runner.py", line 297, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 644, in _locate obj = getattr(obj, part) AttributeError: module 'optimum_benchmark.backends.pytorch' has no attribute 'backend' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1603, in _get_module return importlib.import_module("." + module_name, self.__name__) File "/usr/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "<frozen importlib._bootstrap>", line 1050, in _gcd_import File "<frozen importlib._bootstrap>", line 1027, in _find_and_load File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 688, in _load_unlocked File "<frozen importlib._bootstrap_external>", line 883, in exec_module File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed File "/usr/local/lib/python3.10/dist-packages/transformers/trainer.py", line 215, in <module> from peft import PeftModel File "/usr/local/lib/python3.10/dist-packages/peft/__init__.py", line 22, in <module> from .auto import ( File "/usr/local/lib/python3.10/dist-packages/peft/auto.py", line 32, in <module> from .mapping import MODEL_TYPE_TO_PEFT_MODEL_MAPPING File "/usr/local/lib/python3.10/dist-packages/peft/mapping.py", line 22, in <module> from peft.tuners.xlora.model import XLoraModel File "/usr/local/lib/python3.10/dist-packages/peft/tuners/__init__.py", line 21, in <module> from .lora import LoraConfig, LoraModel, LoftQConfig, LoraRuntimeConfig File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/__init__.py", line 20, in <module> from .model import LoraModel File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/model.py", line 50, in <module> from .awq import dispatch_awq File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/awq.py", line 26, in <module> from awq.modules.linear import WQLinear_GEMM File "/usr/local/lib/python3.10/dist-packages/awq/__init__.py", line 2, in <module> from awq.models.auto import AutoAWQForCausalLM File "/usr/local/lib/python3.10/dist-packages/awq/models/__init__.py", line 1, in <module> from .mpt import MptAWQForCausalLM File "/usr/local/lib/python3.10/dist-packages/awq/models/mpt.py", line 1, in <module> from .base import BaseAWQForCausalLM File "/usr/local/lib/python3.10/dist-packages/awq/models/base.py", line 52, in <module> from awq.quantize.quantizer import AwqQuantizer File "/usr/local/lib/python3.10/dist-packages/awq/quantize/quantizer.py", line 10, in <module> from awq.quantize.scale import apply_scale, apply_clip File "/usr/local/lib/python3.10/dist-packages/awq/quantize/scale.py", line 8, in <module> from transformers.models.llama.modeling_llama import LlamaRMSNorm File "/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py", line 32, in <module> from ...modeling_flash_attention_utils import _flash_attention_forward File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py", line 27, in <module> from flash_attn.bert_padding import index_first_axis, pad_input, unpad_input # noqa File "/usr/local/lib/python3.10/dist-packages/flash_attn/__init__.py", line 3, in <module> from flash_attn.flash_attn_interface import ( File "/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py", line 15, in <module> import flash_attn_2_cuda as flash_attn_gpu ImportError: /usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 650, in _locate obj = import_module(mod) File "/usr/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "<frozen importlib._bootstrap>", line 1050, in _gcd_import File "<frozen importlib._bootstrap>", line 1027, in _find_and_load File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 688, in _load_unlocked File "<frozen importlib._bootstrap_external>", line 883, in exec_module File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 10, in <module> from transformers import ( File "<frozen importlib._bootstrap>", line 1075, in _handle_fromlist File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1593, in __getattr__ module = self._get_module(self._class_to_module[name]) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1605, in _get_module raise RuntimeError( RuntimeError: Failed to import transformers.trainer because of the following error (look up to see its traceback): /usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 69, in run backend_factory: Type[Backend] = get_class(backend_config._target_) File "/usr/local/lib/python3.10/dist-packages/hydra/utils.py", line 40, in get_class raise e File "/usr/local/lib/python3.10/dist-packages/hydra/utils.py", line 31, in get_class cls = _locate(path) File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 658, in _locate raise ImportError( ImportError: Error loading 'optimum_benchmark.backends.pytorch.backend.PyTorchBackend': RuntimeError('Failed to import transformers.trainer because of the following error (look up to see its traceback):\n/usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv')
2025-02-05T16:18:45.323625
03886520-a1ca-4afc-a74c-76bca3bbd085
2025-02-05T16:16:54.711736
1xT4
cuda
unquantized
pytorch
openai-community/gpt2
false
Traceback (most recent call last): File "/llm-perf-backend/llm_perf/common/benchmark_runner.py", line 297, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 149, in load_transformers_model self.load_transformers_model_with_no_weights() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 124, in load_transformers_model_with_no_weights self.pretrained_model = self.automodel_loader.from_pretrained( File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py", line 564, in from_pretrained return model_class.from_pretrained( File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 3826, in from_pretrained config = cls._autoset_attn_implementation( File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 1556, in _autoset_attn_implementation cls._check_and_enable_flash_attn_2( File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 1658, in _check_and_enable_flash_attn_2 raise ImportError(f"{preface} the package flash_attn seems to be not installed. {install_message}") ImportError: FlashAttention2 has been toggled on, but it cannot be used due to the following error: the package flash_attn seems to be not installed. Please refer to the documentation of https://huggingface.co/docs/transformers/perf_infer_gpu_one#flashattention-2 to install Flash Attention 2.
2025-02-05T16:27:33.315082
8d5167da-766b-4715-902c-37597ace0904
2025-02-05T16:26:39.182355
1xT4
cuda
unquantized
pytorch
openai-community/gpt2
false
Traceback (most recent call last): File "/llm-perf-backend/llm_perf/common/benchmark_runner.py", line 297, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 149, in load_transformers_model self.load_transformers_model_with_no_weights() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 124, in load_transformers_model_with_no_weights self.pretrained_model = self.automodel_loader.from_pretrained( File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py", line 564, in from_pretrained return model_class.from_pretrained( File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 3826, in from_pretrained config = cls._autoset_attn_implementation( File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 1556, in _autoset_attn_implementation cls._check_and_enable_flash_attn_2( File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 1658, in _check_and_enable_flash_attn_2 raise ImportError(f"{preface} the package flash_attn seems to be not installed. {install_message}") ImportError: FlashAttention2 has been toggled on, but it cannot be used due to the following error: the package flash_attn seems to be not installed. Please refer to the documentation of https://huggingface.co/docs/transformers/perf_infer_gpu_one#flashattention-2 to install Flash Attention 2.
2025-02-05T16:27:57.180163
8d5167da-766b-4715-902c-37597ace0904
2025-02-05T16:26:39.182355
1xT4
cuda
unquantized
pytorch
facebook/opt-125m
false
Traceback (most recent call last): File "/llm-perf-backend/llm_perf/common/benchmark_runner.py", line 297, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 149, in load_transformers_model self.load_transformers_model_with_no_weights() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 124, in load_transformers_model_with_no_weights self.pretrained_model = self.automodel_loader.from_pretrained( File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py", line 564, in from_pretrained return model_class.from_pretrained( File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 3826, in from_pretrained config = cls._autoset_attn_implementation( File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 1556, in _autoset_attn_implementation cls._check_and_enable_flash_attn_2( File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 1658, in _check_and_enable_flash_attn_2 raise ImportError(f"{preface} the package flash_attn seems to be not installed. {install_message}") ImportError: FlashAttention2 has been toggled on, but it cannot be used due to the following error: the package flash_attn seems to be not installed. Please refer to the documentation of https://huggingface.co/docs/transformers/perf_infer_gpu_one#flashattention-2 to install Flash Attention 2.
2025-02-05T16:28:54.977003
8d5167da-766b-4715-902c-37597ace0904
2025-02-05T16:26:39.182355
1xT4
cuda
unquantized
pytorch
facebook/opt-125m
false
Traceback (most recent call last): File "/llm-perf-backend/llm_perf/common/benchmark_runner.py", line 297, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 149, in load_transformers_model self.load_transformers_model_with_no_weights() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 124, in load_transformers_model_with_no_weights self.pretrained_model = self.automodel_loader.from_pretrained( File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py", line 564, in from_pretrained return model_class.from_pretrained( File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 3826, in from_pretrained config = cls._autoset_attn_implementation( File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 1556, in _autoset_attn_implementation cls._check_and_enable_flash_attn_2( File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 1658, in _check_and_enable_flash_attn_2 raise ImportError(f"{preface} the package flash_attn seems to be not installed. {install_message}") ImportError: FlashAttention2 has been toggled on, but it cannot be used due to the following error: the package flash_attn seems to be not installed. Please refer to the documentation of https://huggingface.co/docs/transformers/perf_infer_gpu_one#flashattention-2 to install Flash Attention 2.
2025-02-05T16:29:17.544956
8d5167da-766b-4715-902c-37597ace0904
2025-02-05T16:26:39.182355
1xT4
cuda
unquantized
pytorch
openai-community/gpt2
false
Traceback (most recent call last): File "/llm-perf-backend/llm_perf/common/benchmark_runner.py", line 297, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 644, in _locate obj = getattr(obj, part) AttributeError: module 'optimum_benchmark.backends.pytorch' has no attribute 'backend' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1603, in _get_module return importlib.import_module("." + module_name, self.__name__) File "/usr/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "<frozen importlib._bootstrap>", line 1050, in _gcd_import File "<frozen importlib._bootstrap>", line 1027, in _find_and_load File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 688, in _load_unlocked File "<frozen importlib._bootstrap_external>", line 883, in exec_module File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed File "/usr/local/lib/python3.10/dist-packages/transformers/trainer.py", line 215, in <module> from peft import PeftModel File "/usr/local/lib/python3.10/dist-packages/peft/__init__.py", line 22, in <module> from .auto import ( File "/usr/local/lib/python3.10/dist-packages/peft/auto.py", line 32, in <module> from .mapping import MODEL_TYPE_TO_PEFT_MODEL_MAPPING File "/usr/local/lib/python3.10/dist-packages/peft/mapping.py", line 22, in <module> from peft.tuners.xlora.model import XLoraModel File "/usr/local/lib/python3.10/dist-packages/peft/tuners/__init__.py", line 21, in <module> from .lora import LoraConfig, LoraModel, LoftQConfig, LoraRuntimeConfig File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/__init__.py", line 20, in <module> from .model import LoraModel File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/model.py", line 50, in <module> from .awq import dispatch_awq File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/awq.py", line 26, in <module> from awq.modules.linear import WQLinear_GEMM File "/usr/local/lib/python3.10/dist-packages/awq/__init__.py", line 2, in <module> from awq.models.auto import AutoAWQForCausalLM File "/usr/local/lib/python3.10/dist-packages/awq/models/__init__.py", line 1, in <module> from .mpt import MptAWQForCausalLM File "/usr/local/lib/python3.10/dist-packages/awq/models/mpt.py", line 1, in <module> from .base import BaseAWQForCausalLM File "/usr/local/lib/python3.10/dist-packages/awq/models/base.py", line 52, in <module> from awq.quantize.quantizer import AwqQuantizer File "/usr/local/lib/python3.10/dist-packages/awq/quantize/quantizer.py", line 10, in <module> from awq.quantize.scale import apply_scale, apply_clip File "/usr/local/lib/python3.10/dist-packages/awq/quantize/scale.py", line 8, in <module> from transformers.models.llama.modeling_llama import LlamaRMSNorm File "/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py", line 32, in <module> from ...modeling_flash_attention_utils import _flash_attention_forward File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py", line 27, in <module> from flash_attn.bert_padding import index_first_axis, pad_input, unpad_input # noqa File "/usr/local/lib/python3.10/dist-packages/flash_attn/__init__.py", line 3, in <module> from flash_attn.flash_attn_interface import ( File "/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py", line 15, in <module> import flash_attn_2_cuda as flash_attn_gpu ImportError: /usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 650, in _locate obj = import_module(mod) File "/usr/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "<frozen importlib._bootstrap>", line 1050, in _gcd_import File "<frozen importlib._bootstrap>", line 1027, in _find_and_load File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 688, in _load_unlocked File "<frozen importlib._bootstrap_external>", line 883, in exec_module File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 10, in <module> from transformers import ( File "<frozen importlib._bootstrap>", line 1075, in _handle_fromlist File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1593, in __getattr__ module = self._get_module(self._class_to_module[name]) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1605, in _get_module raise RuntimeError( RuntimeError: Failed to import transformers.trainer because of the following error (look up to see its traceback): /usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 69, in run backend_factory: Type[Backend] = get_class(backend_config._target_) File "/usr/local/lib/python3.10/dist-packages/hydra/utils.py", line 40, in get_class raise e File "/usr/local/lib/python3.10/dist-packages/hydra/utils.py", line 31, in get_class cls = _locate(path) File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 658, in _locate raise ImportError( ImportError: Error loading 'optimum_benchmark.backends.pytorch.backend.PyTorchBackend': RuntimeError('Failed to import transformers.trainer because of the following error (look up to see its traceback):\n/usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv')
2025-02-05T16:31:42.373752
4b0db4ab-2fe1-4e13-a7c6-93344a8b08a2
2025-02-05T16:30:57.040646
1xT4
cuda
unquantized
pytorch
openai-community/gpt2
false
Traceback (most recent call last): File "/llm-perf-backend/llm_perf/common/benchmark_runner.py", line 297, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 644, in _locate obj = getattr(obj, part) AttributeError: module 'optimum_benchmark.backends.pytorch' has no attribute 'backend' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1603, in _get_module return importlib.import_module("." + module_name, self.__name__) File "/usr/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "<frozen importlib._bootstrap>", line 1050, in _gcd_import File "<frozen importlib._bootstrap>", line 1027, in _find_and_load File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 688, in _load_unlocked File "<frozen importlib._bootstrap_external>", line 883, in exec_module File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed File "/usr/local/lib/python3.10/dist-packages/transformers/trainer.py", line 215, in <module> from peft import PeftModel File "/usr/local/lib/python3.10/dist-packages/peft/__init__.py", line 22, in <module> from .auto import ( File "/usr/local/lib/python3.10/dist-packages/peft/auto.py", line 32, in <module> from .mapping import MODEL_TYPE_TO_PEFT_MODEL_MAPPING File "/usr/local/lib/python3.10/dist-packages/peft/mapping.py", line 22, in <module> from peft.tuners.xlora.model import XLoraModel File "/usr/local/lib/python3.10/dist-packages/peft/tuners/__init__.py", line 21, in <module> from .lora import LoraConfig, LoraModel, LoftQConfig, LoraRuntimeConfig File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/__init__.py", line 20, in <module> from .model import LoraModel File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/model.py", line 50, in <module> from .awq import dispatch_awq File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/awq.py", line 26, in <module> from awq.modules.linear import WQLinear_GEMM File "/usr/local/lib/python3.10/dist-packages/awq/__init__.py", line 2, in <module> from awq.models.auto import AutoAWQForCausalLM File "/usr/local/lib/python3.10/dist-packages/awq/models/__init__.py", line 1, in <module> from .mpt import MptAWQForCausalLM File "/usr/local/lib/python3.10/dist-packages/awq/models/mpt.py", line 1, in <module> from .base import BaseAWQForCausalLM File "/usr/local/lib/python3.10/dist-packages/awq/models/base.py", line 52, in <module> from awq.quantize.quantizer import AwqQuantizer File "/usr/local/lib/python3.10/dist-packages/awq/quantize/quantizer.py", line 10, in <module> from awq.quantize.scale import apply_scale, apply_clip File "/usr/local/lib/python3.10/dist-packages/awq/quantize/scale.py", line 8, in <module> from transformers.models.llama.modeling_llama import LlamaRMSNorm File "/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py", line 32, in <module> from ...modeling_flash_attention_utils import _flash_attention_forward File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py", line 27, in <module> from flash_attn.bert_padding import index_first_axis, pad_input, unpad_input # noqa File "/usr/local/lib/python3.10/dist-packages/flash_attn/__init__.py", line 3, in <module> from flash_attn.flash_attn_interface import ( File "/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py", line 15, in <module> import flash_attn_2_cuda as flash_attn_gpu ImportError: /usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 650, in _locate obj = import_module(mod) File "/usr/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "<frozen importlib._bootstrap>", line 1050, in _gcd_import File "<frozen importlib._bootstrap>", line 1027, in _find_and_load File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 688, in _load_unlocked File "<frozen importlib._bootstrap_external>", line 883, in exec_module File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 10, in <module> from transformers import ( File "<frozen importlib._bootstrap>", line 1075, in _handle_fromlist File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1593, in __getattr__ module = self._get_module(self._class_to_module[name]) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1605, in _get_module raise RuntimeError( RuntimeError: Failed to import transformers.trainer because of the following error (look up to see its traceback): /usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 69, in run backend_factory: Type[Backend] = get_class(backend_config._target_) File "/usr/local/lib/python3.10/dist-packages/hydra/utils.py", line 40, in get_class raise e File "/usr/local/lib/python3.10/dist-packages/hydra/utils.py", line 31, in get_class cls = _locate(path) File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 658, in _locate raise ImportError( ImportError: Error loading 'optimum_benchmark.backends.pytorch.backend.PyTorchBackend': RuntimeError('Failed to import transformers.trainer because of the following error (look up to see its traceback):\n/usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv')
2025-02-06T08:45:39.449306
73a95cce-e60e-4115-8874-4aaa173b2aa3
2025-02-06T08:44:53.179701
1xT4
cuda
unquantized
pytorch
openai-community/gpt2
false
Traceback (most recent call last): File "/llm-perf-backend/llm_perf/common/benchmark_runner.py", line 297, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 644, in _locate obj = getattr(obj, part) AttributeError: module 'optimum_benchmark.backends.pytorch' has no attribute 'backend' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1603, in _get_module return importlib.import_module("." + module_name, self.__name__) File "/usr/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "<frozen importlib._bootstrap>", line 1050, in _gcd_import File "<frozen importlib._bootstrap>", line 1027, in _find_and_load File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 688, in _load_unlocked File "<frozen importlib._bootstrap_external>", line 883, in exec_module File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed File "/usr/local/lib/python3.10/dist-packages/transformers/trainer.py", line 215, in <module> from peft import PeftModel File "/usr/local/lib/python3.10/dist-packages/peft/__init__.py", line 22, in <module> from .auto import ( File "/usr/local/lib/python3.10/dist-packages/peft/auto.py", line 32, in <module> from .mapping import MODEL_TYPE_TO_PEFT_MODEL_MAPPING File "/usr/local/lib/python3.10/dist-packages/peft/mapping.py", line 22, in <module> from peft.tuners.xlora.model import XLoraModel File "/usr/local/lib/python3.10/dist-packages/peft/tuners/__init__.py", line 21, in <module> from .lora import LoraConfig, LoraModel, LoftQConfig, LoraRuntimeConfig File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/__init__.py", line 20, in <module> from .model import LoraModel File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/model.py", line 50, in <module> from .awq import dispatch_awq File "/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/awq.py", line 26, in <module> from awq.modules.linear import WQLinear_GEMM File "/usr/local/lib/python3.10/dist-packages/awq/__init__.py", line 2, in <module> from awq.models.auto import AutoAWQForCausalLM File "/usr/local/lib/python3.10/dist-packages/awq/models/__init__.py", line 1, in <module> from .mpt import MptAWQForCausalLM File "/usr/local/lib/python3.10/dist-packages/awq/models/mpt.py", line 1, in <module> from .base import BaseAWQForCausalLM File "/usr/local/lib/python3.10/dist-packages/awq/models/base.py", line 52, in <module> from awq.quantize.quantizer import AwqQuantizer File "/usr/local/lib/python3.10/dist-packages/awq/quantize/quantizer.py", line 10, in <module> from awq.quantize.scale import apply_scale, apply_clip File "/usr/local/lib/python3.10/dist-packages/awq/quantize/scale.py", line 8, in <module> from transformers.models.llama.modeling_llama import LlamaRMSNorm File "/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py", line 32, in <module> from ...modeling_flash_attention_utils import _flash_attention_forward File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py", line 27, in <module> from flash_attn.bert_padding import index_first_axis, pad_input, unpad_input # noqa File "/usr/local/lib/python3.10/dist-packages/flash_attn/__init__.py", line 3, in <module> from flash_attn.flash_attn_interface import ( File "/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py", line 15, in <module> import flash_attn_2_cuda as flash_attn_gpu ImportError: /usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 650, in _locate obj = import_module(mod) File "/usr/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "<frozen importlib._bootstrap>", line 1050, in _gcd_import File "<frozen importlib._bootstrap>", line 1027, in _find_and_load File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 688, in _load_unlocked File "<frozen importlib._bootstrap_external>", line 883, in exec_module File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 10, in <module> from transformers import ( File "<frozen importlib._bootstrap>", line 1075, in _handle_fromlist File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1593, in __getattr__ module = self._get_module(self._class_to_module[name]) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1605, in _get_module raise RuntimeError( RuntimeError: Failed to import transformers.trainer because of the following error (look up to see its traceback): /usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 69, in run backend_factory: Type[Backend] = get_class(backend_config._target_) File "/usr/local/lib/python3.10/dist-packages/hydra/utils.py", line 40, in get_class raise e File "/usr/local/lib/python3.10/dist-packages/hydra/utils.py", line 31, in get_class cls = _locate(path) File "/usr/local/lib/python3.10/dist-packages/hydra/_internal/utils.py", line 658, in _locate raise ImportError( ImportError: Error loading 'optimum_benchmark.backends.pytorch.backend.PyTorchBackend': RuntimeError('Failed to import transformers.trainer because of the following error (look up to see its traceback):\n/usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv')
2025-02-06T08:45:54.648082
73a95cce-e60e-4115-8874-4aaa173b2aa3
2025-02-06T08:44:53.179701
1xT4
cuda
unquantized
pytorch
bigscience/bloomz-560m
false
Traceback (most recent call last): File "/llm-perf-backend/llm_perf/common/benchmark_runner.py", line 297, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 149, in load_transformers_model self.load_transformers_model_with_no_weights() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 124, in load_transformers_model_with_no_weights self.pretrained_model = self.automodel_loader.from_pretrained( File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py", line 564, in from_pretrained return model_class.from_pretrained( File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 3826, in from_pretrained config = cls._autoset_attn_implementation( File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 1565, in _autoset_attn_implementation config = cls._check_and_enable_sdpa( File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 1731, in _check_and_enable_sdpa raise ValueError( ValueError: BloomForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation="eager"` meanwhile. Example: `model = AutoModel.from_pretrained("openai/whisper-tiny", attn_implementation="eager")`
2025-02-06T09:18:17.866903
1c99483e-b56b-4326-aee9-66f38518c30e
2025-02-06T09:16:35.806219
1xT4
cuda
unquantized
pytorch
openai-community/gpt2
true
2025-02-06T09:20:44.677569
0e3562c9-963a-49c9-be25-cfa94c03003a
2025-02-06T09:19:41.190913
1xT4
cuda
unquantized
pytorch
openai-community/gpt2
true
2025-02-06T09:21:49.942995
0e3562c9-963a-49c9-be25-cfa94c03003a
2025-02-06T09:19:41.190913
1xT4
cuda
unquantized
pytorch
openai-community/gpt2
true
2025-02-06T09:22:56.522260
0e3562c9-963a-49c9-be25-cfa94c03003a
2025-02-06T09:19:41.190913
1xT4
cuda
unquantized
pytorch
openai-community/gpt2
true
2025-02-06T09:24:03.915016
0e3562c9-963a-49c9-be25-cfa94c03003a
2025-02-06T09:19:41.190913
32vCPU-C7i
cpu
unquantized
pytorch
openai-community/gpt2
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T09:56:50.675962
938a1a37-c311-4be1-85f4-dd27b03cce3c
2025-02-06T09:56:41.233415
32vCPU-C7i
cpu
unquantized
pytorch
openai-community/gpt2
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T09:57:01.606291
938a1a37-c311-4be1-85f4-dd27b03cce3c
2025-02-06T09:56:41.233415
32vCPU-C7i
cpu
unquantized
pytorch
openai-community/gpt2
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T09:57:13.683017
938a1a37-c311-4be1-85f4-dd27b03cce3c
2025-02-06T09:56:41.233415
32vCPU-C7i
cpu
unquantized
pytorch
openai-community/gpt2
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T09:57:24.310483
938a1a37-c311-4be1-85f4-dd27b03cce3c
2025-02-06T09:56:41.233415
32vCPU-C7i
cpu
unquantized
pytorch
openai-community/gpt2
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T09:57:36.126144
938a1a37-c311-4be1-85f4-dd27b03cce3c
2025-02-06T09:56:41.233415
32vCPU-C7i
cpu
unquantized
pytorch
openai-community/gpt2
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T09:57:46.794500
938a1a37-c311-4be1-85f4-dd27b03cce3c
2025-02-06T09:56:41.233415
32vCPU-C7i
cpu
unquantized
pytorch
facebook/opt-125m
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T09:57:57.638603
938a1a37-c311-4be1-85f4-dd27b03cce3c
2025-02-06T09:56:41.233415
32vCPU-C7i
cpu
unquantized
pytorch
facebook/opt-125m
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T09:58:11.213695
938a1a37-c311-4be1-85f4-dd27b03cce3c
2025-02-06T09:56:41.233415
32vCPU-C7i
cpu
unquantized
pytorch
facebook/opt-125m
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T09:58:24.607254
938a1a37-c311-4be1-85f4-dd27b03cce3c
2025-02-06T09:56:41.233415
32vCPU-C7i
cpu
unquantized
pytorch
facebook/opt-125m
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T09:58:39.142446
938a1a37-c311-4be1-85f4-dd27b03cce3c
2025-02-06T09:56:41.233415
32vCPU-C7i
cpu
unquantized
pytorch
facebook/opt-125m
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T09:58:54.351007
938a1a37-c311-4be1-85f4-dd27b03cce3c
2025-02-06T09:56:41.233415
32vCPU-C7i
cpu
unquantized
pytorch
facebook/opt-125m
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T09:59:11.381282
938a1a37-c311-4be1-85f4-dd27b03cce3c
2025-02-06T09:56:41.233415
32vCPU-C7i
cpu
unquantized
pytorch
openai-community/gpt2
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:09:31.650354
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
32vCPU-C7i
cpu
unquantized
pytorch
openai-community/gpt2
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:09:41.822548
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
32vCPU-C7i
cpu
unquantized
pytorch
openai-community/gpt2
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:09:50.955328
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
32vCPU-C7i
cpu
unquantized
pytorch
openai-community/gpt2
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:10:02.091578
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
32vCPU-C7i
cpu
unquantized
pytorch
openai-community/gpt2
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:10:13.644524
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
1xT4
cuda
torchao
pytorch
bigscience/bloomz-560m
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model self.process_quantization_config() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config self.quantization_config = AutoQuantizationConfig.from_dict( File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict return target_cls.from_dict(quantization_config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict config = cls(**config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__ self.post_init() File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init raise ValueError( ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7fab04143250>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
2025-02-06T10:10:26.426187
bd27598f-6f94-4e3d-9747-85c8ce835edd
2025-02-06T10:09:56.868153
32vCPU-C7i
cpu
unquantized
pytorch
openai-community/gpt2
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:10:27.799185
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
32vCPU-C7i
cpu
unquantized
pytorch
facebook/opt-125m
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:10:44.632379
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
1xT4
cuda
torchao
pytorch
meta-llama/Llama-3.1-8B-Instruct
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model self.process_quantization_config() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config self.quantization_config = AutoQuantizationConfig.from_dict( File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict return target_cls.from_dict(quantization_config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict config = cls(**config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__ self.post_init() File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init raise ValueError( ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f885846fd90>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
2025-02-06T10:10:54.800330
bd27598f-6f94-4e3d-9747-85c8ce835edd
2025-02-06T10:09:56.868153
32vCPU-C7i
cpu
unquantized
pytorch
facebook/opt-125m
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:11:03.810607
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
32vCPU-C7i
cpu
unquantized
pytorch
facebook/opt-125m
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:11:20.997163
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
32vCPU-C7i
cpu
unquantized
pytorch
facebook/opt-125m
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:11:40.876875
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
1xT4
cuda
torchao
pytorch
distilbert/distilgpt2
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model self.process_quantization_config() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config self.quantization_config = AutoQuantizationConfig.from_dict( File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict return target_cls.from_dict(quantization_config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict config = cls(**config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__ self.post_init() File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init raise ValueError( ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7fb8da1fa710>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
2025-02-06T10:11:48.635299
bd27598f-6f94-4e3d-9747-85c8ce835edd
2025-02-06T10:09:56.868153
32vCPU-C7i
cpu
unquantized
pytorch
facebook/opt-125m
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:12:00.639319
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
1xT4
cuda
torchao
pytorch
distilbert/distilgpt2
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model self.process_quantization_config() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config self.quantization_config = AutoQuantizationConfig.from_dict( File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict return target_cls.from_dict(quantization_config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict config = cls(**config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__ self.post_init() File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init raise ValueError( ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f4c647c4ca0>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
2025-02-06T10:12:14.822852
bd27598f-6f94-4e3d-9747-85c8ce835edd
2025-02-06T10:09:56.868153
32vCPU-C7i
cpu
unquantized
pytorch
facebook/opt-125m
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:12:18.359729
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
32vCPU-C7i
cpu
unquantized
pytorch
bigscience/bloomz-560m
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:12:38.778996
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
1xT4
cuda
torchao
pytorch
meta-llama/Llama-3.1-405B
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model self.process_quantization_config() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config self.quantization_config = AutoQuantizationConfig.from_dict( File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict return target_cls.from_dict(quantization_config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict config = cls(**config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__ self.post_init() File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init raise ValueError( ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f239014fb50>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
2025-02-06T10:12:44.565974
bd27598f-6f94-4e3d-9747-85c8ce835edd
2025-02-06T10:09:56.868153
32vCPU-C7i
cpu
unquantized
pytorch
bigscience/bloomz-560m
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:12:58.240254
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
1xT4
cuda
torchao
pytorch
meta-llama/Llama-3.1-405B
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model self.process_quantization_config() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config self.quantization_config = AutoQuantizationConfig.from_dict( File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict return target_cls.from_dict(quantization_config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict config = cls(**config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__ self.post_init() File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init raise ValueError( ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f4838b1e050>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
2025-02-06T10:13:11.834624
bd27598f-6f94-4e3d-9747-85c8ce835edd
2025-02-06T10:09:56.868153
32vCPU-C7i
cpu
unquantized
pytorch
bigscience/bloomz-560m
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:13:17.293850
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
1xA10
cuda
torchao
pytorch
bigscience/bloomz-560m
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model self.process_quantization_config() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config self.quantization_config = AutoQuantizationConfig.from_dict( File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict return target_cls.from_dict(quantization_config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict config = cls(**config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__ self.post_init() File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init raise ValueError( ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f5ffc2832e0>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
2025-02-06T10:13:19.241876
fc85a78b-b92a-46d9-b36b-715516040a48
2025-02-06T10:12:52.699176
32vCPU-C7i
cpu
unquantized
pytorch
bigscience/bloomz-560m
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:13:36.395832
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
1xT4
cuda
torchao
pytorch
meta-llama/Meta-Llama-3-8B-Instruct
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model self.process_quantization_config() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config self.quantization_config = AutoQuantizationConfig.from_dict( File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict return target_cls.from_dict(quantization_config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict config = cls(**config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__ self.post_init() File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init raise ValueError( ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f096e25bd00>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
2025-02-06T10:13:40.308349
bd27598f-6f94-4e3d-9747-85c8ce835edd
2025-02-06T10:09:56.868153
1xA10
cuda
torchao
pytorch
distilbert/distilgpt2
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model self.process_quantization_config() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config self.quantization_config = AutoQuantizationConfig.from_dict( File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict return target_cls.from_dict(quantization_config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict config = cls(**config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__ self.post_init() File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init raise ValueError( ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f45002ca710>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
2025-02-06T10:13:45.720524
fc85a78b-b92a-46d9-b36b-715516040a48
2025-02-06T10:12:52.699176
32vCPU-C7i
cpu
unquantized
pytorch
bigscience/bloomz-560m
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:13:56.268390
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
1xT4
cuda
torchao
pytorch
meta-llama/Meta-Llama-3-8B-Instruct
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model self.process_quantization_config() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config self.quantization_config = AutoQuantizationConfig.from_dict( File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict return target_cls.from_dict(quantization_config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict config = cls(**config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__ self.post_init() File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init raise ValueError( ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7efff024a050>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
2025-02-06T10:14:07.388246
bd27598f-6f94-4e3d-9747-85c8ce835edd
2025-02-06T10:09:56.868153
1xA10
cuda
torchao
pytorch
distilbert/distilgpt2
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model self.process_quantization_config() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config self.quantization_config = AutoQuantizationConfig.from_dict( File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict return target_cls.from_dict(quantization_config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict config = cls(**config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__ self.post_init() File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init raise ValueError( ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f4b05ab4ca0>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
2025-02-06T10:14:10.815562
fc85a78b-b92a-46d9-b36b-715516040a48
2025-02-06T10:12:52.699176
32vCPU-C7i
cpu
unquantized
pytorch
bigscience/bloomz-560m
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:14:14.324658
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
32vCPU-C7i
cpu
unquantized
pytorch
meta-llama/Llama-3.1-8B-Instruct
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:14:32.663277
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
1xT4
cuda
torchao
pytorch
microsoft/Phi-3-mini-4k-instruct
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model self.process_quantization_config() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config self.quantization_config = AutoQuantizationConfig.from_dict( File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict return target_cls.from_dict(quantization_config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict config = cls(**config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__ self.post_init() File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init raise ValueError( ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f2335b3c280>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
2025-02-06T10:14:35.179953
bd27598f-6f94-4e3d-9747-85c8ce835edd
2025-02-06T10:09:56.868153
1xA10
cuda
torchao
pytorch
meta-llama/Llama-3.1-405B
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model self.process_quantization_config() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config self.quantization_config = AutoQuantizationConfig.from_dict( File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict return target_cls.from_dict(quantization_config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict config = cls(**config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__ self.post_init() File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init raise ValueError( ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f352c733b50>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
2025-02-06T10:14:36.775187
fc85a78b-b92a-46d9-b36b-715516040a48
2025-02-06T10:12:52.699176
32vCPU-C7i
cpu
unquantized
pytorch
meta-llama/Llama-3.1-8B-Instruct
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:14:52.253129
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
1xT4
cuda
torchao
pytorch
microsoft/Phi-3-mini-4k-instruct
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model self.process_quantization_config() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config self.quantization_config = AutoQuantizationConfig.from_dict( File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict return target_cls.from_dict(quantization_config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict config = cls(**config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__ self.post_init() File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init raise ValueError( ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f2e74226b00>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
2025-02-06T10:15:02.590667
bd27598f-6f94-4e3d-9747-85c8ce835edd
2025-02-06T10:09:56.868153
1xA10
cuda
torchao
pytorch
meta-llama/Llama-3.1-405B
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model self.process_quantization_config() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config self.quantization_config = AutoQuantizationConfig.from_dict( File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict return target_cls.from_dict(quantization_config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict config = cls(**config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__ self.post_init() File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init raise ValueError( ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f7bd4552050>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
2025-02-06T10:15:04.568477
fc85a78b-b92a-46d9-b36b-715516040a48
2025-02-06T10:12:52.699176
32vCPU-C7i
cpu
unquantized
pytorch
meta-llama/Llama-3.1-8B-Instruct
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:15:10.538588
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
32vCPU-C7i
cpu
unquantized
pytorch
meta-llama/Llama-3.1-8B-Instruct
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:15:30.458954
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
1xA10
cuda
torchao
pytorch
meta-llama/Meta-Llama-3-8B-Instruct
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model self.process_quantization_config() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config self.quantization_config = AutoQuantizationConfig.from_dict( File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict return target_cls.from_dict(quantization_config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict config = cls(**config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__ self.post_init() File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init raise ValueError( ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f38f025fd00>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
2025-02-06T10:15:34.638005
fc85a78b-b92a-46d9-b36b-715516040a48
2025-02-06T10:12:52.699176
32vCPU-C7i
cpu
unquantized
pytorch
meta-llama/Llama-3.1-8B-Instruct
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:15:46.892472
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
1xT4
cuda
torchao
pytorch
meta-llama/Meta-Llama-3-8B
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model self.process_quantization_config() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config self.quantization_config = AutoQuantizationConfig.from_dict( File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict return target_cls.from_dict(quantization_config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict config = cls(**config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__ self.post_init() File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init raise ValueError( ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f51b4142050>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
2025-02-06T10:15:57.063359
bd27598f-6f94-4e3d-9747-85c8ce835edd
2025-02-06T10:09:56.868153
1xA10
cuda
torchao
pytorch
meta-llama/Meta-Llama-3-8B-Instruct
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model self.process_quantization_config() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config self.quantization_config = AutoQuantizationConfig.from_dict( File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict return target_cls.from_dict(quantization_config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict config = cls(**config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__ self.post_init() File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init raise ValueError( ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f5158816050>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
2025-02-06T10:16:02.322652
fc85a78b-b92a-46d9-b36b-715516040a48
2025-02-06T10:12:52.699176
32vCPU-C7i
cpu
unquantized
pytorch
meta-llama/Llama-3.1-405B
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:16:04.758642
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
32vCPU-C7i
cpu
unquantized
pytorch
meta-llama/Llama-3.1-405B
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:16:21.725048
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
1xT4
cuda
torchao
pytorch
liuhaotian/llava-v1.5-7b
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model self.process_quantization_config() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config self.quantization_config = AutoQuantizationConfig.from_dict( File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict return target_cls.from_dict(quantization_config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict config = cls(**config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__ self.post_init() File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init raise ValueError( ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f6190389fc0>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
2025-02-06T10:16:23.619766
bd27598f-6f94-4e3d-9747-85c8ce835edd
2025-02-06T10:09:56.868153
1xA10
cuda
torchao
pytorch
microsoft/Phi-3-mini-4k-instruct
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model self.process_quantization_config() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config self.quantization_config = AutoQuantizationConfig.from_dict( File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict return target_cls.from_dict(quantization_config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict config = cls(**config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__ self.post_init() File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init raise ValueError( ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7fedf8184280>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
2025-02-06T10:16:29.309172
fc85a78b-b92a-46d9-b36b-715516040a48
2025-02-06T10:12:52.699176
32vCPU-C7i
cpu
unquantized
pytorch
meta-llama/Llama-3.1-405B
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:16:39.801791
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
1xT4
cuda
torchao
pytorch
liuhaotian/llava-v1.5-7b
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model self.process_quantization_config() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config self.quantization_config = AutoQuantizationConfig.from_dict( File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict return target_cls.from_dict(quantization_config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict config = cls(**config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__ self.post_init() File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init raise ValueError( ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7fc4004204c0>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
2025-02-06T10:16:50.135365
bd27598f-6f94-4e3d-9747-85c8ce835edd
2025-02-06T10:09:56.868153
1xA10
cuda
torchao
pytorch
microsoft/Phi-3-mini-4k-instruct
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model self.process_quantization_config() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config self.quantization_config = AutoQuantizationConfig.from_dict( File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict return target_cls.from_dict(quantization_config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict config = cls(**config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__ self.post_init() File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init raise ValueError( ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f047c42eb00>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
2025-02-06T10:16:54.690294
fc85a78b-b92a-46d9-b36b-715516040a48
2025-02-06T10:12:52.699176
32vCPU-C7i
cpu
unquantized
pytorch
meta-llama/Llama-3.1-405B
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:16:58.191029
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
32vCPU-C7i
cpu
unquantized
pytorch
meta-llama/Llama-3.1-405B
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:17:16.353456
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
1xA10
cuda
torchao
pytorch
meta-llama/Meta-Llama-3-8B
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model self.process_quantization_config() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config self.quantization_config = AutoQuantizationConfig.from_dict( File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict return target_cls.from_dict(quantization_config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict config = cls(**config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__ self.post_init() File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init raise ValueError( ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7fe6a00a3be0>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
2025-02-06T10:17:20.742154
fc85a78b-b92a-46d9-b36b-715516040a48
2025-02-06T10:12:52.699176
32vCPU-C7i
cpu
unquantized
pytorch
meta-llama/Llama-3.1-405B
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:17:32.869921
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
1xT4
cuda
torchao
pytorch
trl-internal-testing/dummy-GPT2-correct-vocab
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model self.process_quantization_config() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config self.quantization_config = AutoQuantizationConfig.from_dict( File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict return target_cls.from_dict(quantization_config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict config = cls(**config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__ self.post_init() File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init raise ValueError( ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f20ccbd0e50>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
2025-02-06T10:17:47.243376
bd27598f-6f94-4e3d-9747-85c8ce835edd
2025-02-06T10:09:56.868153
32vCPU-C7i
cpu
unquantized
pytorch
microsoft/Phi-3-mini-4k-instruct
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:17:52.456368
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
1xA10
cuda
torchao
pytorch
liuhaotian/llava-v1.5-7b
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model self.process_quantization_config() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config self.quantization_config = AutoQuantizationConfig.from_dict( File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict return target_cls.from_dict(quantization_config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict config = cls(**config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__ self.post_init() File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init raise ValueError( ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7ff7e0a25fc0>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
2025-02-06T10:18:12.501138
fc85a78b-b92a-46d9-b36b-715516040a48
2025-02-06T10:12:52.699176
1xT4
cuda
torchao
pytorch
datificate/gpt2-small-spanish
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model self.process_quantization_config() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config self.quantization_config = AutoQuantizationConfig.from_dict( File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict return target_cls.from_dict(quantization_config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict config = cls(**config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__ self.post_init() File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init raise ValueError( ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7fbaa352a9e0>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
2025-02-06T10:18:14.516366
bd27598f-6f94-4e3d-9747-85c8ce835edd
2025-02-06T10:09:56.868153
32vCPU-C7i
cpu
unquantized
pytorch
microsoft/Phi-3-mini-4k-instruct
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:18:30.576978
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
1xA10
cuda
torchao
pytorch
liuhaotian/llava-v1.5-7b
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model self.process_quantization_config() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config self.quantization_config = AutoQuantizationConfig.from_dict( File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict return target_cls.from_dict(quantization_config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict config = cls(**config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__ self.post_init() File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init raise ValueError( ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7fd2b035c4c0>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
2025-02-06T10:18:38.645285
fc85a78b-b92a-46d9-b36b-715516040a48
2025-02-06T10:12:52.699176
1xT4
cuda
torchao
pytorch
datificate/gpt2-small-spanish
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model self.process_quantization_config() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config self.quantization_config = AutoQuantizationConfig.from_dict( File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict return target_cls.from_dict(quantization_config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict config = cls(**config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__ self.post_init() File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init raise ValueError( ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f0e567f0dc0>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
2025-02-06T10:18:41.917489
bd27598f-6f94-4e3d-9747-85c8ce835edd
2025-02-06T10:09:56.868153
32vCPU-C7i
cpu
unquantized
pytorch
liuhaotian/llava-v1.5-7b
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:18:49.888665
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
1xA10
cuda
torchao
pytorch
trl-internal-testing/dummy-GPT2-correct-vocab
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model self.process_quantization_config() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config self.quantization_config = AutoQuantizationConfig.from_dict( File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict return target_cls.from_dict(quantization_config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict config = cls(**config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__ self.post_init() File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init raise ValueError( ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f49fc1aaa70>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
2025-02-06T10:19:04.730530
fc85a78b-b92a-46d9-b36b-715516040a48
2025-02-06T10:12:52.699176
32vCPU-C7i
cpu
unquantized
pytorch
liuhaotian/llava-v1.5-7b
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 57, in launch raise RuntimeError(f"Isolated process exited with non-zero code {isolated_process.exitcode}") RuntimeError: Isolated process exited with non-zero code -6
2025-02-06T10:19:07.250478
9fd70c10-e9cd-4ae1-ad3b-2f56445ec3b5
2025-02-06T10:09:24.419768
1xT4
cuda
torchao
pytorch
Qwen/Qwen2-0.5B
false
Traceback (most recent call last): File "/workspace/llm_perf/common/benchmark_runner.py", line 319, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 51, in launch report = launcher.launch(worker=Benchmark.run, worker_args=[config]) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 66, in launch raise ChildProcessError(response["traceback"]) ChildProcessError: Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/launchers/process/launcher.py", line 103, in target report = worker(*worker_args) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/benchmark/base.py", line 78, in run report = scenario.run(backend) File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 129, in run self.run_model_loading_tracking() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/scenarios/inference/scenario.py", line 184, in run_model_loading_tracking self.backend.load() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 69, in load self.load_transformers_model() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 142, in load_transformers_model self.process_quantization_config() File "/usr/local/lib/python3.10/dist-packages/optimum_benchmark/backends/pytorch/backend.py", line 310, in process_quantization_config self.quantization_config = AutoQuantizationConfig.from_dict( File "/usr/local/lib/python3.10/dist-packages/transformers/quantizers/auto.py", line 103, in from_dict return target_cls.from_dict(quantization_config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 102, in from_dict config = cls(**config_dict) File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1269, in __init__ self.post_init() File "/usr/local/lib/python3.10/dist-packages/transformers/utils/quantization_config.py", line 1298, in post_init raise ValueError( ValueError: Unexpected keyword arg: quant_method for API: <function int4_weight_only at 0x7f91153d6b90>, accepted keyword args are: ['group_size', 'layout', 'use_hqq', 'zero_point_domain']
2025-02-06T10:19:09.232979
bd27598f-6f94-4e3d-9747-85c8ce835edd
2025-02-06T10:09:56.868153
README.md exists but content is empty.
Downloads last month
55