LLava-med-api / model_worker_f9540c.log
kairavishal37's picture
Upload folder using huggingface_hub
eab5375 verified
2025-01-28 00:52:04 | INFO | model_worker | args: Namespace(host='0.0.0.0', port=40000, worker_address='http://localhost:40000', controller_address='http://localhost:10000', model_path='.', model_base=None, model_name=None, device='cuda', multi_modal=True, limit_model_concurrency=5, stream_interval=1, no_register=False, load_8bit=False, load_4bit=False)
2025-01-28 00:52:04 | WARNING | model_worker | Multimodal mode is automatically detected with model name, please make sure `llava` is included in the model path.
2025-01-28 00:52:04 | INFO | model_worker | Loading the model . on worker f9540c ...
2025-01-28 00:52:04 | ERROR | stderr | Traceback (most recent call last):
2025-01-28 00:52:04 | ERROR | stderr | File "C:\Users\admin\AppData\Local\Programs\Python\Python310\lib\runpy.py", line 196, in _run_module_as_main
2025-01-28 00:52:04 | ERROR | stderr | return _run_code(code, main_globals, None,
2025-01-28 00:52:04 | ERROR | stderr | File "C:\Users\admin\AppData\Local\Programs\Python\Python310\lib\runpy.py", line 86, in _run_code
2025-01-28 00:52:04 | ERROR | stderr | exec(code, run_globals)
2025-01-28 00:52:04 | ERROR | stderr | File "C:\Users\admin\LLaVA-Med\llava\serve\model_worker.py", line 275, in <module>
2025-01-28 00:52:04 | ERROR | stderr | worker = ModelWorker(args.controller_address,
2025-01-28 00:52:04 | ERROR | stderr | File "C:\Users\admin\LLaVA-Med\llava\serve\model_worker.py", line 65, in __init__
2025-01-28 00:52:04 | ERROR | stderr | self.tokenizer, self.model, self.image_processor, self.context_len = load_pretrained_model(
2025-01-28 00:52:04 | ERROR | stderr | File "C:\Users\admin\LLaVA-Med\llava\model\builder.py", line 57, in load_pretrained_model
2025-01-28 00:52:04 | ERROR | stderr | model = AutoModelForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs)
2025-01-28 00:52:04 | ERROR | stderr | File "C:\Users\admin\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\models\auto\auto_factory.py", line 526, in from_pretrained
2025-01-28 00:52:04 | ERROR | stderr | config, kwargs = AutoConfig.from_pretrained(
2025-01-28 00:52:04 | ERROR | stderr | File "C:\Users\admin\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\models\auto\configuration_auto.py", line 1082, in from_pretrained
2025-01-28 00:52:04 | ERROR | stderr | config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs)
2025-01-28 00:52:04 | ERROR | stderr | File "C:\Users\admin\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\configuration_utils.py", line 644, in get_config_dict
2025-01-28 00:52:04 | ERROR | stderr | config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
2025-01-28 00:52:04 | ERROR | stderr | File "C:\Users\admin\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\configuration_utils.py", line 699, in _get_config_dict
2025-01-28 00:52:04 | ERROR | stderr | resolved_config_file = cached_file(
2025-01-28 00:52:04 | ERROR | stderr | File "C:\Users\admin\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\utils\hub.py", line 360, in cached_file
2025-01-28 00:52:04 | ERROR | stderr | raise EnvironmentError(
2025-01-28 00:52:04 | ERROR | stderr | OSError: . does not appear to have a file named config.json. Checkout 'https://huggingface.co/./None' for available files.