Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': #24

Open
franztao opened this issue May 30, 2023 · 2 comments

Comments

@franztao
Copy link

执行命名如下
torchrun --nproc_per_node=8 /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py
--lm_path /home/jn/th/work/Multimodal-GPT/checkpoints/llama-7b_hf
--tokenizer_path /home/jn/th/work/Multimodal-GPT/checkpoints/llama-7b_hf
--pretrained_path /home/jn/th/work/Multimodal-GPT/checkpoints/OpenFlamingo-9B/checkpoint.pt
--run_name train-my-gpt4
--learning_rate 1e-5
--lr_scheduler cosine
--batch_size 1
--tuning_config /home/jn/th/work/Multimodal-GPT/configs/lora_config.py
--dataset_config /home/jn/th/work/Multimodal-GPT/configs/dataset_config.py
--report_to_wandb

如下错误:

WARNING:torch.distributed.run:


Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed.


init clip vision encoder
init clip vision encoder
init clip vision encoder
init clip vision encoder
init clip vision encoder
init clip vision encoder
init clip vision encoder
init clip vision encoder
init tokenizer
init tokenizer
╭───────────────────── Traceback (most recent call last) ──────────────────────╮
│ /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py │
│ :460 in │
│ │
│ 457 │
│ 458 │
│ 459 if name == "main": │
│ ❱ 460 │ main() │
│ 461 │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py │
│ :167 in main │
│ │
│ 164 │ │ tokenizer_path=args.tokenizer_path if args.tokenizer_path else │
│ 165 │ │ use_media_placement_augmentation=args.use_media_placement_augm │
│ 166 │ │ pretrained_model_path=args.pretrained_path, │
│ ❱ 167 │ │ tuning_config=tuning_config.tuning_config, │
│ 168 │ ) │
│ 169 │ │
│ 170 │ if args.dataset_config is not None: │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/models/builder.py:23 in │
│ create_model_and_transforms │
│ │
│ 20 │ │ │ tokenizer_path=tokenizer_path, │
│ 21 │ │ │ tuning_config=tuning_config, │
│ 22 │ │ │ pretrained_model_path=pretrained_model_path, │
│ ❱ 23 │ │ │ **kwargs, │
│ 24 │ │ ) │
│ 25 │ # TODO: support BLIP2 │
│ 26 │ else: │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/models/open_flamingo/builder. │
│ py:46 in create_model_and_transforms │
│ │
│ 43 │ # set the vision encoder to output the visual features │
│ 44 │ vision_encoder.visual.output_tokens = True │
│ 45 │ print("init tokenizer") │
│ ❱ 46 │ text_tokenizer = LlamaTokenizer.from_pretrained(tokenizer_path) │
│ 47 │ # add Flamingo special tokens to the tokenizer │
│ 48 │ text_tokenizer.add_special_tokens({"additional_special_tokens": [" │
│ 49 │ if text_tokenizer.pad_token is None: │
│ │
│ /opt/conda/lib/python3.7/site-packages/transformers/tokenization_utils_base. │
│ py:1785 in from_pretrained │
│ │
│ 1782 │ │ │ │ │ subfolder=subfolder, │
│ 1783 │ │ │ │ │ _raise_exceptions_for_missing_entries=False, │
│ 1784 │ │ │ │ │ _raise_exceptions_for_connection_errors=False, │
│ ❱ 1785 │ │ │ │ │ _commit_hash=commit_hash, │
│ 1786 │ │ │ │ ) │
│ 1787 │ │ │ │ commit_hash = extract_commit_hash(resolved_vocab_file │
│ 1788 │
│ │
│ /opt/conda/lib/python3.7/site-packages/transformers/utils/hub.py:429 in │
│ cached_file │
│ │
│ 426 │ │ │ proxies=proxies, │
│ 427 │ │ │ resume_download=resume_download, │
│ 428 │ │ │ use_auth_token=use_auth_token, │
│ ❱ 429 │ │ │ local_files_only=local_files_only, │
│ 430 │ │ ) │
│ 431 │ │
│ 432 │ except RepositoryNotFoundError: │
│ │
│ /opt/conda/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py: │
│ 112 in _inner_fn │
│ │
│ 109 │ │ │ kwargs.items(), # Kwargs values │
│ 110 │ │ ): │
│ 111 │ │ │ if arg_name in ["repo_id", "from_id", "to_id"]: │
│ ❱ 112 │ │ │ │ validate_repo_id(arg_value) │
│ 113 │ │ │ │
│ 114 │ │ │ elif arg_name == "token" and arg_value is not None: │
│ 115 │ │ │ │ has_token = True │
│ │
│ /opt/conda/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py: │
│ 161 in validate_repo_id │
│ │
│ 158 │ │
│ 159 │ if repo_id.count("/") > 1: │
│ 160 │ │ raise HFValidationError( │
│ ❱ 161 │ │ │ "Repo id must be in the form 'repo_name' or 'namespace/rep │
│ 162 │ │ │ f" '{repo_id}'. Use repo_type argument if needed." │
│ 163 │ │ ) │
│ 164 │
╰──────────────────────────────────────────────────────────────────────────────╯
HFValidationError: Repo id must be in the form 'repo_name' or
'namespace/repo_name':
'/home/jn/th/work/Multimodal-GPT/checkpoints/llama-7b_hf'. Use
repo_type argument if needed.
╭───────────────────── Traceback (most recent call last) ──────────────────────╮
│ /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py │
│ :460 in │
│ │
│ 457 │
│ 458 │
│ 459 if name == "main": │
│ ❱ 460 │ main() │
│ 461 │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py │
│ :167 in main │
│ │
│ 164 │ │ tokenizer_path=args.tokenizer_path if args.tokenizer_path else │
│ 165 │ │ use_media_placement_augmentation=args.use_media_placement_augm │
│ 166 │ │ pretrained_model_path=args.pretrained_path, │
│ ❱ 167 │ │ tuning_config=tuning_config.tuning_config, │
│ 168 │ ) │
│ 169 │ │
│ 170 │ if args.dataset_config is not None: │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/models/builder.py:23 in │
│ create_model_and_transforms │
│ │
│ 20 │ │ │ tokenizer_path=tokenizer_path, │
│ 21 │ │ │ tuning_config=tuning_config, │
│ 22 │ │ │ pretrained_model_path=pretrained_model_path, │
│ ❱ 23 │ │ │ **kwargs, │
│ 24 │ │ ) │
│ 25 │ # TODO: support BLIP2 │
│ 26 │ else: │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/models/open_flamingo/builder. │
│ py:46 in create_model_and_transforms │
│ │
│ 43 │ # set the vision encoder to output the visual features │
│ 44 │ vision_encoder.visual.output_tokens = True │
│ 45 │ print("init tokenizer") │
│ ❱ 46 │ text_tokenizer = LlamaTokenizer.from_pretrained(tokenizer_path) │
│ 47 │ # add Flamingo special tokens to the tokenizer │
│ 48 │ text_tokenizer.add_special_tokens({"additional_special_tokens": [" │
│ 49 │ if text_tokenizer.pad_token is None: │
│ │
│ /opt/conda/lib/python3.7/site-packages/transformers/tokenization_utils_base. │
│ py:1785 in from_pretrained │
│ │
│ 1782 │ │ │ │ │ subfolder=subfolder, │
│ 1783 │ │ │ │ │ _raise_exceptions_for_missing_entries=False, │
│ 1784 │ │ │ │ │ _raise_exceptions_for_connection_errors=False, │
│ ❱ 1785 │ │ │ │ │ _commit_hash=commit_hash, │
│ 1786 │ │ │ │ ) │
│ 1787 │ │ │ │ commit_hash = extract_commit_hash(resolved_vocab_file │
│ 1788 │
│ │
│ /opt/conda/lib/python3.7/site-packages/transformers/utils/hub.py:429 in │
│ cached_file │
│ │
│ 426 │ │ │ proxies=proxies, │
│ 427 │ │ │ resume_download=resume_download, │
│ 428 │ │ │ use_auth_token=use_auth_token, │
│ ❱ 429 │ │ │ local_files_only=local_files_only, │
│ 430 │ │ ) │
│ 431 │ │
│ 432 │ except RepositoryNotFoundError: │
│ │
│ /opt/conda/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py: │
│ 112 in _inner_fn │
│ │
│ 109 │ │ │ kwargs.items(), # Kwargs values │
│ 110 │ │ ): │
│ 111 │ │ │ if arg_name in ["repo_id", "from_id", "to_id"]: │
│ ❱ 112 │ │ │ │ validate_repo_id(arg_value) │
│ 113 │ │ │ │
│ 114 │ │ │ elif arg_name == "token" and arg_value is not None: │
│ 115 │ │ │ │ has_token = True │
│ │
│ /opt/conda/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py: │
│ 161 in validate_repo_id │
│ │
│ 158 │ │
│ 159 │ if repo_id.count("/") > 1: │
│ 160 │ │ raise HFValidationError( │
│ ❱ 161 │ │ │ "Repo id must be in the form 'repo_name' or 'namespace/rep │
│ 162 │ │ │ f" '{repo_id}'. Use repo_type argument if needed." │
│ 163 │ │ ) │
│ 164 │
╰──────────────────────────────────────────────────────────────────────────────╯
HFValidationError: Repo id must be in the form 'repo_name' or
'namespace/repo_name':
'/home/jn/th/work/Multimodal-GPT/checkpoints/llama-7b_hf'. Use
repo_type argument if needed.
init tokenizer
init tokenizer
init tokenizer
╭───────────────────── Traceback (most recent call last) ──────────────────────╮
│ /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py │
│ :460 in │
│ │
│ 457 │
│ 458 │
│ 459 if name == "main": │
│ ❱ 460 │ main() │
│ 461 │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py │
│ :167 in main │
│ │
│ 164 │ │ tokenizer_path=args.tokenizer_path if args.tokenizer_path else │
│ 165 │ │ use_media_placement_augmentation=args.use_media_placement_augm │
│ 166 │ │ pretrained_model_path=args.pretrained_path, │
│ ❱ 167 │ │ tuning_config=tuning_config.tuning_config, │
│ 168 │ ) │
│ 169 │ │
│ 170 │ if args.dataset_config is not None: │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/models/builder.py:23 in │
│ create_model_and_transforms │
│ │
│ 20 │ │ │ tokenizer_path=tokenizer_path, │
│ 21 │ │ │ tuning_config=tuning_config, │
│ 22 │ │ │ pretrained_model_path=pretrained_model_path, │
│ ❱ 23 │ │ │ **kwargs, │
│ 24 │ │ ) │
│ 25 │ # TODO: support BLIP2 │
│ 26 │ else: │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/models/open_flamingo/builder. │
│ py:46 in create_model_and_transforms │
│ │
│ 43 │ # set the vision encoder to output the visual features │
│ 44 │ vision_encoder.visual.output_tokens = True │
│ 45 │ print("init tokenizer") │
│ ❱ 46 │ text_tokenizer = LlamaTokenizer.from_pretrained(tokenizer_path) │
│ 47 │ # add Flamingo special tokens to the tokenizer │
│ 48 │ text_tokenizer.add_special_tokens({"additional_special_tokens": [" │
│ 49 │ if text_tokenizer.pad_token is None: │
│ │
│ /opt/conda/lib/python3.7/site-packages/transformers/tokenization_utils_base. │
│ py:1785 in from_pretrained │
│ │
│ 1782 │ │ │ │ │ subfolder=subfolder, │
│ 1783 │ │ │ │ │ _raise_exceptions_for_missing_entries=False, │
│ 1784 │ │ │ │ │ _raise_exceptions_for_connection_errors=False, │
│ ❱ 1785 │ │ │ │ │ _commit_hash=commit_hash, │
│ 1786 │ │ │ │ ) │
│ 1787 │ │ │ │ commit_hash = extract_commit_hash(resolved_vocab_file │
│ 1788 │
│ │
│ /opt/conda/lib/python3.7/site-packages/transformers/utils/hub.py:429 in │
│ cached_file │
│ │
│ 426 │ │ │ proxies=proxies, │
│ 427 │ │ │ resume_download=resume_download, │
│ 428 │ │ │ use_auth_token=use_auth_token, │
│ ❱ 429 │ │ │ local_files_only=local_files_only, │
│ 430 │ │ ) │
│ 431 │ │
│ 432 │ except RepositoryNotFoundError: │
│ │
│ /opt/conda/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py: │
│ 112 in _inner_fn │
│ │
│ 109 │ │ │ kwargs.items(), # Kwargs values │
│ 110 │ │ ): │
│ 111 │ │ │ if arg_name in ["repo_id", "from_id", "to_id"]: │
│ ❱ 112 │ │ │ │ validate_repo_id(arg_value) │
│ 113 │ │ │ │
│ 114 │ │ │ elif arg_name == "token" and arg_value is not None: │
│ 115 │ │ │ │ has_token = True │
│ │
│ /opt/conda/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py: │
│ 161 in validate_repo_id │
│ │
│ 158 │ │
│ 159 │ if repo_id.count("/") > 1: │
│ 160 │ │ raise HFValidationError( │
│ ❱ 161 │ │ │ "Repo id must be in the form 'repo_name' or 'namespace/rep │
│ 162 │ │ │ f" '{repo_id}'. Use repo_type argument if needed." │
│ 163 │ │ ) │
│ 164 │
╰──────────────────────────────────────────────────────────────────────────────╯
HFValidationError: Repo id must be in the form 'repo_name' or
'namespace/repo_name':
'/home/jn/th/work/Multimodal-GPT/checkpoints/llama-7b_hf'. Use
repo_type argument if needed.
init tokenizer
╭───────────────────── Traceback (most recent call last) ──────────────────────╮
│ /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py │
│ :460 in │
│ │
│ 457 │
│ 458 │
│ 459 if name == "main": │
│ ❱ 460 │ main() │
│ 461 │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py │
│ :167 in main │
│ │
│ 164 │ │ tokenizer_path=args.tokenizer_path if args.tokenizer_path else │
│ 165 │ │ use_media_placement_augmentation=args.use_media_placement_augm │
│ 166 │ │ pretrained_model_path=args.pretrained_path, │
│ ❱ 167 │ │ tuning_config=tuning_config.tuning_config, │
│ 168 │ ) │
│ 169 │ │
│ 170 │ if args.dataset_config is not None: │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/models/builder.py:23 in │
│ create_model_and_transforms │
│ │
│ 20 │ │ │ tokenizer_path=tokenizer_path, │
│ 21 │ │ │ tuning_config=tuning_config, │
│ 22 │ │ │ pretrained_model_path=pretrained_model_path, │
│ ❱ 23 │ │ │ **kwargs, │
│ 24 │ │ ) │
│ 25 │ # TODO: support BLIP2 │
│ 26 │ else: │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/models/open_flamingo/builder. │
│ py:46 in create_model_and_transforms │
│ │
│ 43 │ # set the vision encoder to output the visual features │
│ 44 │ vision_encoder.visual.output_tokens = True │
│ 45 │ print("init tokenizer") │
│ ❱ 46 │ text_tokenizer = LlamaTokenizer.from_pretrained(tokenizer_path) │
│ 47 │ # add Flamingo special tokens to the tokenizer │
│ 48 │ text_tokenizer.add_special_tokens({"additional_special_tokens": [" │
│ 49 │ if text_tokenizer.pad_token is None: │
│ │
│ /opt/conda/lib/python3.7/site-packages/transformers/tokenization_utils_base. │
│ py:1785 in from_pretrained │
│ │
│ 1782 │ │ │ │ │ subfolder=subfolder, │
│ 1783 │ │ │ │ │ _raise_exceptions_for_missing_entries=False, │
│ 1784 │ │ │ │ │ _raise_exceptions_for_connection_errors=False, │
│ ❱ 1785 │ │ │ │ │ _commit_hash=commit_hash, │
│ 1786 │ │ │ │ ) │
│ 1787 │ │ │ │ commit_hash = extract_commit_hash(resolved_vocab_file │
│ 1788 │
│ │
│ /opt/conda/lib/python3.7/site-packages/transformers/utils/hub.py:429 in │
│ cached_file │
│ │
│ 426 │ │ │ proxies=proxies, │
│ 427 │ │ │ resume_download=resume_download, │
│ 428 │ │ │ use_auth_token=use_auth_token, │
│ ❱ 429 │ │ │ local_files_only=local_files_only, │
│ 430 │ │ ) │
│ 431 │ │
│ 432 │ except RepositoryNotFoundError: │
│ │
│ /opt/conda/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py: │
│ 112 in _inner_fn │
│ │
│ 109 │ │ │ kwargs.items(), # Kwargs values │
│ 110 │ │ ): │
│ 111 │ │ │ if arg_name in ["repo_id", "from_id", "to_id"]: │
│ ❱ 112 │ │ │ │ validate_repo_id(arg_value) │
│ 113 │ │ │ │
│ 114 │ │ │ elif arg_name == "token" and arg_value is not None: │
│ 115 │ │ │ │ has_token = True │
│ │
│ /opt/conda/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py: │
│ 161 in validate_repo_id │
│ │
│ 158 │ │
│ 159 │ if repo_id.count("/") > 1: │
│ 160 │ │ raise HFValidationError( │
│ ❱ 161 │ │ │ "Repo id must be in the form 'repo_name' or 'namespace/rep │
│ 162 │ │ │ f" '{repo_id}'. Use repo_type argument if needed." │
│ 163 │ │ ) │
│ 164 │
╰──────────────────────────────────────────────────────────────────────────────╯
HFValidationError: Repo id must be in the form 'repo_name' or
'namespace/repo_name':
'/home/jn/th/work/Multimodal-GPT/checkpoints/llama-7b_hf'. Use
repo_type argument if needed.
╭───────────────────── Traceback (most recent call last) ──────────────────────╮
│ /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py │
│ :460 in │
│ │
│ 457 │
│ 458 │
│ 459 if name == "main": │
│ ❱ 460 │ main() │
│ 461 │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py │
│ :167 in main │
│ │
│ 164 │ │ tokenizer_path=args.tokenizer_path if args.tokenizer_path else │
│ 165 │ │ use_media_placement_augmentation=args.use_media_placement_augm │
│ 166 │ │ pretrained_model_path=args.pretrained_path, │
│ ❱ 167 │ │ tuning_config=tuning_config.tuning_config, │
│ 168 │ ) │
│ 169 │ │
│ 170 │ if args.dataset_config is not None: │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/models/builder.py:23 in │
│ create_model_and_transforms │
│ │
│ 20 │ │ │ tokenizer_path=tokenizer_path, │
│ 21 │ │ │ tuning_config=tuning_config, │
│ 22 │ │ │ pretrained_model_path=pretrained_model_path, │
│ ❱ 23 │ │ │ **kwargs, │
│ 24 │ │ ) │
│ 25 │ # TODO: support BLIP2 │
│ 26 │ else: │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/models/open_flamingo/builder. │
│ py:46 in create_model_and_transforms │
│ │
│ 43 │ # set the vision encoder to output the visual features │
│ 44 │ vision_encoder.visual.output_tokens = True │
│ 45 │ print("init tokenizer") │
│ ❱ 46 │ text_tokenizer = LlamaTokenizer.from_pretrained(tokenizer_path) │
│ 47 │ # add Flamingo special tokens to the tokenizer │
│ 48 │ text_tokenizer.add_special_tokens({"additional_special_tokens": [" │
│ 49 │ if text_tokenizer.pad_token is None: │
│ │
│ /opt/conda/lib/python3.7/site-packages/transformers/tokenization_utils_base. │
│ py:1785 in from_pretrained │
│ │
│ 1782 │ │ │ │ │ subfolder=subfolder, │
│ 1783 │ │ │ │ │ _raise_exceptions_for_missing_entries=False, │
│ 1784 │ │ │ │ │ _raise_exceptions_for_connection_errors=False, │
│ ❱ 1785 │ │ │ │ │ _commit_hash=commit_hash, │
│ 1786 │ │ │ │ ) │
│ 1787 │ │ │ │ commit_hash = extract_commit_hash(resolved_vocab_file │
│ 1788 │
│ │
│ /opt/conda/lib/python3.7/site-packages/transformers/utils/hub.py:429 in │
│ cached_file │
│ │
│ 426 │ │ │ proxies=proxies, │
│ 427 │ │ │ resume_download=resume_download, │
│ 428 │ │ │ use_auth_token=use_auth_token, │
│ ❱ 429 │ │ │ local_files_only=local_files_only, │
│ 430 │ │ ) │
│ 431 │ │
│ 432 │ except RepositoryNotFoundError: │
│ │
│ /opt/conda/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py: │
│ 112 in _inner_fn │
│ │
│ 109 │ │ │ kwargs.items(), # Kwargs values │
│ 110 │ │ ): │
│ 111 │ │ │ if arg_name in ["repo_id", "from_id", "to_id"]: │
│ ❱ 112 │ │ │ │ validate_repo_id(arg_value) │
│ 113 │ │ │ │
│ 114 │ │ │ elif arg_name == "token" and arg_value is not None: │
│ 115 │ │ │ │ has_token = True │
│ │
│ /opt/conda/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py: │
│ 161 in validate_repo_id │
│ │
│ 158 │ │
│ 159 │ if repo_id.count("/") > 1: │
│ 160 │ │ raise HFValidationError( │
│ ❱ 161 │ │ │ "Repo id must be in the form 'repo_name' or 'namespace/rep │
│ 162 │ │ │ f" '{repo_id}'. Use repo_type argument if needed." │
│ 163 │ │ ) │
│ 164 │
╰──────────────────────────────────────────────────────────────────────────────╯
HFValidationError: Repo id must be in the form 'repo_name' or
'namespace/repo_name':
'/home/jn/th/work/Multimodal-GPT/checkpoints/llama-7b_hf'. Use
repo_type argument if needed.
╭───────────────────── Traceback (most recent call last) ──────────────────────╮
│ /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py │
│ :460 in │
│ │
│ 457 │
│ 458 │
│ 459 if name == "main": │
│ ❱ 460 │ main() │
│ 461 │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py │
│ :167 in main │
│ │
│ 164 │ │ tokenizer_path=args.tokenizer_path if args.tokenizer_path else │
│ 165 │ │ use_media_placement_augmentation=args.use_media_placement_augm │
│ 166 │ │ pretrained_model_path=args.pretrained_path, │
│ ❱ 167 │ │ tuning_config=tuning_config.tuning_config, │
│ 168 │ ) │
│ 169 │ │
│ 170 │ if args.dataset_config is not None: │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/models/builder.py:23 in │
│ create_model_and_transforms │
│ │
│ 20 │ │ │ tokenizer_path=tokenizer_path, │
│ 21 │ │ │ tuning_config=tuning_config, │
│ 22 │ │ │ pretrained_model_path=pretrained_model_path, │
│ ❱ 23 │ │ │ **kwargs, │
│ 24 │ │ ) │
│ 25 │ # TODO: support BLIP2 │
│ 26 │ else: │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/models/open_flamingo/builder. │
│ py:46 in create_model_and_transforms │
│ │
│ 43 │ # set the vision encoder to output the visual features │
│ 44 │ vision_encoder.visual.output_tokens = True │
│ 45 │ print("init tokenizer") │
│ ❱ 46 │ text_tokenizer = LlamaTokenizer.from_pretrained(tokenizer_path) │
│ 47 │ # add Flamingo special tokens to the tokenizer │
│ 48 │ text_tokenizer.add_special_tokens({"additional_special_tokens": [" │
│ 49 │ if text_tokenizer.pad_token is None: │
│ │
│ /opt/conda/lib/python3.7/site-packages/transformers/tokenization_utils_base. │
│ py:1785 in from_pretrained │
│ │
│ 1782 │ │ │ │ │ subfolder=subfolder, │
│ 1783 │ │ │ │ │ _raise_exceptions_for_missing_entries=False, │
│ 1784 │ │ │ │ │ _raise_exceptions_for_connection_errors=False, │
│ ❱ 1785 │ │ │ │ │ _commit_hash=commit_hash, │
│ 1786 │ │ │ │ ) │
│ 1787 │ │ │ │ commit_hash = extract_commit_hash(resolved_vocab_file │
│ 1788 │
│ │
│ /opt/conda/lib/python3.7/site-packages/transformers/utils/hub.py:429 in │
│ cached_file │
│ │
│ 426 │ │ │ proxies=proxies, │
│ 427 │ │ │ resume_download=resume_download, │
│ 428 │ │ │ use_auth_token=use_auth_token, │
│ ❱ 429 │ │ │ local_files_only=local_files_only, │
│ 430 │ │ ) │
│ 431 │ │
│ 432 │ except RepositoryNotFoundError: │
│ │
│ /opt/conda/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py: │
│ 112 in _inner_fn │
│ │
│ 109 │ │ │ kwargs.items(), # Kwargs values │
│ 110 │ │ ): │
│ 111 │ │ │ if arg_name in ["repo_id", "from_id", "to_id"]: │
│ ❱ 112 │ │ │ │ validate_repo_id(arg_value) │
│ 113 │ │ │ │
│ 114 │ │ │ elif arg_name == "token" and arg_value is not None: │
│ 115 │ │ │ │ has_token = True │
│ │
│ /opt/conda/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py: │
│ 161 in validate_repo_id │
│ │
│ 158 │ │
│ 159 │ if repo_id.count("/") > 1: │
│ 160 │ │ raise HFValidationError( │
│ ❱ 161 │ │ │ "Repo id must be in the form 'repo_name' or 'namespace/rep │
│ 162 │ │ │ f" '{repo_id}'. Use repo_type argument if needed." │
│ 163 │ │ ) │
│ 164 │
╰──────────────────────────────────────────────────────────────────────────────╯
HFValidationError: Repo id must be in the form 'repo_name' or
'namespace/repo_name':
'/home/jn/th/work/Multimodal-GPT/checkpoints/llama-7b_hf'. Use
repo_type argument if needed.
init tokenizer
╭───────────────────── Traceback (most recent call last) ──────────────────────╮
│ /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py │
│ :460 in │
│ │
│ 457 │
│ 458 │
│ 459 if name == "main": │
│ ❱ 460 │ main() │
│ 461 │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py │
│ :167 in main │
│ │
│ 164 │ │ tokenizer_path=args.tokenizer_path if args.tokenizer_path else │
│ 165 │ │ use_media_placement_augmentation=args.use_media_placement_augm │
│ 166 │ │ pretrained_model_path=args.pretrained_path, │
│ ❱ 167 │ │ tuning_config=tuning_config.tuning_config, │
│ 168 │ ) │
│ 169 │ │
│ 170 │ if args.dataset_config is not None: │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/models/builder.py:23 in │
│ create_model_and_transforms │
│ │
│ 20 │ │ │ tokenizer_path=tokenizer_path, │
│ 21 │ │ │ tuning_config=tuning_config, │
│ 22 │ │ │ pretrained_model_path=pretrained_model_path, │
│ ❱ 23 │ │ │ **kwargs, │
│ 24 │ │ ) │
│ 25 │ # TODO: support BLIP2 │
│ 26 │ else: │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/models/open_flamingo/builder. │
│ py:46 in create_model_and_transforms │
│ │
│ 43 │ # set the vision encoder to output the visual features │
│ 44 │ vision_encoder.visual.output_tokens = True │
│ 45 │ print("init tokenizer") │
│ ❱ 46 │ text_tokenizer = LlamaTokenizer.from_pretrained(tokenizer_path) │
│ 47 │ # add Flamingo special tokens to the tokenizer │
│ 48 │ text_tokenizer.add_special_tokens({"additional_special_tokens": [" │
│ 49 │ if text_tokenizer.pad_token is None: │
│ │
│ /opt/conda/lib/python3.7/site-packages/transformers/tokenization_utils_base. │
│ py:1785 in from_pretrained │
│ │
│ 1782 │ │ │ │ │ subfolder=subfolder, │
│ 1783 │ │ │ │ │ _raise_exceptions_for_missing_entries=False, │
│ 1784 │ │ │ │ │ _raise_exceptions_for_connection_errors=False, │
│ ❱ 1785 │ │ │ │ │ _commit_hash=commit_hash, │
│ 1786 │ │ │ │ ) │
│ 1787 │ │ │ │ commit_hash = extract_commit_hash(resolved_vocab_file │
│ 1788 │
│ │
│ /opt/conda/lib/python3.7/site-packages/transformers/utils/hub.py:429 in │
│ cached_file │
│ │
│ 426 │ │ │ proxies=proxies, │
│ 427 │ │ │ resume_download=resume_download, │
│ 428 │ │ │ use_auth_token=use_auth_token, │
│ ❱ 429 │ │ │ local_files_only=local_files_only, │
│ 430 │ │ ) │
│ 431 │ │
│ 432 │ except RepositoryNotFoundError: │
│ │
│ /opt/conda/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py: │
│ 112 in _inner_fn │
│ │
│ 109 │ │ │ kwargs.items(), # Kwargs values │
│ 110 │ │ ): │
│ 111 │ │ │ if arg_name in ["repo_id", "from_id", "to_id"]: │
│ ❱ 112 │ │ │ │ validate_repo_id(arg_value) │
│ 113 │ │ │ │
│ 114 │ │ │ elif arg_name == "token" and arg_value is not None: │
│ 115 │ │ │ │ has_token = True │
│ │
│ /opt/conda/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py: │
│ 161 in validate_repo_id │
│ │
│ 158 │ │
│ 159 │ if repo_id.count("/") > 1: │
│ 160 │ │ raise HFValidationError( │
│ ❱ 161 │ │ │ "Repo id must be in the form 'repo_name' or 'namespace/rep │
│ 162 │ │ │ f" '{repo_id}'. Use repo_type argument if needed." │
│ 163 │ │ ) │
│ 164 │
╰──────────────────────────────────────────────────────────────────────────────╯
HFValidationError: Repo id must be in the form 'repo_name' or
'namespace/repo_name':
'/home/jn/th/work/Multimodal-GPT/checkpoints/llama-7b_hf'. Use
repo_type argument if needed.
init tokenizer
╭───────────────────── Traceback (most recent call last) ──────────────────────╮
│ /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py │
│ :460 in │
│ │
│ 457 │
│ 458 │
│ 459 if name == "main": │
│ ❱ 460 │ main() │
│ 461 │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py │
│ :167 in main │
│ │
│ 164 │ │ tokenizer_path=args.tokenizer_path if args.tokenizer_path else │
│ 165 │ │ use_media_placement_augmentation=args.use_media_placement_augm │
│ 166 │ │ pretrained_model_path=args.pretrained_path, │
│ ❱ 167 │ │ tuning_config=tuning_config.tuning_config, │
│ 168 │ ) │
│ 169 │ │
│ 170 │ if args.dataset_config is not None: │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/models/builder.py:23 in │
│ create_model_and_transforms │
│ │
│ 20 │ │ │ tokenizer_path=tokenizer_path, │
│ 21 │ │ │ tuning_config=tuning_config, │
│ 22 │ │ │ pretrained_model_path=pretrained_model_path, │
│ ❱ 23 │ │ │ **kwargs, │
│ 24 │ │ ) │
│ 25 │ # TODO: support BLIP2 │
│ 26 │ else: │
│ │
│ /home/jn/th/work/Multimodal-GPT/mmgpt/models/open_flamingo/builder. │
│ py:46 in create_model_and_transforms │
│ │
│ 43 │ # set the vision encoder to output the visual features │
│ 44 │ vision_encoder.visual.output_tokens = True │
│ 45 │ print("init tokenizer") │
│ ❱ 46 │ text_tokenizer = LlamaTokenizer.from_pretrained(tokenizer_path) │
│ 47 │ # add Flamingo special tokens to the tokenizer │
│ 48 │ text_tokenizer.add_special_tokens({"additional_special_tokens": [" │
│ 49 │ if text_tokenizer.pad_token is None: │
│ │
│ /opt/conda/lib/python3.7/site-packages/transformers/tokenization_utils_base. │
│ py:1785 in from_pretrained │
│ │
│ 1782 │ │ │ │ │ subfolder=subfolder, │
│ 1783 │ │ │ │ │ _raise_exceptions_for_missing_entries=False, │
│ 1784 │ │ │ │ │ _raise_exceptions_for_connection_errors=False, │
│ ❱ 1785 │ │ │ │ │ _commit_hash=commit_hash, │
│ 1786 │ │ │ │ ) │
│ 1787 │ │ │ │ commit_hash = extract_commit_hash(resolved_vocab_file │
│ 1788 │
│ │
│ /opt/conda/lib/python3.7/site-packages/transformers/utils/hub.py:429 in │
│ cached_file │
│ │
│ 426 │ │ │ proxies=proxies, │
│ 427 │ │ │ resume_download=resume_download, │
│ 428 │ │ │ use_auth_token=use_auth_token, │
│ ❱ 429 │ │ │ local_files_only=local_files_only, │
│ 430 │ │ ) │
│ 431 │ │
│ 432 │ except RepositoryNotFoundError: │
│ │
│ /opt/conda/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py: │
│ 112 in _inner_fn │
│ │
│ 109 │ │ │ kwargs.items(), # Kwargs values │
│ 110 │ │ ): │
│ 111 │ │ │ if arg_name in ["repo_id", "from_id", "to_id"]: │
│ ❱ 112 │ │ │ │ validate_repo_id(arg_value) │
│ 113 │ │ │ │
│ 114 │ │ │ elif arg_name == "token" and arg_value is not None: │
│ 115 │ │ │ │ has_token = True │
│ │
│ /opt/conda/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py: │
│ 161 in validate_repo_id │
│ │
│ 158 │ │
│ 159 │ if repo_id.count("/") > 1: │
│ 160 │ │ raise HFValidationError( │
│ ❱ 161 │ │ │ "Repo id must be in the form 'repo_name' or 'namespace/rep │
│ 162 │ │ │ f" '{repo_id}'. Use repo_type argument if needed." │
│ 163 │ │ ) │
│ 164 │
╰──────────────────────────────────────────────────────────────────────────────╯
HFValidationError: Repo id must be in the form 'repo_name' or
'namespace/repo_name':
'/home/jn/th/work/Multimodal-GPT/checkpoints/llama-7b_hf'. Use
repo_type argument if needed.
WARNING:torch.distributed.elastic.multiprocessing.api:Sending process 23886 closing signal SIGTERM
ERROR:torch.distributed.elastic.multiprocessing.api:failed (exitcode: 1) local_rank: 0 (pid: 23882) of binary: /opt/conda/bin/python
Traceback (most recent call last):
File "/opt/conda/bin/torchrun", line 8, in
sys.exit(main())
File "/opt/conda/lib/python3.7/site-packages/torch/distributed/elastic/multiprocessing/errors/init.py", line 346, in wrapper
return f(*args, **kwargs)
File "/opt/conda/lib/python3.7/site-packages/torch/distributed/run.py", line 762, in main
run(args)
File "/opt/conda/lib/python3.7/site-packages/torch/distributed/run.py", line 756, in run
)(*cmd_args)
File "/opt/conda/lib/python3.7/site-packages/torch/distributed/launcher/api.py", line 132, in call
return launch_agent(self._config, self._entrypoint, list(args))
File "/opt/conda/lib/python3.7/site-packages/torch/distributed/launcher/api.py", line 248, in launch_agent
failures=result.failures,
torch.distributed.elastic.multiprocessing.errors.ChildFailedError:

/home/jn/th/work/Multimodal-GPT/mmgpt/train/instruction_finetune.py FAILED

Failures:
[1]:
time : 2023-05-30_13:17:48
host : pipeline0420-0
rank : 1 (local_rank: 1)
exitcode : 1 (pid: 23883)
error_file: <N/A>
traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
[2]:
time : 2023-05-30_13:17:48
host : pipeline0420-0
rank : 2 (local_rank: 2)
exitcode : 1 (pid: 23884)
error_file: <N/A>
traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
[3]:
time : 2023-05-30_13:17:48
host : pipeline0420-0
rank : 3 (local_rank: 3)
exitcode : 1 (pid: 23885)
error_file: <N/A>
traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
[4]:
time : 2023-05-30_13:17:48
host : pipeline0420-0
rank : 5 (local_rank: 5)
exitcode : 1 (pid: 23887)
error_file: <N/A>
traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
[5]:
time : 2023-05-30_13:17:48
host : pipeline0420-0
rank : 6 (local_rank: 6)
exitcode : 1 (pid: 23888)
error_file: <N/A>
traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html
[6]:
time : 2023-05-30_13:17:48
host : pipeline0420-0
rank : 7 (local_rank: 7)
exitcode : 1 (pid: 23889)
error_file: <N/A>
traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html

Root Cause (first observed failure):
[0]:
time : 2023-05-30_13:17:48
host : pipeline0420-0
rank : 0 (local_rank: 0)
exitcode : 1 (pid: 23882)
error_file: <N/A>
traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html

@noahc1510
Copy link

Similar error by running "python app.py" and caught this error: "HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': 'checkpoints/llama-7b_hf/'. Use repo_type argument if needed."

@noahc1510
Copy link

Similar error by running "python app.py" and caught this error: "HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': 'checkpoints/llama-7b_hf/'. Use repo_type argument if needed."

I fixed it by replace app.py:359 from llama_path = "checkpoints/llama-7b_hf/" to llama_path = "checkpoints/llama-7b-hf/"

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants