We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
好像还没运行就end了 感谢回复,我的运行命令和log如下:
CUDA_VISIBLE_DEVICES=0 swift export --ckpt_dir /wangliyao/swift/output/deepseek-vl-1_3b-chat/v4-20240708-172229/checkpoint-2-merged --load_dataset_config true --custom_val_dataset_path toy_dataset.json
run sh: python /root/miniconda3/envs/sd/lib/python3.10/site-packages/swift/cli/export.py --ckpt_dir /wangliyao/swift/output/deepseek-vl-1_3b-chat/v4-20240708-172229/checkpoint-2-merged --load_dataset_config true --custom_val_dataset_path toy_dataset.json [INFO:swift] Successfully registered /root/miniconda3/envs/sd/lib/python3.10/site-packages/swift/llm/data/dataset_info.json [INFO:swift] Start time of running main: 2024-07-08 18:25:36.670228 [INFO:swift] Using val_dataset, ignoring dataset_test_ratio [INFO:swift] ckpt_dir: /wangliyao/swift/output/deepseek-vl-1_3b-chat/v4-20240708-172229/checkpoint-2-merged [INFO:swift] Setting model_info['revision']: master [INFO:swift] Setting self.eval_human: False [INFO:swift] Setting overwrite_generation_config: True [INFO:swift] args: ExportArguments(model_type='deepseek-vl-1_3b-chat', model_id_or_path='deepseek-ai/deepseek-vl-1.3b-chat', model_revision='master', sft_type='full', template_type='deepseek-vl', infer_backend='pt', ckpt_dir='/wangliyao/swift/output/deepseek-vl-1_3b-chat/v4-20240708-172229/checkpoint-2-merged', load_args_from_ckpt_dir=True, load_dataset_config=True, eval_human=False, seed=42, dtype='bf16', dataset=['toy_dataset.json'], val_dataset=['toy_dataset.json'], dataset_seed=42, dataset_test_ratio=0.0, show_dataset_sample=10, save_result=True, system='You are a helpful language and vision assistant. You are able to understand the visual content that the user provides, and assist the user with a variety of tasks using natural language.', tools_prompt='react_en', max_length=None, truncation_strategy='delete', check_dataset_strategy='none', model_name=[None, None], model_author=[None, None], quant_method='awq', quantization_bit=0, hqq_axis=0, hqq_dynamic_config_path=None, bnb_4bit_comp_dtype='bf16', bnb_4bit_quant_type='nf4', bnb_4bit_use_double_quant=True, bnb_4bit_quant_storage=None, max_new_tokens=2048, do_sample=True, temperature=0.3, top_k=20, top_p=0.7, repetition_penalty=1.0, num_beams=1, stop_words=None, rope_scaling=None, use_flash_attn=None, ignore_args_error=False, stream=True, merge_lora=False, merge_device_map='auto', save_safetensors=True, overwrite_generation_config=True, verbose=None, local_repo_path=None, custom_register_path=None, custom_dataset_info=None, device_map_config_path=None, gpu_memory_utilization=0.9, tensor_parallel_size=1, max_model_len=None, disable_custom_all_reduce=True, enforce_eager=False, vllm_enable_lora=False, vllm_max_lora_rank=16, lora_modules=[], image_input_shape=None, image_feature_size=None, self_cognition_sample=0, train_dataset_sample=-1, val_dataset_sample=None, safe_serialization=None, model_cache_dir=None, merge_lora_and_save=None, custom_train_dataset_path=[], custom_val_dataset_path=['toy_dataset.json'], vllm_lora_modules=None, to_peft_format=False, quant_bits=0, quant_n_samples=256, quant_seqlen=2048, quant_device_map='cpu', quant_output_dir=None, push_to_hub=False, hub_model_id=None, hub_token=None, hub_private_repo=False, commit_message='update files') [INFO:swift] Global seed set to 42 [INFO:swift] End time of running main: 2024-07-08 18:25:36.710787
python /root/miniconda3/envs/sd/lib/python3.10/site-packages/swift/cli/export.py --ckpt_dir /wangliyao/swift/output/deepseek-vl-1_3b-chat/v4-20240708-172229/checkpoint-2-merged --load_dataset_config true --custom_val_dataset_path toy_dataset.json
/root/miniconda3/envs/sd/lib/python3.10/site-packages/swift/llm/data/dataset_info.json
The text was updated successfully, but these errors were encountered:
推理请使用swift infer
swift infer
Sorry, something went wrong.
No branches or pull requests
好像还没运行就end了
感谢回复,我的运行命令和log如下:
CUDA_VISIBLE_DEVICES=0 swift export --ckpt_dir /wangliyao/swift/output/deepseek-vl-1_3b-chat/v4-20240708-172229/checkpoint-2-merged --load_dataset_config true --custom_val_dataset_path toy_dataset.json
run sh:
python /root/miniconda3/envs/sd/lib/python3.10/site-packages/swift/cli/export.py --ckpt_dir /wangliyao/swift/output/deepseek-vl-1_3b-chat/v4-20240708-172229/checkpoint-2-merged --load_dataset_config true --custom_val_dataset_path toy_dataset.json
[INFO:swift] Successfully registered
/root/miniconda3/envs/sd/lib/python3.10/site-packages/swift/llm/data/dataset_info.json
[INFO:swift] Start time of running main: 2024-07-08 18:25:36.670228
[INFO:swift] Using val_dataset, ignoring dataset_test_ratio
[INFO:swift] ckpt_dir: /wangliyao/swift/output/deepseek-vl-1_3b-chat/v4-20240708-172229/checkpoint-2-merged
[INFO:swift] Setting model_info['revision']: master
[INFO:swift] Setting self.eval_human: False
[INFO:swift] Setting overwrite_generation_config: True
[INFO:swift] args: ExportArguments(model_type='deepseek-vl-1_3b-chat', model_id_or_path='deepseek-ai/deepseek-vl-1.3b-chat', model_revision='master', sft_type='full', template_type='deepseek-vl', infer_backend='pt', ckpt_dir='/wangliyao/swift/output/deepseek-vl-1_3b-chat/v4-20240708-172229/checkpoint-2-merged', load_args_from_ckpt_dir=True, load_dataset_config=True, eval_human=False, seed=42, dtype='bf16', dataset=['toy_dataset.json'], val_dataset=['toy_dataset.json'], dataset_seed=42, dataset_test_ratio=0.0, show_dataset_sample=10, save_result=True, system='You are a helpful language and vision assistant. You are able to understand the visual content that the user provides, and assist the user with a variety of tasks using natural language.', tools_prompt='react_en', max_length=None, truncation_strategy='delete', check_dataset_strategy='none', model_name=[None, None], model_author=[None, None], quant_method='awq', quantization_bit=0, hqq_axis=0, hqq_dynamic_config_path=None, bnb_4bit_comp_dtype='bf16', bnb_4bit_quant_type='nf4', bnb_4bit_use_double_quant=True, bnb_4bit_quant_storage=None, max_new_tokens=2048, do_sample=True, temperature=0.3, top_k=20, top_p=0.7, repetition_penalty=1.0, num_beams=1, stop_words=None, rope_scaling=None, use_flash_attn=None, ignore_args_error=False, stream=True, merge_lora=False, merge_device_map='auto', save_safetensors=True, overwrite_generation_config=True, verbose=None, local_repo_path=None, custom_register_path=None, custom_dataset_info=None, device_map_config_path=None, gpu_memory_utilization=0.9, tensor_parallel_size=1, max_model_len=None, disable_custom_all_reduce=True, enforce_eager=False, vllm_enable_lora=False, vllm_max_lora_rank=16, lora_modules=[], image_input_shape=None, image_feature_size=None, self_cognition_sample=0, train_dataset_sample=-1, val_dataset_sample=None, safe_serialization=None, model_cache_dir=None, merge_lora_and_save=None, custom_train_dataset_path=[], custom_val_dataset_path=['toy_dataset.json'], vllm_lora_modules=None, to_peft_format=False, quant_bits=0, quant_n_samples=256, quant_seqlen=2048, quant_device_map='cpu', quant_output_dir=None, push_to_hub=False, hub_model_id=None, hub_token=None, hub_private_repo=False, commit_message='update files')
[INFO:swift] Global seed set to 42
[INFO:swift] End time of running main: 2024-07-08 18:25:36.710787
The text was updated successfully, but these errors were encountered: