You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
(baichuan2-13b) root@ubuntu:~/Baichuan2# python cli_demo.py
init model ...
You are using an old version of the checkpointing format that is deprecated (We will also silently ignore gradient_checkpointing_kwargs in case you passed it).Please update to the new format on your modeling file. To use the new format, you need to completely remove the definition of the method _set_gradient_checkpointing in your model.
Loading checkpoint shards: 100%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████| 3/3 [00:13<00:00, 4.51s/it]
We've detected an older driver with an RTX 4000 series GPU. These drivers have issues with P2P. This can affect the multi-gpu inference when using accelerate device_map.Please make sure to update your driver to the latest version which resolves this.
欢迎使用百川大模型,输入进行对话,vim 多行输入,clear 清空历史,CTRL+C 中断生成,stream 开关流式生成,exit 结束。
用户:你好
Baichuan 2:Traceback (most recent call last):
File "/root/Baichuan2/cli_demo.py", line 87, in
main()
File "/root/Baichuan2/cli_demo.py", line 69, in main
for response in model.chat(tokenizer, messages, stream=True):
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/root/.cache/huggingface/modules/transformers_modules/Baichuan2-13B-Chat/modeling_baichuan.py", line 816, in chat
input_ids = build_chat_input(self, tokenizer, messages, generation_config.max_new_tokens)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/root/.cache/huggingface/modules/transformers_modules/Baichuan2-13B-Chat/generation_utils.py", line 31, in build_chat_input
system_tokens = tokenizer.encode(system)
^^^^^^^^^^^^^^^^
AttributeError: 'bool' object has no attribute 'encode'
The text was updated successfully, but these errors were encountered:
(baichuan2-13b) root@ubuntu:~/Baichuan2# python cli_demo.py
init model ...
You are using an old version of the checkpointing format that is deprecated (We will also silently ignore
gradient_checkpointing_kwargs
in case you passed it).Please update to the new format on your modeling file. To use the new format, you need to completely remove the definition of the method_set_gradient_checkpointing
in your model.Loading checkpoint shards: 100%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████| 3/3 [00:13<00:00, 4.51s/it]
We've detected an older driver with an RTX 4000 series GPU. These drivers have issues with P2P. This can affect the multi-gpu inference when using accelerate device_map.Please make sure to update your driver to the latest version which resolves this.
欢迎使用百川大模型,输入进行对话,vim 多行输入,clear 清空历史,CTRL+C 中断生成,stream 开关流式生成,exit 结束。
用户:你好
Baichuan 2:Traceback (most recent call last):
File "/root/Baichuan2/cli_demo.py", line 87, in
main()
File "/root/Baichuan2/cli_demo.py", line 69, in main
for response in model.chat(tokenizer, messages, stream=True):
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/root/.cache/huggingface/modules/transformers_modules/Baichuan2-13B-Chat/modeling_baichuan.py", line 816, in chat
input_ids = build_chat_input(self, tokenizer, messages, generation_config.max_new_tokens)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/root/.cache/huggingface/modules/transformers_modules/Baichuan2-13B-Chat/generation_utils.py", line 31, in build_chat_input
system_tokens = tokenizer.encode(system)
^^^^^^^^^^^^^^^^
AttributeError: 'bool' object has no attribute 'encode'
The text was updated successfully, but these errors were encountered: