PGµç¾º

±¾Õ¾µãʹÓÃCookies£¬¼ÌÐøä¯ÀÀ±íʾÄúͬÒâÎÒÃÇʹÓÃCookies¡£ CookiesºÍÒþ˽Õþ²ß>

¼òÌåÖÐÎÄ
English
Ê×Ò³ > ¹ØÓÚÎÒÃÇ > ÐÂÎÅÖÐÐÄ > ħ´îÉçÇøÉÏÏß¡°Ô´2.0¡±´óÄ£ÐÍchatÌåÑé½Ó¿Ú£¡

ħ´îÉçÇøÉÏÏß¡°Ô´2.0¡±´óÄ£ÐÍchatÌåÑé½Ó¿Ú£¡

2024-01-19

½üÈÕ£¬PGµç¾ºÐÅÏ¢¿ªÔ´¿ÉÉÌÓõĴóÄ£ÐÍ¡°Ô´2.0¡±È«ÃæÉÏÏßħ´îÉçÇø£¬²¢½«³ÖÐøÔÚÉçÇø½øÐÐÄ£ÐͰ汾µü´ú¸üС£

Ŀǰ£¬Óû§¿ÉÔÚħ´îÉçÇøÏÂÔØ102B¡¢51B¡¢2BÈýÖÖ²ÎÊýµÄ¡°Ô´¡±´óÄ£ÐÍ×îа汾£¬Ò²¿ÉÒÔʹÓÃħ´îÉçÇøÌṩµÄYuan2-2B-Janus-ChatbotÌåÑé½Ó¿Ú£¬Ö±½ÓÌåÑé¡°Ô´¡±2B´óÄ£ÐÍÄÜÁ¦¡£Í¬Ê±£¬ÔÚħ´îÉçÇø¿ÉÒÔʹÓÃÔÆËãÁ¦¶Ô¡°Ô´¡±´óÄ£ÐͽøÐпìËÙ¿ª·¢ºÍ²¿Êð¡£

1.png

£¨×¢£ºÃüÃûÖÐhf´ú±íHuggingface¸ñʽ£¬JanusΪ×îа汾Ãû£©

ħ´îÉçÇøÕë¶Ô¡°Ô´¡±´óÄ£ÐÍÌṩÁËÄ£Ð͵IJ¿Êð¡¢ÍÆÀíºÍ΢µ÷µÄ×î¼Ñʵ¼ù£¬¹©¿ª·¢Õ߲ο¼¡£

GitHub¿ªÔ´µØÖ·£º

https://github.com/IEIT-Yuan/Yuan-2.0

ModelScopeÄ£ÐÍ¿ªÔ´µØÖ·£º

https://modelscope.cn/models/YuanLLM/Yuan2-2B-Janus

ModelScope´´¿Õ¼äÔ´chatÌåÑ鵨ַ£º

https://modelscope.cn/studios/YuanLLM/yuan/summary

СģÐÍÓдóÄÜÁ¿

ÔÚħ´î´´¿Õ¼ä£¬CPU¾Í¿ÉÒÔÔËÐÐhost Yuan2-2B-Janus-ChatÄ£ÐÍ¡£

Yuan2-2B-Janus-ChatÄ£ÐÍËäÈ»³ß´çС£¬µ«ÊÇÒѾ­¾ß±¸±È½ÏºÃµÄÎÄѧ´´×÷ÄÜÁ¦£º

2.jpg

³£Ê¶ÎÊ´ðÄÜÁ¦£º

3.jpg

ÊýѧPGµç¾º¹ÙÍøÄÜÁ¦£º

4.jpg

´úÂëÄÜÁ¦£º

5.jpg

6.jpg

640.png

Ä£ÐÍÍÆÀí

ÔÚModelScopeÉçÇøµÄÃâ·ÑËãÁ¦ÉÏ¿ÉÒÔʵÏÖYuan2-2B-Janus-hfÄ£Ð͵ÄÍÆÀí

Ä£ÐÍÁ´½Ó£º

Yuan2-2B-Janus-hf£º

https://modelscope.cn/models/YuanLLM/Yuan2-2B-Janus-hf/summary

ÍÆÀí´úÂ룺

import torch

import os

from modelscope import AutoModelForCausalLM, AutoTokenizer

print("Creat tokenizer...")

tokenizer = AutoTokenizer.from_pretrained('YuanLLM/Yuan2-2B-Janus-hf', 

add_eos_token=False, add_bos_token=False, eos_token='<eod>')

tokenizer.add_tokens(['<sep>', '<pad>', '<mask>', '<predict>', '<FIM_SUFFIX>', '<FIM_PREFIX>', '<FIM_MIDDLE>','<commit_before>','<commit_msg>','<commit_after>','<jupyter_start>','<jupyter_text>','<jupyter_code>','<jupyter_output>','<empty_output>'], special_tokens=True)

print("Creat model...")

model = AutoModelForCausalLM.from_pretrained('YuanLLM/Yuan2-2B-Janus-hf',

device_map='auto', torch_dtype=torch.bfloat16, trust_remote_code=True)

inputs = tokenizer("ÇëÎÊĿǰ×îÏȽøµÄ»úPGµç¾º¹ÙÍøÑ§Ï°Ëã·¨ÓÐÄÄЩ£¿",

 return_tensors="pt")["input_ids"].to("cuda:0")

outputs = model.generate(inputs,do_sample=False,max_length=100)

print(tokenizer.decode(outputs[0]))

ÏÔ´æÕ¼Óãº

7.png

Ä£ÐÍ΢µ÷

ʹÓÃSWIFT¶Ô Yuan2-2B-Janus-Chat ½øÐÐ΢µ÷, ½â¾ö·ÖÀàÎÊÌâ¡£

ÎÒÃÇʹÓõÄÊý¾Ý¼¯hc3-zh°üº¬Á˸ø¶¨ÎÊÌâµÄÈËÀà-ChatGPT»Ø´ðÎı¾¶Ô, ͨ¹ýÕâ¸öÊý¾Ý¼¯£¬¿ÉÒÔѵÁ·Ò»¸öÇø·Ö¶ÔÓ¦»Ø´ðÊÇÀ´×ÔÈËÀ໹ÊÇChatGPTµÄ·ÖÀàÄ£ÐÍ¡£

´úÂ뿪ԴµØÖ·: 

https://github.com/modelscope/swift

΢µ÷½Å±¾:

´úÂë1.png

ѵÁ·¹ý³ÌÒ²Ö§³Ö±¾µØÊý¾Ý¼¯£¬ÐèÒªÖ¸¶¨ÈçϲÎÊý£º

´úÂë2.png

×Ô¶¨ÒåÊý¾Ý¼¯µÄ¸ñʽ¿ÉÒԲο¼£º

https://github.com/modelscope/swift/blob/main/docs/source/LLM/×Ô¶¨ÒåÓëÍØÕ¹.md#×¢²áÊý¾Ý¼¯µÄ·½Ê½

΢µ÷ºóÍÆÀí½Å±¾: £¨ÕâÀïµÄckpt_dirÐèÒªÐÞ¸ÄΪѵÁ·Éú³ÉµÄcheckpointÎļþ¼Ð£©

´úÂë3.png

΢µ÷ºóÉú³ÉÑùÀý£º

ÑùÀý1£ºÄ£ÐÍÕýÈ·Çø·ÖQAÖеĻشðÄÚÈÝÓÉChatGPTÉú³É

8.jpg

ÑùÀý2£ºÄ£ÐÍÕýÈ·Çø·ÖQAÖеĻشðÄÚÈÝÓÉHumanÉú³É

9.jpg

ÊÛǰ×Éѯ

ÊÛºó·þÎñ

Òâ¼û·´À¡

AIStore

»Øµ½¶¥²¿

»Øµ½¶¥²¿

ÊÕÆð
»Øµ½¶¥²¿ »Øµ½¶¥²¿
ÇëÑ¡Ôñ·þÎñÏîÄ¿
ÊÛǰ×Éѯ
ÊÛºó·þÎñ
·ÃÎÊ AIStore

ɨÂë·ÃÎÊAIStore

¡¾ÍøÕ¾µØÍ¼¡¿¡¾sitemap¡¿