@@ -1549,13 +1549,13 @@ def get_default_conv_template(self, model_path: str) -> Conversation:
1549
1549
return get_conv_template ("llama2-chinese" )
1550
1550
1551
1551
1552
- class VigogneInstructAdapter (BaseModelAdapter ):
1553
- """The model adapter for Vigogne-Instruct (e.g., bofenghuang/vigogne-2-7b-instruct )"""
1552
+ class VigogneAdapter (BaseModelAdapter ):
1553
+ """The model adapter for vigogne (e.g., bofenghuang/vigogne-2-7b-chat )"""
1554
1554
1555
1555
use_fast_tokenizer = False
1556
1556
1557
1557
def match (self , model_path : str ):
1558
- return "vigogne" in model_path . lower () and "instruct" in model_path . lower ( )
1558
+ return bool ( re . search ( r"vigogne|vigostral" , model_path , re . I ) )
1559
1559
1560
1560
def load_model (self , model_path : str , from_pretrained_kwargs : dict ):
1561
1561
revision = from_pretrained_kwargs .get ("revision" , "main" )
@@ -1574,35 +1574,11 @@ def load_model(self, model_path: str, from_pretrained_kwargs: dict):
1574
1574
return model , tokenizer
1575
1575
1576
1576
def get_default_conv_template (self , model_path : str ) -> Conversation :
1577
- return get_conv_template ("alpaca" )
1578
-
1579
-
1580
- class VigogneChatAdapter (BaseModelAdapter ):
1581
- """The model adapter for Vigogne-Chat (e.g., bofenghuang/vigogne-7b-chat)"""
1582
-
1583
- use_fast_tokenizer = False
1584
-
1585
- def match (self , model_path : str ):
1586
- return "vigogne" in model_path .lower () and "chat" in model_path .lower ()
1587
-
1588
- def load_model (self , model_path : str , from_pretrained_kwargs : dict ):
1589
- revision = from_pretrained_kwargs .get ("revision" , "main" )
1590
- tokenizer = AutoTokenizer .from_pretrained (
1591
- model_path ,
1592
- use_fast = self .use_fast_tokenizer ,
1593
- trust_remote_code = True ,
1594
- revision = revision ,
1595
- )
1596
- model = AutoModelForCausalLM .from_pretrained (
1597
- model_path ,
1598
- trust_remote_code = True ,
1599
- low_cpu_mem_usage = True ,
1600
- ** from_pretrained_kwargs ,
1601
- ).eval ()
1602
- return model , tokenizer
1603
-
1604
- def get_default_conv_template (self , model_path : str ) -> Conversation :
1605
- return get_conv_template ("vigogne-chat" )
1577
+ if "chat" in model_path .lower ():
1578
+ if "vigostral" in model_path .lower ():
1579
+ return get_conv_template ("vigogne_chat_v3" )
1580
+ return get_conv_template ("vigogne_chat_v2" )
1581
+ return get_conv_template ("vigogne_instruct" )
1606
1582
1607
1583
1608
1584
class OpenLLaMaOpenInstructAdapter (BaseModelAdapter ):
@@ -1746,8 +1722,7 @@ def get_default_conv_template(self, model_path: str) -> Conversation:
1746
1722
register_model_adapter (BGEAdapter )
1747
1723
register_model_adapter (E5Adapter )
1748
1724
register_model_adapter (Lamma2ChineseAdapter )
1749
- register_model_adapter (VigogneInstructAdapter )
1750
- register_model_adapter (VigogneChatAdapter )
1725
+ register_model_adapter (VigogneAdapter )
1751
1726
register_model_adapter (OpenLLaMaOpenInstructAdapter )
1752
1727
register_model_adapter (ReaLMAdapter )
1753
1728
register_model_adapter (PhindCodeLlamaAdapter )
0 commit comments