Skip to content

Commit 55efa3d

Browse files
authored
[chatglm] rename chatglm to causallm (#6444)
* change AutoModelForCalsalLM * update
1 parent 7a3e379 commit 55efa3d

File tree

13 files changed

+50
-80
lines changed

13 files changed

+50
-80
lines changed

examples/benchmark/peft/paddle/benchmark.py

Lines changed: 11 additions & 28 deletions
Original file line numberDiff line numberDiff line change
@@ -22,11 +22,7 @@
2222
from paddlenlp.data import DataCollatorForSeq2Seq
2323
from paddlenlp.peft import LoRAConfig, LoRAModel
2424
from paddlenlp.trainer import PdArgumentParser, TrainingArguments
25-
from paddlenlp.transformers import (
26-
AutoModelForCausalLM,
27-
AutoTokenizer,
28-
ChatGLMForConditionalGeneration,
29-
)
25+
from paddlenlp.transformers import AutoModelForCausalLM, AutoTokenizer
3026

3127
"""
3228
单卡
@@ -83,29 +79,16 @@ def main():
8379
tokenizer = AutoTokenizer.from_pretrained(model_args.model_name_or_path)
8480
if "llama" in model_args.model_name_or_path:
8581
tokenizer.pad_token = tokenizer.unk_token
86-
if "chatglm" in model_args.model_name_or_path:
87-
model = ChatGLMForConditionalGeneration.from_pretrained(
88-
model_args.model_name_or_path,
89-
load_state_as_np=True,
90-
low_cpu_mem_usage=True,
91-
# use_flash_attention=True,
92-
dtype=dtype,
93-
tensor_parallel_degree=training_args.tensor_parallel_degree,
94-
tensor_parallel_rank=training_args.tensor_parallel_rank,
95-
recompute=training_args.recompute,
96-
)
97-
98-
else:
99-
model = AutoModelForCausalLM.from_pretrained(
100-
model_args.model_name_or_path,
101-
load_state_as_np=True,
102-
low_cpu_mem_usage=True,
103-
# use_flash_attention=True,
104-
dtype=dtype,
105-
tensor_parallel_degree=training_args.tensor_parallel_degree,
106-
tensor_parallel_rank=training_args.tensor_parallel_rank,
107-
use_recompute=training_args.recompute,
108-
)
82+
model = AutoModelForCausalLM.from_pretrained(
83+
model_args.model_name_or_path,
84+
load_state_as_np=True,
85+
low_cpu_mem_usage=True,
86+
# use_flash_attention=True,
87+
dtype=dtype,
88+
tensor_parallel_degree=training_args.tensor_parallel_degree,
89+
tensor_parallel_rank=training_args.tensor_parallel_rank,
90+
use_recompute=training_args.recompute,
91+
)
10992

11093
if model_args.lora:
11194
if "llama" in model_args.model_name_or_path:

llm/chatglm/export_generation_model.py

Lines changed: 2 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -18,11 +18,7 @@
1818
import paddle
1919

2020
from paddlenlp.peft import LoRAConfig, LoRAModel
21-
from paddlenlp.transformers import (
22-
ChatGLMConfig,
23-
ChatGLMForConditionalGeneration,
24-
ChatGLMTokenizer,
25-
)
21+
from paddlenlp.transformers import ChatGLMConfig, ChatGLMForCausalLM, ChatGLMTokenizer
2622

2723

2824
def parse_args():
@@ -61,9 +57,7 @@ def main():
6157
config = ChatGLMConfig.from_pretrained(args.model_name_or_path)
6258
dtype = config.dtype if config.dtype is not None else config.paddle_dtype
6359

64-
model = ChatGLMForConditionalGeneration.from_pretrained(
65-
args.model_name_or_path, load_state_as_np=True, dtype=dtype
66-
)
60+
model = ChatGLMForCausalLM.from_pretrained(args.model_name_or_path, load_state_as_np=True, dtype=dtype)
6761
if args.lora_path is not None:
6862
model = LoRAModel.from_pretrained(model, args.lora_path)
6963

llm/chatglm/finetune_generation.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -32,7 +32,7 @@
3232
chatglm_v2_pad_attention_mask,
3333
)
3434
from paddlenlp.trainer import PdArgumentParser, TrainingArguments, get_last_checkpoint
35-
from paddlenlp.transformers import AutoModelForConditionalGeneration, AutoTokenizer
35+
from paddlenlp.transformers import AutoModelForCausalLM, AutoTokenizer
3636
from paddlenlp.utils.log import logger
3737

3838

@@ -101,7 +101,7 @@ def main():
101101
dtype = "float16"
102102

103103
# Load the pretrained language model.
104-
model = AutoModelForConditionalGeneration.from_pretrained(
104+
model = AutoModelForCausalLM.from_pretrained(
105105
model_args.model_name_or_path,
106106
dtype=dtype,
107107
low_cpu_mem_usage=True,

llm/chatglm/predict_generation.py

Lines changed: 2 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -21,11 +21,7 @@
2121
chatglm_pad_attention_mask,
2222
chatglm_postprocess_past_key_value,
2323
)
24-
from paddlenlp.transformers import (
25-
ChatGLMConfig,
26-
ChatGLMForConditionalGeneration,
27-
ChatGLMTokenizer,
28-
)
24+
from paddlenlp.transformers import ChatGLMConfig, ChatGLMForCausalLM, ChatGLMTokenizer
2925

3026

3127
def parse_arguments():
@@ -96,7 +92,7 @@ def __init__(self, args=None, tokenizer=None, model=None, **kwargs):
9692
config = ChatGLMConfig.from_pretrained(args.model_name_or_path)
9793
dtype = config.dtype if config.dtype is not None else config.paddle_dtype
9894

99-
self.model = ChatGLMForConditionalGeneration.from_pretrained(
95+
self.model = ChatGLMForCausalLM.from_pretrained(
10096
args.model_name_or_path,
10197
tensor_parallel_degree=tensor_parallel_degree,
10298
tensor_parallel_rank=tensor_parallel_rank,

llm/chatglm/test_modeling.py

Lines changed: 6 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,7 @@
2525
def run_mp_paddle():
2626
import paddle
2727

28-
from paddlenlp.transformers import ChatGLMForConditionalGeneration
28+
from paddlenlp.transformers import ChatGLMForCausalLM
2929

3030
tensor_parallel_degree = paddle.distributed.get_world_size()
3131
strategy = paddle.distributed.fleet.DistributedStrategy()
@@ -42,7 +42,7 @@ def run_mp_paddle():
4242
tensor_parallel_rank = mp_group.rank
4343

4444
paddle.set_default_dtype("float32")
45-
model = ChatGLMForConditionalGeneration.from_pretrained(
45+
model = ChatGLMForCausalLM.from_pretrained(
4646
"torch_cache", # "/root/paddlejob/workspace/GLM/ChatGLM-6B/",
4747
load_state_as_np=True,
4848
low_cpu_mem_usage=True,
@@ -59,10 +59,10 @@ def run_mp_paddle():
5959
def run_paddle():
6060
import paddle
6161

62-
from paddlenlp.transformers import ChatGLMForConditionalGeneration
62+
from paddlenlp.transformers import ChatGLMForCausalLM
6363

6464
paddle.set_default_dtype("float32")
65-
model = ChatGLMForConditionalGeneration.from_pretrained(
65+
model = ChatGLMForCausalLM.from_pretrained(
6666
"torch_cache", # "/root/paddlejob/workspace/GLM/ChatGLM-6B/",
6767
load_state_as_np=True,
6868
low_cpu_mem_usage=True,
@@ -87,11 +87,10 @@ def run_torch():
8787

8888
def run_generate():
8989
import paddle
90-
91-
from modeling import ChatGLMForConditionalGeneration
90+
from modeling import ChatGLMForCausalLM
9291

9392
paddle.set_default_dtype("float16")
94-
model = ChatGLMForConditionalGeneration.from_pretrained(
93+
model = ChatGLMForCausalLM.from_pretrained(
9594
"torch_cache", load_state_as_np=True, low_cpu_mem_usage=True, dtype="float16"
9695
)
9796
model.eval()

paddlenlp/taskflow/text2text_generation.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@
1414

1515
import paddle
1616

17-
from ..transformers import ChatGLMForConditionalGeneration, ChatGLMTokenizer
17+
from ..transformers import ChatGLMForCausalLM, ChatGLMTokenizer
1818
from ..utils.log import logger
1919
from .task import Task
2020
from .utils import static_mode_guard
@@ -115,7 +115,7 @@ def _construct_model(self, model):
115115
"""
116116
Construct the inference model for the predictor.
117117
"""
118-
model_instance = ChatGLMForConditionalGeneration.from_pretrained(
118+
model_instance = ChatGLMForCausalLM.from_pretrained(
119119
self.model,
120120
load_state_as_np=True,
121121
dtype=self._dtype,

paddlenlp/transformers/chatglm/modeling.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -38,7 +38,7 @@
3838
__all__ = [
3939
"ChatGLMModel",
4040
"ChatGLMPretrainedModel",
41-
"ChatGLMForConditionalGeneration",
41+
"ChatGLMForCausalLM",
4242
]
4343

4444

@@ -743,12 +743,12 @@ def forward(
743743
return BaseModelOutputWithPastAndCrossAttentions(last_hidden_state=logits, past_key_values=new_caches)
744744

745745

746-
class ChatGLMForConditionalGeneration(ChatGLMPretrainedModel):
746+
class ChatGLMForCausalLM(ChatGLMPretrainedModel):
747747
_keys_to_ignore_on_save = [r"lm_head.weight"]
748748
_tied_weights_keys = ["lm_head.weight"]
749749

750750
def __init__(self, config: ChatGLMConfig):
751-
super(ChatGLMForConditionalGeneration, self).__init__(config)
751+
super(ChatGLMForCausalLM, self).__init__(config)
752752

753753
self.config = config
754754
self.max_sequence_length = config.max_sequence_length

paddlenlp/transformers/chatglm_v2/modeling.py

Lines changed: 7 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -30,6 +30,12 @@
3030
)
3131
from .configuration import CHATGLM_V2_PRETRAINED_RESOURCE_FILES_MAP, ChatGLMv2Config
3232

33+
__all__ = [
34+
"ChatGLMv2Model",
35+
"ChatGLMv2PretrainedModel",
36+
"ChatGLMv2ForCausalLM",
37+
]
38+
3339

3440
def assign_kv_heads(num_kv_heads, num_gpus):
3541
# Initialize the assignment list
@@ -745,7 +751,7 @@ def forward(
745751
)
746752

747753

748-
class ChatGLMv2ForConditionalGeneration(ChatGLMv2PretrainedModel):
754+
class ChatGLMv2ForCausalLM(ChatGLMv2PretrainedModel):
749755
def __init__(self, config: ChatGLMv2Config):
750756
super().__init__(config)
751757
self.max_sequence_length = config.max_sequence_length

paddlenlp/transformers/visualglm/modeling.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@
2727
from ...utils.initializer import normal_, ones_, zeros_
2828
from ..activations import ACT2FN
2929
from ..chatglm.configuration import ChatGLMConfig
30-
from ..chatglm.modeling import ChatGLMForConditionalGeneration
30+
from ..chatglm.modeling import ChatGLMForCausalLM
3131
from ..model_outputs import (
3232
BaseModelOutput,
3333
BaseModelOutputWithPastAndCrossAttentions,
@@ -1163,7 +1163,7 @@ def __init__(self, config: VisualGLMConfig):
11631163
self.qformer = VisualGLMQFormerModel(config.qformer_config)
11641164

11651165
self.language_projection = nn.Linear(config.qformer_config.hidden_size, config.text_config.hidden_size)
1166-
self.language_model = ChatGLMForConditionalGeneration(config.text_config)
1166+
self.language_model = ChatGLMForCausalLM(config.text_config)
11671167

11681168
def get_input_embeddings(self) -> nn.Layer:
11691169
return self.vision_model.embeddings.patch_embedding
@@ -1413,7 +1413,7 @@ def forward(
14131413
)
14141414

14151415

1416-
class ChatGLMForConditionalGenerationWithImage(ChatGLMForConditionalGeneration):
1416+
class ChatGLMForConditionalGenerationWithImage(ChatGLMForCausalLM):
14171417
def __init__(self, config: ChatGLMConfig):
14181418
super(ChatGLMForConditionalGenerationWithImage, self).__init__(config)
14191419
self.config = config

tests/peft/test_prefix.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,7 @@
2525
)
2626
from paddlenlp.transformers import (
2727
ChatGLMv2Config,
28-
ChatGLMv2ForConditionalGeneration,
28+
ChatGLMv2ForCausalLM,
2929
LlamaConfig,
3030
LlamaForCausalLM,
3131
)
@@ -116,7 +116,7 @@ def setUpClass(cls):
116116
dtype="float32",
117117
)
118118

119-
cls.model = ChatGLMv2ForConditionalGeneration(cls.config)
119+
cls.model = ChatGLMv2ForCausalLM(cls.config)
120120
cls.prefix_config = PrefixConfig(
121121
num_prefix_tokens=2,
122122
num_attention_heads=cls.model.config.num_attention_heads,

0 commit comments

Comments
 (0)