|
21 | 21 | from paddle import Tensor |
22 | 22 | from paddle.nn import Layer |
23 | 23 |
|
24 | | -from paddlenlp.utils.env import CONFIG_NAME |
25 | | - |
| 24 | +from ...utils.env import CONFIG_NAME |
26 | 25 | from ...utils.log import logger |
27 | 26 | from .. import PretrainedModel, register_base_model |
28 | 27 | from ..model_outputs import ( |
@@ -295,16 +294,14 @@ def init_weights(self, layer): |
295 | 294 | layer.weight.set_value( |
296 | 295 | paddle.tensor.normal( |
297 | 296 | mean=0.0, |
298 | | - std=self.initializer_range |
299 | | - if hasattr(self, "initializer_range") |
300 | | - else self.transformer.config.initializer_range, |
| 297 | + std=self.config.initializer_range, |
301 | 298 | shape=layer.weight.shape, |
302 | 299 | ) |
303 | 300 | ) |
304 | 301 | elif isinstance(layer, nn.LayerNorm): |
305 | 302 | layer.bias.set_value(paddle.zeros_like(layer.bias)) |
306 | 303 | layer.weight.set_value(paddle.full_like(layer.weight, 1.0)) |
307 | | - layer._epsilon = getattr(self, "layer_norm_epsilon", 1e-05) |
| 304 | + layer._epsilon = self.config.layer_norm_epsilon |
308 | 305 | if isinstance(layer, nn.Linear) and layer.bias is not None: |
309 | 306 | layer.bias.set_value(paddle.zeros_like(layer.bias)) |
310 | 307 |
|
@@ -696,7 +693,4 @@ def __getattr__(self, name): |
696 | 693 | try: |
697 | 694 | return super().__getattr__(name) |
698 | 695 | except AttributeError: |
699 | | - try: |
700 | | - return getattr(getattr(self, self.base_model_prefix), name) |
701 | | - except AttributeError: |
702 | | - return getattr(getattr(self, self.base_model_prefix).config, name) |
| 696 | + return getattr(getattr(self, self.base_model_prefix), name) |
0 commit comments