File tree Expand file tree Collapse file tree 2 files changed +4
-4
lines changed
fastdeploy/model_executor Expand file tree Collapse file tree 2 files changed +4
-4
lines changed Original file line number Diff line number Diff line change 21
21
from .utils import get_tensor
22
22
23
23
24
- class ParallelEmbeddingHiddenStatesProjection (nn .Layer ):
24
+ class ParallelEHProjection (nn .Layer ):
25
25
"""
26
26
"Parallelized Embedding Hidden States Projection.
27
27
"""
@@ -45,7 +45,7 @@ def __init__(
45
45
embedding_dim (int): size of hidden state.
46
46
prefix (str): full name of the layer in the state dict
47
47
"""
48
- super (ParallelEmbeddingHiddenStatesProjection , self ).__init__ ()
48
+ super (ParallelEHProjection , self ).__init__ ()
49
49
self .linear_weight_key = prefix + ".weight"
50
50
if with_bias :
51
51
self .linear_bias_key = prefix + ".bias"
Original file line number Diff line number Diff line change 26
26
from paddleformers .utils .log import logger
27
27
28
28
from fastdeploy .config import FDConfig , ModelConfig
29
- from fastdeploy .model_executor .layers .mtp_linear import ParallelEmbeddingHiddenStatesProjection
29
+ from fastdeploy .model_executor .layers .mtp_linear import ParallelEHProjection
30
30
from fastdeploy .model_executor .layers .normalization import RMSNorm
31
31
from fastdeploy .model_executor .models .ernie4_5_moe import Ernie4_5_DecoderLayer
32
32
from fastdeploy .model_executor .models .model_base import ModelForCasualLM
@@ -286,7 +286,7 @@ def __init__(
286
286
prefix = "ernie.mtp_hidden_norm.0" ,
287
287
)
288
288
289
- self .eh_proj = ParallelEmbeddingHiddenStatesProjection (
289
+ self .eh_proj = ParallelEHProjection (
290
290
fd_config = fd_config ,
291
291
num_embeddings = fd_config .model_config .hidden_size ,
292
292
embedding_dim = fd_config .model_config .hidden_size * 2 ,
You can’t perform that action at this time.
0 commit comments