We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
TiedEmbeddingTransformerDecoder
1 parent 543f698 commit 50d3ef1Copy full SHA for 50d3ef1
torchtune/modules/transformer.py
@@ -11,6 +11,7 @@
11
from torch import nn
12
from torchtune.modules import MultiHeadAttention
13
from torchtune.modules.attention_utils import _MaskType
14
+from torchtune.utils._logging import deprecated
15
16
17
class TransformerSelfAttentionLayer(nn.Module):
@@ -619,6 +620,11 @@ def forward(
619
620
return output
621
622
623
+@deprecated(
624
+ msg="Please use torchtune.modules.TransformerDecoder instead. \
625
+If you need an example, see torchtune.models.qwen2._component_builders.py \
626
+on how to use torch.modules.TiedLinear for the output projection."
627
+)
628
class TiedEmbeddingTransformerDecoder(nn.Module):
629
"""
630
Transformer Decoder with tied embedding weight. A key difference between
0 commit comments