diff --git a/combo/models/embeddings.py b/combo/models/embeddings.py index 49f1ab9d6fa114532927897e93148a793038fdf4..d8e9d7a28a7fa36b60d108a30a3286026a327e51 100644 --- a/combo/models/embeddings.py +++ b/combo/models/embeddings.py @@ -107,7 +107,7 @@ class TransformersWordEmbedder(token_embedders.PretrainedTransformerMismatchedEm def __init__(self, model_name: str, - projection_dim: int, + projection_dim: int = 0, projection_activation: Optional[allen_nn.Activation] = lambda x: x, projection_dropout_rate: Optional[float] = 0.0, freeze_transformer: bool = True,