diff --git a/combo/models/embeddings.py b/combo/models/embeddings.py
index 49f1ab9d6fa114532927897e93148a793038fdf4..d8e9d7a28a7fa36b60d108a30a3286026a327e51 100644
--- a/combo/models/embeddings.py
+++ b/combo/models/embeddings.py
@@ -107,7 +107,7 @@ class TransformersWordEmbedder(token_embedders.PretrainedTransformerMismatchedEm
 
     def __init__(self,
                  model_name: str,
-                 projection_dim: int,
+                 projection_dim: int = 0,
                  projection_activation: Optional[allen_nn.Activation] = lambda x: x,
                  projection_dropout_rate: Optional[float] = 0.0,
                  freeze_transformer: bool = True,