From a25a3cca1caaa761eb59785ce9933c1cb691f667 Mon Sep 17 00:00:00 2001 From: Dejiao Z Date: Tue, 3 Dec 2024 05:00:17 +0000 Subject: [PATCH] fixed config typo --- .ipynb_checkpoints/config-checkpoint.json | 25 +++++++++++++++++++++++ config.json | 2 +- 2 files changed, 26 insertions(+), 1 deletion(-) create mode 100644 .ipynb_checkpoints/config-checkpoint.json diff --git a/.ipynb_checkpoints/config-checkpoint.json b/.ipynb_checkpoints/config-checkpoint.json new file mode 100644 index 0000000..f7cc931 --- /dev/null +++ b/.ipynb_checkpoints/config-checkpoint.json @@ -0,0 +1,25 @@ +{ + "_name_or_path": "codesage/codesage-large-v2", + "architectures": [ + "CodeSage" + ], + "auto_map": { + "AutoConfig": "config_codesage.CodeSageConfig", + "AutoTokenizer": "tokenization_codesage.CodeSageTokenizer", + "AutoModel": "modeling_codesage.CodeSageModel", + "AutoModelForMaskedLM": "modeling_codesage.CodeSageForMaskedLM", + "AutoModelForSequenceClassification": "modeling_codesage.CodeSageForSequenceClassification" + }, + "activation_function": "gelu_new", + "attention_dropout_prob": 0.1, + "embedding_dropout_prob": 0.1, + "initializer_range": 0.02, + "layer_norm_epsilon": 1e-05, + "hidden_size": 1024, + "num_attention_heads": 16, + "num_hidden_layers": 24, + "intermediate_size": 8192, + "max_position_embeddings": 2048, + "residual_dropout_prob": 0.1, + "vocab_size": 49154 +} \ No newline at end of file diff --git a/config.json b/config.json index f7cc931..009c88e 100644 --- a/config.json +++ b/config.json @@ -15,7 +15,7 @@ "embedding_dropout_prob": 0.1, "initializer_range": 0.02, "layer_norm_epsilon": 1e-05, - "hidden_size": 1024, + "hidden_size": 2048, "num_attention_heads": 16, "num_hidden_layers": 24, "intermediate_size": 8192,