diff --git a/1_Pooling/config.json b/1_Pooling/config.json new file mode 100644 index 0000000..0ad909d --- /dev/null +++ b/1_Pooling/config.json @@ -0,0 +1,9 @@ +{ + "word_embedding_dimension": 768, + "pooling_mode_cls_token": false, + "pooling_mode_mean_tokens": true, + "pooling_mode_max_tokens": false, + "pooling_mode_mean_sqrt_len_tokens": false, + "pooling_mode_weightedmean_tokens": false, + "pooling_mode_lasttoken": false +} \ No newline at end of file diff --git a/config.json b/config.json new file mode 100644 index 0000000..cb61ee9 --- /dev/null +++ b/config.json @@ -0,0 +1,56 @@ +{ + "activation_function": "swiglu", + "architectures": [ + "NomicBertModel" + ], + "attn_pdrop": 0.0, + "auto_map": { + "AutoConfig": "nomic-ai/nomic-bert-2048--configuration_hf_nomic_bert.NomicBertConfig", + "AutoModel": "nomic-ai/nomic-bert-2048--modeling_hf_nomic_bert.NomicBertModel", + "AutoModelForMaskedLM": "nomic-ai/nomic-bert-2048--modeling_hf_nomic_bert.NomicBertForPreTraining" + }, + "bos_token_id": null, + "causal": false, + "dense_seq_output": true, + "embd_pdrop": 0.0, + "eos_token_id": null, + "fused_bias_fc": true, + "fused_dropout_add_ln": true, + "initializer_range": 0.02, + "layer_norm_epsilon": 1e-12, + "mlp_fc1_bias": false, + "mlp_fc2_bias": false, + "model_type": "nomic_bert", + "n_embd": 768, + "n_head": 12, + "n_inner": 3072, + "n_layer": 12, + "n_positions": 8192, + "pad_vocab_size_multiple": 64, + "parallel_block": false, + "parallel_block_tied_norm": false, + "prenorm": false, + "qkv_proj_bias": false, + "reorder_and_upcast_attn": false, + "resid_pdrop": 0.0, + "rotary_emb_base": 1000, + "rotary_emb_fraction": 1.0, + "rotary_emb_interleaved": false, + "rotary_emb_scale_base": null, + "rotary_scaling_factor": 2, + "scale_attn_by_inverse_layer_idx": false, + "scale_attn_weights": true, + "summary_activation": null, + "summary_first_dropout": 0.1, + "summary_proj_to_labels": true, + "summary_type": "cls_index", + "summary_use_proj": true, + "torch_dtype": "float32", + "transformers_version": "4.34.0", + "type_vocab_size": 2, + "use_cache": true, + "use_flash_attn": true, + "use_rms_norm": false, + "use_xentropy": true, + "vocab_size": 30528 +} diff --git a/config_sentence_transformers.json b/config_sentence_transformers.json new file mode 100644 index 0000000..a7cfe71 --- /dev/null +++ b/config_sentence_transformers.json @@ -0,0 +1,7 @@ +{ + "__version__": { + "sentence_transformers": "2.4.0.dev0", + "transformers": "4.37.2", + "pytorch": "2.1.0+cu121" + } +} \ No newline at end of file diff --git a/configuration.json b/configuration.json new file mode 100644 index 0000000..e23f0c2 --- /dev/null +++ b/configuration.json @@ -0,0 +1 @@ +{"framework": "pytorch", "task": "text-embedding", "allow_remote": true} \ No newline at end of file