From 6f08329513eede8dd80d30b75f288014fcdb9cbc Mon Sep 17 00:00:00 2001 From: xxl <505279206@qq.com> Date: Fri, 10 Jan 2025 10:32:48 +0800 Subject: [PATCH] first commit --- config.json | 36 ++++++++++++++++++++++++++++++++++++ configuration.json | 1 + generation_config.json | 6 ++++++ 3 files changed, 43 insertions(+) create mode 100644 config.json create mode 100644 configuration.json create mode 100644 generation_config.json diff --git a/config.json b/config.json new file mode 100644 index 0000000..c5563ae --- /dev/null +++ b/config.json @@ -0,0 +1,36 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": 128000, + "eos_token_id": 128001, + "hidden_act": "silu", + "hidden_size": 3072, + "initializer_range": 0.02, + "intermediate_size": 8192, + "is_llama_config": true, + "max_position_embeddings": 4096, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 24, + "num_hidden_layers": 28, + "num_key_value_heads": 8, + "pretraining_tp": 2, + "rms_norm_eps": 1e-05, + "rope_interleaved": false, + "rope_scaling": { + "factor": 32.0, + "high_freq_factor": 4.0, + "low_freq_factor": 1.0, + "original_max_position_embeddings": 8192, + "rope_type": "llama3" + }, + "rope_theta": 500000.0, + "tie_word_embeddings": true, + "torch_dtype": "bfloat16", + "transformers_version": "4.44.0", + "use_cache": true, + "vocab_size": 128256 +} diff --git a/configuration.json b/configuration.json new file mode 100644 index 0000000..159097f --- /dev/null +++ b/configuration.json @@ -0,0 +1 @@ +{"framework": "pytorch", "task": "others", "allow_remote": true} \ No newline at end of file diff --git a/generation_config.json b/generation_config.json new file mode 100644 index 0000000..96414da --- /dev/null +++ b/generation_config.json @@ -0,0 +1,6 @@ +{ + "_from_model_config": true, + "bos_token_id": 128000, + "eos_token_id": 128001, + "transformers_version": "4.44.0" +}