diff --git a/README.md b/README.md new file mode 100644 index 0000000000000000000000000000000000000000..bc5f30d6632ac0efdc7be2e9095e9e9579af2e33 --- /dev/null +++ b/README.md @@ -0,0 +1,199 @@ +--- +library_name: transformers +tags: [] +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + +This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] \ No newline at end of file diff --git a/config.json b/config.json new file mode 100644 index 0000000000000000000000000000000000000000..ead5de9526517af9b43f0e81caa170a9bd2184d8 --- /dev/null +++ b/config.json @@ -0,0 +1,35 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": 128000, + "eos_token_id": 128001, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 8192, + "initializer_range": 0.02, + "intermediate_size": 28672, + "max_position_embeddings": 131072, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 64, + "num_hidden_layers": 80, + "num_key_value_heads": 8, + "pretraining_tp": 1, + "rms_norm_eps": 1e-05, + "rope_scaling": { + "factor": 8.0, + "high_freq_factor": 4.0, + "low_freq_factor": 1.0, + "original_max_position_embeddings": 8192, + "rope_type": "llama3" + }, + "rope_theta": 500000.0, + "tie_word_embeddings": false, + "torch_dtype": "float32", + "transformers_version": "4.50.0", + "use_cache": true, + "vocab_size": 128256 +} diff --git a/generation_config.json b/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..6e76d5c5a8017c74ef094453fa12056f1e11d2e7 --- /dev/null +++ b/generation_config.json @@ -0,0 +1,9 @@ +{ + "_from_model_config": true, + "bos_token_id": 128000, + "do_sample": true, + "eos_token_id": 128001, + "temperature": 0.6, + "top_p": 0.9, + "transformers_version": "4.50.0" +} diff --git a/model-00001-of-00062.safetensors b/model-00001-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8947a75a532fb9f57d99dbf10e3b947fd2019594 --- /dev/null +++ b/model-00001-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:743e47bfe154501029c7a1423b1cd511fe1c5a8167dd96c52e9b204da7958753 +size 4806672984 diff --git a/model-00002-of-00062.safetensors b/model-00002-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..acf6edde33181d9321bd7aa1b76bb9d83cd1c364 --- /dev/null +++ b/model-00002-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e29c51177eeb23ff0be8cb891658c58f9b7a1d622037b14ecf5eb2a64e956129 +size 4362142864 diff --git a/model-00003-of-00062.safetensors b/model-00003-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0aa83503ce79953d3f7a3029d20e0583a72dd980 --- /dev/null +++ b/model-00003-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59f3c2e2208dfe5da9c38144e4e309dad6a6e17ce063bf57cdcb6f6c0d256dd2 +size 4362142864 diff --git a/model-00004-of-00062.safetensors b/model-00004-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6a80a3cf03f9868a6109a8d3003004f795018f79 --- /dev/null +++ b/model-00004-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ce06842adbeedf87e8e6cc9604a270e4ae50a2a632672e9ff9a361c472920212 +size 4966188864 diff --git a/model-00005-of-00062.safetensors b/model-00005-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e5e83383ab9d9dfa5a76a5cea448d67a1dd33329 --- /dev/null +++ b/model-00005-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6c3a690e42b8781bc0f2d00f9af5f6ccfe58adb70efd51c13accc2e7c0b8448f +size 4362142864 diff --git a/model-00006-of-00062.safetensors b/model-00006-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f1ed8cc16e1bf23dc91ee3b4a663ce3a3920f927 --- /dev/null +++ b/model-00006-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9cd21bffdf8479b7e7c12b5343dd7e7bef07e36e24d75edb16571602a6716344 +size 4362142864 diff --git a/model-00007-of-00062.safetensors b/model-00007-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cfb1a48b7c390f1d38f384c38ed2713206fb988a --- /dev/null +++ b/model-00007-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ff8ae450cb2464e6463944bafc42a0bf0a8919bfcb99e5f94108c822f1b910e5 +size 4966188864 diff --git a/model-00008-of-00062.safetensors b/model-00008-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3517fec21928c4b497c40c9c00be9db9265795c8 --- /dev/null +++ b/model-00008-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6ef608878ff77d6ba39a3fc2fc881bcf1d6715552feb552f7afa95f8ad470958 +size 4362142864 diff --git a/model-00009-of-00062.safetensors b/model-00009-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5a148ce24a1925a7c420b6b2cf6057c0ed49131e --- /dev/null +++ b/model-00009-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:38a5e534ae8b1f1ee2c7d3744c6ce76cb8ff205963b596eb264778591a26cfe3 +size 4362142880 diff --git a/model-00010-of-00062.safetensors b/model-00010-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cdd17a16db847f4a86a2dcc78eefd51b9278b3d9 --- /dev/null +++ b/model-00010-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b4bef6a13495fbf538db8f0dda42eb83573ffd03ed6dd017edcdcae3cf9f36c0 +size 4966188880 diff --git a/model-00011-of-00062.safetensors b/model-00011-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..35ac256650683cae418542f3c383444fac504409 --- /dev/null +++ b/model-00011-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:94313c5c0ff46846f8e73f2159a88e8591259392d50be93b5a5219ac55fff64c +size 4362142872 diff --git a/model-00012-of-00062.safetensors b/model-00012-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e6e34ab217f33fb32f8479d40de78ff8e471df9b --- /dev/null +++ b/model-00012-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d063033141d26693e2a0acecaf1ed7bb9c14401c374b694e0916a88f296fcd20 +size 4362142872 diff --git a/model-00013-of-00062.safetensors b/model-00013-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0d4442c8e54e59693161e6eaaa378d193ef59b00 --- /dev/null +++ b/model-00013-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d55991dcae8500740ae00eb182bd055e999e1e3f99fcf92a7aaa80e7ec8022a1 +size 4966188880 diff --git a/model-00014-of-00062.safetensors b/model-00014-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..40cd4e66f44829349b37347f067725728c29eb80 --- /dev/null +++ b/model-00014-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:78f45737659bad8faae924f3a62a864389c52f51642a200e8ccb5392527d4e8a +size 4362142872 diff --git a/model-00015-of-00062.safetensors b/model-00015-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a6806d03fe56c8979831e768f2666393f7cdbcdc --- /dev/null +++ b/model-00015-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4f636be0c5f4a58484d2c89902f56eae1fe15a8ca138ff96552c8ba0fc167e33 +size 4362142872 diff --git a/model-00016-of-00062.safetensors b/model-00016-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8d69e4457a73ded900fe6a65e4a6a9ce5ee2587c --- /dev/null +++ b/model-00016-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1833f624a47210b8d32264e0a1130b554d096fa46ec510fa4d0c57423277693f +size 4966188880 diff --git a/model-00017-of-00062.safetensors b/model-00017-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bdd73f235498f3efc41392d2fbbc09129e97064c --- /dev/null +++ b/model-00017-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:57a59dce5484772eb6de57c00c44f2c35f7989ea8a2ff8f4cabadfe07f2de76a +size 4362142872 diff --git a/model-00018-of-00062.safetensors b/model-00018-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b179ad30017ac554144e6711d5749ed4cc4302f4 --- /dev/null +++ b/model-00018-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f9e095299ffe2cf5565f054d2159b633071c118b45fc8e7ec8dbf8c58ec6c12b +size 4362142872 diff --git a/model-00019-of-00062.safetensors b/model-00019-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d033af4131e74c25394483ab7d69ac4c89c17d69 --- /dev/null +++ b/model-00019-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e7fd46ea7fc5683eb25d155dc92087ef7cb6fa084719218d0de70d668506a9bf +size 4966188880 diff --git a/model-00020-of-00062.safetensors b/model-00020-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..358e578adb128d5cca95cca278228a6ff588683c --- /dev/null +++ b/model-00020-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a06b0ccff0a140cc45bc58c8c2b71899b1b519999639d7afe97cdbc6be28fd6 +size 4362142872 diff --git a/model-00021-of-00062.safetensors b/model-00021-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..459f51d5cb24d9ce235b2e0da043f4c0b180e87e --- /dev/null +++ b/model-00021-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:772133f9e6293c3b15d24e73e68fd271d2a4ea97d70d8708155e6b12d1dbb1c7 +size 4362142872 diff --git a/model-00022-of-00062.safetensors b/model-00022-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cfb9079b757d0f525b42ecfafbcfb2cb3bb6683e --- /dev/null +++ b/model-00022-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3c5aee011e00ac3bca6a9bad2081ecc20a62d647167f80005a5350eeadaf6190 +size 4966188880 diff --git a/model-00023-of-00062.safetensors b/model-00023-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..78f337a94e1d6f0a751a82051c47e9b630b1dc4a --- /dev/null +++ b/model-00023-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:562711399d839adc43d0fcf13defe5b2e7c5fe5065a881f60627bd01bd8bc428 +size 4362142872 diff --git a/model-00024-of-00062.safetensors b/model-00024-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..506733be1aba1fe0cd46120b9764b353e7045952 --- /dev/null +++ b/model-00024-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6759b953a363781ffdc354bf7293ed206906d1af43003fd06559690af8919d13 +size 4362142872 diff --git a/model-00025-of-00062.safetensors b/model-00025-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9707df14c0afa7b2644a65e1576860d3d8b3f779 --- /dev/null +++ b/model-00025-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a20501e91da6b154dac0b402e96bb59b866ec55f63ebf1bf44a6c624794c2ce6 +size 4966188880 diff --git a/model-00026-of-00062.safetensors b/model-00026-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..486fe0df44815a8578bad815ded70def1c0bdf4b --- /dev/null +++ b/model-00026-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d9c5a3f2ab8d2f6fbd6e2fec70f5d8705e905908e5519e19c8ef51df48f0de48 +size 4362142872 diff --git a/model-00027-of-00062.safetensors b/model-00027-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f1ceb8fd74e8e0969f9e04f012c2ffec44c5f25f --- /dev/null +++ b/model-00027-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:52d21f39059efb89815d9b7afa5496d5d7139961f597b3c1c8dd69da590a870d +size 4362142872 diff --git a/model-00028-of-00062.safetensors b/model-00028-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a1e976eb11b24716998041a40f5d890d1d55c980 --- /dev/null +++ b/model-00028-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7a7743118c6e6b198a6364af655c4dcc823182ac459b06bab77305c27f95c57 +size 4966188880 diff --git a/model-00029-of-00062.safetensors b/model-00029-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7995f3261180c6cf9a9e20382dae90f050c9c683 --- /dev/null +++ b/model-00029-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:32663d84a873dca6ee3cb93ca050bea08200d6a3269d8aa9a1ab6f7db5e595dd +size 4362142872 diff --git a/model-00030-of-00062.safetensors b/model-00030-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ed9a66ffb4ace3465843a37cd4e4c4c7286f50fd --- /dev/null +++ b/model-00030-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:adf691c6f477c5ec4b804e2d3f787b2efc98c00c11659c0931bc883df0aa23db +size 4362142872 diff --git a/model-00031-of-00062.safetensors b/model-00031-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..530342e70ebceb778efefdd79d6b6025a4a21482 --- /dev/null +++ b/model-00031-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8e7bf69b34428f847a2d093008a9e92c40899cda53d5e4e0681fc2f872e064ca +size 4966188880 diff --git a/model-00032-of-00062.safetensors b/model-00032-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e99688cad288d898c5f735d1957b193d9283a1f5 --- /dev/null +++ b/model-00032-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e3bdeb25e27e050a8975432625da830c968c2fcd86c977023feee5abc9f86b74 +size 4362142872 diff --git a/model-00033-of-00062.safetensors b/model-00033-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..099de631f1650c29b2eb0649bd0762cd7077c4dc --- /dev/null +++ b/model-00033-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b1398611a19e238b7cf6bc740f0e0bfc48fc8b6296d9d91ec0baf6d3928093ea +size 4362142872 diff --git a/model-00034-of-00062.safetensors b/model-00034-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8f9e11574d0290961d5ad3e4ab79ee5d8d1bd130 --- /dev/null +++ b/model-00034-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d9959f19e5024d38d6d491e52fb378ed68ad365996682995d05d3f15ce35749a +size 4966188880 diff --git a/model-00035-of-00062.safetensors b/model-00035-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..757fce7e54773a89c71652c1833fef9c07bfbf0c --- /dev/null +++ b/model-00035-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:37caa8dd0eb676a04b373e1742796dac46ffe38ed72c932694d26c2e83078078 +size 4362142872 diff --git a/model-00036-of-00062.safetensors b/model-00036-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0c73aba804b4f255a06047d62f329c25e01787c5 --- /dev/null +++ b/model-00036-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4520a6e30fa5140976f1bbee3736077b2599c06055285adc7ab3924e2bf49635 +size 4362142872 diff --git a/model-00037-of-00062.safetensors b/model-00037-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..681c1e5a18d439a7e0e2800e14d94053d3df7084 --- /dev/null +++ b/model-00037-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:12f812e8108a475b25c7c26db85a8f6bd3055ae64a38093e4ed41afc4127e066 +size 4966188880 diff --git a/model-00038-of-00062.safetensors b/model-00038-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..29637e9552b285c4e45b612e91dae15d32d683ee --- /dev/null +++ b/model-00038-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a467d580c3c6388cd3459cf0e22a83de37478de07ff212158d5bd6bb76cc77a6 +size 4362142872 diff --git a/model-00039-of-00062.safetensors b/model-00039-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ea98022a1b5fd8517ff7471a5c0d2ea895f21a39 --- /dev/null +++ b/model-00039-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35c4db40322b1e935989bba46e4be3b522112ab5b4df64cafe60f93fb4171116 +size 4362142872 diff --git a/model-00040-of-00062.safetensors b/model-00040-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..be86a5bb754f2060cd4711e55efe3c402f40abc5 --- /dev/null +++ b/model-00040-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6d7d8e33f3ecff493732c03ae180bcf2e9136702c97786e341a31d73655c094a +size 4966188880 diff --git a/model-00041-of-00062.safetensors b/model-00041-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..19870a6bda18a0d2a9b4c83db65a1f56c7a86425 --- /dev/null +++ b/model-00041-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:01a6df437a67821a3c943b7386e8cb06be167848ae84028f839c7692d430949f +size 4362142872 diff --git a/model-00042-of-00062.safetensors b/model-00042-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1b8109b852245f78887cfd2e19e7f6dceb4b991b --- /dev/null +++ b/model-00042-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ea4a5462d989f2d76c3224b1abb9a5bfeaf70bdb8ee105ca01d6c40494083a70 +size 4362142872 diff --git a/model-00043-of-00062.safetensors b/model-00043-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..757942cbd4df6e346f10c2a6f23ef22ba79ca647 --- /dev/null +++ b/model-00043-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dfa84d4f987815f529c478073f2b3ba1cc28dbf88f4c268bc91af21e45b765f8 +size 4966188880 diff --git a/model-00044-of-00062.safetensors b/model-00044-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cfba873f05a694b227dbec9d0e56c7d86537baac --- /dev/null +++ b/model-00044-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:03b19b60ff1e2a020696d2a29ce438649277d1d347dab174f201560a81d2ff54 +size 4362142872 diff --git a/model-00045-of-00062.safetensors b/model-00045-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f9fb5236f8ec3b389c1846f26a37bc753a3dcc98 --- /dev/null +++ b/model-00045-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:124d6dda2f93b0e176bf7c911f16d04f322fb1ad1a68016f80cdf9dde15442e0 +size 4362142872 diff --git a/model-00046-of-00062.safetensors b/model-00046-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d73d300a66a60b131b5326fdc2fcbd474af7216b --- /dev/null +++ b/model-00046-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59a918f35df571317d74d7b6b58e0390aada1c497d6f621d5ec864ca5004d931 +size 4966188880 diff --git a/model-00047-of-00062.safetensors b/model-00047-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f22ae17185dd35449591dcec112c5eb3a945a037 --- /dev/null +++ b/model-00047-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ee1cb835988cff268d6b8ea6f60377ceb001b3cc8f7e5d4e3a4f555795a26b46 +size 4362142872 diff --git a/model-00048-of-00062.safetensors b/model-00048-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e6c73f9adadf2789ffa2dd501a844032e0155a7f --- /dev/null +++ b/model-00048-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b2e38202971301c06fa5622731f23653dde688ef93e4d102f3e56d73453b85a7 +size 4362142872 diff --git a/model-00049-of-00062.safetensors b/model-00049-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..96047781048363b58a5b19f5b6aa67eeadfbf9a4 --- /dev/null +++ b/model-00049-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7dda27e33a903da2523c9b3c69dbd52cfcb54e6e0e543ed05d9f865887cb4130 +size 4966188880 diff --git a/model-00050-of-00062.safetensors b/model-00050-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..344218c69e11bde9ef12991996390208dc3768a6 --- /dev/null +++ b/model-00050-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6f49da9d00c423248b1f6f901543ddb49a890727537a568eb1872947ce1d18fc +size 4362142872 diff --git a/model-00051-of-00062.safetensors b/model-00051-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..da6550dfbb4767ab5b0457b068b780fb238c3046 --- /dev/null +++ b/model-00051-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:285ffda960e1b8642de558a5eb36a2baa81715ee24fce3afa39029912d75de6c +size 4362142872 diff --git a/model-00052-of-00062.safetensors b/model-00052-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e74bc4dbdd8168970aa21e780e875545b5b1e5bf --- /dev/null +++ b/model-00052-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5ca7ec339ec77b1132fe7e00c7425eaaebc59343cd5950a4c0e8e241d1b89c1d +size 4966188880 diff --git a/model-00053-of-00062.safetensors b/model-00053-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7bb5327bc20be122088b5aabb3965cd86eec8d00 --- /dev/null +++ b/model-00053-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d4113a956ce59ea9a02d3aec13d5993cf1d69b34c0d483eb9d080fbf31379f2d +size 4362142872 diff --git a/model-00054-of-00062.safetensors b/model-00054-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c5ec02c5a3fb88e1b5450a7bc453032f7547ee2d --- /dev/null +++ b/model-00054-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c00b0dd279e5182226f5b325dea6364a54815b4bc78d7fc084f514ab76301d94 +size 4362142872 diff --git a/model-00055-of-00062.safetensors b/model-00055-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4114dc55cc4b14798539ffe75cbf9f66f6c7540d --- /dev/null +++ b/model-00055-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d1f55697c00bc6ce69dfeade832490b617d4e03ebe1604fe4791f973ac2150de +size 4966188880 diff --git a/model-00056-of-00062.safetensors b/model-00056-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9c7b68cae9dab3f8601d5b9f0f912f31e6d7e22f --- /dev/null +++ b/model-00056-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f56ee9b62450cc25c369fa5710d38f6a58581087d827583a2975b10de9ac2679 +size 4362142872 diff --git a/model-00057-of-00062.safetensors b/model-00057-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..185c7b4c3ff56efed2781e242f81aa239b32cab6 --- /dev/null +++ b/model-00057-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2198e373cadd542eef97940f73f6821cb7283b2b899c2d3c6bb839adb927bb5e +size 4362142872 diff --git a/model-00058-of-00062.safetensors b/model-00058-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f5cf3084416ca0621dd8848bf2a2e6d00f92e7db --- /dev/null +++ b/model-00058-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1aff7516cc6ad27c648959ff6af8f9b124b01ac0441e7290052b2532af9d6e03 +size 4966188880 diff --git a/model-00059-of-00062.safetensors b/model-00059-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3ba0fea286d6d0f5c15ffd794f650b8b2ec99e0c --- /dev/null +++ b/model-00059-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f8aa187ef068b7b6bf15c5e1df8c0fcd0122389d11dbbfc1af10462f027270f1 +size 4362142872 diff --git a/model-00060-of-00062.safetensors b/model-00060-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f11bf671f4d5dd687b36acfd26517f580081a958 --- /dev/null +++ b/model-00060-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:17d54b7869968a98cf28f865ec4faa932e796ceeac97da5c2af3b6c384dc9fec +size 4362142872 diff --git a/model-00061-of-00062.safetensors b/model-00061-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..492e37bceec3e7c7ee2f301303602703e2ebea22 --- /dev/null +++ b/model-00061-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:248fd4e8717b373803cbde50202ea50e6b667d7601193f982d5caa546735888c +size 4362241496 diff --git a/model-00062-of-00062.safetensors b/model-00062-of-00062.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3ae0bf5a162a1f01091797f92cb56f932a0d0c71 --- /dev/null +++ b/model-00062-of-00062.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4e4016f2c0cfb4accb6a98dccb5b0d6bdb2ed46110de2224d6c6c06e262f5bcb +size 4202692736 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000000000000000000000000000000000000..8aab69e95f0d7da106bad45a4e7c35897ecf3d80 --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1,730 @@ +{ + "metadata": { + "total_size": 282214825984 + }, + "weight_map": { + "lm_head.weight": "model-00062-of-00062.safetensors", + "model.embed_tokens.weight": "model-00001-of-00062.safetensors", + "model.layers.0.input_layernorm.weight": "model-00002-of-00062.safetensors", + "model.layers.0.mlp.down_proj.weight": "model-00002-of-00062.safetensors", + "model.layers.0.mlp.gate_proj.weight": "model-00002-of-00062.safetensors", + "model.layers.0.mlp.up_proj.weight": "model-00002-of-00062.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model-00002-of-00062.safetensors", + "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00062.safetensors", + "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00062.safetensors", + "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00062.safetensors", + "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00062.safetensors", + "model.layers.1.input_layernorm.weight": "model-00003-of-00062.safetensors", + "model.layers.1.mlp.down_proj.weight": "model-00003-of-00062.safetensors", + "model.layers.1.mlp.gate_proj.weight": "model-00002-of-00062.safetensors", + "model.layers.1.mlp.up_proj.weight": "model-00003-of-00062.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model-00003-of-00062.safetensors", + "model.layers.1.self_attn.k_proj.weight": "model-00002-of-00062.safetensors", + "model.layers.1.self_attn.o_proj.weight": "model-00002-of-00062.safetensors", + "model.layers.1.self_attn.q_proj.weight": "model-00002-of-00062.safetensors", + "model.layers.1.self_attn.v_proj.weight": "model-00002-of-00062.safetensors", + "model.layers.10.input_layernorm.weight": "model-00010-of-00062.safetensors", + "model.layers.10.mlp.down_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.10.mlp.gate_proj.weight": "model-00009-of-00062.safetensors", + "model.layers.10.mlp.up_proj.weight": "model-00009-of-00062.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model-00010-of-00062.safetensors", + "model.layers.10.self_attn.k_proj.weight": "model-00009-of-00062.safetensors", + "model.layers.10.self_attn.o_proj.weight": "model-00009-of-00062.safetensors", + "model.layers.10.self_attn.q_proj.weight": "model-00009-of-00062.safetensors", + "model.layers.10.self_attn.v_proj.weight": "model-00009-of-00062.safetensors", + "model.layers.11.input_layernorm.weight": "model-00010-of-00062.safetensors", + "model.layers.11.mlp.down_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.11.mlp.gate_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.11.mlp.up_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model-00010-of-00062.safetensors", + "model.layers.11.self_attn.k_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.11.self_attn.o_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.11.self_attn.q_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.11.self_attn.v_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.12.input_layernorm.weight": "model-00011-of-00062.safetensors", + "model.layers.12.mlp.down_proj.weight": "model-00011-of-00062.safetensors", + "model.layers.12.mlp.gate_proj.weight": "model-00011-of-00062.safetensors", + "model.layers.12.mlp.up_proj.weight": "model-00011-of-00062.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model-00011-of-00062.safetensors", + "model.layers.12.self_attn.k_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.12.self_attn.o_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.12.self_attn.q_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.12.self_attn.v_proj.weight": "model-00010-of-00062.safetensors", + "model.layers.13.input_layernorm.weight": "model-00012-of-00062.safetensors", + "model.layers.13.mlp.down_proj.weight": "model-00012-of-00062.safetensors", + "model.layers.13.mlp.gate_proj.weight": "model-00011-of-00062.safetensors", + "model.layers.13.mlp.up_proj.weight": "model-00012-of-00062.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model-00012-of-00062.safetensors", + "model.layers.13.self_attn.k_proj.weight": "model-00011-of-00062.safetensors", + "model.layers.13.self_attn.o_proj.weight": "model-00011-of-00062.safetensors", + "model.layers.13.self_attn.q_proj.weight": "model-00011-of-00062.safetensors", + "model.layers.13.self_attn.v_proj.weight": "model-00011-of-00062.safetensors", + "model.layers.14.input_layernorm.weight": "model-00013-of-00062.safetensors", + "model.layers.14.mlp.down_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.14.mlp.gate_proj.weight": "model-00012-of-00062.safetensors", + "model.layers.14.mlp.up_proj.weight": "model-00012-of-00062.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model-00013-of-00062.safetensors", + "model.layers.14.self_attn.k_proj.weight": "model-00012-of-00062.safetensors", + "model.layers.14.self_attn.o_proj.weight": "model-00012-of-00062.safetensors", + "model.layers.14.self_attn.q_proj.weight": "model-00012-of-00062.safetensors", + "model.layers.14.self_attn.v_proj.weight": "model-00012-of-00062.safetensors", + "model.layers.15.input_layernorm.weight": "model-00013-of-00062.safetensors", + "model.layers.15.mlp.down_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.15.mlp.gate_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.15.mlp.up_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model-00013-of-00062.safetensors", + "model.layers.15.self_attn.k_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.15.self_attn.o_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.15.self_attn.q_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.15.self_attn.v_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.16.input_layernorm.weight": "model-00014-of-00062.safetensors", + "model.layers.16.mlp.down_proj.weight": "model-00014-of-00062.safetensors", + "model.layers.16.mlp.gate_proj.weight": "model-00014-of-00062.safetensors", + "model.layers.16.mlp.up_proj.weight": "model-00014-of-00062.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model-00014-of-00062.safetensors", + "model.layers.16.self_attn.k_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.16.self_attn.o_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.16.self_attn.q_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.16.self_attn.v_proj.weight": "model-00013-of-00062.safetensors", + "model.layers.17.input_layernorm.weight": "model-00015-of-00062.safetensors", + "model.layers.17.mlp.down_proj.weight": "model-00015-of-00062.safetensors", + "model.layers.17.mlp.gate_proj.weight": "model-00014-of-00062.safetensors", + "model.layers.17.mlp.up_proj.weight": "model-00015-of-00062.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model-00015-of-00062.safetensors", + "model.layers.17.self_attn.k_proj.weight": "model-00014-of-00062.safetensors", + "model.layers.17.self_attn.o_proj.weight": "model-00014-of-00062.safetensors", + "model.layers.17.self_attn.q_proj.weight": "model-00014-of-00062.safetensors", + "model.layers.17.self_attn.v_proj.weight": "model-00014-of-00062.safetensors", + "model.layers.18.input_layernorm.weight": "model-00016-of-00062.safetensors", + "model.layers.18.mlp.down_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.18.mlp.gate_proj.weight": "model-00015-of-00062.safetensors", + "model.layers.18.mlp.up_proj.weight": "model-00015-of-00062.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model-00016-of-00062.safetensors", + "model.layers.18.self_attn.k_proj.weight": "model-00015-of-00062.safetensors", + "model.layers.18.self_attn.o_proj.weight": "model-00015-of-00062.safetensors", + "model.layers.18.self_attn.q_proj.weight": "model-00015-of-00062.safetensors", + "model.layers.18.self_attn.v_proj.weight": "model-00015-of-00062.safetensors", + "model.layers.19.input_layernorm.weight": "model-00016-of-00062.safetensors", + "model.layers.19.mlp.down_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.19.mlp.gate_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.19.mlp.up_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model-00016-of-00062.safetensors", + "model.layers.19.self_attn.k_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.19.self_attn.o_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.19.self_attn.q_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.19.self_attn.v_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.2.input_layernorm.weight": "model-00004-of-00062.safetensors", + "model.layers.2.mlp.down_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.2.mlp.gate_proj.weight": "model-00003-of-00062.safetensors", + "model.layers.2.mlp.up_proj.weight": "model-00003-of-00062.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model-00004-of-00062.safetensors", + "model.layers.2.self_attn.k_proj.weight": "model-00003-of-00062.safetensors", + "model.layers.2.self_attn.o_proj.weight": "model-00003-of-00062.safetensors", + "model.layers.2.self_attn.q_proj.weight": "model-00003-of-00062.safetensors", + "model.layers.2.self_attn.v_proj.weight": "model-00003-of-00062.safetensors", + "model.layers.20.input_layernorm.weight": "model-00017-of-00062.safetensors", + "model.layers.20.mlp.down_proj.weight": "model-00017-of-00062.safetensors", + "model.layers.20.mlp.gate_proj.weight": "model-00017-of-00062.safetensors", + "model.layers.20.mlp.up_proj.weight": "model-00017-of-00062.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model-00017-of-00062.safetensors", + "model.layers.20.self_attn.k_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.20.self_attn.o_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.20.self_attn.q_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.20.self_attn.v_proj.weight": "model-00016-of-00062.safetensors", + "model.layers.21.input_layernorm.weight": "model-00018-of-00062.safetensors", + "model.layers.21.mlp.down_proj.weight": "model-00018-of-00062.safetensors", + "model.layers.21.mlp.gate_proj.weight": "model-00017-of-00062.safetensors", + "model.layers.21.mlp.up_proj.weight": "model-00018-of-00062.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model-00018-of-00062.safetensors", + "model.layers.21.self_attn.k_proj.weight": "model-00017-of-00062.safetensors", + "model.layers.21.self_attn.o_proj.weight": "model-00017-of-00062.safetensors", + "model.layers.21.self_attn.q_proj.weight": "model-00017-of-00062.safetensors", + "model.layers.21.self_attn.v_proj.weight": "model-00017-of-00062.safetensors", + "model.layers.22.input_layernorm.weight": "model-00019-of-00062.safetensors", + "model.layers.22.mlp.down_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.22.mlp.gate_proj.weight": "model-00018-of-00062.safetensors", + "model.layers.22.mlp.up_proj.weight": "model-00018-of-00062.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model-00019-of-00062.safetensors", + "model.layers.22.self_attn.k_proj.weight": "model-00018-of-00062.safetensors", + "model.layers.22.self_attn.o_proj.weight": "model-00018-of-00062.safetensors", + "model.layers.22.self_attn.q_proj.weight": "model-00018-of-00062.safetensors", + "model.layers.22.self_attn.v_proj.weight": "model-00018-of-00062.safetensors", + "model.layers.23.input_layernorm.weight": "model-00019-of-00062.safetensors", + "model.layers.23.mlp.down_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.23.mlp.gate_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.23.mlp.up_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model-00019-of-00062.safetensors", + "model.layers.23.self_attn.k_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.23.self_attn.o_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.23.self_attn.q_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.23.self_attn.v_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.24.input_layernorm.weight": "model-00020-of-00062.safetensors", + "model.layers.24.mlp.down_proj.weight": "model-00020-of-00062.safetensors", + "model.layers.24.mlp.gate_proj.weight": "model-00020-of-00062.safetensors", + "model.layers.24.mlp.up_proj.weight": "model-00020-of-00062.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model-00020-of-00062.safetensors", + "model.layers.24.self_attn.k_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.24.self_attn.o_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.24.self_attn.q_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.24.self_attn.v_proj.weight": "model-00019-of-00062.safetensors", + "model.layers.25.input_layernorm.weight": "model-00021-of-00062.safetensors", + "model.layers.25.mlp.down_proj.weight": "model-00021-of-00062.safetensors", + "model.layers.25.mlp.gate_proj.weight": "model-00020-of-00062.safetensors", + "model.layers.25.mlp.up_proj.weight": "model-00021-of-00062.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model-00021-of-00062.safetensors", + "model.layers.25.self_attn.k_proj.weight": "model-00020-of-00062.safetensors", + "model.layers.25.self_attn.o_proj.weight": "model-00020-of-00062.safetensors", + "model.layers.25.self_attn.q_proj.weight": "model-00020-of-00062.safetensors", + "model.layers.25.self_attn.v_proj.weight": "model-00020-of-00062.safetensors", + "model.layers.26.input_layernorm.weight": "model-00022-of-00062.safetensors", + "model.layers.26.mlp.down_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.26.mlp.gate_proj.weight": "model-00021-of-00062.safetensors", + "model.layers.26.mlp.up_proj.weight": "model-00021-of-00062.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model-00022-of-00062.safetensors", + "model.layers.26.self_attn.k_proj.weight": "model-00021-of-00062.safetensors", + "model.layers.26.self_attn.o_proj.weight": "model-00021-of-00062.safetensors", + "model.layers.26.self_attn.q_proj.weight": "model-00021-of-00062.safetensors", + "model.layers.26.self_attn.v_proj.weight": "model-00021-of-00062.safetensors", + "model.layers.27.input_layernorm.weight": "model-00022-of-00062.safetensors", + "model.layers.27.mlp.down_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.27.mlp.gate_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.27.mlp.up_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model-00022-of-00062.safetensors", + "model.layers.27.self_attn.k_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.27.self_attn.o_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.27.self_attn.q_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.27.self_attn.v_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.28.input_layernorm.weight": "model-00023-of-00062.safetensors", + "model.layers.28.mlp.down_proj.weight": "model-00023-of-00062.safetensors", + "model.layers.28.mlp.gate_proj.weight": "model-00023-of-00062.safetensors", + "model.layers.28.mlp.up_proj.weight": "model-00023-of-00062.safetensors", + "model.layers.28.post_attention_layernorm.weight": "model-00023-of-00062.safetensors", + "model.layers.28.self_attn.k_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.28.self_attn.o_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.28.self_attn.q_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.28.self_attn.v_proj.weight": "model-00022-of-00062.safetensors", + "model.layers.29.input_layernorm.weight": "model-00024-of-00062.safetensors", + "model.layers.29.mlp.down_proj.weight": "model-00024-of-00062.safetensors", + "model.layers.29.mlp.gate_proj.weight": "model-00023-of-00062.safetensors", + "model.layers.29.mlp.up_proj.weight": "model-00024-of-00062.safetensors", + "model.layers.29.post_attention_layernorm.weight": "model-00024-of-00062.safetensors", + "model.layers.29.self_attn.k_proj.weight": "model-00023-of-00062.safetensors", + "model.layers.29.self_attn.o_proj.weight": "model-00023-of-00062.safetensors", + "model.layers.29.self_attn.q_proj.weight": "model-00023-of-00062.safetensors", + "model.layers.29.self_attn.v_proj.weight": "model-00023-of-00062.safetensors", + "model.layers.3.input_layernorm.weight": "model-00004-of-00062.safetensors", + "model.layers.3.mlp.down_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.3.mlp.gate_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.3.mlp.up_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model-00004-of-00062.safetensors", + "model.layers.3.self_attn.k_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.3.self_attn.o_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.3.self_attn.q_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.3.self_attn.v_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.30.input_layernorm.weight": "model-00025-of-00062.safetensors", + "model.layers.30.mlp.down_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.30.mlp.gate_proj.weight": "model-00024-of-00062.safetensors", + "model.layers.30.mlp.up_proj.weight": "model-00024-of-00062.safetensors", + "model.layers.30.post_attention_layernorm.weight": "model-00025-of-00062.safetensors", + "model.layers.30.self_attn.k_proj.weight": "model-00024-of-00062.safetensors", + "model.layers.30.self_attn.o_proj.weight": "model-00024-of-00062.safetensors", + "model.layers.30.self_attn.q_proj.weight": "model-00024-of-00062.safetensors", + "model.layers.30.self_attn.v_proj.weight": "model-00024-of-00062.safetensors", + "model.layers.31.input_layernorm.weight": "model-00025-of-00062.safetensors", + "model.layers.31.mlp.down_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.31.mlp.gate_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.31.mlp.up_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.31.post_attention_layernorm.weight": "model-00025-of-00062.safetensors", + "model.layers.31.self_attn.k_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.31.self_attn.o_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.31.self_attn.q_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.31.self_attn.v_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.32.input_layernorm.weight": "model-00026-of-00062.safetensors", + "model.layers.32.mlp.down_proj.weight": "model-00026-of-00062.safetensors", + "model.layers.32.mlp.gate_proj.weight": "model-00026-of-00062.safetensors", + "model.layers.32.mlp.up_proj.weight": "model-00026-of-00062.safetensors", + "model.layers.32.post_attention_layernorm.weight": "model-00026-of-00062.safetensors", + "model.layers.32.self_attn.k_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.32.self_attn.o_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.32.self_attn.q_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.32.self_attn.v_proj.weight": "model-00025-of-00062.safetensors", + "model.layers.33.input_layernorm.weight": "model-00027-of-00062.safetensors", + "model.layers.33.mlp.down_proj.weight": "model-00027-of-00062.safetensors", + "model.layers.33.mlp.gate_proj.weight": "model-00026-of-00062.safetensors", + "model.layers.33.mlp.up_proj.weight": "model-00027-of-00062.safetensors", + "model.layers.33.post_attention_layernorm.weight": "model-00027-of-00062.safetensors", + "model.layers.33.self_attn.k_proj.weight": "model-00026-of-00062.safetensors", + "model.layers.33.self_attn.o_proj.weight": "model-00026-of-00062.safetensors", + "model.layers.33.self_attn.q_proj.weight": "model-00026-of-00062.safetensors", + "model.layers.33.self_attn.v_proj.weight": "model-00026-of-00062.safetensors", + "model.layers.34.input_layernorm.weight": "model-00028-of-00062.safetensors", + "model.layers.34.mlp.down_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.34.mlp.gate_proj.weight": "model-00027-of-00062.safetensors", + "model.layers.34.mlp.up_proj.weight": "model-00027-of-00062.safetensors", + "model.layers.34.post_attention_layernorm.weight": "model-00028-of-00062.safetensors", + "model.layers.34.self_attn.k_proj.weight": "model-00027-of-00062.safetensors", + "model.layers.34.self_attn.o_proj.weight": "model-00027-of-00062.safetensors", + "model.layers.34.self_attn.q_proj.weight": "model-00027-of-00062.safetensors", + "model.layers.34.self_attn.v_proj.weight": "model-00027-of-00062.safetensors", + "model.layers.35.input_layernorm.weight": "model-00028-of-00062.safetensors", + "model.layers.35.mlp.down_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.35.mlp.gate_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.35.mlp.up_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.35.post_attention_layernorm.weight": "model-00028-of-00062.safetensors", + "model.layers.35.self_attn.k_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.35.self_attn.o_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.35.self_attn.q_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.35.self_attn.v_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.36.input_layernorm.weight": "model-00029-of-00062.safetensors", + "model.layers.36.mlp.down_proj.weight": "model-00029-of-00062.safetensors", + "model.layers.36.mlp.gate_proj.weight": "model-00029-of-00062.safetensors", + "model.layers.36.mlp.up_proj.weight": "model-00029-of-00062.safetensors", + "model.layers.36.post_attention_layernorm.weight": "model-00029-of-00062.safetensors", + "model.layers.36.self_attn.k_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.36.self_attn.o_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.36.self_attn.q_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.36.self_attn.v_proj.weight": "model-00028-of-00062.safetensors", + "model.layers.37.input_layernorm.weight": "model-00030-of-00062.safetensors", + "model.layers.37.mlp.down_proj.weight": "model-00030-of-00062.safetensors", + "model.layers.37.mlp.gate_proj.weight": "model-00029-of-00062.safetensors", + "model.layers.37.mlp.up_proj.weight": "model-00030-of-00062.safetensors", + "model.layers.37.post_attention_layernorm.weight": "model-00030-of-00062.safetensors", + "model.layers.37.self_attn.k_proj.weight": "model-00029-of-00062.safetensors", + "model.layers.37.self_attn.o_proj.weight": "model-00029-of-00062.safetensors", + "model.layers.37.self_attn.q_proj.weight": "model-00029-of-00062.safetensors", + "model.layers.37.self_attn.v_proj.weight": "model-00029-of-00062.safetensors", + "model.layers.38.input_layernorm.weight": "model-00031-of-00062.safetensors", + "model.layers.38.mlp.down_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.38.mlp.gate_proj.weight": "model-00030-of-00062.safetensors", + "model.layers.38.mlp.up_proj.weight": "model-00030-of-00062.safetensors", + "model.layers.38.post_attention_layernorm.weight": "model-00031-of-00062.safetensors", + "model.layers.38.self_attn.k_proj.weight": "model-00030-of-00062.safetensors", + "model.layers.38.self_attn.o_proj.weight": "model-00030-of-00062.safetensors", + "model.layers.38.self_attn.q_proj.weight": "model-00030-of-00062.safetensors", + "model.layers.38.self_attn.v_proj.weight": "model-00030-of-00062.safetensors", + "model.layers.39.input_layernorm.weight": "model-00031-of-00062.safetensors", + "model.layers.39.mlp.down_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.39.mlp.gate_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.39.mlp.up_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.39.post_attention_layernorm.weight": "model-00031-of-00062.safetensors", + "model.layers.39.self_attn.k_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.39.self_attn.o_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.39.self_attn.q_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.39.self_attn.v_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.4.input_layernorm.weight": "model-00005-of-00062.safetensors", + "model.layers.4.mlp.down_proj.weight": "model-00005-of-00062.safetensors", + "model.layers.4.mlp.gate_proj.weight": "model-00005-of-00062.safetensors", + "model.layers.4.mlp.up_proj.weight": "model-00005-of-00062.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model-00005-of-00062.safetensors", + "model.layers.4.self_attn.k_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.4.self_attn.o_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.4.self_attn.q_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.4.self_attn.v_proj.weight": "model-00004-of-00062.safetensors", + "model.layers.40.input_layernorm.weight": "model-00032-of-00062.safetensors", + "model.layers.40.mlp.down_proj.weight": "model-00032-of-00062.safetensors", + "model.layers.40.mlp.gate_proj.weight": "model-00032-of-00062.safetensors", + "model.layers.40.mlp.up_proj.weight": "model-00032-of-00062.safetensors", + "model.layers.40.post_attention_layernorm.weight": "model-00032-of-00062.safetensors", + "model.layers.40.self_attn.k_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.40.self_attn.o_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.40.self_attn.q_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.40.self_attn.v_proj.weight": "model-00031-of-00062.safetensors", + "model.layers.41.input_layernorm.weight": "model-00033-of-00062.safetensors", + "model.layers.41.mlp.down_proj.weight": "model-00033-of-00062.safetensors", + "model.layers.41.mlp.gate_proj.weight": "model-00032-of-00062.safetensors", + "model.layers.41.mlp.up_proj.weight": "model-00033-of-00062.safetensors", + "model.layers.41.post_attention_layernorm.weight": "model-00033-of-00062.safetensors", + "model.layers.41.self_attn.k_proj.weight": "model-00032-of-00062.safetensors", + "model.layers.41.self_attn.o_proj.weight": "model-00032-of-00062.safetensors", + "model.layers.41.self_attn.q_proj.weight": "model-00032-of-00062.safetensors", + "model.layers.41.self_attn.v_proj.weight": "model-00032-of-00062.safetensors", + "model.layers.42.input_layernorm.weight": "model-00034-of-00062.safetensors", + "model.layers.42.mlp.down_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.42.mlp.gate_proj.weight": "model-00033-of-00062.safetensors", + "model.layers.42.mlp.up_proj.weight": "model-00033-of-00062.safetensors", + "model.layers.42.post_attention_layernorm.weight": "model-00034-of-00062.safetensors", + "model.layers.42.self_attn.k_proj.weight": "model-00033-of-00062.safetensors", + "model.layers.42.self_attn.o_proj.weight": "model-00033-of-00062.safetensors", + "model.layers.42.self_attn.q_proj.weight": "model-00033-of-00062.safetensors", + "model.layers.42.self_attn.v_proj.weight": "model-00033-of-00062.safetensors", + "model.layers.43.input_layernorm.weight": "model-00034-of-00062.safetensors", + "model.layers.43.mlp.down_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.43.mlp.gate_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.43.mlp.up_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.43.post_attention_layernorm.weight": "model-00034-of-00062.safetensors", + "model.layers.43.self_attn.k_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.43.self_attn.o_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.43.self_attn.q_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.43.self_attn.v_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.44.input_layernorm.weight": "model-00035-of-00062.safetensors", + "model.layers.44.mlp.down_proj.weight": "model-00035-of-00062.safetensors", + "model.layers.44.mlp.gate_proj.weight": "model-00035-of-00062.safetensors", + "model.layers.44.mlp.up_proj.weight": "model-00035-of-00062.safetensors", + "model.layers.44.post_attention_layernorm.weight": "model-00035-of-00062.safetensors", + "model.layers.44.self_attn.k_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.44.self_attn.o_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.44.self_attn.q_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.44.self_attn.v_proj.weight": "model-00034-of-00062.safetensors", + "model.layers.45.input_layernorm.weight": "model-00036-of-00062.safetensors", + "model.layers.45.mlp.down_proj.weight": "model-00036-of-00062.safetensors", + "model.layers.45.mlp.gate_proj.weight": "model-00035-of-00062.safetensors", + "model.layers.45.mlp.up_proj.weight": "model-00036-of-00062.safetensors", + "model.layers.45.post_attention_layernorm.weight": "model-00036-of-00062.safetensors", + "model.layers.45.self_attn.k_proj.weight": "model-00035-of-00062.safetensors", + "model.layers.45.self_attn.o_proj.weight": "model-00035-of-00062.safetensors", + "model.layers.45.self_attn.q_proj.weight": "model-00035-of-00062.safetensors", + "model.layers.45.self_attn.v_proj.weight": "model-00035-of-00062.safetensors", + "model.layers.46.input_layernorm.weight": "model-00037-of-00062.safetensors", + "model.layers.46.mlp.down_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.46.mlp.gate_proj.weight": "model-00036-of-00062.safetensors", + "model.layers.46.mlp.up_proj.weight": "model-00036-of-00062.safetensors", + "model.layers.46.post_attention_layernorm.weight": "model-00037-of-00062.safetensors", + "model.layers.46.self_attn.k_proj.weight": "model-00036-of-00062.safetensors", + "model.layers.46.self_attn.o_proj.weight": "model-00036-of-00062.safetensors", + "model.layers.46.self_attn.q_proj.weight": "model-00036-of-00062.safetensors", + "model.layers.46.self_attn.v_proj.weight": "model-00036-of-00062.safetensors", + "model.layers.47.input_layernorm.weight": "model-00037-of-00062.safetensors", + "model.layers.47.mlp.down_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.47.mlp.gate_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.47.mlp.up_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.47.post_attention_layernorm.weight": "model-00037-of-00062.safetensors", + "model.layers.47.self_attn.k_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.47.self_attn.o_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.47.self_attn.q_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.47.self_attn.v_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.48.input_layernorm.weight": "model-00038-of-00062.safetensors", + "model.layers.48.mlp.down_proj.weight": "model-00038-of-00062.safetensors", + "model.layers.48.mlp.gate_proj.weight": "model-00038-of-00062.safetensors", + "model.layers.48.mlp.up_proj.weight": "model-00038-of-00062.safetensors", + "model.layers.48.post_attention_layernorm.weight": "model-00038-of-00062.safetensors", + "model.layers.48.self_attn.k_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.48.self_attn.o_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.48.self_attn.q_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.48.self_attn.v_proj.weight": "model-00037-of-00062.safetensors", + "model.layers.49.input_layernorm.weight": "model-00039-of-00062.safetensors", + "model.layers.49.mlp.down_proj.weight": "model-00039-of-00062.safetensors", + "model.layers.49.mlp.gate_proj.weight": "model-00038-of-00062.safetensors", + "model.layers.49.mlp.up_proj.weight": "model-00039-of-00062.safetensors", + "model.layers.49.post_attention_layernorm.weight": "model-00039-of-00062.safetensors", + "model.layers.49.self_attn.k_proj.weight": "model-00038-of-00062.safetensors", + "model.layers.49.self_attn.o_proj.weight": "model-00038-of-00062.safetensors", + "model.layers.49.self_attn.q_proj.weight": "model-00038-of-00062.safetensors", + "model.layers.49.self_attn.v_proj.weight": "model-00038-of-00062.safetensors", + "model.layers.5.input_layernorm.weight": "model-00006-of-00062.safetensors", + "model.layers.5.mlp.down_proj.weight": "model-00006-of-00062.safetensors", + "model.layers.5.mlp.gate_proj.weight": "model-00005-of-00062.safetensors", + "model.layers.5.mlp.up_proj.weight": "model-00006-of-00062.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model-00006-of-00062.safetensors", + "model.layers.5.self_attn.k_proj.weight": "model-00005-of-00062.safetensors", + "model.layers.5.self_attn.o_proj.weight": "model-00005-of-00062.safetensors", + "model.layers.5.self_attn.q_proj.weight": "model-00005-of-00062.safetensors", + "model.layers.5.self_attn.v_proj.weight": "model-00005-of-00062.safetensors", + "model.layers.50.input_layernorm.weight": "model-00040-of-00062.safetensors", + "model.layers.50.mlp.down_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.50.mlp.gate_proj.weight": "model-00039-of-00062.safetensors", + "model.layers.50.mlp.up_proj.weight": "model-00039-of-00062.safetensors", + "model.layers.50.post_attention_layernorm.weight": "model-00040-of-00062.safetensors", + "model.layers.50.self_attn.k_proj.weight": "model-00039-of-00062.safetensors", + "model.layers.50.self_attn.o_proj.weight": "model-00039-of-00062.safetensors", + "model.layers.50.self_attn.q_proj.weight": "model-00039-of-00062.safetensors", + "model.layers.50.self_attn.v_proj.weight": "model-00039-of-00062.safetensors", + "model.layers.51.input_layernorm.weight": "model-00040-of-00062.safetensors", + "model.layers.51.mlp.down_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.51.mlp.gate_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.51.mlp.up_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.51.post_attention_layernorm.weight": "model-00040-of-00062.safetensors", + "model.layers.51.self_attn.k_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.51.self_attn.o_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.51.self_attn.q_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.51.self_attn.v_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.52.input_layernorm.weight": "model-00041-of-00062.safetensors", + "model.layers.52.mlp.down_proj.weight": "model-00041-of-00062.safetensors", + "model.layers.52.mlp.gate_proj.weight": "model-00041-of-00062.safetensors", + "model.layers.52.mlp.up_proj.weight": "model-00041-of-00062.safetensors", + "model.layers.52.post_attention_layernorm.weight": "model-00041-of-00062.safetensors", + "model.layers.52.self_attn.k_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.52.self_attn.o_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.52.self_attn.q_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.52.self_attn.v_proj.weight": "model-00040-of-00062.safetensors", + "model.layers.53.input_layernorm.weight": "model-00042-of-00062.safetensors", + "model.layers.53.mlp.down_proj.weight": "model-00042-of-00062.safetensors", + "model.layers.53.mlp.gate_proj.weight": "model-00041-of-00062.safetensors", + "model.layers.53.mlp.up_proj.weight": "model-00042-of-00062.safetensors", + "model.layers.53.post_attention_layernorm.weight": "model-00042-of-00062.safetensors", + "model.layers.53.self_attn.k_proj.weight": "model-00041-of-00062.safetensors", + "model.layers.53.self_attn.o_proj.weight": "model-00041-of-00062.safetensors", + "model.layers.53.self_attn.q_proj.weight": "model-00041-of-00062.safetensors", + "model.layers.53.self_attn.v_proj.weight": "model-00041-of-00062.safetensors", + "model.layers.54.input_layernorm.weight": "model-00043-of-00062.safetensors", + "model.layers.54.mlp.down_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.54.mlp.gate_proj.weight": "model-00042-of-00062.safetensors", + "model.layers.54.mlp.up_proj.weight": "model-00042-of-00062.safetensors", + "model.layers.54.post_attention_layernorm.weight": "model-00043-of-00062.safetensors", + "model.layers.54.self_attn.k_proj.weight": "model-00042-of-00062.safetensors", + "model.layers.54.self_attn.o_proj.weight": "model-00042-of-00062.safetensors", + "model.layers.54.self_attn.q_proj.weight": "model-00042-of-00062.safetensors", + "model.layers.54.self_attn.v_proj.weight": "model-00042-of-00062.safetensors", + "model.layers.55.input_layernorm.weight": "model-00043-of-00062.safetensors", + "model.layers.55.mlp.down_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.55.mlp.gate_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.55.mlp.up_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.55.post_attention_layernorm.weight": "model-00043-of-00062.safetensors", + "model.layers.55.self_attn.k_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.55.self_attn.o_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.55.self_attn.q_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.55.self_attn.v_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.56.input_layernorm.weight": "model-00044-of-00062.safetensors", + "model.layers.56.mlp.down_proj.weight": "model-00044-of-00062.safetensors", + "model.layers.56.mlp.gate_proj.weight": "model-00044-of-00062.safetensors", + "model.layers.56.mlp.up_proj.weight": "model-00044-of-00062.safetensors", + "model.layers.56.post_attention_layernorm.weight": "model-00044-of-00062.safetensors", + "model.layers.56.self_attn.k_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.56.self_attn.o_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.56.self_attn.q_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.56.self_attn.v_proj.weight": "model-00043-of-00062.safetensors", + "model.layers.57.input_layernorm.weight": "model-00045-of-00062.safetensors", + "model.layers.57.mlp.down_proj.weight": "model-00045-of-00062.safetensors", + "model.layers.57.mlp.gate_proj.weight": "model-00044-of-00062.safetensors", + "model.layers.57.mlp.up_proj.weight": "model-00045-of-00062.safetensors", + "model.layers.57.post_attention_layernorm.weight": "model-00045-of-00062.safetensors", + "model.layers.57.self_attn.k_proj.weight": "model-00044-of-00062.safetensors", + "model.layers.57.self_attn.o_proj.weight": "model-00044-of-00062.safetensors", + "model.layers.57.self_attn.q_proj.weight": "model-00044-of-00062.safetensors", + "model.layers.57.self_attn.v_proj.weight": "model-00044-of-00062.safetensors", + "model.layers.58.input_layernorm.weight": "model-00046-of-00062.safetensors", + "model.layers.58.mlp.down_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.58.mlp.gate_proj.weight": "model-00045-of-00062.safetensors", + "model.layers.58.mlp.up_proj.weight": "model-00045-of-00062.safetensors", + "model.layers.58.post_attention_layernorm.weight": "model-00046-of-00062.safetensors", + "model.layers.58.self_attn.k_proj.weight": "model-00045-of-00062.safetensors", + "model.layers.58.self_attn.o_proj.weight": "model-00045-of-00062.safetensors", + "model.layers.58.self_attn.q_proj.weight": "model-00045-of-00062.safetensors", + "model.layers.58.self_attn.v_proj.weight": "model-00045-of-00062.safetensors", + "model.layers.59.input_layernorm.weight": "model-00046-of-00062.safetensors", + "model.layers.59.mlp.down_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.59.mlp.gate_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.59.mlp.up_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.59.post_attention_layernorm.weight": "model-00046-of-00062.safetensors", + "model.layers.59.self_attn.k_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.59.self_attn.o_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.59.self_attn.q_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.59.self_attn.v_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.6.input_layernorm.weight": "model-00007-of-00062.safetensors", + "model.layers.6.mlp.down_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.6.mlp.gate_proj.weight": "model-00006-of-00062.safetensors", + "model.layers.6.mlp.up_proj.weight": "model-00006-of-00062.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model-00007-of-00062.safetensors", + "model.layers.6.self_attn.k_proj.weight": "model-00006-of-00062.safetensors", + "model.layers.6.self_attn.o_proj.weight": "model-00006-of-00062.safetensors", + "model.layers.6.self_attn.q_proj.weight": "model-00006-of-00062.safetensors", + "model.layers.6.self_attn.v_proj.weight": "model-00006-of-00062.safetensors", + "model.layers.60.input_layernorm.weight": "model-00047-of-00062.safetensors", + "model.layers.60.mlp.down_proj.weight": "model-00047-of-00062.safetensors", + "model.layers.60.mlp.gate_proj.weight": "model-00047-of-00062.safetensors", + "model.layers.60.mlp.up_proj.weight": "model-00047-of-00062.safetensors", + "model.layers.60.post_attention_layernorm.weight": "model-00047-of-00062.safetensors", + "model.layers.60.self_attn.k_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.60.self_attn.o_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.60.self_attn.q_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.60.self_attn.v_proj.weight": "model-00046-of-00062.safetensors", + "model.layers.61.input_layernorm.weight": "model-00048-of-00062.safetensors", + "model.layers.61.mlp.down_proj.weight": "model-00048-of-00062.safetensors", + "model.layers.61.mlp.gate_proj.weight": "model-00047-of-00062.safetensors", + "model.layers.61.mlp.up_proj.weight": "model-00048-of-00062.safetensors", + "model.layers.61.post_attention_layernorm.weight": "model-00048-of-00062.safetensors", + "model.layers.61.self_attn.k_proj.weight": "model-00047-of-00062.safetensors", + "model.layers.61.self_attn.o_proj.weight": "model-00047-of-00062.safetensors", + "model.layers.61.self_attn.q_proj.weight": "model-00047-of-00062.safetensors", + "model.layers.61.self_attn.v_proj.weight": "model-00047-of-00062.safetensors", + "model.layers.62.input_layernorm.weight": "model-00049-of-00062.safetensors", + "model.layers.62.mlp.down_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.62.mlp.gate_proj.weight": "model-00048-of-00062.safetensors", + "model.layers.62.mlp.up_proj.weight": "model-00048-of-00062.safetensors", + "model.layers.62.post_attention_layernorm.weight": "model-00049-of-00062.safetensors", + "model.layers.62.self_attn.k_proj.weight": "model-00048-of-00062.safetensors", + "model.layers.62.self_attn.o_proj.weight": "model-00048-of-00062.safetensors", + "model.layers.62.self_attn.q_proj.weight": "model-00048-of-00062.safetensors", + "model.layers.62.self_attn.v_proj.weight": "model-00048-of-00062.safetensors", + "model.layers.63.input_layernorm.weight": "model-00049-of-00062.safetensors", + "model.layers.63.mlp.down_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.63.mlp.gate_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.63.mlp.up_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.63.post_attention_layernorm.weight": "model-00049-of-00062.safetensors", + "model.layers.63.self_attn.k_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.63.self_attn.o_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.63.self_attn.q_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.63.self_attn.v_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.64.input_layernorm.weight": "model-00050-of-00062.safetensors", + "model.layers.64.mlp.down_proj.weight": "model-00050-of-00062.safetensors", + "model.layers.64.mlp.gate_proj.weight": "model-00050-of-00062.safetensors", + "model.layers.64.mlp.up_proj.weight": "model-00050-of-00062.safetensors", + "model.layers.64.post_attention_layernorm.weight": "model-00050-of-00062.safetensors", + "model.layers.64.self_attn.k_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.64.self_attn.o_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.64.self_attn.q_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.64.self_attn.v_proj.weight": "model-00049-of-00062.safetensors", + "model.layers.65.input_layernorm.weight": "model-00051-of-00062.safetensors", + "model.layers.65.mlp.down_proj.weight": "model-00051-of-00062.safetensors", + "model.layers.65.mlp.gate_proj.weight": "model-00050-of-00062.safetensors", + "model.layers.65.mlp.up_proj.weight": "model-00051-of-00062.safetensors", + "model.layers.65.post_attention_layernorm.weight": "model-00051-of-00062.safetensors", + "model.layers.65.self_attn.k_proj.weight": "model-00050-of-00062.safetensors", + "model.layers.65.self_attn.o_proj.weight": "model-00050-of-00062.safetensors", + "model.layers.65.self_attn.q_proj.weight": "model-00050-of-00062.safetensors", + "model.layers.65.self_attn.v_proj.weight": "model-00050-of-00062.safetensors", + "model.layers.66.input_layernorm.weight": "model-00052-of-00062.safetensors", + "model.layers.66.mlp.down_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.66.mlp.gate_proj.weight": "model-00051-of-00062.safetensors", + "model.layers.66.mlp.up_proj.weight": "model-00051-of-00062.safetensors", + "model.layers.66.post_attention_layernorm.weight": "model-00052-of-00062.safetensors", + "model.layers.66.self_attn.k_proj.weight": "model-00051-of-00062.safetensors", + "model.layers.66.self_attn.o_proj.weight": "model-00051-of-00062.safetensors", + "model.layers.66.self_attn.q_proj.weight": "model-00051-of-00062.safetensors", + "model.layers.66.self_attn.v_proj.weight": "model-00051-of-00062.safetensors", + "model.layers.67.input_layernorm.weight": "model-00052-of-00062.safetensors", + "model.layers.67.mlp.down_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.67.mlp.gate_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.67.mlp.up_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.67.post_attention_layernorm.weight": "model-00052-of-00062.safetensors", + "model.layers.67.self_attn.k_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.67.self_attn.o_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.67.self_attn.q_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.67.self_attn.v_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.68.input_layernorm.weight": "model-00053-of-00062.safetensors", + "model.layers.68.mlp.down_proj.weight": "model-00053-of-00062.safetensors", + "model.layers.68.mlp.gate_proj.weight": "model-00053-of-00062.safetensors", + "model.layers.68.mlp.up_proj.weight": "model-00053-of-00062.safetensors", + "model.layers.68.post_attention_layernorm.weight": "model-00053-of-00062.safetensors", + "model.layers.68.self_attn.k_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.68.self_attn.o_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.68.self_attn.q_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.68.self_attn.v_proj.weight": "model-00052-of-00062.safetensors", + "model.layers.69.input_layernorm.weight": "model-00054-of-00062.safetensors", + "model.layers.69.mlp.down_proj.weight": "model-00054-of-00062.safetensors", + "model.layers.69.mlp.gate_proj.weight": "model-00053-of-00062.safetensors", + "model.layers.69.mlp.up_proj.weight": "model-00054-of-00062.safetensors", + "model.layers.69.post_attention_layernorm.weight": "model-00054-of-00062.safetensors", + "model.layers.69.self_attn.k_proj.weight": "model-00053-of-00062.safetensors", + "model.layers.69.self_attn.o_proj.weight": "model-00053-of-00062.safetensors", + "model.layers.69.self_attn.q_proj.weight": "model-00053-of-00062.safetensors", + "model.layers.69.self_attn.v_proj.weight": "model-00053-of-00062.safetensors", + "model.layers.7.input_layernorm.weight": "model-00007-of-00062.safetensors", + "model.layers.7.mlp.down_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.7.mlp.gate_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.7.mlp.up_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model-00007-of-00062.safetensors", + "model.layers.7.self_attn.k_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.7.self_attn.o_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.7.self_attn.q_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.7.self_attn.v_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.70.input_layernorm.weight": "model-00055-of-00062.safetensors", + "model.layers.70.mlp.down_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.70.mlp.gate_proj.weight": "model-00054-of-00062.safetensors", + "model.layers.70.mlp.up_proj.weight": "model-00054-of-00062.safetensors", + "model.layers.70.post_attention_layernorm.weight": "model-00055-of-00062.safetensors", + "model.layers.70.self_attn.k_proj.weight": "model-00054-of-00062.safetensors", + "model.layers.70.self_attn.o_proj.weight": "model-00054-of-00062.safetensors", + "model.layers.70.self_attn.q_proj.weight": "model-00054-of-00062.safetensors", + "model.layers.70.self_attn.v_proj.weight": "model-00054-of-00062.safetensors", + "model.layers.71.input_layernorm.weight": "model-00055-of-00062.safetensors", + "model.layers.71.mlp.down_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.71.mlp.gate_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.71.mlp.up_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.71.post_attention_layernorm.weight": "model-00055-of-00062.safetensors", + "model.layers.71.self_attn.k_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.71.self_attn.o_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.71.self_attn.q_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.71.self_attn.v_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.72.input_layernorm.weight": "model-00056-of-00062.safetensors", + "model.layers.72.mlp.down_proj.weight": "model-00056-of-00062.safetensors", + "model.layers.72.mlp.gate_proj.weight": "model-00056-of-00062.safetensors", + "model.layers.72.mlp.up_proj.weight": "model-00056-of-00062.safetensors", + "model.layers.72.post_attention_layernorm.weight": "model-00056-of-00062.safetensors", + "model.layers.72.self_attn.k_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.72.self_attn.o_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.72.self_attn.q_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.72.self_attn.v_proj.weight": "model-00055-of-00062.safetensors", + "model.layers.73.input_layernorm.weight": "model-00057-of-00062.safetensors", + "model.layers.73.mlp.down_proj.weight": "model-00057-of-00062.safetensors", + "model.layers.73.mlp.gate_proj.weight": "model-00056-of-00062.safetensors", + "model.layers.73.mlp.up_proj.weight": "model-00057-of-00062.safetensors", + "model.layers.73.post_attention_layernorm.weight": "model-00057-of-00062.safetensors", + "model.layers.73.self_attn.k_proj.weight": "model-00056-of-00062.safetensors", + "model.layers.73.self_attn.o_proj.weight": "model-00056-of-00062.safetensors", + "model.layers.73.self_attn.q_proj.weight": "model-00056-of-00062.safetensors", + "model.layers.73.self_attn.v_proj.weight": "model-00056-of-00062.safetensors", + "model.layers.74.input_layernorm.weight": "model-00058-of-00062.safetensors", + "model.layers.74.mlp.down_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.74.mlp.gate_proj.weight": "model-00057-of-00062.safetensors", + "model.layers.74.mlp.up_proj.weight": "model-00057-of-00062.safetensors", + "model.layers.74.post_attention_layernorm.weight": "model-00058-of-00062.safetensors", + "model.layers.74.self_attn.k_proj.weight": "model-00057-of-00062.safetensors", + "model.layers.74.self_attn.o_proj.weight": "model-00057-of-00062.safetensors", + "model.layers.74.self_attn.q_proj.weight": "model-00057-of-00062.safetensors", + "model.layers.74.self_attn.v_proj.weight": "model-00057-of-00062.safetensors", + "model.layers.75.input_layernorm.weight": "model-00058-of-00062.safetensors", + "model.layers.75.mlp.down_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.75.mlp.gate_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.75.mlp.up_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.75.post_attention_layernorm.weight": "model-00058-of-00062.safetensors", + "model.layers.75.self_attn.k_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.75.self_attn.o_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.75.self_attn.q_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.75.self_attn.v_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.76.input_layernorm.weight": "model-00059-of-00062.safetensors", + "model.layers.76.mlp.down_proj.weight": "model-00059-of-00062.safetensors", + "model.layers.76.mlp.gate_proj.weight": "model-00059-of-00062.safetensors", + "model.layers.76.mlp.up_proj.weight": "model-00059-of-00062.safetensors", + "model.layers.76.post_attention_layernorm.weight": "model-00059-of-00062.safetensors", + "model.layers.76.self_attn.k_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.76.self_attn.o_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.76.self_attn.q_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.76.self_attn.v_proj.weight": "model-00058-of-00062.safetensors", + "model.layers.77.input_layernorm.weight": "model-00060-of-00062.safetensors", + "model.layers.77.mlp.down_proj.weight": "model-00060-of-00062.safetensors", + "model.layers.77.mlp.gate_proj.weight": "model-00059-of-00062.safetensors", + "model.layers.77.mlp.up_proj.weight": "model-00060-of-00062.safetensors", + "model.layers.77.post_attention_layernorm.weight": "model-00060-of-00062.safetensors", + "model.layers.77.self_attn.k_proj.weight": "model-00059-of-00062.safetensors", + "model.layers.77.self_attn.o_proj.weight": "model-00059-of-00062.safetensors", + "model.layers.77.self_attn.q_proj.weight": "model-00059-of-00062.safetensors", + "model.layers.77.self_attn.v_proj.weight": "model-00059-of-00062.safetensors", + "model.layers.78.input_layernorm.weight": "model-00061-of-00062.safetensors", + "model.layers.78.mlp.down_proj.weight": "model-00061-of-00062.safetensors", + "model.layers.78.mlp.gate_proj.weight": "model-00060-of-00062.safetensors", + "model.layers.78.mlp.up_proj.weight": "model-00060-of-00062.safetensors", + "model.layers.78.post_attention_layernorm.weight": "model-00061-of-00062.safetensors", + "model.layers.78.self_attn.k_proj.weight": "model-00060-of-00062.safetensors", + "model.layers.78.self_attn.o_proj.weight": "model-00060-of-00062.safetensors", + "model.layers.78.self_attn.q_proj.weight": "model-00060-of-00062.safetensors", + "model.layers.78.self_attn.v_proj.weight": "model-00060-of-00062.safetensors", + "model.layers.79.input_layernorm.weight": "model-00061-of-00062.safetensors", + "model.layers.79.mlp.down_proj.weight": "model-00061-of-00062.safetensors", + "model.layers.79.mlp.gate_proj.weight": "model-00061-of-00062.safetensors", + "model.layers.79.mlp.up_proj.weight": "model-00061-of-00062.safetensors", + "model.layers.79.post_attention_layernorm.weight": "model-00061-of-00062.safetensors", + "model.layers.79.self_attn.k_proj.weight": "model-00061-of-00062.safetensors", + "model.layers.79.self_attn.o_proj.weight": "model-00061-of-00062.safetensors", + "model.layers.79.self_attn.q_proj.weight": "model-00061-of-00062.safetensors", + "model.layers.79.self_attn.v_proj.weight": "model-00061-of-00062.safetensors", + "model.layers.8.input_layernorm.weight": "model-00008-of-00062.safetensors", + "model.layers.8.mlp.down_proj.weight": "model-00008-of-00062.safetensors", + "model.layers.8.mlp.gate_proj.weight": "model-00008-of-00062.safetensors", + "model.layers.8.mlp.up_proj.weight": "model-00008-of-00062.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model-00008-of-00062.safetensors", + "model.layers.8.self_attn.k_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.8.self_attn.o_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.8.self_attn.q_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.8.self_attn.v_proj.weight": "model-00007-of-00062.safetensors", + "model.layers.9.input_layernorm.weight": "model-00009-of-00062.safetensors", + "model.layers.9.mlp.down_proj.weight": "model-00009-of-00062.safetensors", + "model.layers.9.mlp.gate_proj.weight": "model-00008-of-00062.safetensors", + "model.layers.9.mlp.up_proj.weight": "model-00009-of-00062.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model-00009-of-00062.safetensors", + "model.layers.9.self_attn.k_proj.weight": "model-00008-of-00062.safetensors", + "model.layers.9.self_attn.o_proj.weight": "model-00008-of-00062.safetensors", + "model.layers.9.self_attn.q_proj.weight": "model-00008-of-00062.safetensors", + "model.layers.9.self_attn.v_proj.weight": "model-00008-of-00062.safetensors", + "model.norm.weight": "model-00061-of-00062.safetensors" + } +}