Marin Logo

Model Card for Marin 8B

This is the model card for the Marin 8B SFT model. The Marin Project is a collaborative effort to develop open-source foundation models.

Datasets

Datasets used in Marin 8B Base

Marin 8B Base was trained on a variety of datasets:

And some new datasets:

(We are still uploading these datasets. The first three will be licensed per their original licenses. The fourth--based on rephrased web content--will be licensed under CC-BY-SA 4.0.)

A full report is available on our ReadTheDocs site.

Datasets used in Marin 8B Instruct

Marin 8B Instruct is currently an SFT-only model. It was trained on the following datasets:

It is quite likely that we will release improved versions of this model in the future.

Checkpoints

We release a number of training checkpoints. Other checkpoints may be made available on request.

Base Model Checkpoints

Main Page: marin-community/marin-8b-base

(More checkpoints are being uploaded right now.)

Name Training Tokens Link
deeper-starling 13.7T marin-community/marin-8b-base

main currently refers to deeper-starling. This may change in the future, though we will maintain model compatibility. If you require a specific checkpoint, please use the revision argument.

Instruct Model Checkpoints

Main Page: marin-community/marin-8b-instruct

Name Training Tokens Link
deeper-starling-05-15 5.3B marin-community/marin-8b-instruct

main currently refers to deeper-starling-05-15. This may change in the future, though we will maintain model compatibility. If you require a specific checkpoint, please use the revision argument.

Installation

Marin 8B uses the Llama architecture and as such should work out-of-the-box with the Hugging Face Transformers library and any other library that supports the Llama architecture.

We use a variant of the Llama 3 tokenizer: stanford-crfm/marin-tokenizer.

Inference

You can use Marin with the standard HuggingFace Transformers library:

from transformers import AutoModelForCausalLM, AutoTokenizer
marin = AutoModelForCausalLM.from_pretrained("marin-community/marin-8b-base")
tokenizer = AutoTokenizer.from_pretrained("marin-community/marin-8b-base")
message = ["The Marin wind is"]
inputs = tokenizer(message, return_tensors='pt', return_token_type_ids=False)
response = marin.generate(**inputs, max_new_tokens=100, do_sample=True, top_k=50, top_p=0.95)
print(tokenizer.batch_decode(response, skip_special_tokens=True)[0])

We released a number of checkpoints of this model. To load a specific checkpoint, simply add the argument revision:

marin = AutoModelForCausalLM.from_pretrained("marin-community/marin-8b-base", revision="deeper-starling")

Model Description

  • Developed by: The Marin team at Stanford CRFM.
  • Model type: a Transformer style autoregressive language model.
  • Knowledge Cutoff: ~July 2024
  • Language(s) (NLP): English
  • License: The code and model are released under Apache 2.0.
  • Contact: dlwh at stanford.edu

Model Sources

Evaluation

Base Model Results

We ran a suite of standard benchmarks to compare our model with Llama 3.1 8B, and the open source 7-8B models Olmo 2 7B, and MAP NEO 7B. For all benchmarks, we used LM Eval Harness with the default setup for each task. (These numbers may differ from reported results due to differences in setup. LM Eval Harness is usually somewhat stricter than other harnesses.)

Average AGI Eval LSAT-AR ARC Easy ARC Challenge BBH BoolQ CommonSense QA COPA GPQA HellaSwag 0-shot HellaSwag 10-shot lambada_openai MMLU 5-shot MMLU 0-shot MMLU Pro OpenBookQA PIQA WinoGrande WSC
Marin 8B Base (Starling) 68.3 20.9 86.5 63.1 50.6 85.9 79.1 92.0 30.3 82.3 83.6 74.7 67.6 65.9 36.5 44.2 84.4 74.5 82.1
Llama 3.1 Base 67.0 20.4 85.8 58.9 46.4 84.2 75.2 92.0 32.3 79.4 81.9 74.7 66.4 65.5 33.3 45.8 82.9 74.4 83.5
OLMo 2 Base 66.7 17.4 85.0 60.7 44.4 85.5 75.4 89.0 26.8 80.5 81.7 73.1 63.9 61.9 30.6 46.2 82.5 74.3 86.1
MAP NEO 7B 62.2 23.0 81.1 52.0 42.4 84.7 81.7 82.0 27.8 72.5 73.3 64.6 58.2 56.4 TODO 39.4 79.0 66.1 73.3

Marin 8B Base fares well on most tasks.

Model Details

Please see our technical retrospective for more details on the pretraining process.

Architecture Details

  • Architecture: Llama 3 8B
  • Hidden size: 4096
  • Feedforward size: 14336
  • Number of layers: 32
  • Number of attention heads: 32
  • Number of KV heads: 8

Tokenizer Details

Marin 8B uses a variant of the Llama 3 tokenizer: stanford-crfm/marin-tokenizer. It has the same vocabulary but bundles a chat template into the base tokenizer for convenience.

Training Phases

Pre-training Phases

  • Kestrel (DCLM WSD-S Phase): DCLM+StarCoder+Proofpile2 using WSD-S (0->2.7T tokens)
  • Ocelot (DCLM WSD Phase): Increased batch size, using WSD. (2.7T->3.78T tokens)
  • Jellyfish (First Cooldown): Higher quality data (~Dolmino+Fine Math). (3.78T->4.78T tokens)
  • Phoenix (Reheated): Rapid rewarming + Nemotron-CC (plus Starcoder). (4.78T->11.1T tokens)
  • Starling (Second Cooldown): Another cooldown. We followed a similar process to the first cooldown, but added a few new datasets. (11.1T->12.75T tokens)
  • Deeper Starling: Somewhat more pretraining. (12.75T->13.7T tokens)

All released pre-training checkpoints except Kestrel use an exponential moving average of the model weights.

SFT Phase

SFT was comparably simple, consisting of only one phase for 5.3B tokens.

Bias, Risks, and Limitations

Like any base language model or fine-tuned model without safety filtering, these models can easily be prompted by users to generate harmful and sensitive content. Such content may also be produced unintentionally, especially in cases involving bias, so we recommend that users consider the risks when applying this technology. Additionally, many statements from Marin or any LLM are often inaccurate, so responses should be verified.

Marin 8B has not undergone any safety tuning or evaluation. We strongly recommend that users use this model with caution and consider the risks when applying this technology. In particular, this model is not intended for fully autonomous use.

Model Card Contact

For errors in this model card, please open an issue in this repository. For technical inquiries, please contact dlwh at stanford.edu.

Acknowledgements

The compute for this model was generously provided by Google's TPU Research Cloud.

(We based this model card on Olmo 2's.)

Downloads last month
3,062
Safetensors
Model size
8.03B params
Tensor type
F32
·
Inference Providers NEW
Input a message to start chatting with marin-community/marin-8b-instruct.

Model tree for marin-community/marin-8b-instruct

Finetuned
(1)
this model
Quantizations
1 model

Datasets used to train marin-community/marin-8b-instruct

Spaces using marin-community/marin-8b-instruct 2