luffycodes's picture
Update README.md
6e2f7e2
|
raw
history blame
1.97 kB
metadata
datasets:
  - luffycodes/Tutorbot-Spock-Bio-Dataset
license: apache-2.0
tags:
  - biology
  - chatgpt
  - llama
  - vicuna
  - tutorbot
  - conversation

NOTE: This "diff model" cannot be used directly.
Users have to apply it on top of the original LLaMA weights to get actual Spock weights.
Please find the instructions here: https://github.com/luffycodes/Tutorbot-Spock-Bio.

Spock Model Card

Github details

Please checkout the repo: https://github.com/luffycodes/Tutorbot-Spock-Bio.

Model details

Model type: Spock is an open-source educational tutoring chatbot trained by fine-tuning LLaMA and Vicuna model on synthetic student-tutorbot conversations generated using a specialized prompt.

Model date: Spock was trained between April 2023 and May 2023.

Organizations developing the model: The Spock team with members from Rice University and OpenStax.

Training dataset

700 conversations generated using a specialized prompt from GPT-4. Dataset link: https://huggingface.co/datasets/luffycodes/Tutorbot-Spock-Bio-Dataset

Paper or resources for more information: https://arxiv.org/abs/2305.13272

Code or resources for more information: https://github.com/luffycodes/Tutorbot-Spock-Bio

License: Apache License 2.0

Where to send questions or comments about the model: Shashank Sonkar (ss164@rice.edu)

If you use this work, please cite: Investigating the Role of Feed-Forward Networks in Transformers Using Parallel Attention and Feed-Forward Net Design https://arxiv.org/abs/2305.13297

@misc{sonkar2023investigating,
      title={Investigating the Role of Feed-Forward Networks in Transformers Using Parallel Attention and Feed-Forward Net Design}, 
      author={Shashank Sonkar and Richard G. Baraniuk},
      year={2023},
      eprint={2305.13297},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}