Update README.md
Browse files
README.md
CHANGED
@@ -12,6 +12,11 @@ tags:
|
|
12 |
- text-generation-inference
|
13 |
---
|
14 |
|
|
|
|
|
|
|
|
|
|
|
15 |
|
16 |
```py
|
17 |
pip install transformers
|
@@ -52,4 +57,24 @@ generated_text = generate_text(prompt, max_length=68)
|
|
52 |
|
53 |
# Print the generated text
|
54 |
print(generated_text)
|
55 |
-
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
12 |
- text-generation-inference
|
13 |
---
|
14 |
|
15 |
+
# **Gpt2-Wikitext-9180**
|
16 |
+
|
17 |
+
> **Gpt2-Wikitext-9180**, fine-tuned from GPT-2, is a Transformer-based language model trained on a large English corpus (WikiText) using self-supervised learning. This means it was trained on raw, unlabeled text data, using an automated process to create inputs and labels by predicting the next word in a sentence. No manual annotation was involved, allowing the model to leverage a vast amount of publicly available data.
|
18 |
+
|
19 |
+
## Demo Inference
|
20 |
|
21 |
```py
|
22 |
pip install transformers
|
|
|
57 |
|
58 |
# Print the generated text
|
59 |
print(generated_text)
|
60 |
+
```
|
61 |
+
|
62 |
+
|
63 |
+
---
|
64 |
+
|
65 |
+
### **Intended Use Case**
|
66 |
+
|
67 |
+
* **Text Generation**: Auto-completion, story generation, or dialogue simulation.
|
68 |
+
* **Language Modeling**: Understanding language structure and context for downstream NLP tasks.
|
69 |
+
* **Educational and Research Use**: Exploring fine-tuning techniques, language understanding, or benchmarking language models.
|
70 |
+
* **Prototyping**: Quick deployment of language-based features in applications and interfaces.
|
71 |
+
|
72 |
+
---
|
73 |
+
|
74 |
+
### **Limitations**
|
75 |
+
|
76 |
+
* **Factual Inaccuracy**: May generate plausible-sounding but incorrect or outdated information.
|
77 |
+
* **Bias and Toxicity**: Can reflect biases present in training data (e.g., stereotypes, offensive language).
|
78 |
+
* **Context Length**: Limited context window inherited from GPT-2 architecture.
|
79 |
+
* **Not Real-Time Aware**: Lacks access to current events or updates beyond its training data.
|
80 |
+
* **Lack of Understanding**: Generates text based on patterns, not genuine comprehension or reasoning.
|