prithivMLmods commited on
Commit
dcb1e1a
·
verified ·
1 Parent(s): eb346fe

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +26 -1
README.md CHANGED
@@ -12,6 +12,11 @@ tags:
12
  - text-generation-inference
13
  ---
14
 
 
 
 
 
 
15
 
16
  ```py
17
  pip install transformers
@@ -52,4 +57,24 @@ generated_text = generate_text(prompt, max_length=68)
52
 
53
  # Print the generated text
54
  print(generated_text)
55
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
12
  - text-generation-inference
13
  ---
14
 
15
+ # **Gpt2-Wikitext-9180**
16
+
17
+ > **Gpt2-Wikitext-9180**, fine-tuned from GPT-2, is a Transformer-based language model trained on a large English corpus (WikiText) using self-supervised learning. This means it was trained on raw, unlabeled text data, using an automated process to create inputs and labels by predicting the next word in a sentence. No manual annotation was involved, allowing the model to leverage a vast amount of publicly available data.
18
+
19
+ ## Demo Inference
20
 
21
  ```py
22
  pip install transformers
 
57
 
58
  # Print the generated text
59
  print(generated_text)
60
+ ```
61
+
62
+
63
+ ---
64
+
65
+ ### **Intended Use Case**
66
+
67
+ * **Text Generation**: Auto-completion, story generation, or dialogue simulation.
68
+ * **Language Modeling**: Understanding language structure and context for downstream NLP tasks.
69
+ * **Educational and Research Use**: Exploring fine-tuning techniques, language understanding, or benchmarking language models.
70
+ * **Prototyping**: Quick deployment of language-based features in applications and interfaces.
71
+
72
+ ---
73
+
74
+ ### **Limitations**
75
+
76
+ * **Factual Inaccuracy**: May generate plausible-sounding but incorrect or outdated information.
77
+ * **Bias and Toxicity**: Can reflect biases present in training data (e.g., stereotypes, offensive language).
78
+ * **Context Length**: Limited context window inherited from GPT-2 architecture.
79
+ * **Not Real-Time Aware**: Lacks access to current events or updates beyond its training data.
80
+ * **Lack of Understanding**: Generates text based on patterns, not genuine comprehension or reasoning.