Upload README.md
Browse files
README.md
CHANGED
@@ -18,7 +18,7 @@ By accessing this model, you are agreeing to the Llama 2 terms and conditions of
|
|
18 |
|
19 |
## Usage:
|
20 |
|
21 |
-
CodeLlama-13B-QML is a medium-sized Language Model that requires significant computing resources to perform with inference (response) times suitable for automatic code completion. Therefore,
|
22 |
|
23 |
Large Language Models, including CodeLlama-13B-QML, are not designed to be deployed in isolation but instead should be deployed as part of an overall AI system with additional safety guardrails as required. Developers are expected to deploy system safeguards when building AI systems.
|
24 |
|
@@ -31,8 +31,6 @@ The configuration has been thoroughly tested on Ubuntu 22.04 LTS running NVIDIA
|
|
31 |
|
32 |
## How to run CodeLlama-13B-QML in ollama:
|
33 |
|
34 |
-
Note: These instructions are written for ollama version 0.5.7. Be aware that your computer needs significant computing resources for reasonable inference times.
|
35 |
-
|
36 |
#### 1. Install ollama
|
37 |
https://ollama.com/download
|
38 |
|
|
|
18 |
|
19 |
## Usage:
|
20 |
|
21 |
+
CodeLlama-13B-QML is a medium-sized Language Model that requires significant computing resources to perform with inference (response) times suitable for automatic code completion. Therefore, it should be used with GPU accelerator, either in cloud environment such as AWS, Google Cloud, Microsoft Azure, or local.
|
22 |
|
23 |
Large Language Models, including CodeLlama-13B-QML, are not designed to be deployed in isolation but instead should be deployed as part of an overall AI system with additional safety guardrails as required. Developers are expected to deploy system safeguards when building AI systems.
|
24 |
|
|
|
31 |
|
32 |
## How to run CodeLlama-13B-QML in ollama:
|
33 |
|
|
|
|
|
34 |
#### 1. Install ollama
|
35 |
https://ollama.com/download
|
36 |
|