> Note: You should have at least 8 GB of RAM to run the 3B models, 16 GB to run the 7B models, and 32 GB to run the 13B models.
> Note: You should have at least 8 GB of RAM to run the 3B models, 16 GB to run the 7B models, and 32 GB to run the 13B models.
## Examples
## Customize your own model
### Pull a public model
### Import from GGUF or GGML
```
Ollama supports importing GGUF and GGML file formats in the Modelfile. This means if you have a model that is not in the Ollama library, you can create it, iterate on it, and upload it to the Ollama library to share with others when you are ready.
ollama pull llama2
```
> This command can also be used to update a local model. Only updated changes will be pulled.
1. Create a file named Modelfile, and add a `FROM` instruction with the local filepath to the model you want to import.
### Run a model interactively
```
FROM ./vicuna-33b.Q4_0.gguf
```
```
3. Create the model in Ollama
ollama run llama2
>>> hi
Hello! How can I help you today?
```
For multiline input, you can wrap text with `"""`:
```
ollama create name -f path_to_modelfile
```
```
5. Run the model
>>> """Hello,
... world!
... """
I'm a basic program that prints the famous "Hello, world!" message to the console.
```
### Run a model non-interactively
```
$ ollama run llama2 'tell me a joke'
Sure! Here's a quick one:
Why did the scarecrow win an award? Because he was outstanding in his field!
```
```
$ cat <<EOF >prompts.txt
tell me a joke about llamas
tell me another one
EOF
$ ollama run llama2 <prompts.txt
>>> tell me a joke about llamas
Why did the llama refuse to play hide-and-seek?
nobody likes to be hided!
>>> tell me another one
```
Sure, here's another one:
ollama run name
```
Why did the llama go to the bar?
### Customize a prompt
To have a hay-often good time!
```
### Run a model on contents of a text file
Models from the Ollama library can be customized with a prompt. The example
```
$ ollama run llama2 "summarize this file:" "$(cat README.md)"
Ollama is a lightweight, extensible framework for building and running language models on the local machine. It provides a simple API for creating, running, and managing models, as well as a library of pre-built models that can be easily used in a variety of applications.
```
### Customize a model
Pull a base model:
```
```
ollama pull llama2
ollama pull llama2
...
@@ -138,30 +108,61 @@ ollama run mario
...
@@ -138,30 +108,61 @@ ollama run mario
Hello! It's your friend Mario.
Hello! It's your friend Mario.
```
```
For more examples, see the [examples](./examples) directory. For more information on creating a Modelfile, see the [Modelfile](./docs/modelfile.md) documentation.
For more examples, see the [examples](./examples) directory. For more information on working with a Modelfile, see the [Modelfile](./docs/modelfile.md) documentation.
## CLI Reference
### Create a model
`ollama create` is used to create a model from a Modelfile.
### Listing local models
### Pull a model
```
```
ollama list
ollama pull llama2
```
```
### Removing local models
> This command can also be used to update a local model. Only the diff will be pulled.
### Remove a model
```
```
ollama rm llama2
ollama rm llama2
```
```
## Model packages
### Copy a model
```
ollama cp llama2 my-llama2
```
### Multiline input
For multiline input, you can wrap text with `"""`:
```
>>> """Hello,
... world!
... """
I'm a basic program that prints the famous "Hello, world!" message to the console.
```
### Pass in prompt as arguments
```
$ ollama run llama2 "summarize this file:" "$(cat README.md)"
Ollama is a lightweight, extensible framework for building and running language models on the local machine. It provides a simple API for creating, running, and managing models, as well as a library of pre-built models that can be easily used in a variety of applications.
```
### List models on your computer
### Overview
```
ollama list
```
Ollama bundles model weights, configurations, and data into a single package, defined by a [Modelfile](./docs/modelfile.md).
### Start Ollama
<picture>
`ollama serve` is used when you want to start ollama without running the desktop application.
@@ -204,19 +205,18 @@ curl -X POST http://localhost:11434/api/generate -d '{
...
@@ -204,19 +205,18 @@ curl -X POST http://localhost:11434/api/generate -d '{
}'
}'
```
```
## Community Projects using Ollama
## Community Integrations
| Project | Description |
-[LangChain](https://python.langchain.com/docs/integrations/llms/ollama) and [LangChain.js](https://js.langchain.com/docs/modules/model_io/models/llms/integrations/ollama) with [example](https://js.langchain.com/docs/use_cases/question_answering/local_retrieval_qa)
| [Continue](https://github.com/continuedev/continue) | Embeds Ollama inside Visual Studio Code. The extension lets you highlight code to add to the prompt, ask questions in the sidebar, and generate code inline. |
-[Discollama](https://github.com/mxyng/discollama)(Discord bot inside the Ollama discord channel)
| [LiteLLM](https://github.com/BerriAI/litellm) | Lightweight Python package to simplify LLM API calls. |