various updates across several ai/vm related systems
This commit is contained in:
@@ -123,15 +123,15 @@ mkdir -p $MODEL_DIR
|
||||
|
||||
# LOCALAI_SINGLE_ACTIVE_BACKEND will unload the previous model before loading the next one
|
||||
# LOCALAI_API_KEY will set an API key, omit to run unprotected.
|
||||
# HF_TOKEN will set a login token for Hugging Face
|
||||
# Good for single-gpu systems.
|
||||
# Use the below to generate a quadlet for /etc/containers/systemd/local-ai.container
|
||||
# podman run --rm ghcr.io/containers/podlet --install --description "Local AI" \
|
||||
podman run \
|
||||
-d \
|
||||
-p 8080:8080 \
|
||||
-e LOCALAI_SINGLE_ACTIVE_BACKEND=true \
|
||||
-e HF_TOKEN=$(cat ~/.cache/huggingface/token) \
|
||||
-e LOCALAI_API_KEY=$(cat ~/.localai/token) \
|
||||
-e LOCALAI_SINGLE_ACTIVE_BACKEND=true \
|
||||
--device /dev/dri \
|
||||
--device /dev/kfd \
|
||||
--name local-ai \
|
||||
@@ -146,12 +146,11 @@ quay.io/go-skynet/local-ai:master-hipblas-ffmpeg
|
||||
podman run \
|
||||
-d \
|
||||
-p 8081:8080 \
|
||||
-e HF_TOKEN=$(cat ~/.cache/huggingface/token) \
|
||||
--name local-ai-webui \
|
||||
--network local-ai \
|
||||
-v $MODEL_DIR:/build/models \
|
||||
-v localai-tmp:/tmp/generated \
|
||||
quay.io/go-skynet/local-ai:master-ffmpeg
|
||||
quay.io/go-skynet/local-ai:master-hipblas-ffmpeg
|
||||
```
|
||||
|
||||
## Running Anything LLM
|
||||
|
||||
@@ -6,6 +6,7 @@
|
||||
- [Unsticking models stuck in "Stopping"](#unsticking-models-stuck-in-stopping)
|
||||
- [Run Anything LLM Interface](#run-anything-llm-interface)
|
||||
- [Installing External Service with Nginx and Certbot](#installing-external-service-with-nginx-and-certbot)
|
||||
- [Ollama Models](#ollama-models)
|
||||
- [Custom Models](#custom-models)
|
||||
- [From Existing Model](#from-existing-model)
|
||||
- [From Scratch](#from-scratch)
|
||||
@@ -48,7 +49,15 @@ Note your ollama instance will be available to podman containers via `http://hos
|
||||
## Install and run Ollama with Podman
|
||||
|
||||
```bash
|
||||
# AMD
|
||||
# Use the below to generate a quadlet for /etc/containers/systemd/local-ai.container
|
||||
# podman run --rm ghcr.io/containers/podlet --install --description "Local AI" \
|
||||
podman run -d --device /dev/kfd --device /dev/dri -v ollama:/root/.ollama -p 11434:11434 --name ollama docker.io/ollama/ollama:rocm
|
||||
|
||||
# CPU
|
||||
# Use the below to generate a quadlet for /etc/containers/systemd/local-ai.container
|
||||
# podman run --rm ghcr.io/containers/podlet --install --description "Local AI" \
|
||||
podman run -d -v ollama:/root/.ollama -p 11434:11434 --name ollama docker.io/ollama/ollama
|
||||
```
|
||||
|
||||
## Unsticking models stuck in "Stopping"
|
||||
@@ -245,6 +254,10 @@ Also consider that podman will not restart your containers at boot. You'll need
|
||||
from the podman run commands. Check out the comments above the podman run commands for more info.
|
||||
Also search the web for "podman quadlets" or ask your AI about it!
|
||||
|
||||
## Ollama Models
|
||||
|
||||
<https://ollama.com/library>
|
||||
|
||||
## Custom Models
|
||||
|
||||
<https://www.gpu-mart.com/blog/import-models-from-huggingface-to-ollama>
|
||||
|
||||
Reference in New Issue
Block a user