There are so many large language models that you can run it locally, e.g. llama, mixtral.

Table of Contents

llama.cpp

The most popular open source LLM framework that support run many LLM locally.

git clone https://github.com/ggerganov/llama.cpp.git
cd llama.cpp
make
./examples/chat-13B.sh

Before you can run the examples, you need download the models first.

ollama

Install the ollama first, then run it

ollama run dolphin-mixtral
pulling manifest
pulling bdb11b0699e0...  84% ▕██████████████████████████████████████████        ▏  22 GB/ 26 GB  6.0 MB/s  11m
ollama -h
Large language model runner

Usage:
  ollama [flags]
  ollama [command]

Available Commands:
  serve       Start ollama
  create      Create a model from a Modelfile
  show        Show information for a model
  run         Run a model
  pull        Pull a model from a registry
  push        Push a model to a registry
  list        List models
  cp          Copy a model
  rm          Remove a model
  help        Help about any command

llamafile

One file that include everything, so just download the llamafile, and run it.

 wget -cS https://huggingface.co/jartine/mistral-7b.llamafile/resolve/main/mistral-7b-instruct-v0.1-Q4_K_M-server.llamafile
chmod +x mistral-7b-instruct-v0.1-Q4_K_M-server.llamafile
./mistral-7b-instruct-v0.1-Q4_K_M-server.llamafile

How to run LLM locally
Tagged on:     

Leave a Reply

Your email address will not be published. Required fields are marked *

This site uses Akismet to reduce spam. Learn how your comment data is processed.