Skip to main content

Quick Start

LlamaEdge is a suite of component libraries and command line tools for developers to embed and run LLMs in their own apps. The best way to quickly experience LlamaEdge is to use easy-to-use utilities built on top of it.

Quick Start with Gaia

Gaia is an integrated tool for running open-source LLMs. It is built on LlamaEdge. Following these simple commands, you will be able to get an Internet-accessible chatbot and an OpenAI-compatible API server running on your devices using any open-source model you choose in a few minutes.

Install the Gaia software with a single command on Mac, Linux, or Windows WSL.

curl -sSfL 'https://github.com/GaiaNet-AI/gaianet-node/releases/latest/download/install.sh' | bash

Then, follow the prompt on your screen to set up the environment path. The command line will begin with source.

Use gaianet init to download the model files and vector database files specified in the $HOME/gaianet/config.json file, and it could take a few minutes since the files are large.

gaianet init

The default $HOME/gaianet/config.json runs a Phi 3.5 LLM and a nomic-embed embedding model. You can easily switch to a Llama 3.1 8b LLM by giving a different configuration to gaianet init. Configurations for many more LLMs are available here.

Start running the Gaia node.

gaianet start

Once it starts on your machine, you can simply go to http://localhost:8080. You can open a browser to that URL to see the node information and then chat with the LLM. This node API server also supports v1/chat/completions and v1/embeddings endpoints, fully compatible with OpenAI APIs.

If you are running it on a server or need to access the LLM sevices from the Internet, the Gaia node has automatically set up connection tunneling for you. The script prints the Internet address for the LLM service on the console as follows.

... ... https://0xf63939431ee11267f4855a166e11cc44d24960c0.us.gaianet.network

To stop running the LLM services, you can run the following script.

gaianet stop

If you're looking to configure LLMs further, explore the details here.

Quick start with Moxin

Moxin is a cross-platform LLM client written in Rust, and built on LlamaEdge components. It offers an intuitive UI for running LLMs with just a few clicks.

Download the Moxin app install package for your device from the Moxin website. Here's how to get started on macOS:

  • Download and install the dmg file from https://www.moxin.app/ on your Macbook.
  • Browse model cards and choose one model to download after open the Moxin app. As models are quite large, this may take several minutes.
  • Engage with the model via a simple and interactive chat interface.