you have two main options for getting local models:
# Install Ollama from https://ollama.ai # Then pull any model: ollama pull llama3.2:1b # Llama 3.2 ollama pull mistral # Mistral 7B - visit Hugging Face
- search for compatible models (e.g., Llama 3)
- download using Git LFS:
# Example for Mistral 7B git lfs install git clone https://huggingface.co/mistralai/Mistral-7B-v0.1choose one of these servers:
-
ollama serve
-
mlx omni server (for mac with apple silicon)
pip install mlx-omni-server mlx-omni-server
make your local server accessible using ngrok or localtunnel:
# For Ollama ngrok http 11434 --host-header="localhost:11434" # For MLX Omni Server ngrok http 10240- go to web.fullmoon.app
- open settings
- enter your endpoint details:
- endpoint URL:
https://your-ngrok-url.ngrok.io/v1 - model name: Same as the model you downloaded (e.g.,
llama2,mistral)
- endpoint URL:
a monorepo containing two versions of the fullmoon web app
- a local version using SQLite for storage
- a web version using IndexedDB for client-side storage live on https://web.fullmoon.app
apps/ ├── local/ # SQLite-based version └── web/ # IndexedDB-based version packages/ ├── database/ # Shared database interface └── ui/ # Shared UI components - Node.js 18+
- pnpm 8+
pnpm installFor local version (sqlite):
# Setup database npx prisma migrate dev # Start development server pnpm dev --filter localFor web version (IndexedDB):
pnpm dev --filter web# Build all pnpm build # Build specific app pnpm build --filter local # or pnpm build --filter webMIT