Run LLMs locally with Psyllama

Psyllama makes it easy to get up and running with large language models locally. With support for GGUF models and seamless llama.cpp integration.

🚀

Fast Setup

Get started in minutes with our simple installation process and intuitive CLI interface.

🧠

GGUF Support

Native support for GGUF models with llama.cpp integration for optimal performance.

🔧

Developer Friendly

OpenAI-compatible API makes it easy to integrate with your existing tools and applications.

🔒

Private & Secure

Run models locally on your hardware. Your data stays private and secure.

GPU Accelerated

Full CUDA support for maximum performance on NVIDIA GPUs.

🌐

Cross Platform

Works on Windows, macOS, and Linux with consistent experience across platforms.

Quick start with Psyllama:

# Install Psyllama

curl -fsSL https://psyllama.com/install.sh | sh

# Pull and run a model

psyllama run lexi-8b-uncensored

# Start chatting

psyllama chat lexi-8b-uncensored