Run LLMs locally with Psyllama
Psyllama makes it easy to get up and running with large language models locally. With support for GGUF models and seamless llama.cpp integration.
Fast Setup
Get started in minutes with our simple installation process and intuitive CLI interface.
GGUF Support
Native support for GGUF models with llama.cpp integration for optimal performance.
Developer Friendly
OpenAI-compatible API makes it easy to integrate with your existing tools and applications.
Private & Secure
Run models locally on your hardware. Your data stays private and secure.
GPU Accelerated
Full CUDA support for maximum performance on NVIDIA GPUs.
Cross Platform
Works on Windows, macOS, and Linux with consistent experience across platforms.
Quick start with Psyllama:
# Install Psyllama
curl -fsSL https://psyllama.com/install.sh | sh
# Pull and run a model
psyllama run lexi-8b-uncensored
# Start chatting
psyllama chat lexi-8b-uncensored