agentskit.js
Providers

llamacpp

llama.cpp server — run GGUF models on CPU or GPU with minimal overhead.

import { llamacpp } from '@agentskit/adapters'

const adapter = llamacpp({
  url: 'http://localhost:8080',
})

Options

OptionTypeDefault
urlstringhttp://localhost:8080
fetchtypeof fetchglobal

Why llamacpp

  • Runs everywhere, including Raspberry Pi + embedded.
  • GGUF quantizations from 4-bit to 16-bit.
✎ Edit this page on GitHub·Found a problem? Open an issue →·How to contribute →

On this page