Skip to content

aifoundry-org/NekkoAPI

Repository files navigation

NekkoAPI

OpenAI compatible web API server for llama.cpp.

Disclaimer: this project is in early development stage. Missing features, bugs and poor documentation are expected. This should change in the very near future. Thank you for coming!

Quick Start

You will need: docker, make, git and curl.

This repository uses submodules, therefore you should clone it with:

git clone --recurse-submodules https://github.com/aifoundry-org/NekkoAPI.git
cd NekkoAPI

If you have already cloned it and forgot submodules, you can always initialize them with:

git submodule update --init --recursive

To build NekkoAPI docker image run:

make docker

You can download a few small(ish) (around 5GB total) preconfigured LLM models from Hugging Face with:

make example-models

With models downloaded you can run the Nekko API server to use them with:

make run-example-docker

API will be accesible on http://localhost:8000.

There is a small example using OpenAPI client library that can be used to try out the server in examples/openai-client.py.

Development

WIP

Acknowledgements

This project was made possible thanks to the outstanding work of several open-source contributors and communities. We would like to express our heartfelt gratitude to:

  • Andrei Betlen, for creating and maintaining llama-cpp-python, which served as the foundation for this project. Your efforts in building and maintaining this library have been invaluable.
  • Georgi Gerganov and the ggml authors, for developing llama.cpp and ggml, which are core dependencies of this project. Your innovation and dedication to open-source have greatly enriched the ecosystem and enabled countless projects, including ours.

Thank you!

License

This project is licensed under the terms of the Apache License version 2.0.

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages