This repository contains examples for interacting with LLMs deployed via llama.cpp (or any other OpenAI API-compatible runtime) using the OpenAI Python library.
Ansible playbooks for setting up llama.cpp on IBM Power (ppc64le) can be found here: DahlitzFlorian/ai-on-ibm-power-playbooks
Create a virtual environment, install the dependencies, and start the desired script.
Make sure to adjust the config-file in the src directory before running any scripts.
python -m venv .venv
source .venv/bin/activate
python -m pip install -r requirements.txt
python src/basic_example.py --config-path src/example_config.ini--config-file(Path, default:src/config.ini): Path to the configuration file.--max-tokens(int, default:300): Maximum number of tokens to predict--prompt-file(Path, default:None): Path to a file reading the prompt from (otherwise script asks for input)--timeout(float, default:300.0): Timeout in seconds