The main focus of this project is on inference and a basic LLM web application serving approach, using an open-source model hosted on a third party hardware.
The project is currently in development and is not intended for production use. The project is open source and can be found on GitHub.
basic-inferllamachat-demo.webm
- Clone the repository:
git clone https://github.com/KrzysiekJa/basic-inferllamachat.git - Navigate to the project directory:
cd basic-inferllamachat - Install
uvpackage manager, if not already installed:pip install uv - Create a virtual environment:
uv venv .venv - Activate the virtual environment (
Linux/macOS):source .venv/bin/activate - Install dependencies using command:
uv sync --locked --all-extras - Create a API key for TogetherAI API on https://together.xyz and export it as an environment variable:
export TOGETHER_API_KEY=<your openai api key> - Create a API key for OpenWeatherMap API on https://openweathermap.org and export it as an environment variable:
export OWM_API_KEY=<your owm api key>, if you intend to use weather chatbot - Run the application:
PYTHONPATH=. python app/main.py
* For Windows users:
- activate the virtual environment using command:
.venv\Scripts\activate, - add `PYTHONPATH``variable following instructions from this stackoverflow thread.