A codebase for evaluating Large Language Models (LLMs) on financial research tasks. Contains tools and methodologies to assess LLM performance, accuracy, and reliability in financial analysis.
- Python 3.11+
- uv package manager
-
Clone the repository
git clone https://github.com/financial-datasets/llm-evaluations.git cd llm-evaluations -
Install dependencies
uv sync
-
Set up environment variables
cp .env.example .env # Edit .env with your API keys and configuration -
Run the example
uv run main.py
- Fork the repository
- Create a feature branch
- Make your changes
- Add tests if applicable
- Submit a pull request
This project is licensed under the MIT License