Skip to content

MicroHealthLLC/TueriGuardian

Repository files navigation

Tueri - The Security Toolkit for LLM Interactions

Tueri is a comprehensive tool designed to fortify the security of Large Language Models (LLMs).

What is Tueri?

By offering sanitization, detection of harmful language, prevention of data leakage, and resistance against prompt injection attacks, Tueri ensures that your interactions with LLMs remain safe and secure.

Getting Started

Important Notes:

  • Tueri is designed for easy integration and deployment in production environments. While it's ready to use out-of-the-box, please be informed that we're constantly improving and updating the repository.
  • Base functionality requires a limited number of libraries. As you explore more advanced features, necessary libraries will be automatically installed.
  • Ensure you're using Python version 3.9 or higher. Confirm with: python --version.
  • Library installation issues? Consider upgrading pip: python -m pip install --upgrade pip.

Examples:

  • Deploy Tueri as API

Supported scanners

Prompt scanners

Output scanners

Acknowledgment

We learned the design and reused code from https://github.com/protectai/llm-guard

About

Guardian for LLMs

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages