This document tracks the development roadmap for infs, a provider-agnostic CLI for running AI models from multiple providers through one consistent interface.
- fal.ai execution — Image generation via async queue API
- Replicate execution — Image generation via prediction polling API
- WaveSpeed AI execution — Image and video generation
- OS keychain integration — Credentials stored securely in the OS keychain via
keyringcrate (falls back tocredentials.tomlwhen keychain is unavailable) -
--jsonoutput flag — Machine-readable JSON output for scripting and automation (infs --json ...) - Shell completion scripts — Generate completions for bash, zsh, fish, PowerShell, and elvish (
infs completions <shell>) - Retry logic with exponential backoff — Automatically retries transient network errors and HTTP 5xx responses with capped exponential backoff
- Streaming LLM responses — Stream tokens as they are generated instead of waiting for the full response (
--streamflag) - Paginated model listing — Handle providers with very large model catalogs via pagination (
--pageand--per-pageflags) - File output for image generation — Automatically download and save generated images to a local file (
--outputflag) - File input support — Pass local files (images, audio, etc.) as input to multimodal models (
--fileflag)
- More providers — ElevenLabs (audio), Stability AI (image), and others
- OAuth support — Support providers that use OAuth-based authentication flows
Have a feature suggestion or want to work on one of the planned items? See CONTRIBUTING.md for guidelines on how to contribute.