Pipecat welcomes community-maintained integrations! As our ecosystem grows, we've established a process for any developer to create and maintain their own service integrations while ensuring discoverability for the Pipecat community.
What we support: Community-maintained integrations that live in separate repositories and are maintained by their authors.
What we don't do: The Pipecat team does not code review, test, or maintain community integrations. We provide guidance and list approved integrations for discoverability.
Why this approach: This allows the community to move quickly while keeping the Pipecat core team focused on maintaining the framework itself.
To be listed as an official community integration, follow these steps:
Create your integration following the patterns and examples shown in the "Integration Patterns and Examples" section below.
Your repository must contain these components:
-
Source code - Complete implementation following Pipecat patterns
-
Foundational example - Single file example showing basic usage (see Pipecat examples)
-
README.md - Must include:
- Introduction and explanation of your integration
- Installation instructions
- Usage instructions with Pipecat Pipeline
- How to run your example
- Pipecat version compatibility (e.g., "Tested with Pipecat v0.0.86")
- Company attribution: If you work for the company providing the service, please mention this in your README. This helps build confidence that the integration will be actively maintained.
-
LICENSE - Permissive license (BSD-2 like Pipecat, or equivalent open source terms)
-
Code documentation - Source code with docstrings (we recommend following Pipecat's docstring conventions)
-
Changelog - Maintain a changelog for version updates
Join our Discord: https://discord.gg/pipecat
Submit a pull request to add your integration to our Community Integrations documentation page.
To submit:
- Fork the Pipecat docs repository
- Edit the file
server/services/community-integrations.mdx - Add your integration to the appropriate service category table with:
- Service name
- Link to your repository
- Maintainer GitHub username(s)
- Include a link to your demo video (approx 30-60 seconds) in your PR description showing:
- Core functionality of your integration
- Handling of an interruption (if applicable to service type)
- Submit your pull request
Once your PR is submitted, post in the #community-integrations Discord channel to let us know.
Base class: STTService
Examples:
Base class: SegmentedSTTService
Examples:
- STT services should push
InterimTranscriptionFramesandTranscriptionFrames - If confidence values are available, filter for values >50% confidence
Base class: OpenAILLMService
Examples:
- AzureLLMService
- GrokLLMService - Shows overriding the base class where needed
Requires: Full implementation
Examples:
-
Frame sequence: Output must follow this frame sequence pattern:
LLMFullResponseStartFrame- Signals the start of an LLM responseLLMTextFrame- Contains LLM content, typically streamed as tokensLLMFullResponseEndFrame- Signals the end of an LLM response
-
Context aggregation: Implement context aggregation to collect user and assistant content:
- Aggregators come in pairs with a
user()instance andassistant()instance - Context must adhere to the
LLMContextuniversal format - Aggregators should handle adding messages, function calls, and images to the context
- Aggregators come in pairs with a
Use for: Websocket-based services supporting word/timestamp alignment
Example:
Use for: Websocket-based services without word/timestamp alignment, requiring disconnection on interruption
Example:
Use for: HTTP-based services supporting word/timestamp alignment
Example:
Use for: HTTP-based services without word/timestamp alignment
Example:
- For websocket services, use asyncio WebSocket implementation (required for v13+ support)
- Handle idle service timeouts with keepalives
- TTSServices push both audio (
TTSRawAudioFrame) and text (TTSTextFrame) frames
Pipecat supports telephony provider integration using websocket connections to exchange MediaStreams. These services use a FrameSerializer to serialize and deserialize inputs from the FastAPIWebsocketTransport.
Examples:
- Include hang-up functionality using the provider's native API, ideally using
aiohttp - Support DTMF (dual-tone multi-frequency) events if the provider supports them:
- Deserialize DTMF events from the provider's protocol to
InputDTMFFrame - Use
KeypadEntryenum for valid keypad entries (0-9, *, #, A-D) - Handle invalid DTMF digits gracefully by returning
None
- Deserialize DTMF events from the provider's protocol to
Base class: ImageGenService
Examples:
- Must implement
run_image_genmethod returning anAsyncGenerator
Vision services process images and provide analysis such as descriptions, object detection, or visual question answering.
Base class: VisionService
Example:
- Must implement
run_visionmethod that takes anLLMContextand returns anAsyncGenerator[Frame, None] - The method processes the latest image in the context and yields frames with analysis results
- Typically yields
TextFrameobjects containing descriptions or answers
- STT:
VendorSTTService - LLM:
VendorLLMService - TTS:
- Websocket:
VendorTTSService - HTTP:
VendorHttpTTSService
- Websocket:
- Image:
VendorImageGenService - Vision:
VendorVisionService - Telephony:
VendorFrameSerializer
Enable metrics in your service:
def can_generate_metrics(self) -> bool:
"""Check if this service can generate processing metrics.
Returns:
True, as this service supports metrics.
"""
return TrueSTT, LLM, and TTS services support ServiceUpdateSettingsFrame for dynamic configuration changes. The base STTService has an _update_settings() method that handles settings, and the private _settings Dict is used to store settings and provide access to the subclass.
async def set_language(self, language: Language):
"""Set the recognition language and reconnect.
Args:
language: The language to use for speech recognition.
"""
logger.info(f"Switching STT language to: [{language}]")
self._settings["language"] = language
await self._disconnect()
await self._connect()Note that, in this example, Deepgram requires the websocket connection be disconnected and reconnected to reinitialize the service with the new value. Consider if your service requires reconnection.
Sample rates are set via PipelineParams and passed to each frame processor at initialization. The pattern is to not set the sample rate value in the constructor of a given service. Instead, use the start() method to initialize sample rates from the frame:
async def start(self, frame: StartFrame):
"""Start the service."""
await super().start(frame)
self._settings["output_format"]["sample_rate"] = self.sample_rate
await self._connect()Note that self.sample_rate is a @property set in the TTSService base class, which provides access to the private sample rate value obtained from the StartFrame.
Use Pipecat's tracing decorators:
- STT:
@traced_stt- decorate a function that handlestranscript,is_final,languageas args - LLM:
@traced_llm- decorate the_process_context()method - TTS:
@traced_tts- decorate therun_tts()method
- Use uv for packaging (encouraged)
- Consider releasing to PyPI for easier installation
- Follow semantic versioning principles
- Maintain a changelog
For REST-based communication, use aiohttp. Pipecat includes this as a required dependency, so using it prevents adding an additional dependency to your integration.
- Wrap API calls in appropriate try/catch blocks
- Handle rate limits and network failures gracefully
- Provide meaningful error messages
- When errors occur, raise exceptions AND push
ErrorFrames to notify the pipeline:
from pipecat.frames.frames import ErrorFrame
try:
# Your API call
result = await self._make_api_call()
except Exception as e:
# Push error frame to pipeline
await self.push_error(ErrorFrame(error=f"{self} error: {e}"))
# Raise or handle as appropriate
raise- Your foundational example serves as a valuable integration-level test
- Unit tests are nice to have. As the Pipecat teams provides better guidance, we will encourage unit testing more
Community integrations are community-maintained and not officially supported by the Pipecat team. Users should evaluate these integrations independently. The Pipecat team reserves the right to remove listings that become unmaintained or problematic.
Pipecat evolves rapidly to support the latest AI technologies and patterns. While we strive to minimize breaking changes, they do occur as the framework matures.
We strongly recommend:
- Join our Discord at https://discord.gg/pipecat and monitor the
#announcementschannel for release notifications - Follow our changelog: https://github.com/pipecat-ai/pipecat/blob/main/CHANGELOG.md
- Test your integration against new Pipecat releases promptly
- Update your README with the last tested Pipecat version
This helps ensure your integration remains compatible and your users have clear expectations about version support.
Join our Discord community at https://discord.gg/pipecat and post in the #community-integrations channel for guidance and support.
For additional questions, you can also reach out to us at pipecat-ai@daily.co.