
    qi                     >   U d Z dZeed<   dZeed<   dZeed<   dZeed<   d	Zeed
<   dZeed<   dZ	eed<   dZ
eed<   dZeed<   dZeed<   dZeed<   dZeed<   dZeed<   dZeed<   dZeed<   dZeed<   dZeed <   d!Zeed"<   dZeed#<   d$Zeed%<   eZeed&<   eZeed'<   y())a  STT service latency defaults.

This module contains P99 time-to-final-segment (TTFS) latency values for STT
services. TTFS measures the time from when speech ends to when the final
transcript is received.

These values are used by turn stop strategies to optimize timing. Each STT
service publishes its latency via STTMetadataFrame at pipeline start.

To measure latency for your specific deployment (region, network conditions,
self-hosted instances), use the STT benchmark tool:
https://github.com/pipecat-ai/stt-benchmark

Run the TTFS benchmark for your service and configuration, then pass the
measured value to your STT service constructor:

    stt = DeepgramSTTService(api_key="...", ttfs_p99_latency=0.45)
g      ?DEFAULT_TTFS_P99gzG?ASSEMBLYAI_TTFS_P99gffffff?AWS_TRANSCRIBE_TTFS_P99g?AZURE_TTFS_P99gQ?CARTESIA_TTFS_P99gffffff?DEEPGRAM_TTFS_P99DEEPGRAM_SAGEMAKER_TTFS_P99gGz @ELEVENLABS_TTFS_P99g=
ףp=?ELEVENLABS_REALTIME_TTFS_P99g(\ @FAL_TTFS_P99gףp=
?GLADIA_TTFS_P99gQ?GOOGLE_TTFS_P99g(\?GRADIUM_TTFS_P99gp=
ף?GROQ_TTFS_P99OPENAI_TTFS_P99g(\?OPENAI_REALTIME_TTFS_P99g@SAMBANOVA_TTFS_P99gQ?SARVAM_TTFS_P99SONIOX_TTFS_P99gGz?SPEECHMATICS_TTFS_P99NVIDIA_TTFS_P99WHISPER_TTFS_P99N)__doc__r   float__annotations__r   r   r   r   r   r   r	   r
   r   r   r   r   r   r   r   r   r   r   r   r   r        N/opt/pipecat/venv/lib/python3.12/site-packages/pipecat/services/stt_latency.py<module>r      s   (  %  " U !!%  %  5  5 %) U )! U !&* e *e    % u  "& % &  E    # u # * )* % *r   