Skip to main content
Filter by
Sorted by
Tagged with
1 vote
1 answer
55 views

I am using an AudioStreamRandomizer with playback mode = Sequential to play a few dozen sound files. When I play the audio stream, I get the next sound in the sequence, but the starting index is ...
Nicolas Gaborel's user avatar
0 votes
0 answers
43 views

I built an ai voice agent with TTS->LLM->STT pipeline, it should make outbound calls and interact with customers. How do I utilize amazon contact center with kinesis video streams to manage this ...
Zaki's user avatar
  • 107
1 vote
0 answers
53 views

I'm trying to make a script that would take some text and turn it into a speech using edge-tts and then stream it using sounddevice, so that playing audio begins right after edge-tts finished it's job....
Artem Melnyk's user avatar
0 votes
0 answers
36 views

I have been trying to figure out if it is possible to stream media-file containing more than one audios and single video in it! As per the steps of connection establishment in RTSP, when the two-way ...
mvp-acl's user avatar
1 vote
1 answer
85 views

I am trying to achieve live transcription using openai whisper model in my app but having some issues with processing the audio to get the waveform. @router.websocket("/stt/predict/live") ...
Imisioluwa's user avatar
0 votes
0 answers
44 views

I am making a voice assistant using gTTS but when the full speech output is delivered and played , it take a lot time to respond when searching in the Internet about this I found a GitHub issue: https:...
mrithul e's user avatar
1 vote
1 answer
59 views

I'm serving an MJPEG stream and an MP3 stream from an esp32. The individual jpg frames are compressed by the camera module itself, but the microcontroller is not powerful enough to decode and re-...
Crumml's user avatar
  • 71
0 votes
0 answers
79 views

I am using OpenAI’s real-time API (gpt-4o-realtime-preview-2024-12-17) in a React-based application for live transcription and response generation. However, I am facing an issue where the transcribed ...
Akash Chandrasekar's user avatar
0 votes
1 answer
127 views

I'm using Microsoft.CongnitiveServices.Speech SDK to convert text to speech by creating a custom API in C#. The return datatype is 'AudioDataStream'. However, I need to convert it to Stream or bytes ...
Maryam Mirza's user avatar
0 votes
0 answers
31 views

I am currently facing an issue while trying to capture and store synthesized speech from FreeTTS as a byte array. The goal is to take the transcribed text, generate an audio output using FreeTTS, and ...
Breno Souza's user avatar
0 votes
1 answer
224 views

I'm receiving audio chunks over WebSocket and storing them in an AVAudioPCMBuffer for playback using AVAudioEngine. However, when I start playing the audio, I hear noise instead of the expected sound. ...
yarslvd's user avatar
  • 13
-1 votes
1 answer
73 views

I'm having problem streaming audio in my site with the audio element. <audio controls="controls"> <source src="?" type="audio/mpeg" /> Your browser does not ...
42WaysToAnswerThat's user avatar
0 votes
0 answers
291 views

I am trying to play the audio stream chunks I receive from ElevenLabs in real time. When the next chunk is received, the audio player should preload the audio so that once the first chunk finishes, ...
Swapnil Phadke's user avatar
0 votes
1 answer
446 views

I am going to implement Audio Streaming using OpenAI TTS model. It will get audio data from OpenAI TTS model as streaming and send it to frontend via WebSocket. It plays on the Frontend. Frontend is ...
kevin Escalante's user avatar
0 votes
1 answer
278 views

I'm building a real-time audio streaming app using WebSocket and audio APIs, but I'm facing issues specifically on iOS. The app streams audio chunks received via WebSocket from the ElevenLabs API and ...
b_encora's user avatar
1 vote
1 answer
165 views

I am using an API where, when I make a POST request and send an audio file through form data, it returns a streamable audio. I don't know any method to play the audio immediately without having to ...
nghien_rbc's user avatar
0 votes
0 answers
26 views

I'm using the react-native-live-audio-stream package to record. I followed the instructions for in the npm // yarn add buffer import { Buffer } from 'buffer'; ... LiveAudioStream.on('data', data =&...
Shawn's user avatar
  • 531
0 votes
2 answers
71 views

I have an Angular client and one SpringBoot server. On the server, I have an endpoint that generates an audio file from text. The audio file generation depends on another API that take too long (5 ...
Idorasi Paul's user avatar
0 votes
0 answers
49 views

Only on mobile chrome, the speaker is set to phone call during playing a music. I figured out this is because of gainNode. But gainNode is important for looping without gap and etc. I don't want to ...
account 0298's user avatar
2 votes
1 answer
60 views

I am working on an embedded Linux system (5.10.24), where I want to play m3u8 audio stream with FFMPEG APIs. Here is my code. #include <stdio.h> #include <stdbool.h> #include <alsa/...
wangt13's user avatar
  • 1,307
0 votes
3 answers
144 views

I implemented a StreamingResponse in FastAPI with audio bytes from async generator sources. But besides need to insert some messages for client side audio player (currently, React Native) just in the ...
Ivan Vinitskyi's user avatar
2 votes
1 answer
445 views

I was implemented ChatGPT Real-time API, that allows to chat with bot in real time with a voice among other methods. But I've faced a problem that ChatGPT always starts talking with itself when using ...
Alexy Krivzov's user avatar
-3 votes
1 answer
68 views

Is it possible to provide audio file from HTTP server that is consumed by clients via streaming (loading a file partially)? HTTP server shouldn't care about streaming, it thinks consumers just ...
nickolay.laptev's user avatar
1 vote
1 answer
52 views

The page is exibit in a TV on a reception. When it's open we got a dialog with a button to allow the webpage to play audio. And the page refresh itself after 40 seconds. For the most time of the day, ...
Israel Bessa's user avatar
8 votes
0 answers
319 views

Overview I have a server which has an an open WebSocket connection with one client application. This client application (android app) can send live microphone audio data. What the server has to do in ...
thomaoc's user avatar
  • 44

1
2 3 4 5
53