diff --git a/src/audio.c b/src/audio.c index 83c80d8e1..b04cc6797 100644 --- a/src/audio.c +++ b/src/audio.c @@ -176,6 +176,22 @@ typedef struct MusicData { unsigned int samplesLeft; // Number of samples left to end } MusicData; +// AudioStreamData +typedef struct AudioStreamData AudioStreamData; +struct AudioStreamData { + mal_dsp dsp; // AudioStream data needs to flow through a persistent conversion pipeline. Not doing this will result in glitches between buffer updates. + float volume; + float pitch; + bool playing; + bool paused; + bool isSubBufferProcessed[2]; + unsigned int frameCursorPos; + unsigned int bufferSizeInFrames; + AudioStreamData* next; + AudioStreamData* prev; + unsigned char buffer[1]; +}; + #if defined(AUDIO_STANDALONE) typedef enum { LOG_INFO = 0, LOG_ERROR, LOG_WARNING, LOG_DEBUG, LOG_OTHER } TraceLogType; #endif @@ -236,6 +252,8 @@ static float masterVolume = 1; static mal_mutex soundLock; static SoundData* firstSound; // Sounds are tracked in a linked list. static SoundData* lastSound; +static AudioStreamData* firstAudioStream; +static AudioStreamData* lastAudioStream; static void AppendSound(SoundData* internalSound) { @@ -272,6 +290,42 @@ static void RemoveSound(SoundData* internalSound) mal_mutex_unlock(&context, &soundLock); } +static void AppendAudioStream(AudioStreamData* internalAudioStream) +{ + mal_mutex_lock(&context, &soundLock); + { + if (firstAudioStream == NULL) { + firstAudioStream = internalAudioStream; + } else { + lastAudioStream->next = internalAudioStream; + internalAudioStream->prev = lastAudioStream; + } + + lastAudioStream = internalAudioStream; + } + mal_mutex_unlock(&context, &soundLock); +} + +static void RemoveAudioStream(AudioStreamData* internalAudioStream) +{ + mal_mutex_lock(&context, &soundLock); + { + if (internalAudioStream->prev == NULL) { + firstAudioStream = internalAudioStream->next; + } else { + internalAudioStream->prev->next = internalAudioStream->next; + } + + if (internalAudioStream->next == NULL) { + lastAudioStream = internalAudioStream->prev; + } else { + internalAudioStream->next->prev = internalAudioStream->prev; + } + } + mal_mutex_unlock(&context, &soundLock); +} + + static void OnLog_MAL(mal_context* pContext, mal_device* pDevice, const char* message) { (void)pContext; @@ -342,8 +396,63 @@ static mal_uint32 OnSendAudioDataToDevice(mal_device* pDevice, mal_uint32 frameC } } - // Music. - // TODO: Implement me. + // AudioStreams. These are handling slightly differently to sounds because we do data conversion at mixing time rather than + // load time. + for (AudioStreamData* internalData = firstAudioStream; internalData != NULL; internalData = internalData->next) + { + // Ignore stopped or paused streams. + if (!internalData->playing || internalData->paused) { + continue; + } + + mal_uint32 framesRead = 0; + for (;;) { + if (framesRead > frameCount) { + TraceLog(LOG_DEBUG, "Mixed too many frames from sound"); + break; + } + if (framesRead == frameCount) { + break; + } + + // Just read as much data we can from the stream. + mal_uint32 framesToRead = (frameCount - framesRead); + while (framesToRead > 0) { + float tempBuffer[1024]; // 512 frames for stereo. + + mal_uint32 framesToReadRightNow = framesToRead; + if (framesToReadRightNow > sizeof(tempBuffer)/DEVICE_CHANNELS) { + framesToReadRightNow = sizeof(tempBuffer)/DEVICE_CHANNELS; + } + + mal_uint32 framesJustRead = mal_dsp_read_frames(&internalData->dsp, framesToReadRightNow, tempBuffer); + if (framesJustRead > 0) { + // This is where the real mixing takes place. This can be optimized. This assumes the device and sound are of the same format. + // + // TODO: Implement pitching. + for (mal_uint32 iFrame = 0; iFrame < framesToRead; ++iFrame) { + float* pFrameOut = pFramesOutF + ((framesRead+iFrame) * device.channels); + float* pFrameIn = tempBuffer + (iFrame * device.channels); + + for (mal_uint32 iChannel = 0; iChannel < device.channels; ++iChannel) { + pFrameOut[iChannel] += pFrameIn[iChannel] * masterVolume * internalData->volume; + } + } + + framesToRead -= framesJustRead; + framesRead += framesJustRead; + } else { + break; // Avoid an infinite loop. + } + } + + // If for some reason we weren't able to read every frame we'll need to break from the loop. Not doing this could + // theoretically put us into an infinite loop. + if (framesToRead > 0) { + break; + } + } + } } mal_mutex_unlock(&context, &soundLock); @@ -1375,6 +1484,65 @@ float GetMusicTimePlayed(Music music) return secondsPlayed; } + +static mal_uint32 UpdateAudioStream_OnDSPRead(mal_uint32 frameCount, void* pFramesOut, void* pUserData) +{ + AudioStreamData* internalData = (AudioStreamData*)pUserData; + + mal_uint32 subBufferSizeInFrames = AUDIO_BUFFER_SIZE; + mal_uint32 currentSubBufferIndex = internalData->frameCursorPos / subBufferSizeInFrames; + if (currentSubBufferIndex > 1) { + TraceLog(LOG_DEBUG, "Frame cursor position moved too far forward in audio stream"); + return 0; + } + + // Another thread can update the processed state of buffers so we just take a copy here to try and avoid potential synchronization problems. + bool isSubBufferProcessed[2]; + isSubBufferProcessed[0] = internalData->isSubBufferProcessed[0]; + isSubBufferProcessed[1] = internalData->isSubBufferProcessed[1]; + + mal_uint32 channels = internalData->dsp.config.channelsIn; + mal_uint32 sampleSizeInBytes = mal_get_sample_size_in_bytes(internalData->dsp.config.formatIn); + mal_uint32 frameSizeInBytes = sampleSizeInBytes*channels; + + // Fill out every frame until we find a buffer that's marked as processed. Then fill the remainder with 0. + mal_uint32 framesRead = 0; + while (!isSubBufferProcessed[currentSubBufferIndex]) + { + mal_uint32 totalFramesRemaining = (frameCount - framesRead); + if (totalFramesRemaining == 0) { + break; + } + + mal_uint32 firstFrameIndexOfThisSubBuffer = subBufferSizeInFrames * currentSubBufferIndex; + mal_uint32 framesRemainingInThisSubBuffer = subBufferSizeInFrames - (internalData->frameCursorPos - firstFrameIndexOfThisSubBuffer); + + mal_uint32 framesToRead = totalFramesRemaining; + if (framesToRead > framesRemainingInThisSubBuffer) { + framesToRead = framesRemainingInThisSubBuffer; + } + + memcpy((unsigned char*)pFramesOut + (framesRead*frameSizeInBytes), internalData->buffer + (internalData->frameCursorPos*frameSizeInBytes), framesToRead*frameSizeInBytes); + + framesRead += framesToRead; + internalData->frameCursorPos = (internalData->frameCursorPos + framesToRead) % internalData->bufferSizeInFrames; + + // If we've read to the end of the buffer, mark it as processed. + if (framesToRead == framesRemainingInThisSubBuffer) { + internalData->isSubBufferProcessed[currentSubBufferIndex] = true; + currentSubBufferIndex = (currentSubBufferIndex + 1) % 2; + } + } + + // Zero-fill excess. + mal_uint32 totalFramesRemaining = (frameCount - framesRead); + if (totalFramesRemaining > 0) { + memset((unsigned char*)pFramesOut + (framesRead*frameSizeInBytes), 0, totalFramesRemaining*frameSizeInBytes); + } + + return frameCount; +} + // Init audio stream (to stream audio pcm data) AudioStream InitAudioStream(unsigned int sampleRate, unsigned int sampleSize, unsigned int channels) { @@ -1391,6 +1559,41 @@ AudioStream InitAudioStream(unsigned int sampleRate, unsigned int sampleSize, un stream.channels = 1; // Fallback to mono channel } + +#if USE_MINI_AL + AudioStreamData* internalData = (AudioStreamData*)calloc(1, sizeof(*internalData) + (AUDIO_BUFFER_SIZE*2 * stream.channels*(stream.sampleSize/8))); + if (internalData == NULL) + { + TraceLog(LOG_ERROR, "Failed to allocate buffer for audio stream"); + return stream; + } + + mal_dsp_config config; + memset(&config, 0, sizeof(config)); + config.formatIn = ((stream.sampleSize == 8) ? mal_format_u8 : ((stream.sampleSize == 16) ? mal_format_s16 : mal_format_f32)); + config.channelsIn = stream.channels; + config.sampleRateIn = stream.sampleRate; + config.formatOut = DEVICE_FORMAT; + config.channelsOut = DEVICE_CHANNELS; + config.sampleRateOut = DEVICE_SAMPLE_RATE; + mal_result result = mal_dsp_init(&config, UpdateAudioStream_OnDSPRead, internalData, &internalData->dsp); + if (result != MAL_SUCCESS) + { + TraceLog(LOG_ERROR, "InitAudioStream() : Failed to initialize data conversion pipeline"); + free(internalData); + return stream; + } + + // Buffers should be marked as processed by default so that a call to UpdateAudioStream() immediately after initialization works correctly. + internalData->isSubBufferProcessed[0] = true; + internalData->isSubBufferProcessed[1] = true; + internalData->bufferSizeInFrames = AUDIO_BUFFER_SIZE*2; + internalData->volume = 1; + internalData->pitch = 1; + AppendAudioStream(internalData); + + stream.handle = internalData; +#else // Setup OpenAL format if (stream.channels == 1) { @@ -1435,6 +1638,7 @@ AudioStream InitAudioStream(unsigned int sampleRate, unsigned int sampleSize, un free(pcm); alSourceQueueBuffers(stream.source, MAX_STREAM_BUFFERS, stream.buffers); +#endif TraceLog(LOG_INFO, "[AUD ID %i] Audio stream loaded successfully (%i Hz, %i bit, %s)", stream.source, stream.sampleRate, stream.sampleSize, (stream.channels == 1) ? "Mono" : "Stereo"); @@ -1444,6 +1648,11 @@ AudioStream InitAudioStream(unsigned int sampleRate, unsigned int sampleSize, un // Close audio stream and free memory void CloseAudioStream(AudioStream stream) { +#if USE_MINI_AL + AudioStreamData* internalData = (AudioStreamData*)stream.handle; + RemoveAudioStream(internalData); + free(internalData); +#else // Stop playing channel alSourceStop(stream.source); @@ -1462,7 +1671,8 @@ void CloseAudioStream(AudioStream stream) // Delete source and buffers alDeleteSources(1, &stream.source); alDeleteBuffers(MAX_STREAM_BUFFERS, stream.buffers); - +#endif + TraceLog(LOG_INFO, "[AUD ID %i] Unloaded audio stream data", stream.source); } @@ -1471,6 +1681,67 @@ void CloseAudioStream(AudioStream stream) // NOTE 2: To unqueue a buffer it needs to be processed: IsAudioBufferProcessed() void UpdateAudioStream(AudioStream stream, const void *data, int samplesCount) { +#if USE_MINI_AL + AudioStreamData* internalData = (AudioStreamData*)stream.handle; + if (internalData == NULL) + { + TraceLog(LOG_ERROR, "Invalid audio stream"); + return; + } + + // We need to determine which half of the buffer needs updating. If the stream is not started and the cursor position is + // at the front of the buffer, update the first subbuffer. + if (internalData->isSubBufferProcessed[0] || internalData->isSubBufferProcessed[1]) + { + mal_uint32 subBufferToUpdate; + if (internalData->isSubBufferProcessed[0] && internalData->isSubBufferProcessed[1]) + { + // Both buffers are available for updating. Update the first one and make sure the cursor is moved back to the front. + subBufferToUpdate = 0; + internalData->frameCursorPos = 0; + } + else + { + // Just update whichever sub-buffer is processed. + subBufferToUpdate = (internalData->isSubBufferProcessed[0]) ? 0 : 1; + } + + mal_uint32 subBufferSizeInFrames = AUDIO_BUFFER_SIZE; + unsigned char *subBuffer = internalData->buffer + ((subBufferSizeInFrames * stream.channels * (stream.sampleSize/8)) * subBufferToUpdate); + + // Does this API expect a whole buffer to be updated in one go? Assuming so, but if not will need to change this logic. + if (subBufferSizeInFrames >= (mal_uint32)samplesCount) + { + mal_uint32 framesToWrite = subBufferSizeInFrames; + if (framesToWrite > (mal_uint32)samplesCount) { + framesToWrite = (mal_uint32)samplesCount; + } + + mal_uint32 bytesToWrite = framesToWrite * stream.channels * (stream.sampleSize/8); + memcpy(subBuffer, data, bytesToWrite); + + // Any leftover frames should be filled with zeros. + mal_uint32 leftoverFrameCount = subBufferSizeInFrames - framesToWrite; + if (leftoverFrameCount > 0) { + memset(subBuffer + bytesToWrite, 0, leftoverFrameCount * stream.channels * (stream.sampleSize/8)); + } + + internalData->isSubBufferProcessed[subBufferToUpdate] = false; + } + else + { + TraceLog(LOG_ERROR, "[AUD ID %i] UpdateAudioStream() : Attempting to write too many frames to buffer"); + return; + } + } + else + { + TraceLog(LOG_ERROR, "[AUD ID %i] Audio buffer not available for updating"); + return; + } + + +#else ALuint buffer = 0; alSourceUnqueueBuffers(stream.source, 1, &buffer); @@ -1481,44 +1752,104 @@ void UpdateAudioStream(AudioStream stream, const void *data, int samplesCount) alSourceQueueBuffers(stream.source, 1, &buffer); } else TraceLog(LOG_WARNING, "[AUD ID %i] Audio buffer not available for unqueuing", stream.source); +#endif } // Check if any audio stream buffers requires refill bool IsAudioBufferProcessed(AudioStream stream) { +#if USE_MINI_AL + AudioStreamData* internalData = (AudioStreamData*)stream.handle; + if (internalData == NULL) + { + TraceLog(LOG_ERROR, "Invalid audio stream"); + return false; + } + + return internalData->isSubBufferProcessed[0] || internalData->isSubBufferProcessed[1]; +#else ALint processed = 0; // Determine if music stream is ready to be written alGetSourcei(stream.source, AL_BUFFERS_PROCESSED, &processed); return (processed > 0); +#endif } // Play audio stream void PlayAudioStream(AudioStream stream) { +#if USE_MINI_AL + AudioStreamData* internalData = (AudioStreamData*)stream.handle; + if (internalData == NULL) + { + TraceLog(LOG_ERROR, "Invalid audio stream"); + return; + } + + internalData->playing = true; +#else alSourcePlay(stream.source); +#endif } // Play audio stream void PauseAudioStream(AudioStream stream) { +#if USE_MINI_AL + AudioStreamData* internalData = (AudioStreamData*)stream.handle; + if (internalData == NULL) + { + TraceLog(LOG_ERROR, "Invalid audio stream"); + return; + } + + internalData->paused = true; +#else alSourcePause(stream.source); +#endif } // Resume audio stream playing void ResumeAudioStream(AudioStream stream) { +#if USE_MINI_AL + AudioStreamData* internalData = (AudioStreamData*)stream.handle; + if (internalData == NULL) + { + TraceLog(LOG_ERROR, "Invalid audio stream"); + return; + } + + internalData->paused = false; +#else ALenum state; alGetSourcei(stream.source, AL_SOURCE_STATE, &state); if (state == AL_PAUSED) alSourcePlay(stream.source); +#endif } // Stop audio stream void StopAudioStream(AudioStream stream) { +#if USE_MINI_AL + AudioStreamData* internalData = (AudioStreamData*)stream.handle; + if (internalData == NULL) + { + TraceLog(LOG_ERROR, "Invalid audio stream"); + return; + } + + internalData->playing = 0; + internalData->paused = 0; + internalData->frameCursorPos = 0; + internalData->isSubBufferProcessed[0] = true; + internalData->isSubBufferProcessed[1] = true; +#else alSourceStop(stream.source); +#endif } //---------------------------------------------------------------------------------- diff --git a/src/raylib.h b/src/raylib.h index b21b08785..5e3724c9a 100644 --- a/src/raylib.h +++ b/src/raylib.h @@ -504,6 +504,8 @@ typedef struct AudioStream { unsigned int sampleSize; // Bit depth (bits per sample): 8, 16, 32 (24 not supported) unsigned int channels; // Number of channels (1-mono, 2-stereo) + void* handle; // A pointer to internal data used by the audio system. + int format; // OpenAL audio format specifier unsigned int source; // OpenAL audio source id unsigned int buffers[2]; // OpenAL audio buffers (double buffering)