raylib/src/audio.c

1142 lines
40 KiB
C
Raw Normal View History

/**********************************************************************************************
*
* raylib.audio
*
* Basic functions to manage Audio: InitAudioDevice, LoadAudioFiles, PlayAudioFiles
2014-09-03 18:51:28 +04:00
*
* Uses external lib:
* OpenAL Soft - Audio device management lib (http://kcat.strangesoft.net/openal.html)
* stb_vorbis - Ogg audio files loading (http://www.nothings.org/stb_vorbis/)
2014-09-03 18:51:28 +04:00
*
* Copyright (c) 2014 Ramon Santamaria (@raysan5)
2014-09-03 18:51:28 +04:00
*
* This software is provided "as-is", without any express or implied warranty. In no event
* will the authors be held liable for any damages arising from the use of this software.
*
2014-09-03 18:51:28 +04:00
* Permission is granted to anyone to use this software for any purpose, including commercial
* applications, and to alter it and redistribute it freely, subject to the following restrictions:
*
2014-09-03 18:51:28 +04:00
* 1. The origin of this software must not be misrepresented; you must not claim that you
* wrote the original software. If you use this software in a product, an acknowledgment
* in the product documentation would be appreciated but is not required.
*
* 2. Altered source versions must be plainly marked as such, and must not be misrepresented
* as being the original software.
*
* 3. This notice may not be removed or altered from any source distribution.
*
**********************************************************************************************/
//#define AUDIO_STANDALONE // NOTE: To use the audio module as standalone lib, just uncomment this line
#if defined(AUDIO_STANDALONE)
#include "audio.h"
#else
#include "raylib.h"
#endif
#include "AL/al.h" // OpenAL basic header
#include "AL/alc.h" // OpenAL context header (like OpenGL, OpenAL requires a context to work)
#include <stdlib.h> // Declares malloc() and free() for memory management
#include <string.h> // Required for strcmp()
#include <stdio.h> // Used for .WAV loading
#if defined(AUDIO_STANDALONE)
#include <stdarg.h> // Used for functions with variable number of parameters (TraceLog())
#else
#include "utils.h" // rRES data decompression utility function
// NOTE: Includes Android fopen function map
#endif
//#define STB_VORBIS_HEADER_ONLY
2016-04-26 04:40:19 +03:00
#include "stb_vorbis.h" // OGG loading functions
#define JAR_XM_IMPLEMENTATION
#include "jar_xm.h" // For playing .xm files
//----------------------------------------------------------------------------------
// Defines and Macros
//----------------------------------------------------------------------------------
#define MUSIC_STREAM_BUFFERS 2
2016-05-01 02:05:43 +03:00
#define MAX_AUDIO_CONTEXTS 4
#if defined(PLATFORM_RPI) || defined(PLATFORM_ANDROID)
// NOTE: On RPI and Android should be lower to avoid frame-stalls
#define MUSIC_BUFFER_SIZE 4096*2 // PCM data buffer (short) - 16Kb (RPI)
2015-02-02 02:53:49 +03:00
#else
// NOTE: On HTML5 (emscripten) this is allocated on heap, by default it's only 16MB!...just take care...
#define MUSIC_BUFFER_SIZE 4096*8 // PCM data buffer (short) - 64Kb
#endif
//----------------------------------------------------------------------------------
// Types and Structures Definition
//----------------------------------------------------------------------------------
// Music type (file streaming from memory)
// NOTE: Anything longer than ~10 seconds should be streamed...
typedef struct Music {
stb_vorbis *stream;
2016-04-25 04:18:18 +03:00
jar_xm_context_t *chipctx; // Stores jar_xm context
2014-09-03 18:51:28 +04:00
ALuint buffers[MUSIC_STREAM_BUFFERS];
ALuint source;
ALenum format;
int channels;
int sampleRate;
2014-09-03 18:51:28 +04:00
int totalSamplesLeft;
2016-04-26 06:05:03 +03:00
float totalLengthSeconds;
2014-09-03 18:51:28 +04:00
bool loop;
2016-04-25 04:18:18 +03:00
bool chipTune; // True if chiptune is loaded
} Music;
// Audio Context, used to create custom audio streams that are not bound to a sound file. There can be
// no more than 4 concurrent audio contexts in use. This is due to each active context being tied to
// a dedicated mix channel.
typedef struct AudioContext_t {
unsigned short sampleRate; // default is 48000
2016-05-01 02:05:43 +03:00
unsigned char bitsPerSample; // 16 is default
unsigned char mixChannel; // 0-3 or mixA-mixD, each mix channel can receive up to one dedicated audio stream
unsigned char channels; // 1=mono, 2=stereo
2016-05-01 01:41:46 +03:00
ALenum alFormat; // openAL format specifier
ALuint alSource; // openAL source
ALuint alBuffer[2]; // openAL sample buffer
} AudioContext_t;
#if defined(AUDIO_STANDALONE)
typedef enum { INFO = 0, ERROR, WARNING, DEBUG, OTHER } TraceLogType;
#endif
//----------------------------------------------------------------------------------
// Global Variables Definition
//----------------------------------------------------------------------------------
2016-05-01 02:05:43 +03:00
static AudioContext_t* mixChannelsActive_g[MAX_AUDIO_CONTEXTS]; // What mix channels are currently active
static bool musicEnabled = false;
2016-05-01 02:05:43 +03:00
static Music currentMusic; // Current music loaded
// NOTE: Only one music file playing at a time
//----------------------------------------------------------------------------------
// Module specific Functions Declaration
//----------------------------------------------------------------------------------
static Wave LoadWAV(const char *fileName); // Load WAV file
static Wave LoadOGG(char *fileName); // Load OGG file
static void UnloadWave(Wave wave); // Unload wave data
static bool BufferMusicStream(ALuint buffer); // Fill music buffers with data
static void EmptyMusicStream(void); // Empty music buffers
#if defined(AUDIO_STANDALONE)
const char *GetExtension(const char *fileName); // Get the extension for a filename
void TraceLog(int msgType, const char *text, ...); // Outputs a trace log message (INFO, ERROR, WARNING)
#endif
//----------------------------------------------------------------------------------
// Module Functions Definition - Audio Device initialization and Closing
//----------------------------------------------------------------------------------
// Initialize audio device and context
void InitAudioDevice(void)
{
// Open and initialize a device with default settings
ALCdevice *device = alcOpenDevice(NULL);
2014-09-03 18:51:28 +04:00
if(!device) TraceLog(ERROR, "Audio device could not be opened");
ALCcontext *context = alcCreateContext(device, NULL);
2014-09-03 18:51:28 +04:00
if(context == NULL || alcMakeContextCurrent(context) == ALC_FALSE)
{
if(context != NULL) alcDestroyContext(context);
2014-09-03 18:51:28 +04:00
alcCloseDevice(device);
2014-09-03 18:51:28 +04:00
TraceLog(ERROR, "Could not setup audio context");
}
TraceLog(INFO, "Audio device and context initialized successfully: %s", alcGetString(device, ALC_DEVICE_SPECIFIER));
2014-09-03 18:51:28 +04:00
// Listener definition (just for 2D)
alListener3f(AL_POSITION, 0, 0, 0);
alListener3f(AL_VELOCITY, 0, 0, 0);
alListener3f(AL_ORIENTATION, 0, 0, -1);
}
// Close the audio device for the current context, and destroys the context
void CloseAudioDevice(void)
{
StopMusicStream(); // Stop music streaming and close current stream
ALCdevice *device;
ALCcontext *context = alcGetCurrentContext();
2014-09-03 18:51:28 +04:00
if (context == NULL) TraceLog(WARNING, "Could not get current audio context for closing");
device = alcGetContextsDevice(context);
alcMakeContextCurrent(NULL);
alcDestroyContext(context);
alcCloseDevice(device);
}
// True if call to InitAudioDevice() was successful and CloseAudioDevice() has not been called yet
2016-05-01 02:05:43 +03:00
bool IsAudioDeviceReady(void)
{
ALCcontext *context = alcGetCurrentContext();
if (context == NULL) return false;
else{
ALCdevice *device = alcGetContextsDevice(context);
if (device == NULL) return false;
else return true;
}
}
//----------------------------------------------------------------------------------
// Module Functions Definition - Custom audio output
//----------------------------------------------------------------------------------
// Audio contexts are for outputing custom audio waveforms, This will shut down any other sound sources currently playing
2016-05-01 02:05:43 +03:00
// The mixChannel is what mix channel you want to operate on, 0-3 are the ones available. Each mix channel can only be used one at a time.
// exmple usage is InitAudioContext(48000, 16, 0, 2); // stereo, mixchannel 1, 16bit, 48khz
AudioContext InitAudioContext(unsigned short sampleRate, unsigned char bitsPerSample, unsigned char mixChannel, unsigned char channels)
{
2016-05-01 02:05:43 +03:00
if(mixChannel > MAX_AUDIO_CONTEXTS) return NULL;
if(!IsAudioDeviceReady()) InitAudioDevice();
else StopMusicStream();
if(!mixChannelsActive_g[mixChannel]){
AudioContext_t *ac = malloc(sizeof(AudioContext_t));
ac->sampleRate = sampleRate;
ac->bitsPerSample = bitsPerSample;
ac->mixChannel = mixChannel;
ac->channels = channels;
mixChannelsActive_g[mixChannel] = ac;
2016-05-01 01:41:46 +03:00
// setup openAL format
2016-05-01 02:05:43 +03:00
if (channels == 1)
2016-05-01 01:41:46 +03:00
{
2016-05-01 02:05:43 +03:00
if (bitsPerSample == 8 ) ac->alFormat = AL_FORMAT_MONO8;
else if (bitsPerSample == 16) ac->alFormat = AL_FORMAT_MONO16;
2016-05-01 01:41:46 +03:00
}
2016-05-01 02:05:43 +03:00
else if (channels == 2)
2016-05-01 01:41:46 +03:00
{
2016-05-01 02:05:43 +03:00
if (bitsPerSample == 8 ) ac->alFormat = AL_FORMAT_STEREO8;
else if (bitsPerSample == 16) ac->alFormat = AL_FORMAT_STEREO16;
2016-05-01 01:41:46 +03:00
}
// Create an audio source
alGenSources(1, &ac->alSource);
alSourcef(ac->alSource, AL_PITCH, 1);
alSourcef(ac->alSource, AL_GAIN, 1);
alSource3f(ac->alSource, AL_POSITION, 0, 0, 0);
alSource3f(ac->alSource, AL_VELOCITY, 0, 0, 0);
// Create Buffer
alGenBuffers(2, &ac->alBuffer);
return ac;
}
return NULL;
}
// Frees buffer in audio context
void CloseAudioContext(AudioContext ctx)
{
AudioContext_t *context = (AudioContext_t*)ctx;
if(context){
2016-05-01 01:41:46 +03:00
alDeleteSources(1, &context->alSource);
alDeleteBuffers(2, &context->alBuffer);
mixChannelsActive_g[context->mixChannel] = NULL;
free(context);
2016-05-01 01:41:46 +03:00
ctx = NULL;
}
}
// Pushes more audio data into context mix channel, if none are ever pushed then zeros are fed in
void UpdateAudioContext(AudioContext ctx, void *data, unsigned short *dataLength)
{
2016-05-01 01:41:46 +03:00
AudioContext_t *context = (AudioContext_t*)ctx;
if(!musicEnabled && context && mixChannelsActive_g[context->mixChannel] == context)
{
;
}
}
//----------------------------------------------------------------------------------
// Module Functions Definition - Sounds loading and playing (.WAV)
//----------------------------------------------------------------------------------
// Load sound to memory
Sound LoadSound(char *fileName)
{
2016-01-23 15:22:13 +03:00
Sound sound = { 0 };
Wave wave = { 0 };
// NOTE: The entire file is loaded to memory to play it all at once (no-streaming)
2014-09-03 18:51:28 +04:00
// Audio file loading
// NOTE: Buffer space is allocated inside function, Wave must be freed
2014-09-03 18:51:28 +04:00
if (strcmp(GetExtension(fileName),"wav") == 0) wave = LoadWAV(fileName);
else if (strcmp(GetExtension(fileName),"ogg") == 0) wave = LoadOGG(fileName);
else TraceLog(WARNING, "[%s] Sound extension not recognized, it can't be loaded", fileName);
2014-09-03 18:51:28 +04:00
if (wave.data != NULL)
{
ALenum format = 0;
// The OpenAL format is worked out by looking at the number of channels and the bits per sample
2014-09-03 18:51:28 +04:00
if (wave.channels == 1)
{
if (wave.bitsPerSample == 8 ) format = AL_FORMAT_MONO8;
else if (wave.bitsPerSample == 16) format = AL_FORMAT_MONO16;
2014-09-03 18:51:28 +04:00
}
else if (wave.channels == 2)
{
if (wave.bitsPerSample == 8 ) format = AL_FORMAT_STEREO8;
else if (wave.bitsPerSample == 16) format = AL_FORMAT_STEREO16;
}
2014-09-03 18:51:28 +04:00
// Create an audio source
ALuint source;
alGenSources(1, &source); // Generate pointer to audio source
2014-09-03 18:51:28 +04:00
alSourcef(source, AL_PITCH, 1);
alSourcef(source, AL_GAIN, 1);
alSource3f(source, AL_POSITION, 0, 0, 0);
alSource3f(source, AL_VELOCITY, 0, 0, 0);
alSourcei(source, AL_LOOPING, AL_FALSE);
2014-09-03 18:51:28 +04:00
// Convert loaded data to OpenAL buffer
//----------------------------------------
ALuint buffer;
alGenBuffers(1, &buffer); // Generate pointer to buffer
// Upload sound data to buffer
alBufferData(buffer, format, wave.data, wave.dataSize, wave.sampleRate);
// Attach sound buffer to source
alSourcei(source, AL_BUFFER, buffer);
2015-02-02 02:53:49 +03:00
TraceLog(INFO, "[%s] Sound file loaded successfully (SampleRate: %i, BitRate: %i, Channels: %i)", fileName, wave.sampleRate, wave.bitsPerSample, wave.channels);
2014-09-03 18:51:28 +04:00
// Unallocate WAV data
UnloadWave(wave);
2014-09-03 18:51:28 +04:00
sound.source = source;
sound.buffer = buffer;
}
2014-09-03 18:51:28 +04:00
return sound;
}
// Load sound from wave data
Sound LoadSoundFromWave(Wave wave)
{
2016-01-23 15:22:13 +03:00
Sound sound = { 0 };
if (wave.data != NULL)
{
ALenum format = 0;
// The OpenAL format is worked out by looking at the number of channels and the bits per sample
if (wave.channels == 1)
{
if (wave.bitsPerSample == 8 ) format = AL_FORMAT_MONO8;
else if (wave.bitsPerSample == 16) format = AL_FORMAT_MONO16;
}
else if (wave.channels == 2)
{
if (wave.bitsPerSample == 8 ) format = AL_FORMAT_STEREO8;
else if (wave.bitsPerSample == 16) format = AL_FORMAT_STEREO16;
}
// Create an audio source
ALuint source;
alGenSources(1, &source); // Generate pointer to audio source
alSourcef(source, AL_PITCH, 1);
alSourcef(source, AL_GAIN, 1);
alSource3f(source, AL_POSITION, 0, 0, 0);
alSource3f(source, AL_VELOCITY, 0, 0, 0);
alSourcei(source, AL_LOOPING, AL_FALSE);
// Convert loaded data to OpenAL buffer
//----------------------------------------
ALuint buffer;
alGenBuffers(1, &buffer); // Generate pointer to buffer
// Upload sound data to buffer
alBufferData(buffer, format, wave.data, wave.dataSize, wave.sampleRate);
// Attach sound buffer to source
alSourcei(source, AL_BUFFER, buffer);
// Unallocate WAV data
UnloadWave(wave);
TraceLog(INFO, "[Wave] Sound file loaded successfully (SampleRate: %i, BitRate: %i, Channels: %i)", wave.sampleRate, wave.bitsPerSample, wave.channels);
sound.source = source;
sound.buffer = buffer;
}
return sound;
}
// Load sound to memory from rRES file (raylib Resource)
2016-02-12 14:22:56 +03:00
// TODO: Maybe rresName could be directly a char array with all the data?
Sound LoadSoundFromRES(const char *rresName, int resId)
{
2016-01-23 15:22:13 +03:00
Sound sound = { 0 };
#if defined(AUDIO_STANDALONE)
TraceLog(WARNING, "Sound loading from rRES resource file not supported on standalone mode");
#else
bool found = false;
char id[4]; // rRES file identifier
unsigned char version; // rRES file version and subversion
char useless; // rRES header reserved data
short numRes;
2014-09-03 18:51:28 +04:00
ResInfoHeader infoHeader;
2014-09-03 18:51:28 +04:00
FILE *rresFile = fopen(rresName, "rb");
2015-02-02 02:53:49 +03:00
if (rresFile == NULL)
{
TraceLog(WARNING, "[%s] rRES raylib resource file could not be opened", rresName);
}
else
{
// Read rres file (basic file check - id)
fread(&id[0], sizeof(char), 1, rresFile);
fread(&id[1], sizeof(char), 1, rresFile);
fread(&id[2], sizeof(char), 1, rresFile);
fread(&id[3], sizeof(char), 1, rresFile);
fread(&version, sizeof(char), 1, rresFile);
fread(&useless, sizeof(char), 1, rresFile);
2014-09-03 18:51:28 +04:00
if ((id[0] != 'r') && (id[1] != 'R') && (id[2] != 'E') &&(id[3] != 'S'))
{
TraceLog(WARNING, "[%s] This is not a valid raylib resource file", rresName);
}
else
{
// Read number of resources embedded
fread(&numRes, sizeof(short), 1, rresFile);
2014-09-03 18:51:28 +04:00
for (int i = 0; i < numRes; i++)
{
fread(&infoHeader, sizeof(ResInfoHeader), 1, rresFile);
2014-09-03 18:51:28 +04:00
if (infoHeader.id == resId)
{
found = true;
// Check data is of valid SOUND type
if (infoHeader.type == 1) // SOUND data type
{
// TODO: Check data compression type
// NOTE: We suppose compression type 2 (DEFLATE - default)
2014-09-03 18:51:28 +04:00
// Reading SOUND parameters
Wave wave;
short sampleRate, bps;
char channels, reserved;
2014-09-03 18:51:28 +04:00
fread(&sampleRate, sizeof(short), 1, rresFile); // Sample rate (frequency)
fread(&bps, sizeof(short), 1, rresFile); // Bits per sample
fread(&channels, 1, 1, rresFile); // Channels (1 - mono, 2 - stereo)
fread(&reserved, 1, 1, rresFile); // <reserved>
2014-09-03 18:51:28 +04:00
wave.sampleRate = sampleRate;
wave.dataSize = infoHeader.srcSize;
wave.bitsPerSample = bps;
wave.channels = (short)channels;
2014-09-03 18:51:28 +04:00
unsigned char *data = malloc(infoHeader.size);
fread(data, infoHeader.size, 1, rresFile);
2014-09-03 18:51:28 +04:00
wave.data = DecompressData(data, infoHeader.size, infoHeader.srcSize);
2014-09-03 18:51:28 +04:00
free(data);
2014-09-03 18:51:28 +04:00
// Convert wave to Sound (OpenAL)
ALenum format = 0;
2014-09-03 18:51:28 +04:00
// The OpenAL format is worked out by looking at the number of channels and the bits per sample
2014-09-03 18:51:28 +04:00
if (wave.channels == 1)
{
if (wave.bitsPerSample == 8 ) format = AL_FORMAT_MONO8;
else if (wave.bitsPerSample == 16) format = AL_FORMAT_MONO16;
2014-09-03 18:51:28 +04:00
}
else if (wave.channels == 2)
{
if (wave.bitsPerSample == 8 ) format = AL_FORMAT_STEREO8;
else if (wave.bitsPerSample == 16) format = AL_FORMAT_STEREO16;
}
2014-09-03 18:51:28 +04:00
// Create an audio source
ALuint source;
alGenSources(1, &source); // Generate pointer to audio source
2014-09-03 18:51:28 +04:00
alSourcef(source, AL_PITCH, 1);
alSourcef(source, AL_GAIN, 1);
alSource3f(source, AL_POSITION, 0, 0, 0);
alSource3f(source, AL_VELOCITY, 0, 0, 0);
alSourcei(source, AL_LOOPING, AL_FALSE);
2014-09-03 18:51:28 +04:00
// Convert loaded data to OpenAL buffer
//----------------------------------------
ALuint buffer;
alGenBuffers(1, &buffer); // Generate pointer to buffer
// Upload sound data to buffer
alBufferData(buffer, format, (void*)wave.data, wave.dataSize, wave.sampleRate);
// Attach sound buffer to source
alSourcei(source, AL_BUFFER, buffer);
2015-02-02 02:53:49 +03:00
TraceLog(INFO, "[%s] Sound loaded successfully from resource (SampleRate: %i, BitRate: %i, Channels: %i)", rresName, wave.sampleRate, wave.bitsPerSample, wave.channels);
2014-09-03 18:51:28 +04:00
// Unallocate WAV data
UnloadWave(wave);
sound.source = source;
sound.buffer = buffer;
}
else
{
TraceLog(WARNING, "[%s] Required resource do not seem to be a valid SOUND resource", rresName);
}
}
else
{
// Depending on type, skip the right amount of parameters
switch (infoHeader.type)
{
case 0: fseek(rresFile, 6, SEEK_CUR); break; // IMAGE: Jump 6 bytes of parameters
case 1: fseek(rresFile, 6, SEEK_CUR); break; // SOUND: Jump 6 bytes of parameters
case 2: fseek(rresFile, 5, SEEK_CUR); break; // MODEL: Jump 5 bytes of parameters (TODO: Review)
case 3: break; // TEXT: No parameters
case 4: break; // RAW: No parameters
default: break;
}
2014-09-03 18:51:28 +04:00
// Jump DATA to read next infoHeader
fseek(rresFile, infoHeader.size, SEEK_CUR);
2014-09-03 18:51:28 +04:00
}
}
}
2014-09-03 18:51:28 +04:00
fclose(rresFile);
}
2014-09-03 18:51:28 +04:00
if (!found) TraceLog(WARNING, "[%s] Required resource id [%i] could not be found in the raylib resource file", rresName, resId);
#endif
return sound;
}
// Unload sound
void UnloadSound(Sound sound)
{
alDeleteSources(1, &sound.source);
alDeleteBuffers(1, &sound.buffer);
2015-08-05 20:17:56 +03:00
TraceLog(INFO, "Unloaded sound data");
}
// Play a sound
void PlaySound(Sound sound)
{
alSourcePlay(sound.source); // Play the sound
2014-09-03 18:51:28 +04:00
//TraceLog(INFO, "Playing sound");
// Find the current position of the sound being played
// NOTE: Only work when the entire file is in a single buffer
//int byteOffset;
//alGetSourcei(sound.source, AL_BYTE_OFFSET, &byteOffset);
//
//int sampleRate;
//alGetBufferi(sound.buffer, AL_FREQUENCY, &sampleRate); // AL_CHANNELS, AL_BITS (bps)
2014-09-03 18:51:28 +04:00
//float seconds = (float)byteOffset / sampleRate; // Number of seconds since the beginning of the sound
//or
//float result;
//alGetSourcef(sound.source, AL_SEC_OFFSET, &result); // AL_SAMPLE_OFFSET
}
// Pause a sound
void PauseSound(Sound sound)
{
alSourcePause(sound.source);
}
// Stop reproducing a sound
void StopSound(Sound sound)
{
alSourceStop(sound.source);
}
// Check if a sound is playing
bool SoundIsPlaying(Sound sound)
{
bool playing = false;
ALint state;
2014-09-03 18:51:28 +04:00
alGetSourcei(sound.source, AL_SOURCE_STATE, &state);
if (state == AL_PLAYING) playing = true;
2014-09-03 18:51:28 +04:00
return playing;
}
// Set volume for a sound
void SetSoundVolume(Sound sound, float volume)
{
alSourcef(sound.source, AL_GAIN, volume);
}
// Set pitch for a sound
void SetSoundPitch(Sound sound, float pitch)
{
alSourcef(sound.source, AL_PITCH, pitch);
}
//----------------------------------------------------------------------------------
// Module Functions Definition - Music loading and stream playing (.OGG)
//----------------------------------------------------------------------------------
// Start music playing (open stream)
void PlayMusicStream(char *fileName)
{
if (strcmp(GetExtension(fileName),"ogg") == 0)
{
// Stop current music, clean buffers, unload current stream
StopMusicStream();
2014-09-03 18:51:28 +04:00
// Open audio stream
currentMusic.stream = stb_vorbis_open_filename(fileName, NULL, NULL);
2014-09-03 18:51:28 +04:00
if (currentMusic.stream == NULL)
{
TraceLog(WARNING, "[%s] OGG audio file could not be opened", fileName);
}
else
{
// Get file info
stb_vorbis_info info = stb_vorbis_get_info(currentMusic.stream);
2014-09-03 18:51:28 +04:00
currentMusic.channels = info.channels;
currentMusic.sampleRate = info.sample_rate;
2014-09-03 18:51:28 +04:00
TraceLog(INFO, "[%s] Ogg sample rate: %i", fileName, info.sample_rate);
TraceLog(INFO, "[%s] Ogg channels: %i", fileName, info.channels);
2015-12-03 15:45:06 +03:00
TraceLog(DEBUG, "[%s] Temp memory required: %i", fileName, info.temp_memory_required);
2014-09-03 18:51:28 +04:00
if (info.channels == 2) currentMusic.format = AL_FORMAT_STEREO16;
else currentMusic.format = AL_FORMAT_MONO16;
2014-09-03 18:51:28 +04:00
currentMusic.loop = true; // We loop by default
musicEnabled = true;
2014-09-03 18:51:28 +04:00
// Create an audio source
alGenSources(1, &currentMusic.source); // Generate pointer to audio source
2014-09-03 18:51:28 +04:00
alSourcef(currentMusic.source, AL_PITCH, 1);
alSourcef(currentMusic.source, AL_GAIN, 1);
alSource3f(currentMusic.source, AL_POSITION, 0, 0, 0);
alSource3f(currentMusic.source, AL_VELOCITY, 0, 0, 0);
//alSourcei(currentMusic.source, AL_LOOPING, AL_TRUE); // ERROR: Buffers do not queue!
2014-09-03 18:51:28 +04:00
// Generate two OpenAL buffers
alGenBuffers(2, currentMusic.buffers);
// Fill buffers with music...
BufferMusicStream(currentMusic.buffers[0]);
BufferMusicStream(currentMusic.buffers[1]);
2014-09-03 18:51:28 +04:00
// Queue buffers and start playing
alSourceQueueBuffers(currentMusic.source, 2, currentMusic.buffers);
alSourcePlay(currentMusic.source);
2014-09-03 18:51:28 +04:00
2015-12-03 15:45:06 +03:00
// NOTE: Regularly, we must check if a buffer has been processed and refill it: UpdateMusicStream()
currentMusic.totalSamplesLeft = stb_vorbis_stream_length_in_samples(currentMusic.stream) * currentMusic.channels;
currentMusic.totalLengthSeconds = stb_vorbis_stream_length_in_seconds(currentMusic.stream);
}
}
2016-04-25 04:18:18 +03:00
else if (strcmp(GetExtension(fileName),"xm") == 0)
{
// Stop current music, clean buffers, unload current stream
StopMusicStream();
// new song settings for xm chiptune
2016-04-25 04:18:18 +03:00
currentMusic.chipTune = true;
currentMusic.channels = 2;
currentMusic.sampleRate = 48000;
currentMusic.loop = true;
2016-04-26 06:05:03 +03:00
// only stereo is supported for xm
2016-04-26 04:40:19 +03:00
if(!jar_xm_create_context_from_file(&currentMusic.chipctx, currentMusic.sampleRate, fileName))
2016-04-25 04:18:18 +03:00
{
2016-04-27 02:50:07 +03:00
currentMusic.format = AL_FORMAT_STEREO16;
2016-04-26 04:40:19 +03:00
jar_xm_set_max_loop_count(currentMusic.chipctx, 0); // infinite number of loops
currentMusic.totalSamplesLeft = jar_xm_get_remaining_samples(currentMusic.chipctx);
2016-04-27 02:50:07 +03:00
currentMusic.totalLengthSeconds = ((float)currentMusic.totalSamplesLeft) / ((float)currentMusic.sampleRate);
2016-04-25 04:18:18 +03:00
musicEnabled = true;
2016-04-26 06:05:03 +03:00
2016-04-27 02:50:07 +03:00
TraceLog(INFO, "[%s] XM number of samples: %i", fileName, currentMusic.totalSamplesLeft);
TraceLog(INFO, "[%s] XM track length: %11.6f sec", fileName, currentMusic.totalLengthSeconds);
2016-04-26 06:05:03 +03:00
// Set up OpenAL
alGenSources(1, &currentMusic.source);
alSourcef(currentMusic.source, AL_PITCH, 1);
alSourcef(currentMusic.source, AL_GAIN, 1);
alSource3f(currentMusic.source, AL_POSITION, 0, 0, 0);
alSource3f(currentMusic.source, AL_VELOCITY, 0, 0, 0);
alGenBuffers(2, currentMusic.buffers);
BufferMusicStream(currentMusic.buffers[0]);
BufferMusicStream(currentMusic.buffers[1]);
alSourceQueueBuffers(currentMusic.source, 2, currentMusic.buffers);
alSourcePlay(currentMusic.source);
2016-04-27 02:50:07 +03:00
// NOTE: Regularly, we must check if a buffer has been processed and refill it: UpdateMusicStream()
2016-04-25 04:18:18 +03:00
}
2016-04-27 02:50:07 +03:00
else TraceLog(WARNING, "[%s] XM file could not be opened", fileName);
2016-04-25 04:18:18 +03:00
}
else TraceLog(WARNING, "[%s] Music extension not recognized, it can't be loaded", fileName);
}
// Stop music playing (close stream)
void StopMusicStream(void)
{
if (musicEnabled)
{
2016-04-25 04:18:18 +03:00
alSourceStop(currentMusic.source);
EmptyMusicStream(); // Empty music buffers
alDeleteSources(1, &currentMusic.source);
alDeleteBuffers(2, currentMusic.buffers);
if (currentMusic.chipTune)
{
jar_xm_free_context(currentMusic.chipctx);
}
else
{
stb_vorbis_close(currentMusic.stream);
}
}
2014-09-03 18:51:28 +04:00
musicEnabled = false;
}
// Pause music playing
void PauseMusicStream(void)
{
// Pause music stream if music available!
if (musicEnabled)
{
TraceLog(INFO, "Pausing music stream");
alSourcePause(currentMusic.source);
2015-01-21 02:12:54 +03:00
musicEnabled = false;
}
}
// Resume music playing
void ResumeMusicStream(void)
{
// Resume music playing... if music available!
2015-01-21 02:12:54 +03:00
ALenum state;
alGetSourcei(currentMusic.source, AL_SOURCE_STATE, &state);
2015-02-02 02:53:49 +03:00
2015-01-21 02:12:54 +03:00
if (state == AL_PAUSED)
{
2015-01-21 02:12:54 +03:00
TraceLog(INFO, "Resuming music stream");
alSourcePlay(currentMusic.source);
2015-01-21 02:12:54 +03:00
musicEnabled = true;
}
}
// Check if music is playing
bool MusicIsPlaying(void)
{
bool playing = false;
ALint state;
2014-09-03 18:51:28 +04:00
alGetSourcei(currentMusic.source, AL_SOURCE_STATE, &state);
if (state == AL_PLAYING) playing = true;
2014-09-03 18:51:28 +04:00
return playing;
}
// Set volume for music
void SetMusicVolume(float volume)
{
alSourcef(currentMusic.source, AL_GAIN, volume);
}
// Get current music time length (in seconds)
float GetMusicTimeLength(void)
{
2016-04-25 04:18:18 +03:00
float totalSeconds;
if (currentMusic.chipTune)
{
2016-04-27 02:50:07 +03:00
totalSeconds = currentMusic.totalLengthSeconds;
2016-04-25 04:18:18 +03:00
}
else
{
totalSeconds = stb_vorbis_stream_length_in_seconds(currentMusic.stream);
}
2014-09-03 18:51:28 +04:00
return totalSeconds;
}
// Get current music time played (in seconds)
float GetMusicTimePlayed(void)
{
2016-04-25 04:18:18 +03:00
float secondsPlayed;
if (currentMusic.chipTune)
{
2016-04-25 08:00:40 +03:00
uint64_t samples;
2016-04-26 06:05:03 +03:00
jar_xm_get_position(currentMusic.chipctx, NULL, NULL, NULL, &samples);
secondsPlayed = (float)samples / (currentMusic.sampleRate * currentMusic.channels); // Not sure if this is the correct value
2016-04-25 04:18:18 +03:00
}
else
{
int totalSamples = stb_vorbis_stream_length_in_samples(currentMusic.stream) * currentMusic.channels;
int samplesPlayed = totalSamples - currentMusic.totalSamplesLeft;
secondsPlayed = (float)samplesPlayed / (currentMusic.sampleRate * currentMusic.channels);
}
2014-09-03 18:51:28 +04:00
return secondsPlayed;
}
//----------------------------------------------------------------------------------
// Module specific Functions Definition
//----------------------------------------------------------------------------------
// Fill music buffers with new data from music stream
static bool BufferMusicStream(ALuint buffer)
{
2014-09-03 18:51:28 +04:00
short pcm[MUSIC_BUFFER_SIZE];
2014-09-03 18:51:28 +04:00
int size = 0; // Total size of data steamed (in bytes)
int streamedBytes = 0; // samples of data obtained, channels are not included in calculation
bool active = true; // We can get more data from stream (not finished)
2014-09-03 18:51:28 +04:00
if (musicEnabled)
{
if (currentMusic.chipTune) // There is no end of stream for xmfiles, once the end is reached zeros are generated for non looped chiptunes.
{
int readlen = MUSIC_BUFFER_SIZE / 2;
jar_xm_generate_samples_16bit(currentMusic.chipctx, pcm, readlen); // reads 2*readlen shorts and moves them to buffer+size memory location
size += readlen * currentMusic.channels; // Not sure if this is what it needs
}
else
{
while (size < MUSIC_BUFFER_SIZE)
2016-04-25 04:18:18 +03:00
{
streamedBytes = stb_vorbis_get_samples_short_interleaved(currentMusic.stream, currentMusic.channels, pcm + size, MUSIC_BUFFER_SIZE - size);
if (streamedBytes > 0) size += (streamedBytes*currentMusic.channels);
else break;
2016-04-25 04:18:18 +03:00
}
}
TraceLog(DEBUG, "Streaming music data to buffer. Bytes streamed: %i", size);
}
2014-09-03 18:51:28 +04:00
if (size > 0)
{
alBufferData(buffer, currentMusic.format, pcm, size*sizeof(short), currentMusic.sampleRate);
currentMusic.totalSamplesLeft -= size;
if(currentMusic.totalSamplesLeft <= 0) active = false; // end if no more samples left
}
else
{
active = false;
TraceLog(WARNING, "No more data obtained from stream");
}
2014-09-03 18:51:28 +04:00
return active;
}
// Empty music buffers
static void EmptyMusicStream(void)
{
2014-09-03 18:51:28 +04:00
ALuint buffer = 0;
int queued = 0;
2014-09-03 18:51:28 +04:00
alGetSourcei(currentMusic.source, AL_BUFFERS_QUEUED, &queued);
2014-09-03 18:51:28 +04:00
2015-12-03 15:45:06 +03:00
while (queued > 0)
{
alSourceUnqueueBuffers(currentMusic.source, 1, &buffer);
2014-09-03 18:51:28 +04:00
queued--;
}
}
// Update (re-fill) music buffers if data already processed
2015-10-06 18:13:40 +03:00
void UpdateMusicStream(void)
{
ALuint buffer = 0;
ALint processed = 0;
bool active = true;
2014-09-03 18:51:28 +04:00
if (musicEnabled)
{
// Get the number of already processed buffers (if any)
alGetSourcei(currentMusic.source, AL_BUFFERS_PROCESSED, &processed);
2014-09-03 18:51:28 +04:00
while (processed > 0)
{
// Recover processed buffer for refill
alSourceUnqueueBuffers(currentMusic.source, 1, &buffer);
// Refill buffer
active = BufferMusicStream(buffer);
2014-09-03 18:51:28 +04:00
// If no more data to stream, restart music (if loop)
2014-09-03 18:51:28 +04:00
if ((!active) && (currentMusic.loop))
{
2016-04-26 06:05:03 +03:00
if(currentMusic.chipTune)
{
2016-04-27 02:50:07 +03:00
currentMusic.totalSamplesLeft = currentMusic.totalLengthSeconds * currentMusic.sampleRate;
2016-04-26 06:05:03 +03:00
}
else
{
stb_vorbis_seek_start(currentMusic.stream);
currentMusic.totalSamplesLeft = stb_vorbis_stream_length_in_samples(currentMusic.stream)*currentMusic.channels;
}
2015-12-03 15:45:06 +03:00
active = BufferMusicStream(buffer);
}
2014-09-03 18:51:28 +04:00
// Add refilled buffer to queue again... don't let the music stop!
alSourceQueueBuffers(currentMusic.source, 1, &buffer);
2014-09-03 18:51:28 +04:00
2016-04-27 10:02:11 +03:00
if (alGetError() != AL_NO_ERROR) TraceLog(WARNING, "Error buffering data...");
2014-09-03 18:51:28 +04:00
processed--;
}
2014-09-03 18:51:28 +04:00
ALenum state;
alGetSourcei(currentMusic.source, AL_SOURCE_STATE, &state);
2014-09-03 18:51:28 +04:00
if ((state != AL_PLAYING) && active) alSourcePlay(currentMusic.source);
2014-09-03 18:51:28 +04:00
if (!active) StopMusicStream();
}
}
// Load WAV file into Wave structure
static Wave LoadWAV(const char *fileName)
{
2013-12-01 15:34:31 +04:00
// Basic WAV headers structs
typedef struct {
char chunkID[4];
int chunkSize;
2013-12-01 15:34:31 +04:00
char format[4];
} RiffHeader;
typedef struct {
char subChunkID[4];
int subChunkSize;
2013-12-01 15:34:31 +04:00
short audioFormat;
short numChannels;
int sampleRate;
int byteRate;
2013-12-01 15:34:31 +04:00
short blockAlign;
short bitsPerSample;
} WaveFormat;
typedef struct {
char subChunkID[4];
int subChunkSize;
2013-12-01 15:34:31 +04:00
} WaveData;
2014-09-03 18:51:28 +04:00
2013-12-01 15:34:31 +04:00
RiffHeader riffHeader;
WaveFormat waveFormat;
WaveData waveData;
2014-09-03 18:51:28 +04:00
2016-01-23 15:22:13 +03:00
Wave wave = { 0 };
2013-12-01 15:34:31 +04:00
FILE *wavFile;
2014-09-03 18:51:28 +04:00
wavFile = fopen(fileName, "rb");
2014-09-03 18:51:28 +04:00
if (wavFile == NULL)
{
TraceLog(WARNING, "[%s] WAV file could not be opened", fileName);
wave.data = NULL;
}
else
{
// Read in the first chunk into the struct
fread(&riffHeader, sizeof(RiffHeader), 1, wavFile);
2014-09-03 18:51:28 +04:00
// Check for RIFF and WAVE tags
if (strncmp(riffHeader.chunkID, "RIFF", 4) ||
strncmp(riffHeader.format, "WAVE", 4))
{
TraceLog(WARNING, "[%s] Invalid RIFF or WAVE Header", fileName);
}
else
{
// Read in the 2nd chunk for the wave info
fread(&waveFormat, sizeof(WaveFormat), 1, wavFile);
2014-09-03 18:51:28 +04:00
// Check for fmt tag
if ((waveFormat.subChunkID[0] != 'f') || (waveFormat.subChunkID[1] != 'm') ||
(waveFormat.subChunkID[2] != 't') || (waveFormat.subChunkID[3] != ' '))
{
TraceLog(WARNING, "[%s] Invalid Wave format", fileName);
}
else
{
// Check for extra parameters;
if (waveFormat.subChunkSize > 16) fseek(wavFile, sizeof(short), SEEK_CUR);
2014-09-03 18:51:28 +04:00
// Read in the the last byte of data before the sound file
fread(&waveData, sizeof(WaveData), 1, wavFile);
2014-09-03 18:51:28 +04:00
// Check for data tag
if ((waveData.subChunkID[0] != 'd') || (waveData.subChunkID[1] != 'a') ||
(waveData.subChunkID[2] != 't') || (waveData.subChunkID[3] != 'a'))
{
TraceLog(WARNING, "[%s] Invalid data header", fileName);
}
else
{
// Allocate memory for data
2014-09-03 18:51:28 +04:00
wave.data = (unsigned char *)malloc(sizeof(unsigned char) * waveData.subChunkSize);
// Read in the sound data into the soundData variable
fread(wave.data, waveData.subChunkSize, 1, wavFile);
2014-09-03 18:51:28 +04:00
// Now we set the variables that we need later
wave.dataSize = waveData.subChunkSize;
wave.sampleRate = waveFormat.sampleRate;
wave.channels = waveFormat.numChannels;
wave.bitsPerSample = waveFormat.bitsPerSample;
2014-09-03 18:51:28 +04:00
TraceLog(INFO, "[%s] WAV file loaded successfully (SampleRate: %i, BitRate: %i, Channels: %i)", fileName, wave.sampleRate, wave.bitsPerSample, wave.channels);
}
}
}
fclose(wavFile);
}
2014-09-03 18:51:28 +04:00
return wave;
2013-12-01 15:34:31 +04:00
}
// Load OGG file into Wave structure
// NOTE: Using stb_vorbis library
static Wave LoadOGG(char *fileName)
{
Wave wave;
2014-09-03 18:51:28 +04:00
stb_vorbis *oggFile = stb_vorbis_open_filename(fileName, NULL, NULL);
2014-09-03 18:51:28 +04:00
if (oggFile == NULL)
{
TraceLog(WARNING, "[%s] OGG file could not be opened", fileName);
wave.data = NULL;
}
else
{
stb_vorbis_info info = stb_vorbis_get_info(oggFile);
2014-09-03 18:51:28 +04:00
wave.sampleRate = info.sample_rate;
wave.bitsPerSample = 16;
wave.channels = info.channels;
TraceLog(DEBUG, "[%s] Ogg sample rate: %i", fileName, info.sample_rate);
TraceLog(DEBUG, "[%s] Ogg channels: %i", fileName, info.channels);
2014-09-03 18:51:28 +04:00
int totalSamplesLength = (stb_vorbis_stream_length_in_samples(oggFile) * info.channels);
2014-09-03 18:51:28 +04:00
wave.dataSize = totalSamplesLength*sizeof(short); // Size must be in bytes
2014-09-03 18:51:28 +04:00
TraceLog(DEBUG, "[%s] Samples length: %i", fileName, totalSamplesLength);
2014-09-03 18:51:28 +04:00
float totalSeconds = stb_vorbis_stream_length_in_seconds(oggFile);
2014-09-03 18:51:28 +04:00
TraceLog(DEBUG, "[%s] Total seconds: %f", fileName, totalSeconds);
2014-09-03 18:51:28 +04:00
if (totalSeconds > 10) TraceLog(WARNING, "[%s] Ogg audio lenght is larger than 10 seconds (%f), that's a big file in memory, consider music streaming", fileName, totalSeconds);
2014-09-03 18:51:28 +04:00
int totalSamples = totalSeconds*info.sample_rate*info.channels;
2014-09-03 18:51:28 +04:00
TraceLog(DEBUG, "[%s] Total samples calculated: %i", fileName, totalSamples);
wave.data = malloc(sizeof(short)*totalSamplesLength);
2014-09-03 18:51:28 +04:00
int samplesObtained = stb_vorbis_get_samples_short_interleaved(oggFile, info.channels, wave.data, totalSamplesLength);
2015-02-02 02:53:49 +03:00
TraceLog(DEBUG, "[%s] Samples obtained: %i", fileName, samplesObtained);
TraceLog(INFO, "[%s] OGG file loaded successfully (SampleRate: %i, BitRate: %i, Channels: %i)", fileName, wave.sampleRate, wave.bitsPerSample, wave.channels);
stb_vorbis_close(oggFile);
}
2014-09-03 18:51:28 +04:00
return wave;
}
// Unload Wave data
static void UnloadWave(Wave wave)
{
free(wave.data);
2015-08-05 20:17:56 +03:00
TraceLog(INFO, "Unloaded wave data");
}
// Some required functions for audio standalone module version
#if defined(AUDIO_STANDALONE)
// Get the extension for a filename
const char *GetExtension(const char *fileName)
{
const char *dot = strrchr(fileName, '.');
if(!dot || dot == fileName) return "";
return (dot + 1);
}
// Outputs a trace log message (INFO, ERROR, WARNING)
// NOTE: If a file has been init, output log is written there
void TraceLog(int msgType, const char *text, ...)
{
va_list args;
int traceDebugMsgs = 0;
#ifdef DO_NOT_TRACE_DEBUG_MSGS
traceDebugMsgs = 0;
#endif
switch(msgType)
{
case INFO: fprintf(stdout, "INFO: "); break;
case ERROR: fprintf(stdout, "ERROR: "); break;
case WARNING: fprintf(stdout, "WARNING: "); break;
case DEBUG: if (traceDebugMsgs) fprintf(stdout, "DEBUG: "); break;
default: break;
}
if ((msgType != DEBUG) || ((msgType == DEBUG) && (traceDebugMsgs)))
{
va_start(args, text);
vfprintf(stdout, text, args);
va_end(args);
fprintf(stdout, "\n");
}
if (msgType == ERROR) exit(1); // If ERROR message, exit program
}
#endif