/** * FreeRDP: A Remote Desktop Protocol Implementation * Video Redirection Virtual Channel - Media Container * * Copyright 2010-2011 Vic Lee * Copyright 2012 Hewlett-Packard Development Company, L.P. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #ifdef HAVE_CONFIG_H #include "config.h" #endif #include #include #include #include #ifdef HAVE_UNISTD_H #include #endif #ifndef _WIN32 #include #endif #include #include #include #include #include #include #include #include #include "tsmf_constants.h" #include "tsmf_types.h" #include "tsmf_decoder.h" #include "tsmf_audio.h" #include "tsmf_main.h" #include "tsmf_codec.h" #include "tsmf_media.h" #include #define AUDIO_TOLERANCE 10000000LL struct _TSMF_PRESENTATION { BYTE presentation_id[GUID_SIZE]; const char* audio_name; const char* audio_device; int eos; UINT32 last_x; UINT32 last_y; UINT32 last_width; UINT32 last_height; UINT16 last_num_rects; RDP_RECT* last_rects; UINT32 output_x; UINT32 output_y; UINT32 output_width; UINT32 output_height; UINT16 output_num_rects; RDP_RECT* output_rects; IWTSVirtualChannelCallback* channel_callback; UINT64 audio_start_time; UINT64 audio_end_time; HANDLE mutex; HANDLE thread; LIST* stream_list; }; struct _TSMF_STREAM { UINT32 stream_id; TSMF_PRESENTATION* presentation; ITSMFDecoder* decoder; int major_type; int eos; UINT32 width; UINT32 height; ITSMFAudioDevice* audio; UINT32 sample_rate; UINT32 channels; UINT32 bits_per_sample; /* The end_time of last played sample */ UINT64 last_end_time; /* Next sample should not start before this system time. */ UINT64 next_start_time; BOOL started; HANDLE thread; HANDLE stopEvent; wQueue* sample_list; wQueue* sample_ack_list; }; struct _TSMF_SAMPLE { UINT32 sample_id; UINT64 start_time; UINT64 end_time; UINT64 duration; UINT32 extensions; UINT32 data_size; BYTE* data; UINT32 decoded_size; UINT32 pixfmt; TSMF_STREAM* stream; IWTSVirtualChannelCallback* channel_callback; UINT64 ack_time; }; static LIST* presentation_list = NULL; static UINT64 last_played_audio_time = 0; static HANDLE tsmf_mutex = NULL; static int TERMINATING = 0; static UINT64 get_current_time(void) { struct timeval tp; gettimeofday(&tp, 0); return ((UINT64)tp.tv_sec) * 10000000LL + ((UINT64)tp.tv_usec) * 10LL; } static TSMF_SAMPLE* tsmf_stream_pop_sample(TSMF_STREAM* stream, int sync) { TSMF_STREAM* s; LIST_ITEM* item; TSMF_SAMPLE* sample; BOOL pending = FALSE; TSMF_PRESENTATION* presentation = stream->presentation; if (Queue_Count(stream->sample_list) < 1) return NULL; if (sync) { if (stream->decoder) { if (stream->decoder->GetDecodedData) { if (stream->major_type == TSMF_MAJOR_TYPE_AUDIO) { /* Check if some other stream has earlier sample that needs to be played first */ if (stream->last_end_time > AUDIO_TOLERANCE) { WaitForSingleObject(presentation->mutex, INFINITE); for (item = presentation->stream_list->head; item; item = item->next) { s = (TSMF_STREAM*) item->data; if (s != stream && !s->eos && s->last_end_time && s->last_end_time < stream->last_end_time - AUDIO_TOLERANCE) { pending = TRUE; break; } } ReleaseMutex(presentation->mutex); } } else { if (stream->last_end_time > presentation->audio_end_time) { pending = TRUE; } } } } } if (pending) return NULL; sample = (TSMF_SAMPLE*) Queue_Dequeue(stream->sample_list); if (sample && (sample->end_time > stream->last_end_time)) stream->last_end_time = sample->end_time; return sample; } static void tsmf_sample_free(TSMF_SAMPLE* sample) { if (sample->data) free(sample->data); free(sample); } static void tsmf_sample_ack(TSMF_SAMPLE* sample) { tsmf_playback_ack(sample->channel_callback, sample->sample_id, sample->duration, sample->data_size); } static void tsmf_sample_queue_ack(TSMF_SAMPLE* sample) { TSMF_STREAM* stream = sample->stream; Queue_Enqueue(stream->sample_ack_list, sample); } static void tsmf_stream_process_ack(TSMF_STREAM* stream) { TSMF_SAMPLE* sample; UINT64 ack_time; ack_time = get_current_time(); while ((Queue_Count(stream->sample_ack_list) > 0) && !(WaitForSingleObject(stream->stopEvent, 0) == WAIT_OBJECT_0)) { sample = (TSMF_SAMPLE*) Queue_Peek(stream->sample_ack_list); if (!sample || (sample->ack_time > ack_time)) break; sample = Queue_Dequeue(stream->sample_ack_list); tsmf_sample_ack(sample); tsmf_sample_free(sample); } } TSMF_PRESENTATION* tsmf_presentation_new(const BYTE* guid, IWTSVirtualChannelCallback* pChannelCallback) { TSMF_PRESENTATION* presentation; pthread_t thid = pthread_self(); FILE* fout = NULL; fout = fopen("/tmp/tsmf.tid", "wt"); if (fout) { fprintf(fout, "%d\n", (int) (size_t) thid); fclose(fout); } presentation = tsmf_presentation_find_by_id(guid); if (presentation) { DEBUG_WARN("duplicated presentation id!"); return NULL; } presentation = (TSMF_PRESENTATION*) malloc(sizeof(TSMF_PRESENTATION)); ZeroMemory(presentation, sizeof(TSMF_PRESENTATION)); memcpy(presentation->presentation_id, guid, GUID_SIZE); presentation->channel_callback = pChannelCallback; presentation->mutex = CreateMutex(NULL, FALSE, NULL); presentation->stream_list = list_new(); list_enqueue(presentation_list, presentation); return presentation; } TSMF_PRESENTATION* tsmf_presentation_find_by_id(const BYTE* guid) { LIST_ITEM* item; TSMF_PRESENTATION* presentation; for (item = presentation_list->head; item; item = item->next) { presentation = (TSMF_PRESENTATION*) item->data; if (memcmp(presentation->presentation_id, guid, GUID_SIZE) == 0) return presentation; } return NULL; } static void tsmf_presentation_restore_last_video_frame(TSMF_PRESENTATION* presentation) { RDP_REDRAW_EVENT* revent; if (presentation->last_width && presentation->last_height) { revent = (RDP_REDRAW_EVENT*) freerdp_event_new(TsmfChannel_Class, TsmfChannel_Redraw, NULL, NULL); revent->x = presentation->last_x; revent->y = presentation->last_y; revent->width = presentation->last_width; revent->height = presentation->last_height; if (!tsmf_push_event(presentation->channel_callback, (RDP_EVENT*) revent)) { freerdp_event_free((RDP_EVENT*) revent); } presentation->last_x = 0; presentation->last_y = 0; presentation->last_width = 0; presentation->last_height = 0; } } static void tsmf_sample_playback_video(TSMF_SAMPLE* sample) { UINT64 t; RDP_VIDEO_FRAME_EVENT* vevent; TSMF_STREAM* stream = sample->stream; TSMF_PRESENTATION* presentation = stream->presentation; DEBUG_DVC("MessageId %d EndTime %d data_size %d consumed.", sample->sample_id, (int)sample->end_time, sample->data_size); if (sample->data) { t = get_current_time(); if (stream->next_start_time > t && (sample->end_time >= presentation->audio_start_time || sample->end_time < stream->last_end_time)) { USleep((stream->next_start_time - t) / 10); } stream->next_start_time = t + sample->duration - 50000; if (presentation->last_x != presentation->output_x || presentation->last_y != presentation->output_y || presentation->last_width != presentation->output_width || presentation->last_height != presentation->output_height || presentation->last_num_rects != presentation->output_num_rects || (presentation->last_rects && presentation->output_rects && memcmp(presentation->last_rects, presentation->output_rects, presentation->last_num_rects * sizeof(RDP_RECT)) != 0)) { tsmf_presentation_restore_last_video_frame(presentation); presentation->last_x = presentation->output_x; presentation->last_y = presentation->output_y; presentation->last_width = presentation->output_width; presentation->last_height = presentation->output_height; if (presentation->last_rects) { free(presentation->last_rects); presentation->last_rects = NULL; } presentation->last_num_rects = presentation->output_num_rects; if (presentation->last_num_rects > 0) { presentation->last_rects = malloc(presentation->last_num_rects * sizeof(RDP_RECT)); ZeroMemory(presentation->last_rects, presentation->last_num_rects * sizeof(RDP_RECT)); memcpy(presentation->last_rects, presentation->output_rects, presentation->last_num_rects * sizeof(RDP_RECT)); } } vevent = (RDP_VIDEO_FRAME_EVENT*) freerdp_event_new(TsmfChannel_Class, TsmfChannel_VideoFrame, NULL, NULL); vevent->frame_data = sample->data; vevent->frame_size = sample->decoded_size; vevent->frame_pixfmt = sample->pixfmt; vevent->frame_width = sample->stream->width; vevent->frame_height = sample->stream->height; vevent->x = presentation->output_x; vevent->y = presentation->output_y; vevent->width = presentation->output_width; vevent->height = presentation->output_height; if (presentation->output_num_rects > 0) { vevent->num_visible_rects = presentation->output_num_rects; vevent->visible_rects = (RDP_RECT*) malloc(presentation->output_num_rects * sizeof(RDP_RECT)); ZeroMemory(vevent->visible_rects, presentation->output_num_rects * sizeof(RDP_RECT)); memcpy(vevent->visible_rects, presentation->output_rects, presentation->output_num_rects * sizeof(RDP_RECT)); } /* The frame data ownership is passed to the event object, and is freed after the event is processed. */ sample->data = NULL; sample->decoded_size = 0; if (!tsmf_push_event(sample->channel_callback, (RDP_EVENT*) vevent)) { freerdp_event_free((RDP_EVENT*) vevent); } #if 0 /* Dump a .ppm image for every 30 frames. Assuming the frame is in YUV format, we extract the Y values to create a grayscale image. */ static int frame_id = 0; char buf[100]; FILE * fp; if ((frame_id % 30) == 0) { snprintf(buf, sizeof(buf), "/tmp/FreeRDP_Frame_%d.ppm", frame_id); fp = fopen(buf, "wb"); fwrite("P5\n", 1, 3, fp); snprintf(buf, sizeof(buf), "%d %d\n", sample->stream->width, sample->stream->height); fwrite(buf, 1, strlen(buf), fp); fwrite("255\n", 1, 4, fp); fwrite(sample->data, 1, sample->stream->width * sample->stream->height, fp); fflush(fp); fclose(fp); } frame_id++; #endif } } static void tsmf_sample_playback_audio(TSMF_SAMPLE* sample) { UINT64 latency = 0; TSMF_STREAM* stream = sample->stream; DEBUG_DVC("MessageId %d EndTime %d consumed.", sample->sample_id, (int)sample->end_time); if (sample->stream->audio && sample->data) { sample->stream->audio->Play(sample->stream->audio, sample->data, sample->decoded_size); sample->data = NULL; sample->decoded_size = 0; if (stream->audio && stream->audio->GetLatency) latency = stream->audio->GetLatency(stream->audio); } else { latency = 0; } sample->ack_time = latency + get_current_time(); stream->last_end_time = sample->end_time + latency; stream->presentation->audio_start_time = sample->start_time + latency; stream->presentation->audio_end_time = sample->end_time + latency; } static void tsmf_sample_playback(TSMF_SAMPLE* sample) { BOOL ret = FALSE; UINT32 width; UINT32 height; UINT32 pixfmt = 0; TSMF_STREAM* stream = sample->stream; if (stream->decoder) { if (stream->decoder->DecodeEx) ret = stream->decoder->DecodeEx(stream->decoder, sample->data, sample->data_size, sample->extensions, sample->start_time, sample->end_time, sample->duration); else ret = stream->decoder->Decode(stream->decoder, sample->data, sample->data_size, sample->extensions); } if (!ret) { tsmf_sample_ack(sample); tsmf_sample_free(sample); return; } free(sample->data); sample->data = NULL; if (stream->major_type == TSMF_MAJOR_TYPE_VIDEO) { if (stream->decoder->GetDecodedFormat) { pixfmt = stream->decoder->GetDecodedFormat(stream->decoder); if (pixfmt == ((UINT32) -1)) { tsmf_sample_ack(sample); tsmf_sample_free(sample); return; } sample->pixfmt = pixfmt; } ret = FALSE ; if (stream->decoder->GetDecodedDimension) { ret = stream->decoder->GetDecodedDimension(stream->decoder, &width, &height); if (ret && (width != stream->width || height != stream->height)) { DEBUG_DVC("video dimension changed to %d x %d", width, height); stream->width = width; stream->height = height; } } } if (stream->decoder->GetDecodedData) { sample->data = stream->decoder->GetDecodedData(stream->decoder, &sample->decoded_size); switch (sample->stream->major_type) { case TSMF_MAJOR_TYPE_VIDEO: tsmf_sample_playback_video(sample); tsmf_sample_ack(sample); tsmf_sample_free(sample); break; case TSMF_MAJOR_TYPE_AUDIO: tsmf_sample_playback_audio(sample); tsmf_sample_queue_ack(sample); break; } } else { TSMF_STREAM * stream = sample->stream; UINT64 ack_anticipation_time = get_current_time(); UINT64 currentRunningTime = sample->start_time; UINT32 bufferLevel = 0; if (stream->decoder->GetRunningTime) { currentRunningTime = stream->decoder->GetRunningTime(stream->decoder); } if (stream->decoder->BufferLevel) { bufferLevel = stream->decoder->BufferLevel(stream->decoder); } switch (sample->stream->major_type) { case TSMF_MAJOR_TYPE_VIDEO: { TSMF_PRESENTATION * presentation = sample->stream->presentation; /* * Tell gstreamer that presentation screen area has moved. * So it can render on the new area. */ if (presentation->last_x != presentation->output_x || presentation->last_y != presentation->output_y || presentation->last_width != presentation->output_width || presentation->last_height != presentation->output_height) { presentation->last_x = presentation->output_x; presentation->last_y = presentation->output_y; presentation->last_width = presentation->output_width; presentation->last_height = presentation->output_height; if(stream->decoder->UpdateRenderingArea) { stream->decoder->UpdateRenderingArea(stream->decoder, presentation->output_x, presentation->output_y, presentation->output_width, presentation->output_height, presentation->output_num_rects, presentation->output_rects); } } if ( presentation->last_num_rects != presentation->output_num_rects || (presentation->last_rects && presentation->output_rects && memcmp(presentation->last_rects, presentation->output_rects, presentation->last_num_rects * sizeof(RDP_RECT)) != 0)) { if (presentation->last_rects) { free(presentation->last_rects); presentation->last_rects = NULL; } presentation->last_num_rects = presentation->output_num_rects; if (presentation->last_num_rects > 0) { presentation->last_rects = malloc(presentation->last_num_rects * sizeof(RDP_RECT)); ZeroMemory(presentation->last_rects, presentation->last_num_rects * sizeof(RDP_RECT)); memcpy(presentation->last_rects, presentation->output_rects, presentation->last_num_rects * sizeof(RDP_RECT)); } if(stream->decoder->UpdateRenderingArea) { stream->decoder->UpdateRenderingArea(stream->decoder, presentation->output_x, presentation->output_y, presentation->output_width, presentation->output_height, presentation->output_num_rects, presentation->output_rects); } } if (bufferLevel < 24) { ack_anticipation_time += sample->duration; } else { if (currentRunningTime > sample->start_time) { ack_anticipation_time += sample->duration; } else if(currentRunningTime == 0) { ack_anticipation_time += sample->duration; } else { ack_anticipation_time += (sample->start_time - currentRunningTime); } } break; } case TSMF_MAJOR_TYPE_AUDIO: { last_played_audio_time = currentRunningTime; if (bufferLevel < 2) { ack_anticipation_time += sample->duration; } else { if (currentRunningTime > sample->start_time) { ack_anticipation_time += sample->duration; } else if(currentRunningTime == 0) { ack_anticipation_time += sample->duration; } else { ack_anticipation_time += (sample->start_time - currentRunningTime); } } break; } } sample->ack_time = ack_anticipation_time; tsmf_sample_queue_ack(sample); } } static void* tsmf_stream_playback_func(void* arg) { TSMF_SAMPLE* sample; TSMF_STREAM* stream = (TSMF_STREAM*) arg; TSMF_PRESENTATION* presentation = stream->presentation; DEBUG_DVC("in %d", stream->stream_id); if (stream->major_type == TSMF_MAJOR_TYPE_AUDIO && stream->sample_rate && stream->channels && stream->bits_per_sample) { if (stream->decoder) { if (stream->decoder->GetDecodedData) { stream->audio = tsmf_load_audio_device( presentation->audio_name && presentation->audio_name[0] ? presentation->audio_name : NULL, presentation->audio_device && presentation->audio_device[0] ? presentation->audio_device : NULL); if (stream->audio) { stream->audio->SetFormat(stream->audio, stream->sample_rate, stream->channels, stream->bits_per_sample); } } } } while (!(WaitForSingleObject(stream->stopEvent, 0) == WAIT_OBJECT_0)) { tsmf_stream_process_ack(stream); sample = tsmf_stream_pop_sample(stream, 1); if (sample) tsmf_sample_playback(sample); else USleep(5000); } if (stream->eos || presentation->eos) { while ((sample = tsmf_stream_pop_sample(stream, 1)) != NULL) tsmf_sample_playback(sample); } if (stream->audio) { stream->audio->Free(stream->audio); stream->audio = NULL; } SetEvent(stream->stopEvent); DEBUG_DVC("out %d", stream->stream_id); return NULL; } static void tsmf_stream_start(TSMF_STREAM* stream) { if (!stream->started) { ResumeThread(stream->thread); stream->started = TRUE; } } static void tsmf_stream_stop(TSMF_STREAM* stream) { if (!stream) return; if (!stream->decoder) return; if (stream->started) { SetEvent(stream->stopEvent); stream->started = FALSE; } if (stream->decoder->Control) { stream->decoder->Control(stream->decoder, Control_Flush, NULL); } } static void tsmf_stream_pause(TSMF_STREAM* stream) { if (!stream) return; if (!stream->decoder) return; if (stream->decoder->Control) { stream->decoder->Control(stream->decoder, Control_Pause, NULL); } } static void tsmf_stream_restart(TSMF_STREAM* stream) { if (!stream) return; if (!stream->decoder) return; if (stream->decoder->Control) { stream->decoder->Control(stream->decoder, Control_Restart, NULL); } } static void tsmf_stream_change_volume(TSMF_STREAM* stream, UINT32 newVolume, UINT32 muted) { if (!stream) return; if (stream->decoder != NULL && stream->decoder->ChangeVolume) { stream->decoder->ChangeVolume(stream->decoder, newVolume, muted); } else if (stream->audio != NULL && stream->audio->ChangeVolume) { stream->audio->ChangeVolume(stream->audio, newVolume, muted); } } void tsmf_presentation_volume_changed(TSMF_PRESENTATION* presentation, UINT32 newVolume, UINT32 muted) { LIST_ITEM* item; TSMF_STREAM* stream; for (item = presentation->stream_list->head; item; item = item->next) { stream = (TSMF_STREAM*) item->data; tsmf_stream_change_volume(stream, newVolume, muted); } } void tsmf_presentation_paused(TSMF_PRESENTATION* presentation) { LIST_ITEM* item; TSMF_STREAM* stream; for (item = presentation->stream_list->head; item; item = item->next) { stream = (TSMF_STREAM*) item->data; tsmf_stream_pause(stream); } } void tsmf_presentation_restarted(TSMF_PRESENTATION* presentation) { LIST_ITEM* item; TSMF_STREAM* stream; for (item = presentation->stream_list->head; item; item = item->next) { stream = (TSMF_STREAM*) item->data; tsmf_stream_restart(stream); } } void tsmf_presentation_start(TSMF_PRESENTATION* presentation) { LIST_ITEM* item; TSMF_STREAM* stream; for (item = presentation->stream_list->head; item; item = item->next) { stream = (TSMF_STREAM*) item->data; tsmf_stream_start(stream); } } void tsmf_presentation_stop(TSMF_PRESENTATION* presentation) { LIST_ITEM* item; TSMF_STREAM* stream; tsmf_presentation_flush(presentation); for (item = presentation->stream_list->head; item; item = item->next) { stream = (TSMF_STREAM*) item->data; tsmf_stream_stop(stream); } tsmf_presentation_restore_last_video_frame(presentation); if (presentation->last_rects) { free(presentation->last_rects); presentation->last_rects = NULL; } presentation->last_num_rects = 0; if (presentation->output_rects) { free(presentation->output_rects); presentation->output_rects = NULL; } presentation->output_num_rects = 0; } void tsmf_presentation_set_geometry_info(TSMF_PRESENTATION* presentation, UINT32 x, UINT32 y, UINT32 width, UINT32 height, int num_rects, RDP_RECT* rects) { presentation->output_x = x; presentation->output_y = y; presentation->output_width = width; presentation->output_height = height; if (presentation->output_rects) free(presentation->output_rects); presentation->output_rects = rects; presentation->output_num_rects = num_rects; } void tsmf_presentation_set_audio_device(TSMF_PRESENTATION* presentation, const char* name, const char* device) { presentation->audio_name = name; presentation->audio_device = device; } static void tsmf_stream_flush(TSMF_STREAM* stream) { //TSMF_SAMPLE* sample; /* TODO: free lists */ if (stream->audio) stream->audio->Flush(stream->audio); stream->eos = 0; stream->last_end_time = 0; stream->next_start_time = 0; if (stream->major_type == TSMF_MAJOR_TYPE_AUDIO) { stream->presentation->audio_start_time = 0; stream->presentation->audio_end_time = 0; } } void tsmf_presentation_flush(TSMF_PRESENTATION* presentation) { LIST_ITEM* item; TSMF_STREAM * stream; for (item = presentation->stream_list->head; item; item = item->next) { stream = (TSMF_STREAM*) item->data; tsmf_stream_flush(stream); } presentation->eos = 0; presentation->audio_start_time = 0; presentation->audio_end_time = 0; } void tsmf_presentation_free(TSMF_PRESENTATION* presentation) { TSMF_STREAM* stream; tsmf_presentation_stop(presentation); WaitForSingleObject(presentation->mutex, INFINITE); list_remove(presentation_list, presentation); ReleaseMutex(presentation->mutex); while (list_size(presentation->stream_list) > 0) { stream = (TSMF_STREAM*) list_dequeue(presentation->stream_list); tsmf_stream_free(stream); } list_free(presentation->stream_list); CloseHandle(presentation->mutex); free(presentation); } TSMF_STREAM* tsmf_stream_new(TSMF_PRESENTATION* presentation, UINT32 stream_id) { TSMF_STREAM* stream; stream = tsmf_stream_find_by_id(presentation, stream_id); if (stream) { DEBUG_WARN("duplicated stream id %d!", stream_id); return NULL; } stream = (TSMF_STREAM*) malloc(sizeof(TSMF_STREAM)); ZeroMemory(stream, sizeof(TSMF_STREAM)); stream->stream_id = stream_id; stream->presentation = presentation; stream->started = FALSE; stream->stopEvent = CreateEvent(NULL, TRUE, FALSE, NULL); stream->thread = CreateThread(NULL, 0, (LPTHREAD_START_ROUTINE) tsmf_stream_playback_func, stream, CREATE_SUSPENDED, NULL); stream->sample_list = Queue_New(TRUE, -1, -1); stream->sample_ack_list = Queue_New(TRUE, -1, -1); WaitForSingleObject(presentation->mutex, INFINITE); list_enqueue(presentation->stream_list, stream); ReleaseMutex(presentation->mutex); return stream; } TSMF_STREAM* tsmf_stream_find_by_id(TSMF_PRESENTATION* presentation, UINT32 stream_id) { LIST_ITEM* item; TSMF_STREAM* stream; for (item = presentation->stream_list->head; item; item = item->next) { stream = (TSMF_STREAM*) item->data; if (stream->stream_id == stream_id) return stream; } return NULL; } void tsmf_stream_set_format(TSMF_STREAM* stream, const char* name, wStream* s) { TS_AM_MEDIA_TYPE mediatype; if (stream->decoder) { DEBUG_WARN("duplicated call"); return; } tsmf_codec_parse_media_type(&mediatype, s); if (mediatype.MajorType == TSMF_MAJOR_TYPE_VIDEO) { DEBUG_DVC("video width %d height %d bit_rate %d frame_rate %f codec_data %d", mediatype.Width, mediatype.Height, mediatype.BitRate, (double) mediatype.SamplesPerSecond.Numerator / (double) mediatype.SamplesPerSecond.Denominator, mediatype.ExtraDataSize); } else if (mediatype.MajorType == TSMF_MAJOR_TYPE_AUDIO) { DEBUG_DVC("audio channel %d sample_rate %d bits_per_sample %d codec_data %d", mediatype.Channels, mediatype.SamplesPerSecond.Numerator, mediatype.BitsPerSample, mediatype.ExtraDataSize); stream->sample_rate = mediatype.SamplesPerSecond.Numerator; stream->channels = mediatype.Channels; stream->bits_per_sample = mediatype.BitsPerSample; if (stream->bits_per_sample == 0) stream->bits_per_sample = 16; } stream->major_type = mediatype.MajorType; stream->width = mediatype.Width; stream->height = mediatype.Height; stream->decoder = tsmf_load_decoder(name, &mediatype); } void tsmf_stream_end(TSMF_STREAM* stream) { stream->eos = 1; stream->presentation->eos = 1; } void tsmf_stream_free(TSMF_STREAM* stream) { TSMF_PRESENTATION* presentation = stream->presentation; tsmf_stream_stop(stream); tsmf_stream_flush(stream); WaitForSingleObject(presentation->mutex, INFINITE); list_remove(presentation->stream_list, stream); ReleaseMutex(presentation->mutex); Queue_Free(stream->sample_list); Queue_Free(stream->sample_ack_list); if (stream->decoder) { stream->decoder->Free(stream->decoder); stream->decoder = 0; } SetEvent(stream->thread); free(stream); stream = 0; } void tsmf_stream_push_sample(TSMF_STREAM* stream, IWTSVirtualChannelCallback* pChannelCallback, UINT32 sample_id, UINT64 start_time, UINT64 end_time, UINT64 duration, UINT32 extensions, UINT32 data_size, BYTE* data) { TSMF_SAMPLE* sample; WaitForSingleObject(tsmf_mutex, INFINITE); if (TERMINATING) { ReleaseMutex(tsmf_mutex); return; } ReleaseMutex(tsmf_mutex); sample = (TSMF_SAMPLE*) malloc(sizeof(TSMF_SAMPLE)); ZeroMemory(sample, sizeof(TSMF_SAMPLE)); sample->sample_id = sample_id; sample->start_time = start_time; sample->end_time = end_time; sample->duration = duration; sample->extensions = extensions; sample->stream = stream; sample->channel_callback = pChannelCallback; sample->data_size = data_size; sample->data = malloc(data_size + TSMF_BUFFER_PADDING_SIZE); ZeroMemory(sample->data, data_size + TSMF_BUFFER_PADDING_SIZE); CopyMemory(sample->data, data, data_size); Queue_Enqueue(stream->sample_list, sample); } #ifndef _WIN32 static void tsmf_signal_handler(int s) { LIST_ITEM* p_item; TSMF_PRESENTATION* presentation; LIST_ITEM* s_item; TSMF_STREAM* _stream; WaitForSingleObject(tsmf_mutex, INFINITE); TERMINATING = 1; ReleaseMutex(tsmf_mutex); if (presentation_list) { for (p_item = presentation_list->head; p_item; p_item = p_item->next) { presentation = (TSMF_PRESENTATION*) p_item->data; for (s_item = presentation->stream_list->head; s_item; s_item = s_item->next) { _stream = (TSMF_STREAM*) s_item->data; tsmf_stream_free(_stream); } tsmf_presentation_free(presentation); } } unlink("/tmp/tsmf.tid"); if (s == SIGINT) { signal(s, SIG_DFL); kill(getpid(), s); } else if (s == SIGUSR1) { signal(s, SIG_DFL); } } #endif void tsmf_media_init(void) { #ifndef _WIN32 struct sigaction sigtrap; sigtrap.sa_handler = tsmf_signal_handler; sigemptyset(&sigtrap.sa_mask); sigtrap.sa_flags = 0; sigaction(SIGINT, &sigtrap, 0); sigaction(SIGUSR1, &sigtrap, 0); #endif tsmf_mutex = CreateMutex(NULL, FALSE, NULL); if (presentation_list == NULL) presentation_list = list_new(); }