mirror of https://github.com/FreeRDP/FreeRDP
video: mutualize things in the common channel code
This commit is contained in:
parent
0743559d9c
commit
b8e3b232de
|
@ -31,12 +31,18 @@
|
||||||
#include <winpr/stream.h>
|
#include <winpr/stream.h>
|
||||||
#include <winpr/cmdline.h>
|
#include <winpr/cmdline.h>
|
||||||
#include <winpr/collections.h>
|
#include <winpr/collections.h>
|
||||||
|
#include <winpr/interlocked.h>
|
||||||
|
#include <winpr/sysinfo.h>
|
||||||
|
|
||||||
#include <freerdp/addin.h>
|
#include <freerdp/addin.h>
|
||||||
|
#include <freerdp/primitives.h>
|
||||||
#include <freerdp/client/video.h>
|
#include <freerdp/client/video.h>
|
||||||
#include <freerdp/channels/log.h>
|
#include <freerdp/channels/log.h>
|
||||||
|
#include <freerdp/codec/h264.h>
|
||||||
|
#include <freerdp/codec/yuv.h>
|
||||||
|
|
||||||
#define TAG CHANNELS_TAG("video.client")
|
|
||||||
|
#define TAG CHANNELS_TAG("video")
|
||||||
|
|
||||||
#include "video_main.h"
|
#include "video_main.h"
|
||||||
|
|
||||||
|
@ -73,6 +79,64 @@ struct _VIDEO_PLUGIN
|
||||||
};
|
};
|
||||||
typedef struct _VIDEO_PLUGIN VIDEO_PLUGIN;
|
typedef struct _VIDEO_PLUGIN VIDEO_PLUGIN;
|
||||||
|
|
||||||
|
|
||||||
|
#define XF_VIDEO_UNLIMITED_RATE 31
|
||||||
|
|
||||||
|
BYTE MFVideoFormat_H264[] = {'H', '2', '6', '4',
|
||||||
|
0x00, 0x00,
|
||||||
|
0x10, 0x00,
|
||||||
|
0x80, 0x00,
|
||||||
|
0x00, 0xAA, 0x00, 0x38, 0x9B, 0x71};
|
||||||
|
|
||||||
|
typedef struct _PresentationContext PresentationContext;
|
||||||
|
typedef struct _VideoFrame VideoFrame;
|
||||||
|
|
||||||
|
|
||||||
|
/** @brief private data for the channel */
|
||||||
|
struct _VideoClientContextPriv
|
||||||
|
{
|
||||||
|
VideoClientContext *video;
|
||||||
|
GeometryClientContext *geometry;
|
||||||
|
wQueue *frames;
|
||||||
|
CRITICAL_SECTION framesLock;
|
||||||
|
wBufferPool *surfacePool;
|
||||||
|
UINT32 publishedFrames;
|
||||||
|
UINT32 droppedFrames;
|
||||||
|
UINT32 lastSentRate;
|
||||||
|
UINT64 nextFeedbackTime;
|
||||||
|
PresentationContext *currentPresentation;
|
||||||
|
};
|
||||||
|
|
||||||
|
/** @brief */
|
||||||
|
struct _VideoFrame
|
||||||
|
{
|
||||||
|
UINT64 publishTime;
|
||||||
|
UINT64 hnsDuration;
|
||||||
|
UINT32 x, y, w, h;
|
||||||
|
BYTE *surfaceData;
|
||||||
|
PresentationContext *presentation;
|
||||||
|
};
|
||||||
|
|
||||||
|
struct _PresentationContext
|
||||||
|
{
|
||||||
|
VideoClientContext *video;
|
||||||
|
BYTE PresentationId;
|
||||||
|
UINT32 SourceWidth, SourceHeight;
|
||||||
|
UINT32 ScaledWidth, ScaledHeight;
|
||||||
|
MAPPED_GEOMETRY *geometry;
|
||||||
|
|
||||||
|
UINT64 startTimeStamp;
|
||||||
|
UINT64 publishOffset;
|
||||||
|
H264_CONTEXT *h264;
|
||||||
|
YUV_CONTEXT *yuv;
|
||||||
|
wStream *currentSample;
|
||||||
|
UINT64 lastPublishTime, nextPublishTime;
|
||||||
|
volatile LONG refCounter;
|
||||||
|
BYTE *surfaceData;
|
||||||
|
VideoSurface *surface;
|
||||||
|
};
|
||||||
|
|
||||||
|
|
||||||
static const char *video_command_name(BYTE cmd)
|
static const char *video_command_name(BYTE cmd)
|
||||||
{
|
{
|
||||||
switch(cmd)
|
switch(cmd)
|
||||||
|
@ -86,10 +150,336 @@ static const char *video_command_name(BYTE cmd)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static BOOL yuv_to_rgb(PresentationContext *presentation, BYTE *dest)
|
||||||
|
{
|
||||||
|
const BYTE* pYUVPoint[3];
|
||||||
|
H264_CONTEXT *h264 = presentation->h264;
|
||||||
|
|
||||||
|
BYTE** ppYUVData;
|
||||||
|
ppYUVData = h264->pYUVData;
|
||||||
|
|
||||||
|
pYUVPoint[0] = ppYUVData[0];
|
||||||
|
pYUVPoint[1] = ppYUVData[1];
|
||||||
|
pYUVPoint[2] = ppYUVData[2];
|
||||||
|
|
||||||
|
if (!yuv_context_decode(presentation->yuv, pYUVPoint, h264->iStride, PIXEL_FORMAT_BGRX32, dest, h264->width * 4))
|
||||||
|
{
|
||||||
|
WLog_ERR(TAG, "error in yuv_to_rgb conversion");
|
||||||
|
return FALSE;
|
||||||
|
}
|
||||||
|
|
||||||
|
return TRUE;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
static void video_client_context_set_geometry(VideoClientContext *video, GeometryClientContext *geometry)
|
||||||
|
{
|
||||||
|
video->priv->geometry = geometry;
|
||||||
|
}
|
||||||
|
|
||||||
|
VideoClientContextPriv *VideoClientContextPriv_new(VideoClientContext *video)
|
||||||
|
{
|
||||||
|
VideoClientContextPriv *ret = calloc(1, sizeof(*ret));
|
||||||
|
if (!ret)
|
||||||
|
return NULL;
|
||||||
|
|
||||||
|
ret->frames = Queue_New(TRUE, 10, 2);
|
||||||
|
if (!ret->frames)
|
||||||
|
{
|
||||||
|
WLog_ERR(TAG, "unable to allocate frames queue");
|
||||||
|
goto error_frames;
|
||||||
|
}
|
||||||
|
|
||||||
|
ret->surfacePool = BufferPool_New(FALSE, 0, 16);
|
||||||
|
if (!ret->surfacePool)
|
||||||
|
{
|
||||||
|
WLog_ERR(TAG, "unable to create surface pool");
|
||||||
|
goto error_surfacePool;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!InitializeCriticalSectionAndSpinCount(&ret->framesLock, 4000))
|
||||||
|
{
|
||||||
|
WLog_ERR(TAG, "unable to initialize frames lock");
|
||||||
|
goto error_spinlock;
|
||||||
|
}
|
||||||
|
|
||||||
|
ret->video = video;
|
||||||
|
|
||||||
|
/* don't set to unlimited so that we have the chance to send a feedback in
|
||||||
|
* the first second (for servers that want feedback directly)
|
||||||
|
*/
|
||||||
|
ret->lastSentRate = 30;
|
||||||
|
return ret;
|
||||||
|
|
||||||
|
error_spinlock:
|
||||||
|
BufferPool_Free(ret->surfacePool);
|
||||||
|
error_surfacePool:
|
||||||
|
Queue_Free(ret->frames);
|
||||||
|
error_frames:
|
||||||
|
free(ret);
|
||||||
|
return NULL;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
static PresentationContext *PresentationContext_new(VideoClientContext *video, BYTE PresentationId,
|
||||||
|
UINT32 x, UINT32 y, UINT32 width, UINT32 height)
|
||||||
|
{
|
||||||
|
VideoClientContextPriv *priv = video->priv;
|
||||||
|
PresentationContext *ret = calloc(1, sizeof(*ret));
|
||||||
|
if (!ret)
|
||||||
|
return NULL;
|
||||||
|
|
||||||
|
ret->video = video;
|
||||||
|
ret->PresentationId = PresentationId;
|
||||||
|
|
||||||
|
ret->h264 = h264_context_new(FALSE);
|
||||||
|
if (!ret->h264)
|
||||||
|
{
|
||||||
|
WLog_ERR(TAG, "unable to create a h264 context");
|
||||||
|
goto error_h264;
|
||||||
|
}
|
||||||
|
h264_context_reset(ret->h264, width, height);
|
||||||
|
|
||||||
|
ret->currentSample = Stream_New(NULL, 4096);
|
||||||
|
if (!ret->currentSample)
|
||||||
|
{
|
||||||
|
WLog_ERR(TAG, "unable to create current packet stream");
|
||||||
|
goto error_currentSample;
|
||||||
|
}
|
||||||
|
|
||||||
|
ret->surfaceData = BufferPool_Take(priv->surfacePool, width * height * 4);
|
||||||
|
if (!ret->surfaceData)
|
||||||
|
{
|
||||||
|
WLog_ERR(TAG, "unable to allocate surfaceData");
|
||||||
|
goto error_surfaceData;
|
||||||
|
}
|
||||||
|
|
||||||
|
ret->surface = video->createSurface(video, ret->surfaceData, x, y, width, height);
|
||||||
|
if (!ret->surface)
|
||||||
|
{
|
||||||
|
WLog_ERR(TAG, "unable to create surface");
|
||||||
|
goto error_surface;
|
||||||
|
}
|
||||||
|
|
||||||
|
ret->yuv = yuv_context_new(FALSE);
|
||||||
|
if (!ret->yuv)
|
||||||
|
{
|
||||||
|
WLog_ERR(TAG, "unable to create YUV decoder");
|
||||||
|
goto error_yuv;
|
||||||
|
}
|
||||||
|
|
||||||
|
yuv_context_reset(ret->yuv, width, height);
|
||||||
|
ret->refCounter = 1;
|
||||||
|
return ret;
|
||||||
|
|
||||||
|
error_yuv:
|
||||||
|
video->deleteSurface(video, ret->surface);
|
||||||
|
error_surface:
|
||||||
|
BufferPool_Return(priv->surfacePool, ret->surfaceData);
|
||||||
|
error_surfaceData:
|
||||||
|
Stream_Free(ret->currentSample, TRUE);
|
||||||
|
error_currentSample:
|
||||||
|
h264_context_free(ret->h264);
|
||||||
|
error_h264:
|
||||||
|
free(ret);
|
||||||
|
return NULL;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
static void PresentationContext_unref(PresentationContext *c)
|
||||||
|
{
|
||||||
|
VideoClientContextPriv *priv;
|
||||||
|
|
||||||
|
if (!c)
|
||||||
|
return;
|
||||||
|
|
||||||
|
if (InterlockedDecrement(&c->refCounter) != 0)
|
||||||
|
return;
|
||||||
|
|
||||||
|
priv = c->video->priv;
|
||||||
|
if (c->geometry)
|
||||||
|
{
|
||||||
|
c->geometry->MappedGeometryUpdate = NULL;
|
||||||
|
c->geometry->MappedGeometryClear = NULL;
|
||||||
|
c->geometry->custom = NULL;
|
||||||
|
}
|
||||||
|
|
||||||
|
h264_context_free(c->h264);
|
||||||
|
Stream_Free(c->currentSample, TRUE);
|
||||||
|
c->video->deleteSurface(c->video, c->surface);
|
||||||
|
BufferPool_Return(priv->surfacePool, c->surfaceData);
|
||||||
|
yuv_context_free(c->yuv);
|
||||||
|
free(c);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
static void VideoFrame_free(VideoFrame **pframe)
|
||||||
|
{
|
||||||
|
VideoFrame *frame = *pframe;
|
||||||
|
|
||||||
|
PresentationContext_unref(frame->presentation);
|
||||||
|
BufferPool_Return(frame->presentation->video->priv->surfacePool, frame->surfaceData);
|
||||||
|
free(frame);
|
||||||
|
*pframe = NULL;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
static void VideoClientContextPriv_free(VideoClientContextPriv *priv)
|
||||||
|
{
|
||||||
|
EnterCriticalSection(&priv->framesLock);
|
||||||
|
while (Queue_Count(priv->frames))
|
||||||
|
{
|
||||||
|
VideoFrame *frame = Queue_Dequeue(priv->frames);
|
||||||
|
if (frame)
|
||||||
|
VideoFrame_free(&frame);
|
||||||
|
}
|
||||||
|
|
||||||
|
Queue_Free(priv->frames);
|
||||||
|
LeaveCriticalSection(&priv->framesLock);
|
||||||
|
|
||||||
|
DeleteCriticalSection(&priv->framesLock);
|
||||||
|
|
||||||
|
if (priv->currentPresentation)
|
||||||
|
PresentationContext_unref(priv->currentPresentation);
|
||||||
|
|
||||||
|
BufferPool_Free(priv->surfacePool);
|
||||||
|
free(priv);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
static UINT video_control_send_presentation_response(VideoClientContext *context, TSMM_PRESENTATION_RESPONSE *resp)
|
||||||
|
{
|
||||||
|
BYTE buf[12];
|
||||||
|
wStream *s;
|
||||||
|
VIDEO_PLUGIN* video = (VIDEO_PLUGIN *)context->handle;
|
||||||
|
IWTSVirtualChannel* channel;
|
||||||
|
UINT ret;
|
||||||
|
|
||||||
|
s = Stream_New(buf, 12);
|
||||||
|
if (!s)
|
||||||
|
return CHANNEL_RC_NO_MEMORY;
|
||||||
|
|
||||||
|
Stream_Write_UINT32(s, 12); /* cbSize */
|
||||||
|
Stream_Write_UINT32(s, TSMM_PACKET_TYPE_PRESENTATION_RESPONSE); /* PacketType */
|
||||||
|
Stream_Write_UINT8(s, resp->PresentationId);
|
||||||
|
Stream_Zero(s, 3);
|
||||||
|
Stream_SealLength(s);
|
||||||
|
|
||||||
|
channel = video->control_callback->channel_callback->channel;
|
||||||
|
ret = channel->Write(channel, 12, buf, NULL);
|
||||||
|
Stream_Free(s, FALSE);
|
||||||
|
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
|
|
||||||
|
static BOOL video_onMappedGeometryUpdate(MAPPED_GEOMETRY *geometry)
|
||||||
|
{
|
||||||
|
WLog_DBG(TAG, "geometry updated");
|
||||||
|
return TRUE;
|
||||||
|
}
|
||||||
|
|
||||||
|
static BOOL video_onMappedGeometryClear(MAPPED_GEOMETRY *geometry)
|
||||||
|
{
|
||||||
|
PresentationContext *presentation = (PresentationContext *)geometry->custom;
|
||||||
|
|
||||||
|
presentation->geometry = NULL;
|
||||||
|
return TRUE;
|
||||||
|
}
|
||||||
|
|
||||||
|
static UINT video_PresentationRequest(VideoClientContext* video, TSMM_PRESENTATION_REQUEST *req)
|
||||||
|
{
|
||||||
|
VideoClientContextPriv *priv = video->priv;
|
||||||
|
PresentationContext *presentation;
|
||||||
|
UINT ret = CHANNEL_RC_OK;
|
||||||
|
|
||||||
|
presentation = priv->currentPresentation;
|
||||||
|
|
||||||
|
if (req->Command == TSMM_START_PRESENTATION)
|
||||||
|
{
|
||||||
|
MAPPED_GEOMETRY *geom;
|
||||||
|
TSMM_PRESENTATION_RESPONSE resp;
|
||||||
|
|
||||||
|
if (memcmp(req->VideoSubtypeId, MFVideoFormat_H264, 16) != 0)
|
||||||
|
{
|
||||||
|
WLog_ERR(TAG, "not a H264 video, ignoring request");
|
||||||
|
return CHANNEL_RC_OK;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (presentation)
|
||||||
|
{
|
||||||
|
if (presentation->PresentationId == req->PresentationId)
|
||||||
|
{
|
||||||
|
WLog_ERR(TAG, "ignoring start request for existing presentation %d", req->PresentationId);
|
||||||
|
return CHANNEL_RC_OK;
|
||||||
|
}
|
||||||
|
|
||||||
|
WLog_ERR(TAG, "releasing current presentation %d", req->PresentationId);
|
||||||
|
PresentationContext_unref(presentation);
|
||||||
|
presentation = priv->currentPresentation = NULL;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!priv->geometry)
|
||||||
|
{
|
||||||
|
WLog_ERR(TAG, "geometry channel not ready, ignoring request");
|
||||||
|
return CHANNEL_RC_OK;
|
||||||
|
}
|
||||||
|
|
||||||
|
geom = HashTable_GetItemValue(priv->geometry->geometries, &(req->GeometryMappingId));
|
||||||
|
if (!geom)
|
||||||
|
{
|
||||||
|
WLog_ERR(TAG, "geometry mapping 0x%"PRIx64" not registered", req->GeometryMappingId);
|
||||||
|
return CHANNEL_RC_OK;
|
||||||
|
}
|
||||||
|
|
||||||
|
WLog_DBG(TAG, "creating presentation 0x%x", req->PresentationId);
|
||||||
|
presentation = PresentationContext_new(video, req->PresentationId,
|
||||||
|
geom->topLevelLeft + geom->left + geom->geometry.boundingRect.x,
|
||||||
|
geom->topLevelTop + geom->top + geom->geometry.boundingRect.y,
|
||||||
|
req->SourceWidth, req->SourceHeight);
|
||||||
|
if (!presentation)
|
||||||
|
{
|
||||||
|
WLog_ERR(TAG, "unable to create presentation video");
|
||||||
|
return CHANNEL_RC_NO_MEMORY;
|
||||||
|
}
|
||||||
|
|
||||||
|
priv->currentPresentation = presentation;
|
||||||
|
presentation->video = video;
|
||||||
|
presentation->geometry = geom;
|
||||||
|
presentation->SourceWidth = req->SourceWidth;
|
||||||
|
presentation->SourceHeight = req->SourceHeight;
|
||||||
|
presentation->ScaledWidth = req->ScaledWidth;
|
||||||
|
presentation->ScaledHeight = req->ScaledHeight;
|
||||||
|
|
||||||
|
geom->custom = presentation;
|
||||||
|
geom->MappedGeometryUpdate = video_onMappedGeometryUpdate;
|
||||||
|
geom->MappedGeometryClear = video_onMappedGeometryClear;
|
||||||
|
|
||||||
|
/* send back response */
|
||||||
|
resp.PresentationId = req->PresentationId;
|
||||||
|
ret = video_control_send_presentation_response(video, &resp);
|
||||||
|
}
|
||||||
|
else if (req->Command == TSMM_STOP_PRESENTATION)
|
||||||
|
{
|
||||||
|
WLog_DBG(TAG, "stopping presentation 0x%x", req->PresentationId);
|
||||||
|
if (!presentation)
|
||||||
|
{
|
||||||
|
WLog_ERR(TAG, "unknown presentation to stop %d", req->PresentationId);
|
||||||
|
return CHANNEL_RC_OK;
|
||||||
|
}
|
||||||
|
|
||||||
|
priv->currentPresentation = NULL;
|
||||||
|
priv->droppedFrames = 0;
|
||||||
|
priv->publishedFrames = 0;
|
||||||
|
PresentationContext_unref(presentation);
|
||||||
|
}
|
||||||
|
|
||||||
|
return CHANNEL_RC_OK;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
static UINT video_read_tsmm_presentation_req(VideoClientContext *context, wStream *s)
|
static UINT video_read_tsmm_presentation_req(VideoClientContext *context, wStream *s)
|
||||||
{
|
{
|
||||||
TSMM_PRESENTATION_REQUEST req;
|
TSMM_PRESENTATION_REQUEST req;
|
||||||
UINT ret = CHANNEL_RC_OK;
|
|
||||||
|
|
||||||
if (Stream_GetRemainingLength(s) < 60)
|
if (Stream_GetRemainingLength(s) < 60)
|
||||||
{
|
{
|
||||||
|
@ -129,36 +519,9 @@ static UINT video_read_tsmm_presentation_req(VideoClientContext *context, wStrea
|
||||||
req.SourceWidth, req.SourceHeight, req.ScaledWidth, req.ScaledHeight,
|
req.SourceWidth, req.SourceHeight, req.ScaledWidth, req.ScaledHeight,
|
||||||
req.hnsTimestampOffset, req.GeometryMappingId);
|
req.hnsTimestampOffset, req.GeometryMappingId);
|
||||||
|
|
||||||
if (context->PresentationRequest)
|
return video_PresentationRequest(context, &req);
|
||||||
ret = context->PresentationRequest(context, &req);
|
|
||||||
|
|
||||||
return ret;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static UINT video_control_send_presentation_response(VideoClientContext *context, TSMM_PRESENTATION_RESPONSE *resp)
|
|
||||||
{
|
|
||||||
BYTE buf[12];
|
|
||||||
wStream *s;
|
|
||||||
VIDEO_PLUGIN* video = (VIDEO_PLUGIN *)context->handle;
|
|
||||||
IWTSVirtualChannel* channel;
|
|
||||||
UINT ret;
|
|
||||||
|
|
||||||
s = Stream_New(buf, 12);
|
|
||||||
if (!s)
|
|
||||||
return CHANNEL_RC_NO_MEMORY;
|
|
||||||
|
|
||||||
Stream_Write_UINT32(s, 12); /* cbSize */
|
|
||||||
Stream_Write_UINT32(s, TSMM_PACKET_TYPE_PRESENTATION_RESPONSE); /* PacketType */
|
|
||||||
Stream_Write_UINT8(s, resp->PresentationId);
|
|
||||||
Stream_Zero(s, 3);
|
|
||||||
Stream_SealLength(s);
|
|
||||||
|
|
||||||
channel = video->control_callback->channel_callback->channel;
|
|
||||||
ret = channel->Write(channel, 12, buf, NULL);
|
|
||||||
Stream_Free(s, FALSE);
|
|
||||||
|
|
||||||
return ret;
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -248,6 +611,263 @@ static UINT video_control_send_client_notification(VideoClientContext *context,
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void video_timer(VideoClientContext *video, UINT64 now)
|
||||||
|
{
|
||||||
|
PresentationContext *presentation;
|
||||||
|
VideoClientContextPriv *priv = video->priv;
|
||||||
|
VideoFrame *peekFrame, *frame = NULL;
|
||||||
|
|
||||||
|
EnterCriticalSection(&priv->framesLock);
|
||||||
|
do
|
||||||
|
{
|
||||||
|
peekFrame = (VideoFrame *)Queue_Peek(priv->frames);
|
||||||
|
if (!peekFrame)
|
||||||
|
break;
|
||||||
|
|
||||||
|
if (peekFrame->publishTime > now)
|
||||||
|
break;
|
||||||
|
|
||||||
|
if (frame)
|
||||||
|
{
|
||||||
|
/* free skipped frame */
|
||||||
|
WLog_DBG(TAG, "dropping frame @%"PRIu64, frame->publishTime);
|
||||||
|
priv->droppedFrames++;
|
||||||
|
VideoFrame_free(&frame);
|
||||||
|
}
|
||||||
|
frame = peekFrame;
|
||||||
|
Queue_Dequeue(priv->frames);
|
||||||
|
}
|
||||||
|
while (1);
|
||||||
|
LeaveCriticalSection(&priv->framesLock);
|
||||||
|
|
||||||
|
if (!frame)
|
||||||
|
goto treat_feedback;
|
||||||
|
|
||||||
|
presentation = frame->presentation;
|
||||||
|
|
||||||
|
priv->publishedFrames++;
|
||||||
|
memcpy(presentation->surfaceData, frame->surfaceData, frame->w * frame->h * 4);
|
||||||
|
|
||||||
|
video->showSurface(video, presentation->surface);
|
||||||
|
|
||||||
|
PresentationContext_unref(presentation);
|
||||||
|
BufferPool_Return(priv->surfacePool, frame->surfaceData);
|
||||||
|
free(frame);
|
||||||
|
|
||||||
|
treat_feedback:
|
||||||
|
if (priv->nextFeedbackTime < now)
|
||||||
|
{
|
||||||
|
/* we can compute some feedback only if we have some published frames and
|
||||||
|
* a current presentation
|
||||||
|
*/
|
||||||
|
if (priv->publishedFrames && priv->currentPresentation)
|
||||||
|
{
|
||||||
|
UINT32 computedRate;
|
||||||
|
|
||||||
|
InterlockedIncrement(&priv->currentPresentation->refCounter);
|
||||||
|
|
||||||
|
if (priv->droppedFrames)
|
||||||
|
{
|
||||||
|
/**
|
||||||
|
* some dropped frames, looks like we're asking too many frames per seconds,
|
||||||
|
* try lowering rate. We go directly from unlimited rate to 24 frames/seconds
|
||||||
|
* otherwise we lower rate by 2 frames by seconds
|
||||||
|
*/
|
||||||
|
if (priv->lastSentRate == XF_VIDEO_UNLIMITED_RATE)
|
||||||
|
computedRate = 24;
|
||||||
|
else
|
||||||
|
{
|
||||||
|
computedRate = priv->lastSentRate - 2;
|
||||||
|
if (!computedRate)
|
||||||
|
computedRate = 2;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
else
|
||||||
|
{
|
||||||
|
/**
|
||||||
|
* we treat all frames ok, so either ask the server to send more,
|
||||||
|
* or stay unlimited
|
||||||
|
*/
|
||||||
|
if (priv->lastSentRate == XF_VIDEO_UNLIMITED_RATE)
|
||||||
|
computedRate = XF_VIDEO_UNLIMITED_RATE; /* stay unlimited */
|
||||||
|
else
|
||||||
|
{
|
||||||
|
computedRate = priv->lastSentRate + 2;
|
||||||
|
if (computedRate > XF_VIDEO_UNLIMITED_RATE)
|
||||||
|
computedRate = XF_VIDEO_UNLIMITED_RATE;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if (computedRate != priv->lastSentRate)
|
||||||
|
{
|
||||||
|
TSMM_CLIENT_NOTIFICATION notif;
|
||||||
|
notif.PresentationId = priv->currentPresentation->PresentationId;
|
||||||
|
notif.NotificationType = TSMM_CLIENT_NOTIFICATION_TYPE_FRAMERATE_OVERRIDE;
|
||||||
|
if (computedRate == XF_VIDEO_UNLIMITED_RATE)
|
||||||
|
{
|
||||||
|
notif.FramerateOverride.Flags = 0x01;
|
||||||
|
notif.FramerateOverride.DesiredFrameRate = 0x00;
|
||||||
|
}
|
||||||
|
else
|
||||||
|
{
|
||||||
|
notif.FramerateOverride.Flags = 0x02;
|
||||||
|
notif.FramerateOverride.DesiredFrameRate = computedRate;
|
||||||
|
}
|
||||||
|
|
||||||
|
video_control_send_client_notification(video, ¬if);
|
||||||
|
priv->lastSentRate = computedRate;
|
||||||
|
|
||||||
|
WLog_DBG(TAG, "server notified with rate %d published=%d dropped=%d", priv->lastSentRate,
|
||||||
|
priv->publishedFrames, priv->droppedFrames);
|
||||||
|
}
|
||||||
|
|
||||||
|
PresentationContext_unref(priv->currentPresentation);
|
||||||
|
}
|
||||||
|
|
||||||
|
WLog_DBG(TAG, "currentRate=%d published=%d dropped=%d", priv->lastSentRate,
|
||||||
|
priv->publishedFrames, priv->droppedFrames);
|
||||||
|
|
||||||
|
priv->droppedFrames = 0;
|
||||||
|
priv->publishedFrames = 0;
|
||||||
|
priv->nextFeedbackTime = now + 1000;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
static UINT video_VideoData(VideoClientContext* context, TSMM_VIDEO_DATA *data)
|
||||||
|
{
|
||||||
|
VideoClientContextPriv *priv = context->priv;
|
||||||
|
PresentationContext *presentation;
|
||||||
|
int status;
|
||||||
|
|
||||||
|
presentation = priv->currentPresentation;
|
||||||
|
if (!presentation)
|
||||||
|
{
|
||||||
|
WLog_ERR(TAG, "no current presentation");
|
||||||
|
return CHANNEL_RC_OK;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (presentation->PresentationId != data->PresentationId)
|
||||||
|
{
|
||||||
|
WLog_ERR(TAG, "current presentation id=%d doesn't match data id=%d", presentation->PresentationId,
|
||||||
|
data->PresentationId);
|
||||||
|
return CHANNEL_RC_OK;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!Stream_EnsureRemainingCapacity(presentation->currentSample, data->cbSample))
|
||||||
|
{
|
||||||
|
WLog_ERR(TAG, "unable to expand the current packet");
|
||||||
|
return CHANNEL_RC_NO_MEMORY;
|
||||||
|
}
|
||||||
|
|
||||||
|
Stream_Write(presentation->currentSample, data->pSample, data->cbSample);
|
||||||
|
|
||||||
|
if (data->CurrentPacketIndex == data->PacketsInSample)
|
||||||
|
{
|
||||||
|
H264_CONTEXT *h264 = presentation->h264;
|
||||||
|
UINT64 startTime = GetTickCount64(), timeAfterH264;
|
||||||
|
MAPPED_GEOMETRY *geom = presentation->geometry;
|
||||||
|
|
||||||
|
Stream_SealLength(presentation->currentSample);
|
||||||
|
Stream_SetPosition(presentation->currentSample, 0);
|
||||||
|
|
||||||
|
status = h264->subsystem->Decompress(h264, Stream_Pointer(presentation->currentSample),
|
||||||
|
Stream_Length(presentation->currentSample));
|
||||||
|
if (status == 0)
|
||||||
|
return CHANNEL_RC_OK;
|
||||||
|
|
||||||
|
if (status < 0)
|
||||||
|
return CHANNEL_RC_OK;
|
||||||
|
|
||||||
|
timeAfterH264 = GetTickCount64();
|
||||||
|
if (data->SampleNumber == 1)
|
||||||
|
{
|
||||||
|
presentation->lastPublishTime = startTime;
|
||||||
|
}
|
||||||
|
|
||||||
|
presentation->lastPublishTime += (data->hnsDuration / 10000);
|
||||||
|
if (presentation->lastPublishTime <= timeAfterH264 + 10)
|
||||||
|
{
|
||||||
|
int dropped = 0;
|
||||||
|
|
||||||
|
/* if the frame is to be published in less than 10 ms, let's consider it's now */
|
||||||
|
yuv_to_rgb(presentation, presentation->surfaceData);
|
||||||
|
|
||||||
|
context->showSurface(context, presentation->surface);
|
||||||
|
|
||||||
|
priv->publishedFrames++;
|
||||||
|
|
||||||
|
/* cleanup previously scheduled frames */
|
||||||
|
EnterCriticalSection(&priv->framesLock);
|
||||||
|
while (Queue_Count(priv->frames) > 0)
|
||||||
|
{
|
||||||
|
VideoFrame *frame = Queue_Dequeue(priv->frames);
|
||||||
|
if (frame)
|
||||||
|
{
|
||||||
|
priv->droppedFrames++;
|
||||||
|
VideoFrame_free(&frame);
|
||||||
|
dropped++;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
LeaveCriticalSection(&priv->framesLock);
|
||||||
|
|
||||||
|
if (dropped)
|
||||||
|
WLog_DBG(TAG, "showing frame (%d dropped)", dropped);
|
||||||
|
}
|
||||||
|
else
|
||||||
|
{
|
||||||
|
BOOL enqueueResult;
|
||||||
|
VideoFrame *frame = calloc(1, sizeof(*frame));
|
||||||
|
if (!frame)
|
||||||
|
{
|
||||||
|
WLog_ERR(TAG, "unable to create frame");
|
||||||
|
return CHANNEL_RC_NO_MEMORY;
|
||||||
|
}
|
||||||
|
|
||||||
|
frame->presentation = presentation;
|
||||||
|
frame->publishTime = presentation->lastPublishTime;
|
||||||
|
frame->x = geom->topLevelLeft + geom->left + geom->geometry.boundingRect.x;
|
||||||
|
frame->y = geom->topLevelTop + geom->top + geom->geometry.boundingRect.y;
|
||||||
|
frame->w = presentation->SourceWidth;
|
||||||
|
frame->h = presentation->SourceHeight;
|
||||||
|
|
||||||
|
frame->surfaceData = BufferPool_Take(priv->surfacePool, frame->w * frame->h * 4);
|
||||||
|
if (!frame->surfaceData)
|
||||||
|
{
|
||||||
|
WLog_ERR(TAG, "unable to allocate frame data");
|
||||||
|
free(frame);
|
||||||
|
return CHANNEL_RC_NO_MEMORY;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!yuv_to_rgb(presentation, frame->surfaceData))
|
||||||
|
{
|
||||||
|
WLog_ERR(TAG, "error during YUV->RGB conversion");
|
||||||
|
BufferPool_Return(priv->surfacePool, frame->surfaceData);
|
||||||
|
free(frame);
|
||||||
|
return CHANNEL_RC_NO_MEMORY;
|
||||||
|
}
|
||||||
|
|
||||||
|
InterlockedIncrement(&presentation->refCounter);
|
||||||
|
|
||||||
|
EnterCriticalSection(&priv->framesLock);
|
||||||
|
enqueueResult = Queue_Enqueue(priv->frames, frame);
|
||||||
|
LeaveCriticalSection(&priv->framesLock);
|
||||||
|
|
||||||
|
if (!enqueueResult)
|
||||||
|
{
|
||||||
|
WLog_ERR(TAG, "unable to enqueue frame");
|
||||||
|
VideoFrame_free(&frame);
|
||||||
|
return CHANNEL_RC_NO_MEMORY;
|
||||||
|
}
|
||||||
|
|
||||||
|
WLog_DBG(TAG, "scheduling frame in %"PRIu32" ms", (frame->publishTime-startTime));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return CHANNEL_RC_OK;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
static UINT video_data_on_data_received(IWTSVirtualChannelCallback* pChannelCallback, wStream *s)
|
static UINT video_data_on_data_received(IWTSVirtualChannelCallback* pChannelCallback, wStream *s)
|
||||||
{
|
{
|
||||||
|
@ -295,15 +915,14 @@ static UINT video_data_on_data_received(IWTSVirtualChannelCallback* pChannelCall
|
||||||
Stream_Read_UINT32(s, data.cbSample);
|
Stream_Read_UINT32(s, data.cbSample);
|
||||||
data.pSample = Stream_Pointer(s);
|
data.pSample = Stream_Pointer(s);
|
||||||
|
|
||||||
|
/*
|
||||||
WLog_DBG(TAG, "videoData: id:%"PRIu8" version:%"PRIu8" flags:0x%"PRIx8" timestamp=%"PRIu64" duration=%"PRIu64
|
WLog_DBG(TAG, "videoData: id:%"PRIu8" version:%"PRIu8" flags:0x%"PRIx8" timestamp=%"PRIu64" duration=%"PRIu64
|
||||||
" curPacketIndex:%"PRIu16" packetInSample:%"PRIu16" sampleNumber:%"PRIu32" cbSample:%"PRIu32"",
|
" curPacketIndex:%"PRIu16" packetInSample:%"PRIu16" sampleNumber:%"PRIu32" cbSample:%"PRIu32"",
|
||||||
data.PresentationId, data.Version, data.Flags, data.hnsTimestamp, data.hnsDuration,
|
data.PresentationId, data.Version, data.Flags, data.hnsTimestamp, data.hnsDuration,
|
||||||
data.CurrentPacketIndex, data.PacketsInSample, data.SampleNumber, data.cbSample);
|
data.CurrentPacketIndex, data.PacketsInSample, data.SampleNumber, data.cbSample);
|
||||||
|
*/
|
||||||
|
|
||||||
if (context->VideoData)
|
return video_VideoData(context, &data);
|
||||||
return context->VideoData(context, &data);
|
|
||||||
|
|
||||||
return CHANNEL_RC_OK;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
@ -441,6 +1060,9 @@ static UINT video_plugin_terminated(IWTSPlugin* pPlugin)
|
||||||
{
|
{
|
||||||
VIDEO_PLUGIN* video = (VIDEO_PLUGIN*) pPlugin;
|
VIDEO_PLUGIN* video = (VIDEO_PLUGIN*) pPlugin;
|
||||||
|
|
||||||
|
if (video->context)
|
||||||
|
VideoClientContextPriv_free(video->context->priv);
|
||||||
|
|
||||||
free(video->control_callback);
|
free(video->control_callback);
|
||||||
free(video->data_callback);
|
free(video->data_callback);
|
||||||
free(video->wtsPlugin.pInterface);
|
free(video->wtsPlugin.pInterface);
|
||||||
|
@ -468,7 +1090,8 @@ UINT DVCPluginEntry(IDRDYNVC_ENTRY_POINTS* pEntryPoints)
|
||||||
{
|
{
|
||||||
UINT error = CHANNEL_RC_OK;
|
UINT error = CHANNEL_RC_OK;
|
||||||
VIDEO_PLUGIN* videoPlugin;
|
VIDEO_PLUGIN* videoPlugin;
|
||||||
VideoClientContext* context;
|
VideoClientContext* videoContext;
|
||||||
|
VideoClientContextPriv *priv;
|
||||||
|
|
||||||
videoPlugin = (VIDEO_PLUGIN*) pEntryPoints->GetPlugin(pEntryPoints, "video");
|
videoPlugin = (VIDEO_PLUGIN*) pEntryPoints->GetPlugin(pEntryPoints, "video");
|
||||||
if (!videoPlugin)
|
if (!videoPlugin)
|
||||||
|
@ -485,20 +1108,30 @@ UINT DVCPluginEntry(IDRDYNVC_ENTRY_POINTS* pEntryPoints)
|
||||||
videoPlugin->wtsPlugin.Disconnected = NULL;
|
videoPlugin->wtsPlugin.Disconnected = NULL;
|
||||||
videoPlugin->wtsPlugin.Terminated = video_plugin_terminated;
|
videoPlugin->wtsPlugin.Terminated = video_plugin_terminated;
|
||||||
|
|
||||||
context = (VideoClientContext*) calloc(1, sizeof(VideoClientContext));
|
videoContext = (VideoClientContext*) calloc(1, sizeof(VideoClientContext));
|
||||||
if (!context)
|
if (!videoContext)
|
||||||
{
|
{
|
||||||
WLog_ERR(TAG, "calloc failed!");
|
WLog_ERR(TAG, "calloc failed!");
|
||||||
free(videoPlugin);
|
free(videoPlugin);
|
||||||
return CHANNEL_RC_NO_MEMORY;
|
return CHANNEL_RC_NO_MEMORY;
|
||||||
}
|
}
|
||||||
|
|
||||||
context->handle = (void*) videoPlugin;
|
priv = VideoClientContextPriv_new(videoContext);
|
||||||
context->PresentationResponse = video_control_send_presentation_response;
|
if (!priv)
|
||||||
context->ClientNotification = video_control_send_client_notification;
|
{
|
||||||
|
WLog_ERR(TAG, "VideoClientContextPriv_new failed!");
|
||||||
|
free(videoContext);
|
||||||
|
free(videoPlugin);
|
||||||
|
return CHANNEL_RC_NO_MEMORY;
|
||||||
|
}
|
||||||
|
|
||||||
videoPlugin->wtsPlugin.pInterface = (void*) context;
|
videoContext->handle = (void*) videoPlugin;
|
||||||
videoPlugin->context = context;
|
videoContext->priv = priv;
|
||||||
|
videoContext->timer = video_timer;
|
||||||
|
videoContext->setGeometry = video_client_context_set_geometry;
|
||||||
|
|
||||||
|
videoPlugin->wtsPlugin.pInterface = (void*) videoContext;
|
||||||
|
videoPlugin->context = videoContext;
|
||||||
|
|
||||||
error = pEntryPoints->RegisterPlugin(pEntryPoints, "video", (IWTSPlugin*) videoPlugin);
|
error = pEntryPoints->RegisterPlugin(pEntryPoints, "video", (IWTSPlugin*) videoPlugin);
|
||||||
}
|
}
|
||||||
|
|
|
@ -1280,12 +1280,12 @@ static BOOL xf_post_connect(freerdp* instance)
|
||||||
return FALSE;
|
return FALSE;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!(xfc->xfVideo = xf_video_new(xfc)))
|
/* if (!(xfc->xfVideo = xf_video_new(xfc)))
|
||||||
{
|
{
|
||||||
xf_clipboard_free(xfc->clipboard);
|
xf_clipboard_free(xfc->clipboard);
|
||||||
xf_disp_free(xfc->xfDisp);
|
xf_disp_free(xfc->xfDisp);
|
||||||
return FALSE;
|
return FALSE;
|
||||||
}
|
}*/
|
||||||
|
|
||||||
EventArgsInit(&e, "xfreerdp");
|
EventArgsInit(&e, "xfreerdp");
|
||||||
e.width = settings->DesktopWidth;
|
e.width = settings->DesktopWidth;
|
||||||
|
@ -1323,12 +1323,6 @@ static void xf_post_disconnect(freerdp* instance)
|
||||||
else
|
else
|
||||||
xf_DestroyDummyWindow(xfc, xfc->drawable);
|
xf_DestroyDummyWindow(xfc, xfc->drawable);
|
||||||
|
|
||||||
if (xfc->xfVideo)
|
|
||||||
{
|
|
||||||
xf_video_free(xfc->xfVideo);
|
|
||||||
xfc->xfVideo = NULL;
|
|
||||||
}
|
|
||||||
|
|
||||||
xf_window_free(xfc);
|
xf_window_free(xfc);
|
||||||
xf_keyboard_free(xfc);
|
xf_keyboard_free(xfc);
|
||||||
}
|
}
|
||||||
|
|
|
@ -16,648 +16,98 @@
|
||||||
* See the License for the specific language governing permissions and
|
* See the License for the specific language governing permissions and
|
||||||
* limitations under the License.
|
* limitations under the License.
|
||||||
*/
|
*/
|
||||||
#include <winpr/sysinfo.h>
|
|
||||||
#include <winpr/interlocked.h>
|
|
||||||
|
|
||||||
#include <freerdp/client/geometry.h>
|
#include <freerdp/client/geometry.h>
|
||||||
#include <freerdp/client/video.h>
|
#include <freerdp/client/video.h>
|
||||||
#include <freerdp/primitives.h>
|
|
||||||
#include <freerdp/codec/h264.h>
|
|
||||||
#include <freerdp/codec/yuv.h>
|
|
||||||
|
|
||||||
#include "xf_video.h"
|
#include "xf_video.h"
|
||||||
|
|
||||||
#define TAG CLIENT_TAG("video")
|
#define TAG CLIENT_TAG("video")
|
||||||
#define XF_VIDEO_UNLIMITED_RATE 31
|
|
||||||
|
|
||||||
|
typedef struct
|
||||||
BYTE MFVideoFormat_H264[] = {'H', '2', '6', '4',
|
|
||||||
0x00, 0x00,
|
|
||||||
0x10, 0x00,
|
|
||||||
0x80, 0x00,
|
|
||||||
0x00, 0xAA, 0x00, 0x38, 0x9B, 0x71};
|
|
||||||
|
|
||||||
typedef struct _xfPresentationContext xfPresentationContext;
|
|
||||||
|
|
||||||
struct _xfVideoContext
|
|
||||||
{
|
{
|
||||||
xfContext *xfc;
|
VideoSurface base;
|
||||||
wQueue *frames;
|
XImage *image;
|
||||||
CRITICAL_SECTION framesLock;
|
} xfVideoSurface;
|
||||||
wBufferPool *surfacePool;
|
|
||||||
UINT32 publishedFrames;
|
|
||||||
UINT32 droppedFrames;
|
|
||||||
UINT32 lastSentRate;
|
|
||||||
UINT64 nextFeedbackTime;
|
|
||||||
xfPresentationContext *currentPresentation;
|
|
||||||
};
|
|
||||||
|
|
||||||
struct _xfVideoFrame
|
|
||||||
{
|
|
||||||
UINT64 publishTime;
|
|
||||||
UINT64 hnsDuration;
|
|
||||||
UINT32 x, y, w, h;
|
|
||||||
BYTE *surfaceData;
|
|
||||||
xfPresentationContext *presentation;
|
|
||||||
};
|
|
||||||
typedef struct _xfVideoFrame xfVideoFrame;
|
|
||||||
|
|
||||||
struct _xfPresentationContext
|
|
||||||
{
|
|
||||||
xfContext *xfc;
|
|
||||||
xfVideoContext *xfVideo;
|
|
||||||
BYTE PresentationId;
|
|
||||||
UINT32 SourceWidth, SourceHeight;
|
|
||||||
UINT32 ScaledWidth, ScaledHeight;
|
|
||||||
MAPPED_GEOMETRY *geometry;
|
|
||||||
|
|
||||||
UINT64 startTimeStamp;
|
|
||||||
UINT64 publishOffset;
|
|
||||||
H264_CONTEXT *h264;
|
|
||||||
YUV_CONTEXT *yuv;
|
|
||||||
wStream *currentSample;
|
|
||||||
BYTE *surfaceData;
|
|
||||||
XImage *surface;
|
|
||||||
UINT64 lastPublishTime, nextPublishTime;
|
|
||||||
volatile LONG refCounter;
|
|
||||||
};
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
static void xfPresentationContext_unref(xfPresentationContext *c)
|
|
||||||
{
|
|
||||||
xfVideoContext *xfVideo;
|
|
||||||
|
|
||||||
if (!c)
|
|
||||||
return;
|
|
||||||
|
|
||||||
if (InterlockedDecrement(&c->refCounter) != 0)
|
|
||||||
return;
|
|
||||||
|
|
||||||
xfVideo = c->xfVideo;
|
|
||||||
if (c->geometry)
|
|
||||||
{
|
|
||||||
c->geometry->MappedGeometryUpdate = NULL;
|
|
||||||
c->geometry->MappedGeometryClear = NULL;
|
|
||||||
c->geometry->custom = NULL;
|
|
||||||
}
|
|
||||||
|
|
||||||
h264_context_free(c->h264);
|
|
||||||
Stream_Free(c->currentSample, TRUE);
|
|
||||||
XFree(c->surface);
|
|
||||||
BufferPool_Return(xfVideo->surfacePool, c->surfaceData);
|
|
||||||
yuv_context_free(c->yuv);
|
|
||||||
free(c);
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
static xfPresentationContext *xfPresentationContext_new(xfContext *xfc, BYTE PresentationId, UINT32 width, UINT32 height)
|
|
||||||
{
|
|
||||||
xfVideoContext *xfVideo = xfc->xfVideo;
|
|
||||||
xfPresentationContext *ret = calloc(1, sizeof(*ret));
|
|
||||||
if (!ret)
|
|
||||||
return NULL;
|
|
||||||
|
|
||||||
ret->xfc = xfc;
|
|
||||||
ret->PresentationId = PresentationId;
|
|
||||||
|
|
||||||
ret->h264 = h264_context_new(FALSE);
|
|
||||||
if (!ret->h264)
|
|
||||||
{
|
|
||||||
WLog_ERR(TAG, "unable to create a h264 context");
|
|
||||||
goto error_h264;
|
|
||||||
}
|
|
||||||
h264_context_reset(ret->h264, width, height);
|
|
||||||
|
|
||||||
ret->currentSample = Stream_New(NULL, 4096);
|
|
||||||
if (!ret->currentSample)
|
|
||||||
{
|
|
||||||
WLog_ERR(TAG, "unable to create current packet stream");
|
|
||||||
goto error_currentSample;
|
|
||||||
}
|
|
||||||
|
|
||||||
ret->surfaceData = BufferPool_Take(xfVideo->surfacePool, width * height * 4);
|
|
||||||
if (!ret->surfaceData)
|
|
||||||
{
|
|
||||||
WLog_ERR(TAG, "unable to allocate surfaceData");
|
|
||||||
goto error_surfaceData;
|
|
||||||
}
|
|
||||||
|
|
||||||
ret->surface = XCreateImage(xfc->display, xfc->visual, xfc->depth, ZPixmap, 0,
|
|
||||||
(char *)ret->surfaceData, width, height, 8, width * 4);
|
|
||||||
if (!ret->surface)
|
|
||||||
{
|
|
||||||
WLog_ERR(TAG, "unable to create surface");
|
|
||||||
goto error_surface;
|
|
||||||
}
|
|
||||||
|
|
||||||
ret->yuv = yuv_context_new(FALSE);
|
|
||||||
if (!ret->yuv)
|
|
||||||
{
|
|
||||||
WLog_ERR(TAG, "unable to create YUV decoder");
|
|
||||||
goto error_yuv;
|
|
||||||
}
|
|
||||||
|
|
||||||
yuv_context_reset(ret->yuv, width, height);
|
|
||||||
ret->refCounter = 1;
|
|
||||||
return ret;
|
|
||||||
|
|
||||||
error_yuv:
|
|
||||||
XFree(ret->surface);
|
|
||||||
error_surface:
|
|
||||||
BufferPool_Return(xfVideo->surfacePool, ret->surfaceData);
|
|
||||||
error_surfaceData:
|
|
||||||
Stream_Free(ret->currentSample, TRUE);
|
|
||||||
error_currentSample:
|
|
||||||
h264_context_free(ret->h264);
|
|
||||||
error_h264:
|
|
||||||
free(ret);
|
|
||||||
return NULL;
|
|
||||||
}
|
|
||||||
|
|
||||||
xfVideoContext *xf_video_new(xfContext *xfc)
|
|
||||||
{
|
|
||||||
xfVideoContext *ret = calloc(1, sizeof(xfVideoContext));
|
|
||||||
if (!ret)
|
|
||||||
return NULL;
|
|
||||||
|
|
||||||
ret->frames = Queue_New(TRUE, 10, 2);
|
|
||||||
if (!ret->frames)
|
|
||||||
{
|
|
||||||
WLog_ERR(TAG, "unable to allocate frames queue");
|
|
||||||
goto error_frames;
|
|
||||||
}
|
|
||||||
|
|
||||||
ret->surfacePool = BufferPool_New(FALSE, 0, 16);
|
|
||||||
if (!ret->surfacePool)
|
|
||||||
{
|
|
||||||
WLog_ERR(TAG, "unable to create surface pool");
|
|
||||||
goto error_surfacePool;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (!InitializeCriticalSectionAndSpinCount(&ret->framesLock, 4000))
|
|
||||||
{
|
|
||||||
WLog_ERR(TAG, "unable to initialize frames lock");
|
|
||||||
goto error_spinlock;
|
|
||||||
}
|
|
||||||
|
|
||||||
ret->xfc = xfc;
|
|
||||||
|
|
||||||
/* don't set to unlimited so that we have the chance to send a feedback in
|
|
||||||
* the first second (for servers that want feedback directly)
|
|
||||||
*/
|
|
||||||
ret->lastSentRate = 30;
|
|
||||||
return ret;
|
|
||||||
|
|
||||||
error_spinlock:
|
|
||||||
BufferPool_Free(ret->surfacePool);
|
|
||||||
error_surfacePool:
|
|
||||||
Queue_Free(ret->frames);
|
|
||||||
error_frames:
|
|
||||||
free(ret);
|
|
||||||
return NULL;
|
|
||||||
}
|
|
||||||
|
|
||||||
void xf_video_geometry_init(xfContext *xfc, GeometryClientContext *geom)
|
void xf_video_geometry_init(xfContext *xfc, GeometryClientContext *geom)
|
||||||
{
|
{
|
||||||
xfc->geometry = geom;
|
xfc->geometry = geom;
|
||||||
}
|
if (xfc->video)
|
||||||
|
|
||||||
static BOOL xf_video_onMappedGeometryUpdate(MAPPED_GEOMETRY *geometry)
|
|
||||||
{
|
|
||||||
return TRUE;
|
|
||||||
}
|
|
||||||
|
|
||||||
static BOOL xf_video_onMappedGeometryClear(MAPPED_GEOMETRY *geometry)
|
|
||||||
{
|
|
||||||
xfPresentationContext *presentation = (xfPresentationContext *)geometry->custom;
|
|
||||||
|
|
||||||
presentation->geometry = NULL;
|
|
||||||
return TRUE;
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
static UINT xf_video_PresentationRequest(VideoClientContext* context, TSMM_PRESENTATION_REQUEST *req)
|
|
||||||
{
|
|
||||||
xfContext *xfc = context->custom;
|
|
||||||
xfVideoContext *xfVideo = xfc->xfVideo;
|
|
||||||
xfPresentationContext *presentation;
|
|
||||||
UINT ret = CHANNEL_RC_OK;
|
|
||||||
|
|
||||||
presentation = xfVideo->currentPresentation;
|
|
||||||
|
|
||||||
if (req->Command == TSMM_START_PRESENTATION)
|
|
||||||
{
|
{
|
||||||
MAPPED_GEOMETRY *geom;
|
VideoClientContext *video = xfc->video;
|
||||||
TSMM_PRESENTATION_RESPONSE resp;
|
video->setGeometry(video, xfc->geometry);
|
||||||
|
|
||||||
if (memcmp(req->VideoSubtypeId, MFVideoFormat_H264, 16) != 0)
|
|
||||||
{
|
|
||||||
WLog_ERR(TAG, "not a H264 video, ignoring request");
|
|
||||||
return CHANNEL_RC_OK;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (presentation)
|
|
||||||
{
|
|
||||||
if (presentation->PresentationId == req->PresentationId)
|
|
||||||
{
|
|
||||||
WLog_ERR(TAG, "ignoring start request for existing presentation %d", req->PresentationId);
|
|
||||||
return CHANNEL_RC_OK;
|
|
||||||
}
|
|
||||||
|
|
||||||
WLog_ERR(TAG, "releasing current presentation %d", req->PresentationId);
|
|
||||||
xfPresentationContext_unref(presentation);
|
|
||||||
presentation = xfVideo->currentPresentation = NULL;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (!xfc->geometry)
|
|
||||||
{
|
|
||||||
WLog_ERR(TAG, "geometry channel not ready, ignoring request");
|
|
||||||
return CHANNEL_RC_OK;
|
|
||||||
}
|
|
||||||
|
|
||||||
geom = HashTable_GetItemValue(xfc->geometry->geometries, &(req->GeometryMappingId));
|
|
||||||
if (!geom)
|
|
||||||
{
|
|
||||||
WLog_ERR(TAG, "geometry mapping 0x%"PRIx64" not registered", req->GeometryMappingId);
|
|
||||||
return CHANNEL_RC_OK;
|
|
||||||
}
|
|
||||||
|
|
||||||
WLog_DBG(TAG, "creating presentation 0x%x", req->PresentationId);
|
|
||||||
presentation = xfPresentationContext_new(xfc, req->PresentationId, req->SourceWidth, req->SourceHeight);
|
|
||||||
if (!presentation)
|
|
||||||
{
|
|
||||||
WLog_ERR(TAG, "unable to create presentation context");
|
|
||||||
return CHANNEL_RC_NO_MEMORY;
|
|
||||||
}
|
|
||||||
|
|
||||||
xfVideo->currentPresentation = presentation;
|
|
||||||
presentation->xfVideo = xfVideo;
|
|
||||||
presentation->geometry = geom;
|
|
||||||
presentation->SourceWidth = req->SourceWidth;
|
|
||||||
presentation->SourceHeight = req->SourceHeight;
|
|
||||||
presentation->ScaledWidth = req->ScaledWidth;
|
|
||||||
presentation->ScaledHeight = req->ScaledHeight;
|
|
||||||
|
|
||||||
geom->custom = presentation;
|
|
||||||
geom->MappedGeometryUpdate = xf_video_onMappedGeometryUpdate;
|
|
||||||
geom->MappedGeometryClear = xf_video_onMappedGeometryClear;
|
|
||||||
|
|
||||||
/* send back response */
|
|
||||||
resp.PresentationId = req->PresentationId;
|
|
||||||
ret = context->PresentationResponse(context, &resp);
|
|
||||||
}
|
}
|
||||||
else if (req->Command == TSMM_STOP_PRESENTATION)
|
}
|
||||||
|
|
||||||
|
static VideoSurface *xfVideoCreateSurface(VideoClientContext *video, BYTE *data, UINT32 x, UINT32 y,
|
||||||
|
UINT32 width, UINT32 height)
|
||||||
|
{
|
||||||
|
xfContext *xfc = (xfContext *)video->custom;
|
||||||
|
xfVideoSurface *ret = calloc(1, sizeof(*ret));
|
||||||
|
if (!ret)
|
||||||
|
return NULL;
|
||||||
|
|
||||||
|
ret->base.data = data;
|
||||||
|
ret->base.x = x;
|
||||||
|
ret->base.y = y;
|
||||||
|
ret->base.w = width;
|
||||||
|
ret->base.h = height;
|
||||||
|
|
||||||
|
ret->image = XCreateImage(xfc->display, xfc->visual, xfc->depth, ZPixmap, 0,
|
||||||
|
(char *)data, width, height, 8, width * 4);
|
||||||
|
if (!ret->image)
|
||||||
{
|
{
|
||||||
WLog_DBG(TAG, "stopping presentation 0x%x", req->PresentationId);
|
WLog_ERR(TAG, "unable to create surface image");
|
||||||
if (!presentation)
|
free(ret);
|
||||||
{
|
return NULL;
|
||||||
WLog_ERR(TAG, "unknown presentation to stop %d", req->PresentationId);
|
|
||||||
return CHANNEL_RC_OK;
|
|
||||||
}
|
|
||||||
|
|
||||||
xfVideo->currentPresentation = NULL;
|
|
||||||
xfVideo->droppedFrames = 0;
|
|
||||||
xfVideo->publishedFrames = 0;
|
|
||||||
xfPresentationContext_unref(presentation);
|
|
||||||
}
|
}
|
||||||
|
return &ret->base;
|
||||||
return CHANNEL_RC_OK;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
static BOOL yuv_to_rgb(xfPresentationContext *presentation, BYTE *dest)
|
static BOOL xfVideoShowSurface(VideoClientContext *video, xfVideoSurface *surface)
|
||||||
{
|
{
|
||||||
const BYTE* pYUVPoint[3];
|
xfContext *xfc = video->custom;
|
||||||
H264_CONTEXT *h264 = presentation->h264;
|
|
||||||
|
|
||||||
BYTE** ppYUVData;
|
XPutImage(xfc->display, xfc->drawable, xfc->gc, surface->image,
|
||||||
ppYUVData = h264->pYUVData;
|
|
||||||
|
|
||||||
pYUVPoint[0] = ppYUVData[0];
|
|
||||||
pYUVPoint[1] = ppYUVData[1];
|
|
||||||
pYUVPoint[2] = ppYUVData[2];
|
|
||||||
|
|
||||||
if (!yuv_context_decode(presentation->yuv, pYUVPoint, h264->iStride, PIXEL_FORMAT_BGRX32, dest, h264->width * 4))
|
|
||||||
{
|
|
||||||
WLog_ERR(TAG, "error in yuv_to_rgb conversion");
|
|
||||||
return FALSE;
|
|
||||||
}
|
|
||||||
|
|
||||||
return TRUE;
|
|
||||||
}
|
|
||||||
|
|
||||||
static void xf_video_frame_free(xfVideoFrame **pframe)
|
|
||||||
{
|
|
||||||
xfVideoFrame *frame = *pframe;
|
|
||||||
|
|
||||||
xfPresentationContext_unref(frame->presentation);
|
|
||||||
BufferPool_Return(frame->presentation->xfVideo->surfacePool, frame->surfaceData);
|
|
||||||
free(frame);
|
|
||||||
*pframe = NULL;
|
|
||||||
}
|
|
||||||
|
|
||||||
static void xf_video_timer(xfContext *xfc, TimerEventArgs *timer)
|
|
||||||
{
|
|
||||||
xfVideoContext *xfVideo = xfc->xfVideo;
|
|
||||||
xfPresentationContext *presentation;
|
|
||||||
xfVideoFrame *peekFrame, *frame = NULL;
|
|
||||||
|
|
||||||
EnterCriticalSection(&xfVideo->framesLock);
|
|
||||||
do
|
|
||||||
{
|
|
||||||
peekFrame = (xfVideoFrame *)Queue_Peek(xfVideo->frames);
|
|
||||||
if (!peekFrame)
|
|
||||||
break;
|
|
||||||
|
|
||||||
if (peekFrame->publishTime > timer->now)
|
|
||||||
break;
|
|
||||||
|
|
||||||
if (frame)
|
|
||||||
{
|
|
||||||
/* free skipped frame */
|
|
||||||
WLog_DBG(TAG, "dropping frame @%"PRIu64, frame->publishTime);
|
|
||||||
xfVideo->droppedFrames++;
|
|
||||||
xf_video_frame_free(&frame);
|
|
||||||
}
|
|
||||||
frame = peekFrame;
|
|
||||||
Queue_Dequeue(xfVideo->frames);
|
|
||||||
}
|
|
||||||
while (1);
|
|
||||||
LeaveCriticalSection(&xfVideo->framesLock);
|
|
||||||
|
|
||||||
if (!frame)
|
|
||||||
goto treat_feedback;
|
|
||||||
|
|
||||||
presentation = frame->presentation;
|
|
||||||
|
|
||||||
xfVideo->publishedFrames++;
|
|
||||||
memcpy(presentation->surfaceData, frame->surfaceData, frame->w * frame->h * 4);
|
|
||||||
|
|
||||||
XPutImage(xfc->display, xfc->drawable, xfc->gc, presentation->surface,
|
|
||||||
0, 0,
|
0, 0,
|
||||||
frame->x, frame->y, frame->w, frame->h);
|
surface->base.x, surface->base.y, surface->base.w, surface->base.h);
|
||||||
|
return TRUE;
|
||||||
xfPresentationContext_unref(presentation);
|
|
||||||
BufferPool_Return(xfVideo->surfacePool, frame->surfaceData);
|
|
||||||
free(frame);
|
|
||||||
|
|
||||||
treat_feedback:
|
|
||||||
if (xfVideo->nextFeedbackTime < timer->now)
|
|
||||||
{
|
|
||||||
/* we can compute some feedback only if we have some published frames and
|
|
||||||
* a current presentation
|
|
||||||
*/
|
|
||||||
if (xfVideo->publishedFrames && xfVideo->currentPresentation)
|
|
||||||
{
|
|
||||||
UINT32 computedRate;
|
|
||||||
|
|
||||||
InterlockedIncrement(&xfVideo->currentPresentation->refCounter);
|
|
||||||
|
|
||||||
if (xfVideo->droppedFrames)
|
|
||||||
{
|
|
||||||
/**
|
|
||||||
* some dropped frames, looks like we're asking too many frames per seconds,
|
|
||||||
* try lowering rate. We go directly from unlimited rate to 24 frames/seconds
|
|
||||||
* otherwise we lower rate by 2 frames by seconds
|
|
||||||
*/
|
|
||||||
if (xfVideo->lastSentRate == XF_VIDEO_UNLIMITED_RATE)
|
|
||||||
computedRate = 24;
|
|
||||||
else
|
|
||||||
{
|
|
||||||
computedRate = xfVideo->lastSentRate - 2;
|
|
||||||
if (!computedRate)
|
|
||||||
computedRate = 2;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
else
|
|
||||||
{
|
|
||||||
/**
|
|
||||||
* we treat all frames ok, so either ask the server to send more,
|
|
||||||
* or stay unlimited
|
|
||||||
*/
|
|
||||||
if (xfVideo->lastSentRate == XF_VIDEO_UNLIMITED_RATE)
|
|
||||||
computedRate = XF_VIDEO_UNLIMITED_RATE; /* stay unlimited */
|
|
||||||
else
|
|
||||||
{
|
|
||||||
computedRate = xfVideo->lastSentRate + 2;
|
|
||||||
if (computedRate > XF_VIDEO_UNLIMITED_RATE)
|
|
||||||
computedRate = XF_VIDEO_UNLIMITED_RATE;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
if (computedRate != xfVideo->lastSentRate)
|
|
||||||
{
|
|
||||||
TSMM_CLIENT_NOTIFICATION notif;
|
|
||||||
notif.PresentationId = xfVideo->currentPresentation->PresentationId;
|
|
||||||
notif.NotificationType = TSMM_CLIENT_NOTIFICATION_TYPE_FRAMERATE_OVERRIDE;
|
|
||||||
if (computedRate == XF_VIDEO_UNLIMITED_RATE)
|
|
||||||
{
|
|
||||||
notif.FramerateOverride.Flags = 0x01;
|
|
||||||
notif.FramerateOverride.DesiredFrameRate = 0x00;
|
|
||||||
}
|
|
||||||
else
|
|
||||||
{
|
|
||||||
notif.FramerateOverride.Flags = 0x02;
|
|
||||||
notif.FramerateOverride.DesiredFrameRate = computedRate;
|
|
||||||
}
|
|
||||||
|
|
||||||
xfVideo->xfc->video->ClientNotification(xfVideo->xfc->video, ¬if);
|
|
||||||
xfVideo->lastSentRate = computedRate;
|
|
||||||
|
|
||||||
WLog_DBG(TAG, "server notified with rate %d published=%d dropped=%d", xfVideo->lastSentRate,
|
|
||||||
xfVideo->publishedFrames, xfVideo->droppedFrames);
|
|
||||||
}
|
|
||||||
|
|
||||||
xfPresentationContext_unref(xfVideo->currentPresentation);
|
|
||||||
}
|
|
||||||
|
|
||||||
WLog_DBG(TAG, "currentRate=%d published=%d dropped=%d", xfVideo->lastSentRate,
|
|
||||||
xfVideo->publishedFrames, xfVideo->droppedFrames);
|
|
||||||
|
|
||||||
xfVideo->droppedFrames = 0;
|
|
||||||
xfVideo->publishedFrames = 0;
|
|
||||||
xfVideo->nextFeedbackTime = timer->now + 1000;
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static UINT xf_video_VideoData(VideoClientContext* context, TSMM_VIDEO_DATA *data)
|
static BOOL xfVideoDeleteSurface(VideoClientContext *video, xfVideoSurface *surface)
|
||||||
{
|
{
|
||||||
xfContext *xfc = context->custom;
|
XFree(surface->image);
|
||||||
xfVideoContext *xfVideo = xfc->xfVideo;
|
free(surface);
|
||||||
xfPresentationContext *presentation;
|
|
||||||
int status;
|
|
||||||
|
|
||||||
presentation = xfVideo->currentPresentation;
|
return TRUE;
|
||||||
if (!presentation)
|
|
||||||
{
|
|
||||||
WLog_ERR(TAG, "no current presentation");
|
|
||||||
return CHANNEL_RC_OK;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (presentation->PresentationId != data->PresentationId)
|
|
||||||
{
|
|
||||||
WLog_ERR(TAG, "current presentation id=%d doesn't match data id=%d", presentation->PresentationId,
|
|
||||||
data->PresentationId);
|
|
||||||
return CHANNEL_RC_OK;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (!Stream_EnsureRemainingCapacity(presentation->currentSample, data->cbSample))
|
|
||||||
{
|
|
||||||
WLog_ERR(TAG, "unable to expand the current packet");
|
|
||||||
return CHANNEL_RC_NO_MEMORY;
|
|
||||||
}
|
|
||||||
|
|
||||||
Stream_Write(presentation->currentSample, data->pSample, data->cbSample);
|
|
||||||
|
|
||||||
if (data->CurrentPacketIndex == data->PacketsInSample)
|
|
||||||
{
|
|
||||||
H264_CONTEXT *h264 = presentation->h264;
|
|
||||||
UINT64 startTime = GetTickCount64(), timeAfterH264;
|
|
||||||
MAPPED_GEOMETRY *geom = presentation->geometry;
|
|
||||||
|
|
||||||
Stream_SealLength(presentation->currentSample);
|
|
||||||
Stream_SetPosition(presentation->currentSample, 0);
|
|
||||||
|
|
||||||
status = h264->subsystem->Decompress(h264, Stream_Pointer(presentation->currentSample),
|
|
||||||
Stream_Length(presentation->currentSample));
|
|
||||||
if (status == 0)
|
|
||||||
return CHANNEL_RC_OK;
|
|
||||||
|
|
||||||
if (status < 0)
|
|
||||||
return CHANNEL_RC_OK;
|
|
||||||
|
|
||||||
timeAfterH264 = GetTickCount64();
|
|
||||||
if (data->SampleNumber == 1)
|
|
||||||
{
|
|
||||||
presentation->lastPublishTime = startTime;
|
|
||||||
}
|
|
||||||
|
|
||||||
presentation->lastPublishTime += (data->hnsDuration / 10000);
|
|
||||||
if (presentation->lastPublishTime <= timeAfterH264 + 10)
|
|
||||||
{
|
|
||||||
int dropped = 0;
|
|
||||||
|
|
||||||
/* if the frame is to be published in less than 10 ms, let's consider it's now */
|
|
||||||
yuv_to_rgb(presentation, presentation->surfaceData);
|
|
||||||
|
|
||||||
XPutImage(xfc->display, xfc->drawable, xfc->gc, presentation->surface,
|
|
||||||
0, 0,
|
|
||||||
geom->topLevelLeft + geom->left + geom->geometry.boundingRect.x,
|
|
||||||
geom->topLevelTop + geom->top + geom->geometry.boundingRect.y,
|
|
||||||
presentation->SourceWidth, presentation->SourceHeight);
|
|
||||||
|
|
||||||
xfVideo->publishedFrames++;
|
|
||||||
|
|
||||||
/* cleanup previously scheduled frames */
|
|
||||||
EnterCriticalSection(&xfVideo->framesLock);
|
|
||||||
while (Queue_Count(xfVideo->frames) > 0)
|
|
||||||
{
|
|
||||||
xfVideoFrame *frame = Queue_Dequeue(xfVideo->frames);
|
|
||||||
if (frame)
|
|
||||||
{
|
|
||||||
xfVideo->droppedFrames++;
|
|
||||||
xf_video_frame_free(&frame);
|
|
||||||
dropped++;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
LeaveCriticalSection(&xfVideo->framesLock);
|
|
||||||
|
|
||||||
if (dropped)
|
|
||||||
WLog_DBG(TAG, "showing frame (%d dropped)", dropped);
|
|
||||||
}
|
|
||||||
else
|
|
||||||
{
|
|
||||||
BOOL enqueueResult;
|
|
||||||
xfVideoFrame *frame = calloc(1, sizeof(*frame));
|
|
||||||
if (!frame)
|
|
||||||
{
|
|
||||||
WLog_ERR(TAG, "unable to create frame");
|
|
||||||
return CHANNEL_RC_NO_MEMORY;
|
|
||||||
}
|
|
||||||
|
|
||||||
frame->presentation = presentation;
|
|
||||||
frame->publishTime = presentation->lastPublishTime;
|
|
||||||
frame->x = geom->topLevelLeft + geom->left + geom->geometry.boundingRect.x;
|
|
||||||
frame->y = geom->topLevelTop + geom->top + geom->geometry.boundingRect.y;
|
|
||||||
frame->w = presentation->SourceWidth;
|
|
||||||
frame->h = presentation->SourceHeight;
|
|
||||||
|
|
||||||
frame->surfaceData = BufferPool_Take(xfVideo->surfacePool, frame->w * frame->h * 4);
|
|
||||||
if (!frame->surfaceData)
|
|
||||||
{
|
|
||||||
WLog_ERR(TAG, "unable to allocate frame data");
|
|
||||||
free(frame);
|
|
||||||
return CHANNEL_RC_NO_MEMORY;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (!yuv_to_rgb(presentation, frame->surfaceData))
|
|
||||||
{
|
|
||||||
WLog_ERR(TAG, "error during YUV->RGB conversion");
|
|
||||||
BufferPool_Return(xfVideo->surfacePool, frame->surfaceData);
|
|
||||||
free(frame);
|
|
||||||
return CHANNEL_RC_NO_MEMORY;
|
|
||||||
}
|
|
||||||
|
|
||||||
InterlockedIncrement(&presentation->refCounter);
|
|
||||||
|
|
||||||
EnterCriticalSection(&xfVideo->framesLock);
|
|
||||||
enqueueResult = Queue_Enqueue(xfVideo->frames, frame);
|
|
||||||
LeaveCriticalSection(&xfVideo->framesLock);
|
|
||||||
|
|
||||||
if (!enqueueResult)
|
|
||||||
{
|
|
||||||
WLog_ERR(TAG, "unable to enqueue frame");
|
|
||||||
xf_video_frame_free(&frame);
|
|
||||||
return CHANNEL_RC_NO_MEMORY;
|
|
||||||
}
|
|
||||||
|
|
||||||
WLog_DBG(TAG, "scheduling frame in %"PRIu32" ms", (frame->publishTime-startTime));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return CHANNEL_RC_OK;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
void xf_video_free(xfVideoContext *xfVideo)
|
|
||||||
{
|
|
||||||
EnterCriticalSection(&xfVideo->framesLock);
|
|
||||||
while (Queue_Count(xfVideo->frames))
|
|
||||||
{
|
|
||||||
xfVideoFrame *frame = Queue_Dequeue(xfVideo->frames);
|
|
||||||
if (frame)
|
|
||||||
xf_video_frame_free(&frame);
|
|
||||||
}
|
|
||||||
|
|
||||||
Queue_Free(xfVideo->frames);
|
|
||||||
LeaveCriticalSection(&xfVideo->framesLock);
|
|
||||||
|
|
||||||
DeleteCriticalSection(&xfVideo->framesLock);
|
|
||||||
|
|
||||||
if (xfVideo->currentPresentation)
|
|
||||||
xfPresentationContext_unref(xfVideo->currentPresentation);
|
|
||||||
|
|
||||||
BufferPool_Free(xfVideo->surfacePool);
|
|
||||||
free(xfVideo);
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
void xf_video_control_init(xfContext *xfc, VideoClientContext *video)
|
void xf_video_control_init(xfContext *xfc, VideoClientContext *video)
|
||||||
{
|
{
|
||||||
xfc->video = video;
|
xfc->video = video;
|
||||||
video->custom = xfc;
|
video->custom = xfc;
|
||||||
video->PresentationRequest = xf_video_PresentationRequest;
|
|
||||||
|
video->createSurface = xfVideoCreateSurface;
|
||||||
|
video->showSurface = (pcVideoShowSurface)xfVideoShowSurface;
|
||||||
|
video->deleteSurface = (pcVideoDeleteSurface)xfVideoDeleteSurface;
|
||||||
|
video->setGeometry(video, xfc->geometry);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
void xf_video_control_uninit(xfContext *xfc, VideoClientContext *video)
|
void xf_video_control_uninit(xfContext *xfc, VideoClientContext *video)
|
||||||
{
|
{
|
||||||
video->VideoData = NULL;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void xf_video_timer(xfContext *xfc, TimerEventArgs *timer)
|
||||||
|
{
|
||||||
|
xfc->video->timer(xfc->video, timer->now);
|
||||||
|
}
|
||||||
|
|
||||||
void xf_video_data_init(xfContext *xfc, VideoClientContext *video)
|
void xf_video_data_init(xfContext *xfc, VideoClientContext *video)
|
||||||
{
|
{
|
||||||
video->VideoData = xf_video_VideoData;
|
|
||||||
PubSub_SubscribeTimer(xfc->context.pubSub, (pTimerEventHandler)xf_video_timer);
|
PubSub_SubscribeTimer(xfc->context.pubSub, (pTimerEventHandler)xf_video_timer);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -171,6 +171,7 @@ static COMMAND_LINE_ARGUMENT_A args[] =
|
||||||
{ "v", COMMAND_LINE_VALUE_REQUIRED, "<server>[:port]", NULL, NULL, -1, NULL, "Server hostname" },
|
{ "v", COMMAND_LINE_VALUE_REQUIRED, "<server>[:port]", NULL, NULL, -1, NULL, "Server hostname" },
|
||||||
{ "vc", COMMAND_LINE_VALUE_REQUIRED, "<channel>[,<options>]", NULL, NULL, -1, NULL, "Static virtual channel" },
|
{ "vc", COMMAND_LINE_VALUE_REQUIRED, "<channel>[,<options>]", NULL, NULL, -1, NULL, "Static virtual channel" },
|
||||||
{ "version", COMMAND_LINE_VALUE_FLAG | COMMAND_LINE_PRINT_VERSION, NULL, NULL, NULL, -1, NULL, "Print version" },
|
{ "version", COMMAND_LINE_VALUE_FLAG | COMMAND_LINE_PRINT_VERSION, NULL, NULL, NULL, -1, NULL, "Print version" },
|
||||||
|
{ "video", COMMAND_LINE_VALUE_FLAG, NULL, NULL, NULL, -1, NULL, "Video optimized remoting channel" },
|
||||||
{ "vmconnect", COMMAND_LINE_VALUE_OPTIONAL, "<vmid>", NULL, NULL, -1, NULL, "Hyper-V console (use port 2179, disable negotiation)" },
|
{ "vmconnect", COMMAND_LINE_VALUE_OPTIONAL, "<vmid>", NULL, NULL, -1, NULL, "Hyper-V console (use port 2179, disable negotiation)" },
|
||||||
{ "w", COMMAND_LINE_VALUE_REQUIRED, "<width>", "1024", NULL, -1, NULL, "Width" },
|
{ "w", COMMAND_LINE_VALUE_REQUIRED, "<width>", "1024", NULL, -1, NULL, "Width" },
|
||||||
{ "wallpaper", COMMAND_LINE_VALUE_BOOL, NULL, BoolValueTrue, NULL, -1, NULL, "Enable wallpaper" },
|
{ "wallpaper", COMMAND_LINE_VALUE_BOOL, NULL, BoolValueTrue, NULL, -1, NULL, "Enable wallpaper" },
|
||||||
|
|
|
@ -2,7 +2,7 @@
|
||||||
* FreeRDP: A Remote Desktop Protocol Implementation
|
* FreeRDP: A Remote Desktop Protocol Implementation
|
||||||
* Video Optimized Remoting Virtual Channel Extension
|
* Video Optimized Remoting Virtual Channel Extension
|
||||||
*
|
*
|
||||||
* Copyright 2017 David Fort <contact@hardening-consulting.com>
|
* Copyright 2018 David Fort <contact@hardening-consulting.com>
|
||||||
*
|
*
|
||||||
* Licensed under the Apache License, Version 2.0 (the "License");
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
* you may not use this file except in compliance with the License.
|
* you may not use this file except in compliance with the License.
|
||||||
|
|
|
@ -28,12 +28,12 @@
|
||||||
*/
|
*/
|
||||||
typedef struct _geometry_client_context GeometryClientContext;
|
typedef struct _geometry_client_context GeometryClientContext;
|
||||||
|
|
||||||
|
|
||||||
typedef struct _MAPPED_GEOMETRY MAPPED_GEOMETRY;
|
typedef struct _MAPPED_GEOMETRY MAPPED_GEOMETRY;
|
||||||
typedef BOOL (*pcMappedGeometryAdded)(GeometryClientContext* context, MAPPED_GEOMETRY *geometry);
|
typedef BOOL (*pcMappedGeometryAdded)(GeometryClientContext* context, MAPPED_GEOMETRY *geometry);
|
||||||
typedef BOOL (*pcMappedGeometryUpdate)(MAPPED_GEOMETRY *geometry);
|
typedef BOOL (*pcMappedGeometryUpdate)(MAPPED_GEOMETRY *geometry);
|
||||||
typedef BOOL (*pcMappedGeometryClear)(MAPPED_GEOMETRY *geometry);
|
typedef BOOL (*pcMappedGeometryClear)(MAPPED_GEOMETRY *geometry);
|
||||||
|
|
||||||
|
/** @brief a geometry record tracked by the geometry channel */
|
||||||
struct _MAPPED_GEOMETRY
|
struct _MAPPED_GEOMETRY
|
||||||
{
|
{
|
||||||
UINT64 mappingId;
|
UINT64 mappingId;
|
||||||
|
@ -47,6 +47,7 @@ struct _MAPPED_GEOMETRY
|
||||||
pcMappedGeometryClear MappedGeometryClear;
|
pcMappedGeometryClear MappedGeometryClear;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
/** @brief the geometry context for client channel */
|
||||||
struct _geometry_client_context
|
struct _geometry_client_context
|
||||||
{
|
{
|
||||||
wHashTable *geometries;
|
wHashTable *geometries;
|
||||||
|
|
|
@ -20,26 +20,42 @@
|
||||||
#ifndef FREERDP_CHANNELS_CLIENT_VIDEO_H
|
#ifndef FREERDP_CHANNELS_CLIENT_VIDEO_H
|
||||||
#define FREERDP_CHANNELS_CLIENT_VIDEO_H
|
#define FREERDP_CHANNELS_CLIENT_VIDEO_H
|
||||||
|
|
||||||
|
#include <freerdp/client/geometry.h>
|
||||||
#include <freerdp/channels/video.h>
|
#include <freerdp/channels/video.h>
|
||||||
|
|
||||||
typedef struct _video_client_context VideoClientContext;
|
typedef struct _VideoClientContext VideoClientContext;
|
||||||
|
typedef struct _VideoClientContextPriv VideoClientContextPriv;
|
||||||
|
typedef struct _VideoSurface VideoSurface;
|
||||||
|
|
||||||
typedef UINT (*pcPresentationRequest)(VideoClientContext* context, TSMM_PRESENTATION_REQUEST *req);
|
|
||||||
typedef UINT (*pcPresentationResponse)(VideoClientContext* context, TSMM_PRESENTATION_RESPONSE *resp);
|
/** @brief an implementation of surface used by the video channel */
|
||||||
typedef UINT (*pcVideoClientNotification)(VideoClientContext *context, TSMM_CLIENT_NOTIFICATION *notif);
|
struct _VideoSurface
|
||||||
typedef UINT (*pcVideoData)(VideoClientContext* context, TSMM_VIDEO_DATA *data);
|
{
|
||||||
|
UINT32 x, y, w, h;
|
||||||
|
BYTE *data;
|
||||||
|
};
|
||||||
|
|
||||||
|
typedef void (*pcVideoTimer)(VideoClientContext *video, UINT64 now);
|
||||||
|
typedef void (*pcVideoSetGeometry)(VideoClientContext *video, GeometryClientContext *geometry);
|
||||||
|
typedef VideoSurface *(*pcVideoCreateSurface)(VideoClientContext *video, BYTE *data, UINT32 x, UINT32 y,
|
||||||
|
UINT32 width, UINT32 height);
|
||||||
|
typedef BOOL (*pcVideoShowSurface)(VideoClientContext *video, VideoSurface *surface);
|
||||||
|
typedef BOOL (*pcVideoDeleteSurface)(VideoClientContext *video, VideoSurface *surface);
|
||||||
|
|
||||||
/** @brief context for the video (MS-RDPEVOR) channel */
|
/** @brief context for the video (MS-RDPEVOR) channel */
|
||||||
struct _video_client_context
|
struct _VideoClientContext
|
||||||
{
|
{
|
||||||
void* handle;
|
void* handle;
|
||||||
void* custom;
|
void* custom;
|
||||||
|
VideoClientContextPriv *priv;
|
||||||
|
|
||||||
pcPresentationRequest PresentationRequest;
|
pcVideoSetGeometry setGeometry;
|
||||||
pcPresentationResponse PresentationResponse;
|
pcVideoTimer timer;
|
||||||
pcVideoClientNotification ClientNotification;
|
pcVideoCreateSurface createSurface;
|
||||||
pcVideoData VideoData;
|
pcVideoShowSurface showSurface;
|
||||||
|
pcVideoDeleteSurface deleteSurface;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
||||||
#endif /* FREERDP_CHANNELS_CLIENT_VIDEO_H */
|
#endif /* FREERDP_CHANNELS_CLIENT_VIDEO_H */
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue