1346 lines
39 KiB
C
1346 lines
39 KiB
C
/*
|
|
* abstract_jb: common implementation-independent jitterbuffer stuff
|
|
*
|
|
* Copyright (C) 2005, Attractel OOD
|
|
*
|
|
* Contributors:
|
|
* Slav Klenov <slav@securax.org>
|
|
*
|
|
* See http://www.asterisk.org for more information about
|
|
* the Asterisk project. Please do not directly contact
|
|
* any of the maintainers of this project for assistance;
|
|
* the project provides a web site, mailing lists and IRC
|
|
* channels for your use.
|
|
*
|
|
* This program is free software, distributed under the terms of
|
|
* the GNU General Public License Version 2. See the LICENSE file
|
|
* at the top of the source tree.
|
|
*
|
|
* A license has been granted to Digium (via disclaimer) for the use of
|
|
* this code.
|
|
*/
|
|
|
|
/*! \file
|
|
*
|
|
* \brief Common implementation-independent jitterbuffer stuff.
|
|
*
|
|
* \author Slav Klenov <slav@securax.org>
|
|
*
|
|
*
|
|
*/
|
|
|
|
/*** MODULEINFO
|
|
<support_level>core</support_level>
|
|
***/
|
|
|
|
#include "asterisk.h"
|
|
|
|
#include "asterisk/frame.h"
|
|
#include "asterisk/channel.h"
|
|
#include "asterisk/term.h"
|
|
#include "asterisk/utils.h"
|
|
#include "asterisk/pbx.h"
|
|
#include "asterisk/timing.h"
|
|
#include "asterisk/rtp_engine.h"
|
|
#include "asterisk/format_cache.h"
|
|
|
|
#include "asterisk/abstract_jb.h"
|
|
#include "fixedjitterbuf.h"
|
|
#include "jitterbuf.h"
|
|
|
|
/*! Internal jb flags */
|
|
enum {
|
|
JB_USE = (1 << 0),
|
|
JB_TIMEBASE_INITIALIZED = (1 << 1),
|
|
JB_CREATED = (1 << 2)
|
|
};
|
|
|
|
/*! The maximum size we allow the early frame buffer to get */
|
|
#define MAXIMUM_EARLY_FRAME_COUNT 200
|
|
|
|
|
|
/* Implementation functions */
|
|
/* fixed */
|
|
static void *jb_create_fixed(struct ast_jb_conf *general_config);
|
|
static void jb_destroy_fixed(void *jb);
|
|
static int jb_put_first_fixed(void *jb, struct ast_frame *fin, long now);
|
|
static int jb_put_fixed(void *jb, struct ast_frame *fin, long now);
|
|
static int jb_get_fixed(void *jb, struct ast_frame **fout, long now, long interpl);
|
|
static long jb_next_fixed(void *jb);
|
|
static int jb_remove_fixed(void *jb, struct ast_frame **fout);
|
|
static void jb_force_resynch_fixed(void *jb);
|
|
static void jb_empty_and_reset_fixed(void *jb);
|
|
static int jb_is_late_fixed(void *jb, long ts);
|
|
/* adaptive */
|
|
static void * jb_create_adaptive(struct ast_jb_conf *general_config);
|
|
static void jb_destroy_adaptive(void *jb);
|
|
static int jb_put_first_adaptive(void *jb, struct ast_frame *fin, long now);
|
|
static int jb_put_adaptive(void *jb, struct ast_frame *fin, long now);
|
|
static int jb_get_adaptive(void *jb, struct ast_frame **fout, long now, long interpl);
|
|
static long jb_next_adaptive(void *jb);
|
|
static int jb_remove_adaptive(void *jb, struct ast_frame **fout);
|
|
static void jb_force_resynch_adaptive(void *jb);
|
|
static void jb_empty_and_reset_adaptive(void *jb);
|
|
static int jb_is_late_adaptive(void *jb, long ts);
|
|
|
|
/* Available jb implementations */
|
|
static const struct ast_jb_impl avail_impl[] = {
|
|
{
|
|
.name = "fixed",
|
|
.type = AST_JB_FIXED,
|
|
.create = jb_create_fixed,
|
|
.destroy = jb_destroy_fixed,
|
|
.put_first = jb_put_first_fixed,
|
|
.put = jb_put_fixed,
|
|
.get = jb_get_fixed,
|
|
.next = jb_next_fixed,
|
|
.remove = jb_remove_fixed,
|
|
.force_resync = jb_force_resynch_fixed,
|
|
.empty_and_reset = jb_empty_and_reset_fixed,
|
|
.is_late = jb_is_late_fixed,
|
|
},
|
|
{
|
|
.name = "adaptive",
|
|
.type = AST_JB_ADAPTIVE,
|
|
.create = jb_create_adaptive,
|
|
.destroy = jb_destroy_adaptive,
|
|
.put_first = jb_put_first_adaptive,
|
|
.put = jb_put_adaptive,
|
|
.get = jb_get_adaptive,
|
|
.next = jb_next_adaptive,
|
|
.remove = jb_remove_adaptive,
|
|
.force_resync = jb_force_resynch_adaptive,
|
|
.empty_and_reset = jb_empty_and_reset_adaptive,
|
|
.is_late = jb_is_late_adaptive,
|
|
}
|
|
};
|
|
|
|
static int default_impl = 0;
|
|
|
|
/* Translations between impl and abstract return codes */
|
|
static const int fixed_to_abstract_code[] =
|
|
{AST_JB_IMPL_OK, AST_JB_IMPL_DROP, AST_JB_IMPL_INTERP, AST_JB_IMPL_NOFRAME};
|
|
static const int adaptive_to_abstract_code[] =
|
|
{AST_JB_IMPL_OK, AST_JB_IMPL_NOFRAME, AST_JB_IMPL_NOFRAME, AST_JB_IMPL_INTERP, AST_JB_IMPL_DROP, AST_JB_IMPL_OK};
|
|
|
|
/* JB_GET actions (used only for the frames log) */
|
|
static const char * const jb_get_actions[] = {"Delivered", "Dropped", "Interpolated", "No"};
|
|
|
|
/*! \brief Macros for the frame log files */
|
|
#define jb_framelog(...) do { \
|
|
if (jb->logfile) { \
|
|
fprintf(jb->logfile, __VA_ARGS__); \
|
|
fflush(jb->logfile); \
|
|
} \
|
|
} while (0)
|
|
|
|
|
|
/* Internal utility functions */
|
|
static void jb_choose_impl(struct ast_channel *chan);
|
|
static void jb_get_and_deliver(struct ast_channel *chan);
|
|
static int create_jb(struct ast_channel *chan, struct ast_frame *first_frame);
|
|
static long get_now(struct ast_jb *jb, struct timeval *tv);
|
|
|
|
|
|
/* Interface ast jb functions impl */
|
|
|
|
|
|
static void jb_choose_impl(struct ast_channel *chan)
|
|
{
|
|
struct ast_jb *jb = ast_channel_jb(chan);
|
|
struct ast_jb_conf *jbconf = &jb->conf;
|
|
const struct ast_jb_impl *test_impl;
|
|
int i, avail_impl_count = ARRAY_LEN(avail_impl);
|
|
|
|
jb->impl = &avail_impl[default_impl];
|
|
|
|
if (ast_strlen_zero(jbconf->impl)) {
|
|
return;
|
|
}
|
|
|
|
for (i = 0; i < avail_impl_count; i++) {
|
|
test_impl = &avail_impl[i];
|
|
if (!strcasecmp(jbconf->impl, test_impl->name)) {
|
|
jb->impl = test_impl;
|
|
return;
|
|
}
|
|
}
|
|
}
|
|
|
|
int ast_jb_do_usecheck(struct ast_channel *c0, struct ast_channel *c1)
|
|
{
|
|
struct ast_jb *jb0 = ast_channel_jb(c0);
|
|
struct ast_jb *jb1 = ast_channel_jb(c1);
|
|
struct ast_jb_conf *conf0 = &jb0->conf;
|
|
struct ast_jb_conf *conf1 = &jb1->conf;
|
|
int c0_wants_jitter = ast_channel_tech(c0)->properties & AST_CHAN_TP_WANTSJITTER;
|
|
int c0_creates_jitter = ast_channel_tech(c0)->properties & AST_CHAN_TP_CREATESJITTER;
|
|
int c0_jb_enabled = ast_test_flag(conf0, AST_JB_ENABLED);
|
|
int c0_force_jb = ast_test_flag(conf0, AST_JB_FORCED);
|
|
int c0_jb_timebase_initialized = ast_test_flag(jb0, JB_TIMEBASE_INITIALIZED);
|
|
int c0_jb_created = ast_test_flag(jb0, JB_CREATED);
|
|
int c1_wants_jitter = ast_channel_tech(c1)->properties & AST_CHAN_TP_WANTSJITTER;
|
|
int c1_creates_jitter = ast_channel_tech(c1)->properties & AST_CHAN_TP_CREATESJITTER;
|
|
int c1_jb_enabled = ast_test_flag(conf1, AST_JB_ENABLED);
|
|
int c1_force_jb = ast_test_flag(conf1, AST_JB_FORCED);
|
|
int c1_jb_timebase_initialized = ast_test_flag(jb1, JB_TIMEBASE_INITIALIZED);
|
|
int c1_jb_created = ast_test_flag(jb1, JB_CREATED);
|
|
int inuse = 0;
|
|
|
|
/* Determine whether audio going to c0 needs a jitter buffer */
|
|
if (((!c0_wants_jitter && c1_creates_jitter) || (c0_force_jb && c1_creates_jitter)) && c0_jb_enabled) {
|
|
ast_set_flag(jb0, JB_USE);
|
|
if (!c0_jb_timebase_initialized) {
|
|
if (c1_jb_timebase_initialized) {
|
|
memcpy(&jb0->timebase, &jb1->timebase, sizeof(struct timeval));
|
|
} else {
|
|
gettimeofday(&jb0->timebase, NULL);
|
|
}
|
|
ast_set_flag(jb0, JB_TIMEBASE_INITIALIZED);
|
|
}
|
|
|
|
if (!c0_jb_created) {
|
|
jb_choose_impl(c0);
|
|
}
|
|
|
|
inuse = 1;
|
|
}
|
|
|
|
/* Determine whether audio going to c1 needs a jitter buffer */
|
|
if (((!c1_wants_jitter && c0_creates_jitter) || (c1_force_jb && c0_creates_jitter)) && c1_jb_enabled) {
|
|
ast_set_flag(jb1, JB_USE);
|
|
if (!c1_jb_timebase_initialized) {
|
|
if (c0_jb_timebase_initialized) {
|
|
memcpy(&jb1->timebase, &jb0->timebase, sizeof(struct timeval));
|
|
} else {
|
|
gettimeofday(&jb1->timebase, NULL);
|
|
}
|
|
ast_set_flag(jb1, JB_TIMEBASE_INITIALIZED);
|
|
}
|
|
|
|
if (!c1_jb_created) {
|
|
jb_choose_impl(c1);
|
|
}
|
|
|
|
inuse = 1;
|
|
}
|
|
|
|
return inuse;
|
|
}
|
|
|
|
int ast_jb_get_when_to_wakeup(struct ast_channel *c0, struct ast_channel *c1, int time_left)
|
|
{
|
|
struct ast_jb *jb0 = ast_channel_jb(c0);
|
|
struct ast_jb *jb1 = ast_channel_jb(c1);
|
|
int c0_use_jb = ast_test_flag(jb0, JB_USE);
|
|
int c0_jb_is_created = ast_test_flag(jb0, JB_CREATED);
|
|
int c1_use_jb = ast_test_flag(jb1, JB_USE);
|
|
int c1_jb_is_created = ast_test_flag(jb1, JB_CREATED);
|
|
int wait, wait0, wait1;
|
|
struct timeval tv_now;
|
|
|
|
if (time_left == 0) {
|
|
/* No time left - the bridge will be retried */
|
|
/* TODO: Test disable this */
|
|
/*return 0;*/
|
|
}
|
|
|
|
if (time_left < 0) {
|
|
time_left = INT_MAX;
|
|
}
|
|
|
|
gettimeofday(&tv_now, NULL);
|
|
|
|
wait0 = (c0_use_jb && c0_jb_is_created) ? jb0->next - get_now(jb0, &tv_now) : time_left;
|
|
wait1 = (c1_use_jb && c1_jb_is_created) ? jb1->next - get_now(jb1, &tv_now) : time_left;
|
|
|
|
wait = wait0 < wait1 ? wait0 : wait1;
|
|
wait = wait < time_left ? wait : time_left;
|
|
|
|
if (wait == INT_MAX) {
|
|
wait = -1;
|
|
} else if (wait < 1) {
|
|
/* don't let wait=0, because this can cause the pbx thread to loop without any sleeping at all */
|
|
wait = 1;
|
|
}
|
|
|
|
return wait;
|
|
}
|
|
|
|
|
|
int ast_jb_put(struct ast_channel *chan, struct ast_frame *f)
|
|
{
|
|
struct ast_jb *jb = ast_channel_jb(chan);
|
|
const struct ast_jb_impl *jbimpl = jb->impl;
|
|
void *jbobj = jb->jbobj;
|
|
struct ast_frame *frr;
|
|
long now = 0;
|
|
|
|
if (!ast_test_flag(jb, JB_USE))
|
|
return -1;
|
|
|
|
if (f->frametype != AST_FRAME_VOICE) {
|
|
if (f->frametype == AST_FRAME_DTMF && ast_test_flag(jb, JB_CREATED)) {
|
|
jb_framelog("JB_PUT {now=%ld}: Received DTMF frame. Force resynching jb...\n", now);
|
|
jbimpl->force_resync(jbobj);
|
|
}
|
|
|
|
return -1;
|
|
}
|
|
|
|
/* We consider an enabled jitterbuffer should receive frames with valid timing info. */
|
|
if (!ast_test_flag(f, AST_FRFLAG_HAS_TIMING_INFO) || f->len < 2 || f->ts < 0) {
|
|
ast_log(LOG_WARNING, "%s received frame with invalid timing info: "
|
|
"has_timing_info=%u, len=%ld, ts=%ld, src=%s\n",
|
|
ast_channel_name(chan), ast_test_flag(f, AST_FRFLAG_HAS_TIMING_INFO), f->len, f->ts, f->src);
|
|
return -1;
|
|
}
|
|
|
|
frr = ast_frdup(f);
|
|
|
|
if (!frr) {
|
|
ast_log(LOG_ERROR, "Failed to isolate frame for the jitterbuffer on channel '%s'\n", ast_channel_name(chan));
|
|
return -1;
|
|
}
|
|
|
|
if (!ast_test_flag(jb, JB_CREATED)) {
|
|
if (create_jb(chan, frr)) {
|
|
ast_frfree(frr);
|
|
/* Disable the jitterbuffer */
|
|
ast_clear_flag(jb, JB_USE);
|
|
return -1;
|
|
}
|
|
|
|
ast_set_flag(jb, JB_CREATED);
|
|
return 0;
|
|
} else {
|
|
now = get_now(jb, NULL);
|
|
if (jbimpl->put(jbobj, frr, now) != AST_JB_IMPL_OK) {
|
|
jb_framelog("JB_PUT {now=%ld}: Dropped frame with ts=%ld and len=%ld\n", now, frr->ts, frr->len);
|
|
ast_frfree(frr);
|
|
/*return -1;*/
|
|
/* TODO: Check this fix - should return 0 here, because the dropped frame shouldn't
|
|
be delivered at all */
|
|
return 0;
|
|
}
|
|
|
|
jb->next = jbimpl->next(jbobj);
|
|
|
|
jb_framelog("JB_PUT {now=%ld}: Queued frame with ts=%ld and len=%ld\n", now, frr->ts, frr->len);
|
|
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
|
|
void ast_jb_get_and_deliver(struct ast_channel *c0, struct ast_channel *c1)
|
|
{
|
|
struct ast_jb *jb0 = ast_channel_jb(c0);
|
|
struct ast_jb *jb1 = ast_channel_jb(c1);
|
|
int c0_use_jb = ast_test_flag(jb0, JB_USE);
|
|
int c0_jb_is_created = ast_test_flag(jb0, JB_CREATED);
|
|
int c1_use_jb = ast_test_flag(jb1, JB_USE);
|
|
int c1_jb_is_created = ast_test_flag(jb1, JB_CREATED);
|
|
|
|
if (c0_use_jb && c0_jb_is_created)
|
|
jb_get_and_deliver(c0);
|
|
|
|
if (c1_use_jb && c1_jb_is_created)
|
|
jb_get_and_deliver(c1);
|
|
}
|
|
|
|
|
|
static void jb_get_and_deliver(struct ast_channel *chan)
|
|
{
|
|
struct ast_jb *jb = ast_channel_jb(chan);
|
|
const struct ast_jb_impl *jbimpl = jb->impl;
|
|
void *jbobj = jb->jbobj;
|
|
struct ast_frame *f, finterp = { .frametype = AST_FRAME_VOICE, };
|
|
long now;
|
|
int interpolation_len, res;
|
|
|
|
now = get_now(jb, NULL);
|
|
jb->next = jbimpl->next(jbobj);
|
|
if (now < jb->next) {
|
|
jb_framelog("\tJB_GET {now=%ld}: now < next=%ld\n", now, jb->next);
|
|
return;
|
|
}
|
|
|
|
while (now >= jb->next) {
|
|
interpolation_len = ast_format_get_default_ms(jb->last_format);
|
|
|
|
res = jbimpl->get(jbobj, &f, now, interpolation_len);
|
|
|
|
switch (res) {
|
|
case AST_JB_IMPL_OK:
|
|
/* deliver the frame */
|
|
ast_write(chan, f);
|
|
case AST_JB_IMPL_DROP:
|
|
jb_framelog("\tJB_GET {now=%ld}: %s frame with ts=%ld and len=%ld\n",
|
|
now, jb_get_actions[res], f->ts, f->len);
|
|
ao2_replace(jb->last_format, f->subclass.format);
|
|
ast_frfree(f);
|
|
break;
|
|
case AST_JB_IMPL_INTERP:
|
|
/* interpolate a frame */
|
|
f = &finterp;
|
|
f->subclass.format = jb->last_format;
|
|
f->samples = interpolation_len * 8;
|
|
f->src = "JB interpolation";
|
|
f->delivery = ast_tvadd(jb->timebase, ast_samp2tv(jb->next, 1000));
|
|
f->offset = AST_FRIENDLY_OFFSET;
|
|
/* deliver the interpolated frame */
|
|
ast_write(chan, f);
|
|
jb_framelog("\tJB_GET {now=%ld}: Interpolated frame with len=%d\n", now, interpolation_len);
|
|
break;
|
|
case AST_JB_IMPL_NOFRAME:
|
|
ast_log(LOG_WARNING,
|
|
"AST_JB_IMPL_NOFRAME is returned from the %s jb when now=%ld >= next=%ld, jbnext=%ld!\n",
|
|
jbimpl->name, now, jb->next, jbimpl->next(jbobj));
|
|
jb_framelog("\tJB_GET {now=%ld}: No frame for now!?\n", now);
|
|
return;
|
|
default:
|
|
ast_log(LOG_ERROR, "This should never happen!\n");
|
|
ast_assert("JB type unknown" == NULL);
|
|
break;
|
|
}
|
|
|
|
jb->next = jbimpl->next(jbobj);
|
|
}
|
|
}
|
|
|
|
|
|
static int create_jb(struct ast_channel *chan, struct ast_frame *frr)
|
|
{
|
|
struct ast_jb *jb = ast_channel_jb(chan);
|
|
struct ast_jb_conf *jbconf = &jb->conf;
|
|
const struct ast_jb_impl *jbimpl = jb->impl;
|
|
void *jbobj;
|
|
long now;
|
|
char logfile_pathname[20 + AST_JB_IMPL_NAME_SIZE + 2*AST_CHANNEL_NAME + 1];
|
|
char name1[AST_CHANNEL_NAME], name2[AST_CHANNEL_NAME], *tmp;
|
|
int res;
|
|
|
|
jbobj = jb->jbobj = jbimpl->create(jbconf);
|
|
if (!jbobj) {
|
|
ast_log(LOG_WARNING, "Failed to create jitterbuffer on channel '%s'\n", ast_channel_name(chan));
|
|
return -1;
|
|
}
|
|
|
|
now = get_now(jb, NULL);
|
|
res = jbimpl->put_first(jbobj, frr, now);
|
|
|
|
/* The result of putting the first frame should not differ from OK. However, its possible
|
|
some implementations (i.e. adaptive's when resynch_threshold is specified) to drop it. */
|
|
if (res != AST_JB_IMPL_OK) {
|
|
ast_log(LOG_WARNING, "Failed to put first frame in the jitterbuffer on channel '%s'\n", ast_channel_name(chan));
|
|
/*
|
|
jbimpl->destroy(jbobj);
|
|
return -1;
|
|
*/
|
|
}
|
|
|
|
/* Init next */
|
|
jb->next = jbimpl->next(jbobj);
|
|
|
|
/* Init last format for a first time. */
|
|
jb->last_format = ao2_bump(frr->subclass.format);
|
|
|
|
/* Create a frame log file */
|
|
if (ast_test_flag(jbconf, AST_JB_LOG)) {
|
|
struct ast_channel *bridged = ast_channel_bridge_peer(chan);
|
|
char safe_logfile[30] = "/tmp/logfile-XXXXXX";
|
|
int safe_fd;
|
|
|
|
snprintf(name2, sizeof(name2), "%s", ast_channel_name(chan));
|
|
while ((tmp = strchr(name2, '/'))) {
|
|
*tmp = '#';
|
|
}
|
|
|
|
/* We should always have bridged chan if a jitterbuffer is in use */
|
|
ast_assert(bridged != NULL);
|
|
|
|
snprintf(name1, sizeof(name1), "%s", ast_channel_name(bridged));
|
|
while ((tmp = strchr(name1, '/'))) {
|
|
*tmp = '#';
|
|
}
|
|
|
|
snprintf(logfile_pathname, sizeof(logfile_pathname),
|
|
"/tmp/ast_%s_jb_%s--%s.log", jbimpl->name, name1, name2);
|
|
unlink(logfile_pathname);
|
|
safe_fd = mkstemp(safe_logfile);
|
|
if (safe_fd < 0 || link(safe_logfile, logfile_pathname) || unlink(safe_logfile) || !(jb->logfile = fdopen(safe_fd, "w+b"))) {
|
|
ast_log(LOG_ERROR, "Failed to create frame log file with pathname '%s': %s\n", logfile_pathname, strerror(errno));
|
|
jb->logfile = NULL;
|
|
if (safe_fd > -1) {
|
|
close(safe_fd);
|
|
}
|
|
}
|
|
|
|
if (res == AST_JB_IMPL_OK) {
|
|
jb_framelog("JB_PUT_FIRST {now=%ld}: Queued frame with ts=%ld and len=%ld\n",
|
|
now, frr->ts, frr->len);
|
|
} else {
|
|
jb_framelog("JB_PUT_FIRST {now=%ld}: Dropped frame with ts=%ld and len=%ld\n",
|
|
now, frr->ts, frr->len);
|
|
}
|
|
|
|
ast_channel_cleanup(bridged);
|
|
}
|
|
|
|
ast_verb(3, "%s jitterbuffer created on channel %s\n", jbimpl->name, ast_channel_name(chan));
|
|
|
|
/* Free the frame if it has not been queued in the jb */
|
|
if (res != AST_JB_IMPL_OK) {
|
|
ast_frfree(frr);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
|
|
void ast_jb_destroy(struct ast_channel *chan)
|
|
{
|
|
struct ast_jb *jb = ast_channel_jb(chan);
|
|
const struct ast_jb_impl *jbimpl = jb->impl;
|
|
void *jbobj = jb->jbobj;
|
|
struct ast_frame *f;
|
|
|
|
if (jb->logfile) {
|
|
fclose(jb->logfile);
|
|
jb->logfile = NULL;
|
|
}
|
|
|
|
ao2_cleanup(jb->last_format);
|
|
|
|
if (ast_test_flag(jb, JB_CREATED)) {
|
|
/* Remove and free all frames still queued in jb */
|
|
while (jbimpl->remove(jbobj, &f) == AST_JB_IMPL_OK) {
|
|
ast_frfree(f);
|
|
}
|
|
|
|
jbimpl->destroy(jbobj);
|
|
jb->jbobj = NULL;
|
|
|
|
ast_clear_flag(jb, JB_CREATED);
|
|
|
|
ast_verb(3, "%s jitterbuffer destroyed on channel %s\n", jbimpl->name, ast_channel_name(chan));
|
|
}
|
|
}
|
|
|
|
|
|
static long get_now(struct ast_jb *jb, struct timeval *when)
|
|
{
|
|
struct timeval now;
|
|
|
|
if (!when) {
|
|
when = &now;
|
|
gettimeofday(when, NULL);
|
|
}
|
|
|
|
return ast_tvdiff_ms(*when, jb->timebase);
|
|
}
|
|
|
|
|
|
int ast_jb_read_conf(struct ast_jb_conf *conf, const char *varname, const char *value)
|
|
{
|
|
int prefixlen = sizeof(AST_JB_CONF_PREFIX) - 1;
|
|
const char *name;
|
|
int tmp;
|
|
|
|
if (strncasecmp(AST_JB_CONF_PREFIX, varname, prefixlen)) {
|
|
return -1;
|
|
}
|
|
|
|
name = varname + prefixlen;
|
|
|
|
if (!strcasecmp(name, AST_JB_CONF_ENABLE)) {
|
|
ast_set2_flag(conf, ast_true(value), AST_JB_ENABLED);
|
|
} else if (!strcasecmp(name, AST_JB_CONF_FORCE)) {
|
|
ast_set2_flag(conf, ast_true(value), AST_JB_FORCED);
|
|
} else if (!strcasecmp(name, AST_JB_CONF_MAX_SIZE)) {
|
|
if ((tmp = atoi(value)) > 0)
|
|
conf->max_size = tmp;
|
|
} else if (!strcasecmp(name, AST_JB_CONF_RESYNCH_THRESHOLD)) {
|
|
if ((tmp = atoi(value)) > 0)
|
|
conf->resync_threshold = tmp;
|
|
} else if (!strcasecmp(name, AST_JB_CONF_IMPL)) {
|
|
if (!ast_strlen_zero(value))
|
|
snprintf(conf->impl, sizeof(conf->impl), "%s", value);
|
|
} else if (!strcasecmp(name, AST_JB_CONF_TARGET_EXTRA)) {
|
|
if (sscanf(value, "%30d", &tmp) == 1) {
|
|
conf->target_extra = tmp;
|
|
}
|
|
} else if (!strcasecmp(name, AST_JB_CONF_LOG)) {
|
|
ast_set2_flag(conf, ast_true(value), AST_JB_LOG);
|
|
} else if (!strcasecmp(name, AST_JB_CONF_SYNC_VIDEO)) {
|
|
ast_set2_flag(conf, ast_true(value), AST_JB_SYNC_VIDEO);
|
|
} else {
|
|
return -1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
void ast_jb_enable_for_channel(struct ast_channel *chan)
|
|
{
|
|
struct ast_jb_conf conf = ast_channel_jb(chan)->conf;
|
|
if (ast_test_flag(&conf, AST_JB_ENABLED)) {
|
|
ast_jb_create_framehook(chan, &conf, 1);
|
|
}
|
|
}
|
|
|
|
void ast_jb_configure(struct ast_channel *chan, const struct ast_jb_conf *conf)
|
|
{
|
|
memcpy(&ast_channel_jb(chan)->conf, conf, sizeof(*conf));
|
|
}
|
|
|
|
|
|
void ast_jb_get_config(const struct ast_channel *chan, struct ast_jb_conf *conf)
|
|
{
|
|
memcpy(conf, &ast_channel_jb((struct ast_channel *) chan)->conf, sizeof(*conf));
|
|
}
|
|
|
|
void ast_jb_empty_and_reset(struct ast_channel *c0, struct ast_channel *c1)
|
|
{
|
|
struct ast_jb *jb0 = ast_channel_jb(c0);
|
|
struct ast_jb *jb1 = ast_channel_jb(c1);
|
|
int c0_use_jb = ast_test_flag(jb0, JB_USE);
|
|
int c0_jb_is_created = ast_test_flag(jb0, JB_CREATED);
|
|
int c1_use_jb = ast_test_flag(jb1, JB_USE);
|
|
int c1_jb_is_created = ast_test_flag(jb1, JB_CREATED);
|
|
|
|
if (c0_use_jb && c0_jb_is_created && jb0->impl->empty_and_reset) {
|
|
jb0->impl->empty_and_reset(jb0->jbobj);
|
|
}
|
|
|
|
if (c1_use_jb && c1_jb_is_created && jb1->impl->empty_and_reset) {
|
|
jb1->impl->empty_and_reset(jb1->jbobj);
|
|
}
|
|
}
|
|
|
|
/* Implementation functions */
|
|
|
|
/* fixed */
|
|
static void * jb_create_fixed(struct ast_jb_conf *general_config)
|
|
{
|
|
struct fixed_jb_conf conf;
|
|
|
|
conf.jbsize = general_config->max_size;
|
|
conf.resync_threshold = general_config->resync_threshold;
|
|
|
|
return fixed_jb_new(&conf);
|
|
}
|
|
|
|
static void jb_destroy_fixed(void *jb)
|
|
{
|
|
struct fixed_jb *fixedjb = (struct fixed_jb *) jb;
|
|
|
|
/* Ensure the fixed jb is empty - otherwise it will raise an ASSERT */
|
|
jb_empty_and_reset_fixed(jb);
|
|
|
|
/* destroy the jb */
|
|
fixed_jb_destroy(fixedjb);
|
|
}
|
|
|
|
|
|
static int jb_put_first_fixed(void *jb, struct ast_frame *fin, long now)
|
|
{
|
|
struct fixed_jb *fixedjb = (struct fixed_jb *) jb;
|
|
int res;
|
|
|
|
res = fixed_jb_put_first(fixedjb, fin, fin->len, fin->ts, now);
|
|
|
|
return fixed_to_abstract_code[res];
|
|
}
|
|
|
|
|
|
static int jb_put_fixed(void *jb, struct ast_frame *fin, long now)
|
|
{
|
|
struct fixed_jb *fixedjb = (struct fixed_jb *) jb;
|
|
int res;
|
|
|
|
res = fixed_jb_put(fixedjb, fin, fin->len, fin->ts, now);
|
|
|
|
return fixed_to_abstract_code[res];
|
|
}
|
|
|
|
|
|
static int jb_get_fixed(void *jb, struct ast_frame **fout, long now, long interpl)
|
|
{
|
|
struct fixed_jb *fixedjb = (struct fixed_jb *) jb;
|
|
struct fixed_jb_frame frame = { .data = &ast_null_frame };
|
|
int res;
|
|
|
|
res = fixed_jb_get(fixedjb, &frame, now, interpl);
|
|
*fout = frame.data;
|
|
|
|
return fixed_to_abstract_code[res];
|
|
}
|
|
|
|
|
|
static long jb_next_fixed(void *jb)
|
|
{
|
|
struct fixed_jb *fixedjb = (struct fixed_jb *) jb;
|
|
|
|
return fixed_jb_next(fixedjb);
|
|
}
|
|
|
|
|
|
static int jb_remove_fixed(void *jb, struct ast_frame **fout)
|
|
{
|
|
struct fixed_jb *fixedjb = (struct fixed_jb *) jb;
|
|
struct fixed_jb_frame frame;
|
|
int res;
|
|
|
|
res = fixed_jb_remove(fixedjb, &frame);
|
|
*fout = frame.data;
|
|
|
|
return fixed_to_abstract_code[res];
|
|
}
|
|
|
|
|
|
static void jb_force_resynch_fixed(void *jb)
|
|
{
|
|
struct fixed_jb *fixedjb = (struct fixed_jb *) jb;
|
|
|
|
fixed_jb_set_force_resynch(fixedjb);
|
|
}
|
|
|
|
static void jb_empty_and_reset_fixed(void *jb)
|
|
{
|
|
struct fixed_jb *fixedjb = jb;
|
|
struct fixed_jb_frame f;
|
|
|
|
while (fixed_jb_remove(fixedjb, &f) == FIXED_JB_OK) {
|
|
ast_frfree(f.data);
|
|
}
|
|
}
|
|
|
|
static int jb_is_late_fixed(void *jb, long ts)
|
|
{
|
|
return fixed_jb_is_late(jb, ts);
|
|
}
|
|
|
|
/* adaptive */
|
|
|
|
static void *jb_create_adaptive(struct ast_jb_conf *general_config)
|
|
{
|
|
jb_conf jbconf;
|
|
jitterbuf *adaptivejb;
|
|
|
|
adaptivejb = jb_new();
|
|
if (adaptivejb) {
|
|
jbconf.max_jitterbuf = general_config->max_size;
|
|
jbconf.resync_threshold = general_config->resync_threshold;
|
|
jbconf.max_contig_interp = 10;
|
|
jbconf.target_extra = general_config->target_extra;
|
|
jb_setconf(adaptivejb, &jbconf);
|
|
}
|
|
|
|
return adaptivejb;
|
|
}
|
|
|
|
|
|
static void jb_destroy_adaptive(void *jb)
|
|
{
|
|
jitterbuf *adaptivejb = (jitterbuf *) jb;
|
|
|
|
jb_destroy(adaptivejb);
|
|
}
|
|
|
|
|
|
static int jb_put_first_adaptive(void *jb, struct ast_frame *fin, long now)
|
|
{
|
|
return jb_put_adaptive(jb, fin, now);
|
|
}
|
|
|
|
|
|
static int jb_put_adaptive(void *jb, struct ast_frame *fin, long now)
|
|
{
|
|
jitterbuf *adaptivejb = (jitterbuf *) jb;
|
|
int res;
|
|
|
|
res = jb_put(adaptivejb, fin, JB_TYPE_VOICE, fin->len, fin->ts, now);
|
|
|
|
return adaptive_to_abstract_code[res];
|
|
}
|
|
|
|
|
|
static int jb_get_adaptive(void *jb, struct ast_frame **fout, long now, long interpl)
|
|
{
|
|
jitterbuf *adaptivejb = (jitterbuf *) jb;
|
|
jb_frame frame = { .data = &ast_null_frame };
|
|
int res;
|
|
|
|
res = jb_get(adaptivejb, &frame, now, interpl);
|
|
*fout = frame.data;
|
|
|
|
return adaptive_to_abstract_code[res];
|
|
}
|
|
|
|
|
|
static long jb_next_adaptive(void *jb)
|
|
{
|
|
jitterbuf *adaptivejb = (jitterbuf *) jb;
|
|
|
|
return jb_next(adaptivejb);
|
|
}
|
|
|
|
|
|
static int jb_remove_adaptive(void *jb, struct ast_frame **fout)
|
|
{
|
|
jitterbuf *adaptivejb = (jitterbuf *) jb;
|
|
jb_frame frame;
|
|
int res;
|
|
|
|
res = jb_getall(adaptivejb, &frame);
|
|
*fout = frame.data;
|
|
|
|
return adaptive_to_abstract_code[res];
|
|
}
|
|
|
|
|
|
static void jb_force_resynch_adaptive(void *jb)
|
|
{
|
|
}
|
|
|
|
static void jb_empty_and_reset_adaptive(void *jb)
|
|
{
|
|
jitterbuf *adaptivejb = jb;
|
|
jb_frame f;
|
|
|
|
while (jb_getall(adaptivejb, &f) == JB_OK) {
|
|
ast_frfree(f.data);
|
|
}
|
|
|
|
jb_reset(adaptivejb);
|
|
}
|
|
|
|
const struct ast_jb_impl *ast_jb_get_impl(enum ast_jb_type type)
|
|
{
|
|
int i;
|
|
for (i = 0; i < ARRAY_LEN(avail_impl); i++) {
|
|
if (avail_impl[i].type == type) {
|
|
return &avail_impl[i];
|
|
}
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
static int jb_is_late_adaptive(void *jb, long ts)
|
|
{
|
|
return jb_is_late(jb, ts);
|
|
}
|
|
|
|
#define DEFAULT_TIMER_INTERVAL 20
|
|
#define DEFAULT_SIZE 200
|
|
#define DEFAULT_TARGET_EXTRA 40
|
|
#define DEFAULT_RESYNC 1000
|
|
#define DEFAULT_TYPE AST_JB_FIXED
|
|
|
|
struct jb_stream_sync {
|
|
unsigned int timestamp;
|
|
struct timeval ntp;
|
|
};
|
|
|
|
struct jb_framedata {
|
|
const struct ast_jb_impl *jb_impl;
|
|
struct ast_jb_conf jb_conf;
|
|
struct timeval start_tv;
|
|
struct ast_format *last_format;
|
|
struct ast_timer *timer;
|
|
int timer_interval; /* ms between deliveries */
|
|
int timer_fd;
|
|
int first;
|
|
int audio_stream_id;
|
|
struct jb_stream_sync audio_stream_sync;
|
|
int video_stream_id;
|
|
struct jb_stream_sync video_stream_sync;
|
|
AST_LIST_HEAD_NOLOCK(, ast_frame) early_frames;
|
|
unsigned int early_frame_count;
|
|
struct timeval last_audio_ntp_timestamp;
|
|
int audio_flowing;
|
|
void *jb_obj;
|
|
};
|
|
|
|
static void jb_framedata_destroy(struct jb_framedata *framedata)
|
|
{
|
|
struct ast_frame *frame;
|
|
|
|
if (framedata->timer) {
|
|
ast_timer_close(framedata->timer);
|
|
framedata->timer = NULL;
|
|
}
|
|
if (framedata->jb_impl && framedata->jb_obj) {
|
|
struct ast_frame *f;
|
|
while (framedata->jb_impl->remove(framedata->jb_obj, &f) == AST_JB_IMPL_OK) {
|
|
ast_frfree(f);
|
|
}
|
|
framedata->jb_impl->destroy(framedata->jb_obj);
|
|
framedata->jb_obj = NULL;
|
|
}
|
|
ao2_cleanup(framedata->last_format);
|
|
while ((frame = AST_LIST_REMOVE_HEAD(&framedata->early_frames, frame_list))) {
|
|
ast_frfree(frame);
|
|
}
|
|
ast_free(framedata);
|
|
}
|
|
|
|
void ast_jb_conf_default(struct ast_jb_conf *conf)
|
|
{
|
|
ast_clear_flag(conf, AST_FLAGS_ALL);
|
|
conf->max_size = DEFAULT_SIZE;
|
|
conf->resync_threshold = DEFAULT_RESYNC;
|
|
ast_copy_string(conf->impl, "fixed", sizeof(conf->impl));
|
|
conf->target_extra = DEFAULT_TARGET_EXTRA;
|
|
}
|
|
|
|
static void datastore_destroy_cb(void *data) {
|
|
ast_free(data);
|
|
ast_debug(1, "JITTERBUFFER datastore destroyed\n");
|
|
}
|
|
|
|
static const struct ast_datastore_info jb_datastore = {
|
|
.type = "jitterbuffer",
|
|
.destroy = datastore_destroy_cb
|
|
};
|
|
|
|
static void hook_destroy_cb(void *framedata)
|
|
{
|
|
ast_debug(1, "JITTERBUFFER hook destroyed\n");
|
|
jb_framedata_destroy((struct jb_framedata *) framedata);
|
|
}
|
|
|
|
static struct timeval jitterbuffer_frame_get_ntp_timestamp(const struct jb_stream_sync *stream_sync, const struct ast_frame *frame)
|
|
{
|
|
int timestamp_diff;
|
|
unsigned int rate;
|
|
|
|
/* It's possible for us to receive frames before we receive the information allowing
|
|
* us to do NTP/RTP timestamp calculations. Since the information isn't available we
|
|
* can't generate one and give an empty timestamp.
|
|
*/
|
|
if (ast_tvzero(stream_sync->ntp)) {
|
|
return ast_tv(0, 0);
|
|
}
|
|
|
|
/* Convert the Asterisk timestamp into an RTP timestamp, and then based on the difference we can
|
|
* determine how many samples are in the frame and how long has elapsed since the synchronization
|
|
* RTP and NTP timestamps were received giving us the NTP timestamp for this frame.
|
|
*/
|
|
if (frame->frametype == AST_FRAME_VOICE) {
|
|
rate = ast_rtp_get_rate(frame->subclass.format);
|
|
timestamp_diff = (frame->ts * (rate / 1000)) - stream_sync->timestamp;
|
|
} else {
|
|
/* Video is special - internally we reference it as 1000 to preserve the RTP timestamp but
|
|
* it is actually 90000, this is why we can just directly subtract the timestamp.
|
|
*/
|
|
rate = 90000;
|
|
timestamp_diff = frame->ts - stream_sync->timestamp;
|
|
}
|
|
|
|
if (timestamp_diff < 0) {
|
|
/* It's possible for us to be asked for an NTP timestamp from before our latest
|
|
* RTCP SR report. To handle this we subtract so we go back in time.
|
|
*/
|
|
return ast_tvsub(stream_sync->ntp, ast_samp2tv(abs(timestamp_diff), rate));
|
|
} else {
|
|
return ast_tvadd(stream_sync->ntp, ast_samp2tv(timestamp_diff, rate));
|
|
}
|
|
}
|
|
|
|
static struct ast_frame *hook_event_cb(struct ast_channel *chan, struct ast_frame *frame, enum ast_framehook_event event, void *data)
|
|
{
|
|
struct jb_framedata *framedata = data;
|
|
struct timeval now_tv;
|
|
int64_t relative_frame_start;
|
|
int putframe = 0; /* signifies if audio frame was placed into the buffer or not */
|
|
|
|
switch (event) {
|
|
case AST_FRAMEHOOK_EVENT_READ:
|
|
break;
|
|
case AST_FRAMEHOOK_EVENT_ATTACHED:
|
|
case AST_FRAMEHOOK_EVENT_DETACHED:
|
|
case AST_FRAMEHOOK_EVENT_WRITE:
|
|
return frame;
|
|
}
|
|
|
|
if (ast_channel_fdno(chan) == AST_JITTERBUFFER_FD && framedata->timer) {
|
|
if (ast_timer_ack(framedata->timer, 1) < 0) {
|
|
ast_log(LOG_ERROR, "Failed to acknowledge timer in jitter buffer\n");
|
|
return frame;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* If the frame has been requeued (for instance when the translate core returns
|
|
* more than one frame) then if the frame is late we want to immediately return
|
|
* it. Otherwise attempt to insert it into the jitterbuffer.
|
|
*
|
|
* If the frame is requeued and late then in all likely hood it's a frame that
|
|
* that was previously retrieved from the jitterbuffer, passed to the translate
|
|
* core, and then put back into the channel read queue. Even if it had not been
|
|
* in the jitterbuffer prior to now it needs to be the next frame "out".
|
|
*
|
|
* However late arriving frames that have not been requeued (i.e. regular frames)
|
|
* need to be passed to the jitterbuffer so they can be appropriately dropped. As
|
|
* well any requeued frames that are not late should be put into the jitterbuffer.
|
|
*/
|
|
if (!frame || (ast_test_flag(frame, AST_FRFLAG_REQUEUED) &&
|
|
framedata->jb_impl->is_late(framedata->jb_obj, frame->ts))) {
|
|
return frame;
|
|
}
|
|
|
|
if (ast_test_flag(&framedata->jb_conf, AST_JB_SYNC_VIDEO)) {
|
|
if (frame->frametype == AST_FRAME_VOICE) {
|
|
/* Store the stream identifier for the audio stream so we can associate the incoming RTCP SR
|
|
* with the correct stream sync structure.
|
|
*/
|
|
framedata->audio_stream_id = frame->stream_num;
|
|
} else if (frame->frametype == AST_FRAME_RTCP && frame->subclass.integer == AST_RTP_RTCP_SR) {
|
|
struct ast_rtp_rtcp_report *rtcp_report = frame->data.ptr;
|
|
struct jb_stream_sync *stream_sync = NULL;
|
|
|
|
/* Determine which stream this RTCP is in regards to */
|
|
if (framedata->audio_stream_id == frame->stream_num) {
|
|
stream_sync = &framedata->audio_stream_sync;
|
|
} else if (framedata->video_stream_id == frame->stream_num) {
|
|
stream_sync = &framedata->video_stream_sync;
|
|
}
|
|
|
|
if (stream_sync) {
|
|
/* Store the RTP and NTP timestamp mapping so we can derive an NTP timestamp for each frame */
|
|
stream_sync->timestamp = rtcp_report->sender_information.rtp_timestamp;
|
|
stream_sync->ntp = rtcp_report->sender_information.ntp_timestamp;
|
|
}
|
|
} else if (frame->frametype == AST_FRAME_VIDEO) {
|
|
/* If a video frame is late according to the audio timestamp don't stash it away, just return it.
|
|
* If however it is ahead then we keep it until such time as the audio catches up.
|
|
*/
|
|
struct ast_frame *jbframe;
|
|
|
|
framedata->video_stream_id = frame->stream_num;
|
|
|
|
/* If no timing information is available we can't store this away, so just let it through now */
|
|
if (!ast_test_flag(frame, AST_FRFLAG_HAS_TIMING_INFO)) {
|
|
return frame;
|
|
}
|
|
|
|
/* To ensure that the video starts when the audio starts we only start allowing frames through once
|
|
* audio starts flowing.
|
|
*/
|
|
if (framedata->audio_flowing) {
|
|
struct timeval video_timestamp;
|
|
|
|
video_timestamp = jitterbuffer_frame_get_ntp_timestamp(&framedata->video_stream_sync, frame);
|
|
if (ast_tvdiff_ms(framedata->last_audio_ntp_timestamp, video_timestamp) >= 0) {
|
|
return frame;
|
|
}
|
|
}
|
|
|
|
/* To prevent the early frame buffer from growing uncontrolled we impose a maximum count that it can
|
|
* get to. If this is reached then we drop a video frame, which should cause the receiver to ask for a
|
|
* new key frame.
|
|
*/
|
|
if (framedata->early_frame_count == MAXIMUM_EARLY_FRAME_COUNT) {
|
|
jbframe = AST_LIST_REMOVE_HEAD(&framedata->early_frames, frame_list);
|
|
framedata->early_frame_count--;
|
|
ast_frfree(jbframe);
|
|
}
|
|
|
|
jbframe = ast_frisolate(frame);
|
|
if (!jbframe) {
|
|
/* If we can't isolate the frame the safest thing we can do is return it, even if the A/V sync
|
|
* may be off.
|
|
*/
|
|
return frame;
|
|
}
|
|
|
|
AST_LIST_INSERT_TAIL(&framedata->early_frames, jbframe, frame_list);
|
|
framedata->early_frame_count++;
|
|
return &ast_null_frame;
|
|
}
|
|
}
|
|
|
|
now_tv = ast_tvnow();
|
|
relative_frame_start = ast_tvdiff_ms(now_tv, framedata->start_tv);
|
|
if (relative_frame_start < 0) {
|
|
/*
|
|
* The only way for this to happen is if the system time has
|
|
* stepped backwards between the time framedata->start_tv was
|
|
* set and now. Think an ntpd or systemd-timesyncd adjustment.
|
|
*
|
|
* Just pass the frame through.
|
|
*/
|
|
return frame;
|
|
}
|
|
|
|
if (frame->frametype == AST_FRAME_VOICE) {
|
|
int res;
|
|
struct ast_frame *jbframe;
|
|
|
|
if (!ast_test_flag(frame, AST_FRFLAG_HAS_TIMING_INFO) || frame->len < 2 || frame->ts < 0) {
|
|
/* only frames with timing info can enter the jitterbuffer */
|
|
return frame;
|
|
}
|
|
|
|
jbframe = ast_frisolate(frame);
|
|
ao2_replace(framedata->last_format, frame->subclass.format);
|
|
|
|
if (frame->len && (frame->len != framedata->timer_interval)) {
|
|
framedata->timer_interval = frame->len;
|
|
ast_timer_set_rate(framedata->timer, 1000 / framedata->timer_interval);
|
|
}
|
|
if (!framedata->first) {
|
|
framedata->first = 1;
|
|
res = framedata->jb_impl->put_first(framedata->jb_obj, jbframe, relative_frame_start);
|
|
} else {
|
|
res = framedata->jb_impl->put(framedata->jb_obj, jbframe, relative_frame_start);
|
|
}
|
|
|
|
if (res == AST_JB_IMPL_OK) {
|
|
if (jbframe != frame) {
|
|
ast_frfree(frame);
|
|
}
|
|
frame = &ast_null_frame;
|
|
} else if (jbframe != frame) {
|
|
ast_frfree(jbframe);
|
|
}
|
|
putframe = 1;
|
|
}
|
|
|
|
if (frame->frametype == AST_FRAME_NULL) {
|
|
int res;
|
|
long next = framedata->jb_impl->next(framedata->jb_obj);
|
|
|
|
/* If relative_frame_start is earlier than the next expected output frame
|
|
* from the jitterbuffer we may choose to pass on retrieving
|
|
* a frame during this read iteration. The only exception
|
|
* to this rule is when an audio frame is placed into the buffer
|
|
* and the time for the next frame to come out of the buffer is
|
|
* at least within the timer_interval of the next output frame. By
|
|
* doing this we are able to feed off the timing of the input frames
|
|
* and only rely on our jitterbuffer timer when frames are dropped.
|
|
* During testing, this hybrid form of timing gave more reliable results. */
|
|
if (relative_frame_start < next) {
|
|
long int diff = next - relative_frame_start;
|
|
if (!putframe) {
|
|
return frame;
|
|
} else if (diff >= framedata->timer_interval) {
|
|
return frame;
|
|
}
|
|
}
|
|
|
|
ast_frfree(frame);
|
|
frame = &ast_null_frame;
|
|
res = framedata->jb_impl->get(framedata->jb_obj, &frame, relative_frame_start, framedata->timer_interval);
|
|
switch (res) {
|
|
case AST_JB_IMPL_OK:
|
|
/* got it, and pass it through */
|
|
break;
|
|
case AST_JB_IMPL_DROP:
|
|
ast_frfree(frame);
|
|
frame = &ast_null_frame;
|
|
break;
|
|
case AST_JB_IMPL_INTERP:
|
|
if (framedata->last_format) {
|
|
struct ast_frame tmp = { 0, };
|
|
|
|
tmp.frametype = AST_FRAME_VOICE;
|
|
tmp.subclass.format = framedata->last_format;
|
|
/* example: 8000hz / (1000 / 20ms) = 160 samples */
|
|
tmp.samples = ast_format_get_sample_rate(framedata->last_format) / (1000 / framedata->timer_interval);
|
|
tmp.delivery = ast_tvadd(framedata->start_tv, ast_samp2tv(next, 1000));
|
|
tmp.offset = AST_FRIENDLY_OFFSET;
|
|
tmp.src = "func_jitterbuffer interpolation";
|
|
ast_frfree(frame);
|
|
frame = ast_frdup(&tmp);
|
|
break;
|
|
}
|
|
/* else fall through */
|
|
case AST_JB_IMPL_NOFRAME:
|
|
ast_frfree(frame);
|
|
frame = &ast_null_frame;
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (frame->frametype == AST_FRAME_CONTROL) {
|
|
struct ast_frame *early_frame;
|
|
|
|
switch(frame->subclass.integer) {
|
|
case AST_CONTROL_HOLD:
|
|
case AST_CONTROL_UNHOLD:
|
|
case AST_CONTROL_T38_PARAMETERS:
|
|
case AST_CONTROL_SRCUPDATE:
|
|
case AST_CONTROL_SRCCHANGE:
|
|
framedata->jb_impl->force_resync(framedata->jb_obj);
|
|
/* Since we are resyncing go ahead and clear out the video frames too */
|
|
while ((early_frame = AST_LIST_REMOVE_HEAD(&framedata->early_frames, frame_list))) {
|
|
ast_frfree(early_frame);
|
|
}
|
|
framedata->audio_flowing = 0;
|
|
framedata->early_frame_count = 0;
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
|
|
/* If a voice frame is being passed through see if we need to add any additional frames to it */
|
|
if (ast_test_flag(&framedata->jb_conf, AST_JB_SYNC_VIDEO) && frame->frametype == AST_FRAME_VOICE) {
|
|
AST_LIST_HEAD_NOLOCK(, ast_frame) additional_frames;
|
|
struct ast_frame *early_frame;
|
|
|
|
/* We store the last NTP timestamp for the audio given to the core so that subsequents frames which
|
|
* are late can be passed immediately through (this will occur for video frames which are returned here)
|
|
*/
|
|
framedata->last_audio_ntp_timestamp = jitterbuffer_frame_get_ntp_timestamp(&framedata->audio_stream_sync, frame);
|
|
framedata->audio_flowing = 1;
|
|
|
|
AST_LIST_HEAD_INIT_NOLOCK(&additional_frames);
|
|
|
|
AST_LIST_TRAVERSE_SAFE_BEGIN(&framedata->early_frames, early_frame, frame_list) {
|
|
struct timeval early_timestamp = jitterbuffer_frame_get_ntp_timestamp(&framedata->video_stream_sync, early_frame);
|
|
int diff = ast_tvdiff_ms(framedata->last_audio_ntp_timestamp, early_timestamp);
|
|
|
|
/* If this frame is from the past we need to include it with the audio frame that is going
|
|
* out.
|
|
*/
|
|
if (diff >= 0) {
|
|
AST_LIST_REMOVE_CURRENT(frame_list);
|
|
framedata->early_frame_count--;
|
|
AST_LIST_INSERT_TAIL(&additional_frames, early_frame, frame_list);
|
|
}
|
|
}
|
|
AST_LIST_TRAVERSE_SAFE_END;
|
|
|
|
/* Append any additional frames we may want to include (such as video) */
|
|
AST_LIST_NEXT(frame, frame_list) = AST_LIST_FIRST(&additional_frames);
|
|
}
|
|
|
|
return frame;
|
|
}
|
|
|
|
/* set defaults */
|
|
static int jb_framedata_init(struct jb_framedata *framedata, struct ast_jb_conf *jb_conf)
|
|
{
|
|
int jb_impl_type = DEFAULT_TYPE;
|
|
/* Initialize defaults */
|
|
framedata->timer_fd = -1;
|
|
memcpy(&framedata->jb_conf, jb_conf, sizeof(*jb_conf));
|
|
|
|
/* Figure out implementation type from the configuration implementation string */
|
|
if (!ast_strlen_zero(jb_conf->impl)) {
|
|
if (!strcasecmp(jb_conf->impl, "fixed")) {
|
|
jb_impl_type = AST_JB_FIXED;
|
|
} else if (!strcasecmp(jb_conf->impl, "adaptive")) {
|
|
jb_impl_type = AST_JB_ADAPTIVE;
|
|
} else {
|
|
ast_log(LOG_WARNING, "Unknown Jitterbuffer type %s. Failed to create jitterbuffer.\n", jb_conf->impl);
|
|
return -1;
|
|
}
|
|
}
|
|
|
|
if (!(framedata->jb_impl = ast_jb_get_impl(jb_impl_type))) {
|
|
return -1;
|
|
}
|
|
|
|
if (!(framedata->timer = ast_timer_open())) {
|
|
return -1;
|
|
}
|
|
|
|
framedata->audio_stream_id = -1;
|
|
framedata->video_stream_id = -1;
|
|
AST_LIST_HEAD_INIT_NOLOCK(&framedata->early_frames);
|
|
framedata->timer_fd = ast_timer_fd(framedata->timer);
|
|
framedata->timer_interval = DEFAULT_TIMER_INTERVAL;
|
|
ast_timer_set_rate(framedata->timer, 1000 / framedata->timer_interval);
|
|
framedata->start_tv = ast_tvnow();
|
|
|
|
framedata->jb_obj = framedata->jb_impl->create(&framedata->jb_conf);
|
|
return 0;
|
|
}
|
|
|
|
|
|
void ast_jb_create_framehook(struct ast_channel *chan, struct ast_jb_conf *jb_conf, int prefer_existing)
|
|
{
|
|
struct jb_framedata *framedata;
|
|
struct ast_datastore *datastore = NULL;
|
|
struct ast_framehook_interface interface = {
|
|
.version = AST_FRAMEHOOK_INTERFACE_VERSION,
|
|
.event_cb = hook_event_cb,
|
|
.destroy_cb = hook_destroy_cb,
|
|
};
|
|
int i = 0;
|
|
|
|
/* If disabled, strip any existing jitterbuffer and don't replace it. */
|
|
if (!strcasecmp(jb_conf->impl, "disabled")) {
|
|
int *id;
|
|
ast_channel_lock(chan);
|
|
if ((datastore = ast_channel_datastore_find(chan, &jb_datastore, NULL))) {
|
|
id = datastore->data;
|
|
ast_framehook_detach(chan, *id);
|
|
ast_channel_datastore_remove(chan, datastore);
|
|
ast_datastore_free(datastore);
|
|
}
|
|
ast_channel_unlock(chan);
|
|
return;
|
|
}
|
|
|
|
if (!(framedata = ast_calloc(1, sizeof(*framedata)))) {
|
|
return;
|
|
}
|
|
|
|
if (jb_framedata_init(framedata, jb_conf)) {
|
|
jb_framedata_destroy(framedata);
|
|
return;
|
|
}
|
|
|
|
interface.data = framedata;
|
|
|
|
ast_channel_lock(chan);
|
|
i = ast_framehook_attach(chan, &interface);
|
|
if (i >= 0) {
|
|
int *id;
|
|
if ((datastore = ast_channel_datastore_find(chan, &jb_datastore, NULL))) {
|
|
/* There is already a jitterbuffer on the channel. */
|
|
if (prefer_existing) {
|
|
/* We prefer the existing jitterbuffer, so remove the new one and keep the old one. */
|
|
ast_framehook_detach(chan, i);
|
|
ast_channel_unlock(chan);
|
|
return;
|
|
}
|
|
/* We prefer the new jitterbuffer, so strip the old one. */
|
|
id = datastore->data;
|
|
ast_framehook_detach(chan, *id);
|
|
ast_channel_datastore_remove(chan, datastore);
|
|
ast_datastore_free(datastore);
|
|
}
|
|
|
|
if (!(datastore = ast_datastore_alloc(&jb_datastore, NULL))) {
|
|
ast_framehook_detach(chan, i);
|
|
ast_channel_unlock(chan);
|
|
return;
|
|
}
|
|
|
|
if (!(id = ast_calloc(1, sizeof(int)))) {
|
|
ast_datastore_free(datastore);
|
|
ast_framehook_detach(chan, i);
|
|
ast_channel_unlock(chan);
|
|
return;
|
|
}
|
|
|
|
*id = i; /* Store off the id. The channel is still locked so it is safe to access this ptr. */
|
|
datastore->data = id;
|
|
ast_channel_datastore_add(chan, datastore);
|
|
|
|
ast_channel_set_fd(chan, AST_JITTERBUFFER_FD, framedata->timer_fd);
|
|
} else {
|
|
jb_framedata_destroy(framedata);
|
|
framedata = NULL;
|
|
}
|
|
ast_channel_unlock(chan);
|
|
}
|