asterisk/main/abstract_jb.c

1346 lines
39 KiB
C

/*
* abstract_jb: common implementation-independent jitterbuffer stuff
*
* Copyright (C) 2005, Attractel OOD
*
* Contributors:
* Slav Klenov <slav@securax.org>
*
* See http://www.asterisk.org for more information about
* the Asterisk project. Please do not directly contact
* any of the maintainers of this project for assistance;
* the project provides a web site, mailing lists and IRC
* channels for your use.
*
* This program is free software, distributed under the terms of
* the GNU General Public License Version 2. See the LICENSE file
* at the top of the source tree.
*
* A license has been granted to Digium (via disclaimer) for the use of
* this code.
*/
/*! \file
*
* \brief Common implementation-independent jitterbuffer stuff.
*
* \author Slav Klenov <slav@securax.org>
*
*
*/
/*** MODULEINFO
<support_level>core</support_level>
***/
#include "asterisk.h"
#include "asterisk/frame.h"
#include "asterisk/channel.h"
#include "asterisk/term.h"
#include "asterisk/utils.h"
#include "asterisk/pbx.h"
#include "asterisk/timing.h"
#include "asterisk/rtp_engine.h"
#include "asterisk/format_cache.h"
#include "asterisk/abstract_jb.h"
#include "fixedjitterbuf.h"
#include "jitterbuf.h"
/*! Internal jb flags */
enum {
JB_USE = (1 << 0),
JB_TIMEBASE_INITIALIZED = (1 << 1),
JB_CREATED = (1 << 2)
};
/*! The maximum size we allow the early frame buffer to get */
#define MAXIMUM_EARLY_FRAME_COUNT 200
/* Implementation functions */
/* fixed */
static void *jb_create_fixed(struct ast_jb_conf *general_config);
static void jb_destroy_fixed(void *jb);
static int jb_put_first_fixed(void *jb, struct ast_frame *fin, long now);
static int jb_put_fixed(void *jb, struct ast_frame *fin, long now);
static int jb_get_fixed(void *jb, struct ast_frame **fout, long now, long interpl);
static long jb_next_fixed(void *jb);
static int jb_remove_fixed(void *jb, struct ast_frame **fout);
static void jb_force_resynch_fixed(void *jb);
static void jb_empty_and_reset_fixed(void *jb);
static int jb_is_late_fixed(void *jb, long ts);
/* adaptive */
static void * jb_create_adaptive(struct ast_jb_conf *general_config);
static void jb_destroy_adaptive(void *jb);
static int jb_put_first_adaptive(void *jb, struct ast_frame *fin, long now);
static int jb_put_adaptive(void *jb, struct ast_frame *fin, long now);
static int jb_get_adaptive(void *jb, struct ast_frame **fout, long now, long interpl);
static long jb_next_adaptive(void *jb);
static int jb_remove_adaptive(void *jb, struct ast_frame **fout);
static void jb_force_resynch_adaptive(void *jb);
static void jb_empty_and_reset_adaptive(void *jb);
static int jb_is_late_adaptive(void *jb, long ts);
/* Available jb implementations */
static const struct ast_jb_impl avail_impl[] = {
{
.name = "fixed",
.type = AST_JB_FIXED,
.create = jb_create_fixed,
.destroy = jb_destroy_fixed,
.put_first = jb_put_first_fixed,
.put = jb_put_fixed,
.get = jb_get_fixed,
.next = jb_next_fixed,
.remove = jb_remove_fixed,
.force_resync = jb_force_resynch_fixed,
.empty_and_reset = jb_empty_and_reset_fixed,
.is_late = jb_is_late_fixed,
},
{
.name = "adaptive",
.type = AST_JB_ADAPTIVE,
.create = jb_create_adaptive,
.destroy = jb_destroy_adaptive,
.put_first = jb_put_first_adaptive,
.put = jb_put_adaptive,
.get = jb_get_adaptive,
.next = jb_next_adaptive,
.remove = jb_remove_adaptive,
.force_resync = jb_force_resynch_adaptive,
.empty_and_reset = jb_empty_and_reset_adaptive,
.is_late = jb_is_late_adaptive,
}
};
static int default_impl = 0;
/* Translations between impl and abstract return codes */
static const int fixed_to_abstract_code[] =
{AST_JB_IMPL_OK, AST_JB_IMPL_DROP, AST_JB_IMPL_INTERP, AST_JB_IMPL_NOFRAME};
static const int adaptive_to_abstract_code[] =
{AST_JB_IMPL_OK, AST_JB_IMPL_NOFRAME, AST_JB_IMPL_NOFRAME, AST_JB_IMPL_INTERP, AST_JB_IMPL_DROP, AST_JB_IMPL_OK};
/* JB_GET actions (used only for the frames log) */
static const char * const jb_get_actions[] = {"Delivered", "Dropped", "Interpolated", "No"};
/*! \brief Macros for the frame log files */
#define jb_framelog(...) do { \
if (jb->logfile) { \
fprintf(jb->logfile, __VA_ARGS__); \
fflush(jb->logfile); \
} \
} while (0)
/* Internal utility functions */
static void jb_choose_impl(struct ast_channel *chan);
static void jb_get_and_deliver(struct ast_channel *chan);
static int create_jb(struct ast_channel *chan, struct ast_frame *first_frame);
static long get_now(struct ast_jb *jb, struct timeval *tv);
/* Interface ast jb functions impl */
static void jb_choose_impl(struct ast_channel *chan)
{
struct ast_jb *jb = ast_channel_jb(chan);
struct ast_jb_conf *jbconf = &jb->conf;
const struct ast_jb_impl *test_impl;
int i, avail_impl_count = ARRAY_LEN(avail_impl);
jb->impl = &avail_impl[default_impl];
if (ast_strlen_zero(jbconf->impl)) {
return;
}
for (i = 0; i < avail_impl_count; i++) {
test_impl = &avail_impl[i];
if (!strcasecmp(jbconf->impl, test_impl->name)) {
jb->impl = test_impl;
return;
}
}
}
int ast_jb_do_usecheck(struct ast_channel *c0, struct ast_channel *c1)
{
struct ast_jb *jb0 = ast_channel_jb(c0);
struct ast_jb *jb1 = ast_channel_jb(c1);
struct ast_jb_conf *conf0 = &jb0->conf;
struct ast_jb_conf *conf1 = &jb1->conf;
int c0_wants_jitter = ast_channel_tech(c0)->properties & AST_CHAN_TP_WANTSJITTER;
int c0_creates_jitter = ast_channel_tech(c0)->properties & AST_CHAN_TP_CREATESJITTER;
int c0_jb_enabled = ast_test_flag(conf0, AST_JB_ENABLED);
int c0_force_jb = ast_test_flag(conf0, AST_JB_FORCED);
int c0_jb_timebase_initialized = ast_test_flag(jb0, JB_TIMEBASE_INITIALIZED);
int c0_jb_created = ast_test_flag(jb0, JB_CREATED);
int c1_wants_jitter = ast_channel_tech(c1)->properties & AST_CHAN_TP_WANTSJITTER;
int c1_creates_jitter = ast_channel_tech(c1)->properties & AST_CHAN_TP_CREATESJITTER;
int c1_jb_enabled = ast_test_flag(conf1, AST_JB_ENABLED);
int c1_force_jb = ast_test_flag(conf1, AST_JB_FORCED);
int c1_jb_timebase_initialized = ast_test_flag(jb1, JB_TIMEBASE_INITIALIZED);
int c1_jb_created = ast_test_flag(jb1, JB_CREATED);
int inuse = 0;
/* Determine whether audio going to c0 needs a jitter buffer */
if (((!c0_wants_jitter && c1_creates_jitter) || (c0_force_jb && c1_creates_jitter)) && c0_jb_enabled) {
ast_set_flag(jb0, JB_USE);
if (!c0_jb_timebase_initialized) {
if (c1_jb_timebase_initialized) {
memcpy(&jb0->timebase, &jb1->timebase, sizeof(struct timeval));
} else {
gettimeofday(&jb0->timebase, NULL);
}
ast_set_flag(jb0, JB_TIMEBASE_INITIALIZED);
}
if (!c0_jb_created) {
jb_choose_impl(c0);
}
inuse = 1;
}
/* Determine whether audio going to c1 needs a jitter buffer */
if (((!c1_wants_jitter && c0_creates_jitter) || (c1_force_jb && c0_creates_jitter)) && c1_jb_enabled) {
ast_set_flag(jb1, JB_USE);
if (!c1_jb_timebase_initialized) {
if (c0_jb_timebase_initialized) {
memcpy(&jb1->timebase, &jb0->timebase, sizeof(struct timeval));
} else {
gettimeofday(&jb1->timebase, NULL);
}
ast_set_flag(jb1, JB_TIMEBASE_INITIALIZED);
}
if (!c1_jb_created) {
jb_choose_impl(c1);
}
inuse = 1;
}
return inuse;
}
int ast_jb_get_when_to_wakeup(struct ast_channel *c0, struct ast_channel *c1, int time_left)
{
struct ast_jb *jb0 = ast_channel_jb(c0);
struct ast_jb *jb1 = ast_channel_jb(c1);
int c0_use_jb = ast_test_flag(jb0, JB_USE);
int c0_jb_is_created = ast_test_flag(jb0, JB_CREATED);
int c1_use_jb = ast_test_flag(jb1, JB_USE);
int c1_jb_is_created = ast_test_flag(jb1, JB_CREATED);
int wait, wait0, wait1;
struct timeval tv_now;
if (time_left == 0) {
/* No time left - the bridge will be retried */
/* TODO: Test disable this */
/*return 0;*/
}
if (time_left < 0) {
time_left = INT_MAX;
}
gettimeofday(&tv_now, NULL);
wait0 = (c0_use_jb && c0_jb_is_created) ? jb0->next - get_now(jb0, &tv_now) : time_left;
wait1 = (c1_use_jb && c1_jb_is_created) ? jb1->next - get_now(jb1, &tv_now) : time_left;
wait = wait0 < wait1 ? wait0 : wait1;
wait = wait < time_left ? wait : time_left;
if (wait == INT_MAX) {
wait = -1;
} else if (wait < 1) {
/* don't let wait=0, because this can cause the pbx thread to loop without any sleeping at all */
wait = 1;
}
return wait;
}
int ast_jb_put(struct ast_channel *chan, struct ast_frame *f)
{
struct ast_jb *jb = ast_channel_jb(chan);
const struct ast_jb_impl *jbimpl = jb->impl;
void *jbobj = jb->jbobj;
struct ast_frame *frr;
long now = 0;
if (!ast_test_flag(jb, JB_USE))
return -1;
if (f->frametype != AST_FRAME_VOICE) {
if (f->frametype == AST_FRAME_DTMF && ast_test_flag(jb, JB_CREATED)) {
jb_framelog("JB_PUT {now=%ld}: Received DTMF frame. Force resynching jb...\n", now);
jbimpl->force_resync(jbobj);
}
return -1;
}
/* We consider an enabled jitterbuffer should receive frames with valid timing info. */
if (!ast_test_flag(f, AST_FRFLAG_HAS_TIMING_INFO) || f->len < 2 || f->ts < 0) {
ast_log(LOG_WARNING, "%s received frame with invalid timing info: "
"has_timing_info=%u, len=%ld, ts=%ld, src=%s\n",
ast_channel_name(chan), ast_test_flag(f, AST_FRFLAG_HAS_TIMING_INFO), f->len, f->ts, f->src);
return -1;
}
frr = ast_frdup(f);
if (!frr) {
ast_log(LOG_ERROR, "Failed to isolate frame for the jitterbuffer on channel '%s'\n", ast_channel_name(chan));
return -1;
}
if (!ast_test_flag(jb, JB_CREATED)) {
if (create_jb(chan, frr)) {
ast_frfree(frr);
/* Disable the jitterbuffer */
ast_clear_flag(jb, JB_USE);
return -1;
}
ast_set_flag(jb, JB_CREATED);
return 0;
} else {
now = get_now(jb, NULL);
if (jbimpl->put(jbobj, frr, now) != AST_JB_IMPL_OK) {
jb_framelog("JB_PUT {now=%ld}: Dropped frame with ts=%ld and len=%ld\n", now, frr->ts, frr->len);
ast_frfree(frr);
/*return -1;*/
/* TODO: Check this fix - should return 0 here, because the dropped frame shouldn't
be delivered at all */
return 0;
}
jb->next = jbimpl->next(jbobj);
jb_framelog("JB_PUT {now=%ld}: Queued frame with ts=%ld and len=%ld\n", now, frr->ts, frr->len);
return 0;
}
}
void ast_jb_get_and_deliver(struct ast_channel *c0, struct ast_channel *c1)
{
struct ast_jb *jb0 = ast_channel_jb(c0);
struct ast_jb *jb1 = ast_channel_jb(c1);
int c0_use_jb = ast_test_flag(jb0, JB_USE);
int c0_jb_is_created = ast_test_flag(jb0, JB_CREATED);
int c1_use_jb = ast_test_flag(jb1, JB_USE);
int c1_jb_is_created = ast_test_flag(jb1, JB_CREATED);
if (c0_use_jb && c0_jb_is_created)
jb_get_and_deliver(c0);
if (c1_use_jb && c1_jb_is_created)
jb_get_and_deliver(c1);
}
static void jb_get_and_deliver(struct ast_channel *chan)
{
struct ast_jb *jb = ast_channel_jb(chan);
const struct ast_jb_impl *jbimpl = jb->impl;
void *jbobj = jb->jbobj;
struct ast_frame *f, finterp = { .frametype = AST_FRAME_VOICE, };
long now;
int interpolation_len, res;
now = get_now(jb, NULL);
jb->next = jbimpl->next(jbobj);
if (now < jb->next) {
jb_framelog("\tJB_GET {now=%ld}: now < next=%ld\n", now, jb->next);
return;
}
while (now >= jb->next) {
interpolation_len = ast_format_get_default_ms(jb->last_format);
res = jbimpl->get(jbobj, &f, now, interpolation_len);
switch (res) {
case AST_JB_IMPL_OK:
/* deliver the frame */
ast_write(chan, f);
case AST_JB_IMPL_DROP:
jb_framelog("\tJB_GET {now=%ld}: %s frame with ts=%ld and len=%ld\n",
now, jb_get_actions[res], f->ts, f->len);
ao2_replace(jb->last_format, f->subclass.format);
ast_frfree(f);
break;
case AST_JB_IMPL_INTERP:
/* interpolate a frame */
f = &finterp;
f->subclass.format = jb->last_format;
f->samples = interpolation_len * 8;
f->src = "JB interpolation";
f->delivery = ast_tvadd(jb->timebase, ast_samp2tv(jb->next, 1000));
f->offset = AST_FRIENDLY_OFFSET;
/* deliver the interpolated frame */
ast_write(chan, f);
jb_framelog("\tJB_GET {now=%ld}: Interpolated frame with len=%d\n", now, interpolation_len);
break;
case AST_JB_IMPL_NOFRAME:
ast_log(LOG_WARNING,
"AST_JB_IMPL_NOFRAME is returned from the %s jb when now=%ld >= next=%ld, jbnext=%ld!\n",
jbimpl->name, now, jb->next, jbimpl->next(jbobj));
jb_framelog("\tJB_GET {now=%ld}: No frame for now!?\n", now);
return;
default:
ast_log(LOG_ERROR, "This should never happen!\n");
ast_assert("JB type unknown" == NULL);
break;
}
jb->next = jbimpl->next(jbobj);
}
}
static int create_jb(struct ast_channel *chan, struct ast_frame *frr)
{
struct ast_jb *jb = ast_channel_jb(chan);
struct ast_jb_conf *jbconf = &jb->conf;
const struct ast_jb_impl *jbimpl = jb->impl;
void *jbobj;
long now;
char logfile_pathname[20 + AST_JB_IMPL_NAME_SIZE + 2*AST_CHANNEL_NAME + 1];
char name1[AST_CHANNEL_NAME], name2[AST_CHANNEL_NAME], *tmp;
int res;
jbobj = jb->jbobj = jbimpl->create(jbconf);
if (!jbobj) {
ast_log(LOG_WARNING, "Failed to create jitterbuffer on channel '%s'\n", ast_channel_name(chan));
return -1;
}
now = get_now(jb, NULL);
res = jbimpl->put_first(jbobj, frr, now);
/* The result of putting the first frame should not differ from OK. However, its possible
some implementations (i.e. adaptive's when resynch_threshold is specified) to drop it. */
if (res != AST_JB_IMPL_OK) {
ast_log(LOG_WARNING, "Failed to put first frame in the jitterbuffer on channel '%s'\n", ast_channel_name(chan));
/*
jbimpl->destroy(jbobj);
return -1;
*/
}
/* Init next */
jb->next = jbimpl->next(jbobj);
/* Init last format for a first time. */
jb->last_format = ao2_bump(frr->subclass.format);
/* Create a frame log file */
if (ast_test_flag(jbconf, AST_JB_LOG)) {
struct ast_channel *bridged = ast_channel_bridge_peer(chan);
char safe_logfile[30] = "/tmp/logfile-XXXXXX";
int safe_fd;
snprintf(name2, sizeof(name2), "%s", ast_channel_name(chan));
while ((tmp = strchr(name2, '/'))) {
*tmp = '#';
}
/* We should always have bridged chan if a jitterbuffer is in use */
ast_assert(bridged != NULL);
snprintf(name1, sizeof(name1), "%s", ast_channel_name(bridged));
while ((tmp = strchr(name1, '/'))) {
*tmp = '#';
}
snprintf(logfile_pathname, sizeof(logfile_pathname),
"/tmp/ast_%s_jb_%s--%s.log", jbimpl->name, name1, name2);
unlink(logfile_pathname);
safe_fd = mkstemp(safe_logfile);
if (safe_fd < 0 || link(safe_logfile, logfile_pathname) || unlink(safe_logfile) || !(jb->logfile = fdopen(safe_fd, "w+b"))) {
ast_log(LOG_ERROR, "Failed to create frame log file with pathname '%s': %s\n", logfile_pathname, strerror(errno));
jb->logfile = NULL;
if (safe_fd > -1) {
close(safe_fd);
}
}
if (res == AST_JB_IMPL_OK) {
jb_framelog("JB_PUT_FIRST {now=%ld}: Queued frame with ts=%ld and len=%ld\n",
now, frr->ts, frr->len);
} else {
jb_framelog("JB_PUT_FIRST {now=%ld}: Dropped frame with ts=%ld and len=%ld\n",
now, frr->ts, frr->len);
}
ast_channel_cleanup(bridged);
}
ast_verb(3, "%s jitterbuffer created on channel %s\n", jbimpl->name, ast_channel_name(chan));
/* Free the frame if it has not been queued in the jb */
if (res != AST_JB_IMPL_OK) {
ast_frfree(frr);
}
return 0;
}
void ast_jb_destroy(struct ast_channel *chan)
{
struct ast_jb *jb = ast_channel_jb(chan);
const struct ast_jb_impl *jbimpl = jb->impl;
void *jbobj = jb->jbobj;
struct ast_frame *f;
if (jb->logfile) {
fclose(jb->logfile);
jb->logfile = NULL;
}
ao2_cleanup(jb->last_format);
if (ast_test_flag(jb, JB_CREATED)) {
/* Remove and free all frames still queued in jb */
while (jbimpl->remove(jbobj, &f) == AST_JB_IMPL_OK) {
ast_frfree(f);
}
jbimpl->destroy(jbobj);
jb->jbobj = NULL;
ast_clear_flag(jb, JB_CREATED);
ast_verb(3, "%s jitterbuffer destroyed on channel %s\n", jbimpl->name, ast_channel_name(chan));
}
}
static long get_now(struct ast_jb *jb, struct timeval *when)
{
struct timeval now;
if (!when) {
when = &now;
gettimeofday(when, NULL);
}
return ast_tvdiff_ms(*when, jb->timebase);
}
int ast_jb_read_conf(struct ast_jb_conf *conf, const char *varname, const char *value)
{
int prefixlen = sizeof(AST_JB_CONF_PREFIX) - 1;
const char *name;
int tmp;
if (strncasecmp(AST_JB_CONF_PREFIX, varname, prefixlen)) {
return -1;
}
name = varname + prefixlen;
if (!strcasecmp(name, AST_JB_CONF_ENABLE)) {
ast_set2_flag(conf, ast_true(value), AST_JB_ENABLED);
} else if (!strcasecmp(name, AST_JB_CONF_FORCE)) {
ast_set2_flag(conf, ast_true(value), AST_JB_FORCED);
} else if (!strcasecmp(name, AST_JB_CONF_MAX_SIZE)) {
if ((tmp = atoi(value)) > 0)
conf->max_size = tmp;
} else if (!strcasecmp(name, AST_JB_CONF_RESYNCH_THRESHOLD)) {
if ((tmp = atoi(value)) > 0)
conf->resync_threshold = tmp;
} else if (!strcasecmp(name, AST_JB_CONF_IMPL)) {
if (!ast_strlen_zero(value))
snprintf(conf->impl, sizeof(conf->impl), "%s", value);
} else if (!strcasecmp(name, AST_JB_CONF_TARGET_EXTRA)) {
if (sscanf(value, "%30d", &tmp) == 1) {
conf->target_extra = tmp;
}
} else if (!strcasecmp(name, AST_JB_CONF_LOG)) {
ast_set2_flag(conf, ast_true(value), AST_JB_LOG);
} else if (!strcasecmp(name, AST_JB_CONF_SYNC_VIDEO)) {
ast_set2_flag(conf, ast_true(value), AST_JB_SYNC_VIDEO);
} else {
return -1;
}
return 0;
}
void ast_jb_enable_for_channel(struct ast_channel *chan)
{
struct ast_jb_conf conf = ast_channel_jb(chan)->conf;
if (ast_test_flag(&conf, AST_JB_ENABLED)) {
ast_jb_create_framehook(chan, &conf, 1);
}
}
void ast_jb_configure(struct ast_channel *chan, const struct ast_jb_conf *conf)
{
memcpy(&ast_channel_jb(chan)->conf, conf, sizeof(*conf));
}
void ast_jb_get_config(const struct ast_channel *chan, struct ast_jb_conf *conf)
{
memcpy(conf, &ast_channel_jb((struct ast_channel *) chan)->conf, sizeof(*conf));
}
void ast_jb_empty_and_reset(struct ast_channel *c0, struct ast_channel *c1)
{
struct ast_jb *jb0 = ast_channel_jb(c0);
struct ast_jb *jb1 = ast_channel_jb(c1);
int c0_use_jb = ast_test_flag(jb0, JB_USE);
int c0_jb_is_created = ast_test_flag(jb0, JB_CREATED);
int c1_use_jb = ast_test_flag(jb1, JB_USE);
int c1_jb_is_created = ast_test_flag(jb1, JB_CREATED);
if (c0_use_jb && c0_jb_is_created && jb0->impl->empty_and_reset) {
jb0->impl->empty_and_reset(jb0->jbobj);
}
if (c1_use_jb && c1_jb_is_created && jb1->impl->empty_and_reset) {
jb1->impl->empty_and_reset(jb1->jbobj);
}
}
/* Implementation functions */
/* fixed */
static void * jb_create_fixed(struct ast_jb_conf *general_config)
{
struct fixed_jb_conf conf;
conf.jbsize = general_config->max_size;
conf.resync_threshold = general_config->resync_threshold;
return fixed_jb_new(&conf);
}
static void jb_destroy_fixed(void *jb)
{
struct fixed_jb *fixedjb = (struct fixed_jb *) jb;
/* Ensure the fixed jb is empty - otherwise it will raise an ASSERT */
jb_empty_and_reset_fixed(jb);
/* destroy the jb */
fixed_jb_destroy(fixedjb);
}
static int jb_put_first_fixed(void *jb, struct ast_frame *fin, long now)
{
struct fixed_jb *fixedjb = (struct fixed_jb *) jb;
int res;
res = fixed_jb_put_first(fixedjb, fin, fin->len, fin->ts, now);
return fixed_to_abstract_code[res];
}
static int jb_put_fixed(void *jb, struct ast_frame *fin, long now)
{
struct fixed_jb *fixedjb = (struct fixed_jb *) jb;
int res;
res = fixed_jb_put(fixedjb, fin, fin->len, fin->ts, now);
return fixed_to_abstract_code[res];
}
static int jb_get_fixed(void *jb, struct ast_frame **fout, long now, long interpl)
{
struct fixed_jb *fixedjb = (struct fixed_jb *) jb;
struct fixed_jb_frame frame = { .data = &ast_null_frame };
int res;
res = fixed_jb_get(fixedjb, &frame, now, interpl);
*fout = frame.data;
return fixed_to_abstract_code[res];
}
static long jb_next_fixed(void *jb)
{
struct fixed_jb *fixedjb = (struct fixed_jb *) jb;
return fixed_jb_next(fixedjb);
}
static int jb_remove_fixed(void *jb, struct ast_frame **fout)
{
struct fixed_jb *fixedjb = (struct fixed_jb *) jb;
struct fixed_jb_frame frame;
int res;
res = fixed_jb_remove(fixedjb, &frame);
*fout = frame.data;
return fixed_to_abstract_code[res];
}
static void jb_force_resynch_fixed(void *jb)
{
struct fixed_jb *fixedjb = (struct fixed_jb *) jb;
fixed_jb_set_force_resynch(fixedjb);
}
static void jb_empty_and_reset_fixed(void *jb)
{
struct fixed_jb *fixedjb = jb;
struct fixed_jb_frame f;
while (fixed_jb_remove(fixedjb, &f) == FIXED_JB_OK) {
ast_frfree(f.data);
}
}
static int jb_is_late_fixed(void *jb, long ts)
{
return fixed_jb_is_late(jb, ts);
}
/* adaptive */
static void *jb_create_adaptive(struct ast_jb_conf *general_config)
{
jb_conf jbconf;
jitterbuf *adaptivejb;
adaptivejb = jb_new();
if (adaptivejb) {
jbconf.max_jitterbuf = general_config->max_size;
jbconf.resync_threshold = general_config->resync_threshold;
jbconf.max_contig_interp = 10;
jbconf.target_extra = general_config->target_extra;
jb_setconf(adaptivejb, &jbconf);
}
return adaptivejb;
}
static void jb_destroy_adaptive(void *jb)
{
jitterbuf *adaptivejb = (jitterbuf *) jb;
jb_destroy(adaptivejb);
}
static int jb_put_first_adaptive(void *jb, struct ast_frame *fin, long now)
{
return jb_put_adaptive(jb, fin, now);
}
static int jb_put_adaptive(void *jb, struct ast_frame *fin, long now)
{
jitterbuf *adaptivejb = (jitterbuf *) jb;
int res;
res = jb_put(adaptivejb, fin, JB_TYPE_VOICE, fin->len, fin->ts, now);
return adaptive_to_abstract_code[res];
}
static int jb_get_adaptive(void *jb, struct ast_frame **fout, long now, long interpl)
{
jitterbuf *adaptivejb = (jitterbuf *) jb;
jb_frame frame = { .data = &ast_null_frame };
int res;
res = jb_get(adaptivejb, &frame, now, interpl);
*fout = frame.data;
return adaptive_to_abstract_code[res];
}
static long jb_next_adaptive(void *jb)
{
jitterbuf *adaptivejb = (jitterbuf *) jb;
return jb_next(adaptivejb);
}
static int jb_remove_adaptive(void *jb, struct ast_frame **fout)
{
jitterbuf *adaptivejb = (jitterbuf *) jb;
jb_frame frame;
int res;
res = jb_getall(adaptivejb, &frame);
*fout = frame.data;
return adaptive_to_abstract_code[res];
}
static void jb_force_resynch_adaptive(void *jb)
{
}
static void jb_empty_and_reset_adaptive(void *jb)
{
jitterbuf *adaptivejb = jb;
jb_frame f;
while (jb_getall(adaptivejb, &f) == JB_OK) {
ast_frfree(f.data);
}
jb_reset(adaptivejb);
}
const struct ast_jb_impl *ast_jb_get_impl(enum ast_jb_type type)
{
int i;
for (i = 0; i < ARRAY_LEN(avail_impl); i++) {
if (avail_impl[i].type == type) {
return &avail_impl[i];
}
}
return NULL;
}
static int jb_is_late_adaptive(void *jb, long ts)
{
return jb_is_late(jb, ts);
}
#define DEFAULT_TIMER_INTERVAL 20
#define DEFAULT_SIZE 200
#define DEFAULT_TARGET_EXTRA 40
#define DEFAULT_RESYNC 1000
#define DEFAULT_TYPE AST_JB_FIXED
struct jb_stream_sync {
unsigned int timestamp;
struct timeval ntp;
};
struct jb_framedata {
const struct ast_jb_impl *jb_impl;
struct ast_jb_conf jb_conf;
struct timeval start_tv;
struct ast_format *last_format;
struct ast_timer *timer;
int timer_interval; /* ms between deliveries */
int timer_fd;
int first;
int audio_stream_id;
struct jb_stream_sync audio_stream_sync;
int video_stream_id;
struct jb_stream_sync video_stream_sync;
AST_LIST_HEAD_NOLOCK(, ast_frame) early_frames;
unsigned int early_frame_count;
struct timeval last_audio_ntp_timestamp;
int audio_flowing;
void *jb_obj;
};
static void jb_framedata_destroy(struct jb_framedata *framedata)
{
struct ast_frame *frame;
if (framedata->timer) {
ast_timer_close(framedata->timer);
framedata->timer = NULL;
}
if (framedata->jb_impl && framedata->jb_obj) {
struct ast_frame *f;
while (framedata->jb_impl->remove(framedata->jb_obj, &f) == AST_JB_IMPL_OK) {
ast_frfree(f);
}
framedata->jb_impl->destroy(framedata->jb_obj);
framedata->jb_obj = NULL;
}
ao2_cleanup(framedata->last_format);
while ((frame = AST_LIST_REMOVE_HEAD(&framedata->early_frames, frame_list))) {
ast_frfree(frame);
}
ast_free(framedata);
}
void ast_jb_conf_default(struct ast_jb_conf *conf)
{
ast_clear_flag(conf, AST_FLAGS_ALL);
conf->max_size = DEFAULT_SIZE;
conf->resync_threshold = DEFAULT_RESYNC;
ast_copy_string(conf->impl, "fixed", sizeof(conf->impl));
conf->target_extra = DEFAULT_TARGET_EXTRA;
}
static void datastore_destroy_cb(void *data) {
ast_free(data);
ast_debug(1, "JITTERBUFFER datastore destroyed\n");
}
static const struct ast_datastore_info jb_datastore = {
.type = "jitterbuffer",
.destroy = datastore_destroy_cb
};
static void hook_destroy_cb(void *framedata)
{
ast_debug(1, "JITTERBUFFER hook destroyed\n");
jb_framedata_destroy((struct jb_framedata *) framedata);
}
static struct timeval jitterbuffer_frame_get_ntp_timestamp(const struct jb_stream_sync *stream_sync, const struct ast_frame *frame)
{
int timestamp_diff;
unsigned int rate;
/* It's possible for us to receive frames before we receive the information allowing
* us to do NTP/RTP timestamp calculations. Since the information isn't available we
* can't generate one and give an empty timestamp.
*/
if (ast_tvzero(stream_sync->ntp)) {
return ast_tv(0, 0);
}
/* Convert the Asterisk timestamp into an RTP timestamp, and then based on the difference we can
* determine how many samples are in the frame and how long has elapsed since the synchronization
* RTP and NTP timestamps were received giving us the NTP timestamp for this frame.
*/
if (frame->frametype == AST_FRAME_VOICE) {
rate = ast_rtp_get_rate(frame->subclass.format);
timestamp_diff = (frame->ts * (rate / 1000)) - stream_sync->timestamp;
} else {
/* Video is special - internally we reference it as 1000 to preserve the RTP timestamp but
* it is actually 90000, this is why we can just directly subtract the timestamp.
*/
rate = 90000;
timestamp_diff = frame->ts - stream_sync->timestamp;
}
if (timestamp_diff < 0) {
/* It's possible for us to be asked for an NTP timestamp from before our latest
* RTCP SR report. To handle this we subtract so we go back in time.
*/
return ast_tvsub(stream_sync->ntp, ast_samp2tv(abs(timestamp_diff), rate));
} else {
return ast_tvadd(stream_sync->ntp, ast_samp2tv(timestamp_diff, rate));
}
}
static struct ast_frame *hook_event_cb(struct ast_channel *chan, struct ast_frame *frame, enum ast_framehook_event event, void *data)
{
struct jb_framedata *framedata = data;
struct timeval now_tv;
int64_t relative_frame_start;
int putframe = 0; /* signifies if audio frame was placed into the buffer or not */
switch (event) {
case AST_FRAMEHOOK_EVENT_READ:
break;
case AST_FRAMEHOOK_EVENT_ATTACHED:
case AST_FRAMEHOOK_EVENT_DETACHED:
case AST_FRAMEHOOK_EVENT_WRITE:
return frame;
}
if (ast_channel_fdno(chan) == AST_JITTERBUFFER_FD && framedata->timer) {
if (ast_timer_ack(framedata->timer, 1) < 0) {
ast_log(LOG_ERROR, "Failed to acknowledge timer in jitter buffer\n");
return frame;
}
}
/*
* If the frame has been requeued (for instance when the translate core returns
* more than one frame) then if the frame is late we want to immediately return
* it. Otherwise attempt to insert it into the jitterbuffer.
*
* If the frame is requeued and late then in all likely hood it's a frame that
* that was previously retrieved from the jitterbuffer, passed to the translate
* core, and then put back into the channel read queue. Even if it had not been
* in the jitterbuffer prior to now it needs to be the next frame "out".
*
* However late arriving frames that have not been requeued (i.e. regular frames)
* need to be passed to the jitterbuffer so they can be appropriately dropped. As
* well any requeued frames that are not late should be put into the jitterbuffer.
*/
if (!frame || (ast_test_flag(frame, AST_FRFLAG_REQUEUED) &&
framedata->jb_impl->is_late(framedata->jb_obj, frame->ts))) {
return frame;
}
if (ast_test_flag(&framedata->jb_conf, AST_JB_SYNC_VIDEO)) {
if (frame->frametype == AST_FRAME_VOICE) {
/* Store the stream identifier for the audio stream so we can associate the incoming RTCP SR
* with the correct stream sync structure.
*/
framedata->audio_stream_id = frame->stream_num;
} else if (frame->frametype == AST_FRAME_RTCP && frame->subclass.integer == AST_RTP_RTCP_SR) {
struct ast_rtp_rtcp_report *rtcp_report = frame->data.ptr;
struct jb_stream_sync *stream_sync = NULL;
/* Determine which stream this RTCP is in regards to */
if (framedata->audio_stream_id == frame->stream_num) {
stream_sync = &framedata->audio_stream_sync;
} else if (framedata->video_stream_id == frame->stream_num) {
stream_sync = &framedata->video_stream_sync;
}
if (stream_sync) {
/* Store the RTP and NTP timestamp mapping so we can derive an NTP timestamp for each frame */
stream_sync->timestamp = rtcp_report->sender_information.rtp_timestamp;
stream_sync->ntp = rtcp_report->sender_information.ntp_timestamp;
}
} else if (frame->frametype == AST_FRAME_VIDEO) {
/* If a video frame is late according to the audio timestamp don't stash it away, just return it.
* If however it is ahead then we keep it until such time as the audio catches up.
*/
struct ast_frame *jbframe;
framedata->video_stream_id = frame->stream_num;
/* If no timing information is available we can't store this away, so just let it through now */
if (!ast_test_flag(frame, AST_FRFLAG_HAS_TIMING_INFO)) {
return frame;
}
/* To ensure that the video starts when the audio starts we only start allowing frames through once
* audio starts flowing.
*/
if (framedata->audio_flowing) {
struct timeval video_timestamp;
video_timestamp = jitterbuffer_frame_get_ntp_timestamp(&framedata->video_stream_sync, frame);
if (ast_tvdiff_ms(framedata->last_audio_ntp_timestamp, video_timestamp) >= 0) {
return frame;
}
}
/* To prevent the early frame buffer from growing uncontrolled we impose a maximum count that it can
* get to. If this is reached then we drop a video frame, which should cause the receiver to ask for a
* new key frame.
*/
if (framedata->early_frame_count == MAXIMUM_EARLY_FRAME_COUNT) {
jbframe = AST_LIST_REMOVE_HEAD(&framedata->early_frames, frame_list);
framedata->early_frame_count--;
ast_frfree(jbframe);
}
jbframe = ast_frisolate(frame);
if (!jbframe) {
/* If we can't isolate the frame the safest thing we can do is return it, even if the A/V sync
* may be off.
*/
return frame;
}
AST_LIST_INSERT_TAIL(&framedata->early_frames, jbframe, frame_list);
framedata->early_frame_count++;
return &ast_null_frame;
}
}
now_tv = ast_tvnow();
relative_frame_start = ast_tvdiff_ms(now_tv, framedata->start_tv);
if (relative_frame_start < 0) {
/*
* The only way for this to happen is if the system time has
* stepped backwards between the time framedata->start_tv was
* set and now. Think an ntpd or systemd-timesyncd adjustment.
*
* Just pass the frame through.
*/
return frame;
}
if (frame->frametype == AST_FRAME_VOICE) {
int res;
struct ast_frame *jbframe;
if (!ast_test_flag(frame, AST_FRFLAG_HAS_TIMING_INFO) || frame->len < 2 || frame->ts < 0) {
/* only frames with timing info can enter the jitterbuffer */
return frame;
}
jbframe = ast_frisolate(frame);
ao2_replace(framedata->last_format, frame->subclass.format);
if (frame->len && (frame->len != framedata->timer_interval)) {
framedata->timer_interval = frame->len;
ast_timer_set_rate(framedata->timer, 1000 / framedata->timer_interval);
}
if (!framedata->first) {
framedata->first = 1;
res = framedata->jb_impl->put_first(framedata->jb_obj, jbframe, relative_frame_start);
} else {
res = framedata->jb_impl->put(framedata->jb_obj, jbframe, relative_frame_start);
}
if (res == AST_JB_IMPL_OK) {
if (jbframe != frame) {
ast_frfree(frame);
}
frame = &ast_null_frame;
} else if (jbframe != frame) {
ast_frfree(jbframe);
}
putframe = 1;
}
if (frame->frametype == AST_FRAME_NULL) {
int res;
long next = framedata->jb_impl->next(framedata->jb_obj);
/* If relative_frame_start is earlier than the next expected output frame
* from the jitterbuffer we may choose to pass on retrieving
* a frame during this read iteration. The only exception
* to this rule is when an audio frame is placed into the buffer
* and the time for the next frame to come out of the buffer is
* at least within the timer_interval of the next output frame. By
* doing this we are able to feed off the timing of the input frames
* and only rely on our jitterbuffer timer when frames are dropped.
* During testing, this hybrid form of timing gave more reliable results. */
if (relative_frame_start < next) {
long int diff = next - relative_frame_start;
if (!putframe) {
return frame;
} else if (diff >= framedata->timer_interval) {
return frame;
}
}
ast_frfree(frame);
frame = &ast_null_frame;
res = framedata->jb_impl->get(framedata->jb_obj, &frame, relative_frame_start, framedata->timer_interval);
switch (res) {
case AST_JB_IMPL_OK:
/* got it, and pass it through */
break;
case AST_JB_IMPL_DROP:
ast_frfree(frame);
frame = &ast_null_frame;
break;
case AST_JB_IMPL_INTERP:
if (framedata->last_format) {
struct ast_frame tmp = { 0, };
tmp.frametype = AST_FRAME_VOICE;
tmp.subclass.format = framedata->last_format;
/* example: 8000hz / (1000 / 20ms) = 160 samples */
tmp.samples = ast_format_get_sample_rate(framedata->last_format) / (1000 / framedata->timer_interval);
tmp.delivery = ast_tvadd(framedata->start_tv, ast_samp2tv(next, 1000));
tmp.offset = AST_FRIENDLY_OFFSET;
tmp.src = "func_jitterbuffer interpolation";
ast_frfree(frame);
frame = ast_frdup(&tmp);
break;
}
/* else fall through */
case AST_JB_IMPL_NOFRAME:
ast_frfree(frame);
frame = &ast_null_frame;
break;
}
}
if (frame->frametype == AST_FRAME_CONTROL) {
struct ast_frame *early_frame;
switch(frame->subclass.integer) {
case AST_CONTROL_HOLD:
case AST_CONTROL_UNHOLD:
case AST_CONTROL_T38_PARAMETERS:
case AST_CONTROL_SRCUPDATE:
case AST_CONTROL_SRCCHANGE:
framedata->jb_impl->force_resync(framedata->jb_obj);
/* Since we are resyncing go ahead and clear out the video frames too */
while ((early_frame = AST_LIST_REMOVE_HEAD(&framedata->early_frames, frame_list))) {
ast_frfree(early_frame);
}
framedata->audio_flowing = 0;
framedata->early_frame_count = 0;
break;
default:
break;
}
}
/* If a voice frame is being passed through see if we need to add any additional frames to it */
if (ast_test_flag(&framedata->jb_conf, AST_JB_SYNC_VIDEO) && frame->frametype == AST_FRAME_VOICE) {
AST_LIST_HEAD_NOLOCK(, ast_frame) additional_frames;
struct ast_frame *early_frame;
/* We store the last NTP timestamp for the audio given to the core so that subsequents frames which
* are late can be passed immediately through (this will occur for video frames which are returned here)
*/
framedata->last_audio_ntp_timestamp = jitterbuffer_frame_get_ntp_timestamp(&framedata->audio_stream_sync, frame);
framedata->audio_flowing = 1;
AST_LIST_HEAD_INIT_NOLOCK(&additional_frames);
AST_LIST_TRAVERSE_SAFE_BEGIN(&framedata->early_frames, early_frame, frame_list) {
struct timeval early_timestamp = jitterbuffer_frame_get_ntp_timestamp(&framedata->video_stream_sync, early_frame);
int diff = ast_tvdiff_ms(framedata->last_audio_ntp_timestamp, early_timestamp);
/* If this frame is from the past we need to include it with the audio frame that is going
* out.
*/
if (diff >= 0) {
AST_LIST_REMOVE_CURRENT(frame_list);
framedata->early_frame_count--;
AST_LIST_INSERT_TAIL(&additional_frames, early_frame, frame_list);
}
}
AST_LIST_TRAVERSE_SAFE_END;
/* Append any additional frames we may want to include (such as video) */
AST_LIST_NEXT(frame, frame_list) = AST_LIST_FIRST(&additional_frames);
}
return frame;
}
/* set defaults */
static int jb_framedata_init(struct jb_framedata *framedata, struct ast_jb_conf *jb_conf)
{
int jb_impl_type = DEFAULT_TYPE;
/* Initialize defaults */
framedata->timer_fd = -1;
memcpy(&framedata->jb_conf, jb_conf, sizeof(*jb_conf));
/* Figure out implementation type from the configuration implementation string */
if (!ast_strlen_zero(jb_conf->impl)) {
if (!strcasecmp(jb_conf->impl, "fixed")) {
jb_impl_type = AST_JB_FIXED;
} else if (!strcasecmp(jb_conf->impl, "adaptive")) {
jb_impl_type = AST_JB_ADAPTIVE;
} else {
ast_log(LOG_WARNING, "Unknown Jitterbuffer type %s. Failed to create jitterbuffer.\n", jb_conf->impl);
return -1;
}
}
if (!(framedata->jb_impl = ast_jb_get_impl(jb_impl_type))) {
return -1;
}
if (!(framedata->timer = ast_timer_open())) {
return -1;
}
framedata->audio_stream_id = -1;
framedata->video_stream_id = -1;
AST_LIST_HEAD_INIT_NOLOCK(&framedata->early_frames);
framedata->timer_fd = ast_timer_fd(framedata->timer);
framedata->timer_interval = DEFAULT_TIMER_INTERVAL;
ast_timer_set_rate(framedata->timer, 1000 / framedata->timer_interval);
framedata->start_tv = ast_tvnow();
framedata->jb_obj = framedata->jb_impl->create(&framedata->jb_conf);
return 0;
}
void ast_jb_create_framehook(struct ast_channel *chan, struct ast_jb_conf *jb_conf, int prefer_existing)
{
struct jb_framedata *framedata;
struct ast_datastore *datastore = NULL;
struct ast_framehook_interface interface = {
.version = AST_FRAMEHOOK_INTERFACE_VERSION,
.event_cb = hook_event_cb,
.destroy_cb = hook_destroy_cb,
};
int i = 0;
/* If disabled, strip any existing jitterbuffer and don't replace it. */
if (!strcasecmp(jb_conf->impl, "disabled")) {
int *id;
ast_channel_lock(chan);
if ((datastore = ast_channel_datastore_find(chan, &jb_datastore, NULL))) {
id = datastore->data;
ast_framehook_detach(chan, *id);
ast_channel_datastore_remove(chan, datastore);
ast_datastore_free(datastore);
}
ast_channel_unlock(chan);
return;
}
if (!(framedata = ast_calloc(1, sizeof(*framedata)))) {
return;
}
if (jb_framedata_init(framedata, jb_conf)) {
jb_framedata_destroy(framedata);
return;
}
interface.data = framedata;
ast_channel_lock(chan);
i = ast_framehook_attach(chan, &interface);
if (i >= 0) {
int *id;
if ((datastore = ast_channel_datastore_find(chan, &jb_datastore, NULL))) {
/* There is already a jitterbuffer on the channel. */
if (prefer_existing) {
/* We prefer the existing jitterbuffer, so remove the new one and keep the old one. */
ast_framehook_detach(chan, i);
ast_channel_unlock(chan);
return;
}
/* We prefer the new jitterbuffer, so strip the old one. */
id = datastore->data;
ast_framehook_detach(chan, *id);
ast_channel_datastore_remove(chan, datastore);
ast_datastore_free(datastore);
}
if (!(datastore = ast_datastore_alloc(&jb_datastore, NULL))) {
ast_framehook_detach(chan, i);
ast_channel_unlock(chan);
return;
}
if (!(id = ast_calloc(1, sizeof(int)))) {
ast_datastore_free(datastore);
ast_framehook_detach(chan, i);
ast_channel_unlock(chan);
return;
}
*id = i; /* Store off the id. The channel is still locked so it is safe to access this ptr. */
datastore->data = id;
ast_channel_datastore_add(chan, datastore);
ast_channel_set_fd(chan, AST_JITTERBUFFER_FD, framedata->timer_fd);
} else {
jb_framedata_destroy(framedata);
framedata = NULL;
}
ast_channel_unlock(chan);
}