mirror of
https://github.com/librempeg/librempeg
synced 2024-11-23 11:39:47 +00:00
d92f38c179
This is possible now that the next-API is gone. Signed-off-by: Andreas Rheinhardt <andreas.rheinhardt@outlook.com> Signed-off-by: James Almer <jamrial@gmail.com>
400 lines
13 KiB
C
400 lines
13 KiB
C
/*
|
|
* Pulseaudio input
|
|
* Copyright (c) 2011 Luca Barbato <lu_zero@gentoo.org>
|
|
* Copyright 2004-2006 Lennart Poettering
|
|
* Copyright (c) 2014 Michael Niedermayer <michaelni@gmx.at>
|
|
*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
#include <pulse/rtclock.h>
|
|
#include <pulse/error.h>
|
|
|
|
#include "libavutil/internal.h"
|
|
#include "libavutil/opt.h"
|
|
#include "libavutil/time.h"
|
|
|
|
#include "libavformat/avformat.h"
|
|
#include "libavformat/internal.h"
|
|
#include "pulse_audio_common.h"
|
|
#include "timefilter.h"
|
|
|
|
#define DEFAULT_CODEC_ID AV_NE(AV_CODEC_ID_PCM_S16BE, AV_CODEC_ID_PCM_S16LE)
|
|
|
|
typedef struct PulseData {
|
|
AVClass *class;
|
|
char *server;
|
|
char *name;
|
|
char *stream_name;
|
|
int sample_rate;
|
|
int channels;
|
|
int frame_size;
|
|
int fragment_size;
|
|
|
|
pa_threaded_mainloop *mainloop;
|
|
pa_context *context;
|
|
pa_stream *stream;
|
|
size_t pa_frame_size;
|
|
|
|
TimeFilter *timefilter;
|
|
int last_period;
|
|
int wallclock;
|
|
} PulseData;
|
|
|
|
|
|
#define CHECK_SUCCESS_GOTO(rerror, expression, label) \
|
|
do { \
|
|
if (!(expression)) { \
|
|
rerror = AVERROR_EXTERNAL; \
|
|
goto label; \
|
|
} \
|
|
} while (0)
|
|
|
|
#define CHECK_DEAD_GOTO(p, rerror, label) \
|
|
do { \
|
|
if (!(p)->context || !PA_CONTEXT_IS_GOOD(pa_context_get_state((p)->context)) || \
|
|
!(p)->stream || !PA_STREAM_IS_GOOD(pa_stream_get_state((p)->stream))) { \
|
|
rerror = AVERROR_EXTERNAL; \
|
|
goto label; \
|
|
} \
|
|
} while (0)
|
|
|
|
static void context_state_cb(pa_context *c, void *userdata) {
|
|
PulseData *p = userdata;
|
|
|
|
switch (pa_context_get_state(c)) {
|
|
case PA_CONTEXT_READY:
|
|
case PA_CONTEXT_TERMINATED:
|
|
case PA_CONTEXT_FAILED:
|
|
pa_threaded_mainloop_signal(p->mainloop, 0);
|
|
break;
|
|
}
|
|
}
|
|
|
|
static void stream_state_cb(pa_stream *s, void * userdata) {
|
|
PulseData *p = userdata;
|
|
|
|
switch (pa_stream_get_state(s)) {
|
|
case PA_STREAM_READY:
|
|
case PA_STREAM_FAILED:
|
|
case PA_STREAM_TERMINATED:
|
|
pa_threaded_mainloop_signal(p->mainloop, 0);
|
|
break;
|
|
}
|
|
}
|
|
|
|
static void stream_request_cb(pa_stream *s, size_t length, void *userdata) {
|
|
PulseData *p = userdata;
|
|
|
|
pa_threaded_mainloop_signal(p->mainloop, 0);
|
|
}
|
|
|
|
static void stream_latency_update_cb(pa_stream *s, void *userdata) {
|
|
PulseData *p = userdata;
|
|
|
|
pa_threaded_mainloop_signal(p->mainloop, 0);
|
|
}
|
|
|
|
static av_cold int pulse_close(AVFormatContext *s)
|
|
{
|
|
PulseData *pd = s->priv_data;
|
|
|
|
if (pd->mainloop)
|
|
pa_threaded_mainloop_stop(pd->mainloop);
|
|
|
|
if (pd->stream)
|
|
pa_stream_unref(pd->stream);
|
|
pd->stream = NULL;
|
|
|
|
if (pd->context) {
|
|
pa_context_disconnect(pd->context);
|
|
pa_context_unref(pd->context);
|
|
}
|
|
pd->context = NULL;
|
|
|
|
if (pd->mainloop)
|
|
pa_threaded_mainloop_free(pd->mainloop);
|
|
pd->mainloop = NULL;
|
|
|
|
ff_timefilter_destroy(pd->timefilter);
|
|
pd->timefilter = NULL;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static av_cold int pulse_read_header(AVFormatContext *s)
|
|
{
|
|
PulseData *pd = s->priv_data;
|
|
AVStream *st;
|
|
char *device = NULL;
|
|
int ret;
|
|
enum AVCodecID codec_id =
|
|
s->audio_codec_id == AV_CODEC_ID_NONE ? DEFAULT_CODEC_ID : s->audio_codec_id;
|
|
const pa_sample_spec ss = { ff_codec_id_to_pulse_format(codec_id),
|
|
pd->sample_rate,
|
|
pd->channels };
|
|
|
|
pa_buffer_attr attr = { -1 };
|
|
pa_channel_map cmap;
|
|
const pa_buffer_attr *queried_attr;
|
|
|
|
pa_channel_map_init_extend(&cmap, pd->channels, PA_CHANNEL_MAP_WAVEEX);
|
|
|
|
st = avformat_new_stream(s, NULL);
|
|
|
|
if (!st) {
|
|
av_log(s, AV_LOG_ERROR, "Cannot add stream\n");
|
|
return AVERROR(ENOMEM);
|
|
}
|
|
|
|
attr.fragsize = pd->fragment_size;
|
|
|
|
if (s->url[0] != '\0' && strcmp(s->url, "default"))
|
|
device = s->url;
|
|
|
|
if (!(pd->mainloop = pa_threaded_mainloop_new())) {
|
|
pulse_close(s);
|
|
return AVERROR_EXTERNAL;
|
|
}
|
|
|
|
if (!(pd->context = pa_context_new(pa_threaded_mainloop_get_api(pd->mainloop), pd->name))) {
|
|
pulse_close(s);
|
|
return AVERROR_EXTERNAL;
|
|
}
|
|
|
|
pa_context_set_state_callback(pd->context, context_state_cb, pd);
|
|
|
|
if (pa_context_connect(pd->context, pd->server, 0, NULL) < 0) {
|
|
pulse_close(s);
|
|
return AVERROR(pa_context_errno(pd->context));
|
|
}
|
|
|
|
pa_threaded_mainloop_lock(pd->mainloop);
|
|
|
|
if (pa_threaded_mainloop_start(pd->mainloop) < 0) {
|
|
ret = -1;
|
|
goto unlock_and_fail;
|
|
}
|
|
|
|
for (;;) {
|
|
pa_context_state_t state;
|
|
|
|
state = pa_context_get_state(pd->context);
|
|
|
|
if (state == PA_CONTEXT_READY)
|
|
break;
|
|
|
|
if (!PA_CONTEXT_IS_GOOD(state)) {
|
|
ret = AVERROR(pa_context_errno(pd->context));
|
|
goto unlock_and_fail;
|
|
}
|
|
|
|
/* Wait until the context is ready */
|
|
pa_threaded_mainloop_wait(pd->mainloop);
|
|
}
|
|
|
|
if (!(pd->stream = pa_stream_new(pd->context, pd->stream_name, &ss, &cmap))) {
|
|
ret = AVERROR(pa_context_errno(pd->context));
|
|
goto unlock_and_fail;
|
|
}
|
|
|
|
pa_stream_set_state_callback(pd->stream, stream_state_cb, pd);
|
|
pa_stream_set_read_callback(pd->stream, stream_request_cb, pd);
|
|
pa_stream_set_write_callback(pd->stream, stream_request_cb, pd);
|
|
pa_stream_set_latency_update_callback(pd->stream, stream_latency_update_cb, pd);
|
|
|
|
ret = pa_stream_connect_record(pd->stream, device, &attr,
|
|
PA_STREAM_INTERPOLATE_TIMING
|
|
| (pd->fragment_size == -1 ? PA_STREAM_ADJUST_LATENCY : 0)
|
|
|PA_STREAM_AUTO_TIMING_UPDATE);
|
|
|
|
if (ret < 0) {
|
|
ret = AVERROR(pa_context_errno(pd->context));
|
|
goto unlock_and_fail;
|
|
}
|
|
|
|
for (;;) {
|
|
pa_stream_state_t state;
|
|
|
|
state = pa_stream_get_state(pd->stream);
|
|
|
|
if (state == PA_STREAM_READY)
|
|
break;
|
|
|
|
if (!PA_STREAM_IS_GOOD(state)) {
|
|
ret = AVERROR(pa_context_errno(pd->context));
|
|
goto unlock_and_fail;
|
|
}
|
|
|
|
/* Wait until the stream is ready */
|
|
pa_threaded_mainloop_wait(pd->mainloop);
|
|
}
|
|
|
|
/* Query actual fragment size */
|
|
queried_attr = pa_stream_get_buffer_attr(pd->stream);
|
|
if (!queried_attr || queried_attr->fragsize > INT_MAX/100) {
|
|
ret = AVERROR_EXTERNAL;
|
|
goto unlock_and_fail;
|
|
}
|
|
pd->fragment_size = queried_attr->fragsize;
|
|
pd->pa_frame_size = pa_frame_size(&ss);
|
|
|
|
pa_threaded_mainloop_unlock(pd->mainloop);
|
|
|
|
/* take real parameters */
|
|
st->codecpar->codec_type = AVMEDIA_TYPE_AUDIO;
|
|
st->codecpar->codec_id = codec_id;
|
|
st->codecpar->sample_rate = pd->sample_rate;
|
|
st->codecpar->channels = pd->channels;
|
|
avpriv_set_pts_info(st, 64, 1, 1000000); /* 64 bits pts in us */
|
|
|
|
pd->timefilter = ff_timefilter_new(1000000.0 / pd->sample_rate,
|
|
pd->fragment_size / pd->pa_frame_size, 1.5E-6);
|
|
|
|
if (!pd->timefilter) {
|
|
pulse_close(s);
|
|
return AVERROR(ENOMEM);
|
|
}
|
|
|
|
return 0;
|
|
|
|
unlock_and_fail:
|
|
pa_threaded_mainloop_unlock(pd->mainloop);
|
|
|
|
pulse_close(s);
|
|
return ret;
|
|
}
|
|
|
|
static int pulse_read_packet(AVFormatContext *s, AVPacket *pkt)
|
|
{
|
|
PulseData *pd = s->priv_data;
|
|
int ret;
|
|
size_t read_length;
|
|
const void *read_data = NULL;
|
|
int64_t dts;
|
|
pa_usec_t latency;
|
|
int negative;
|
|
ptrdiff_t pos = 0;
|
|
|
|
pa_threaded_mainloop_lock(pd->mainloop);
|
|
|
|
CHECK_DEAD_GOTO(pd, ret, unlock_and_fail);
|
|
|
|
while (pos < pd->fragment_size) {
|
|
int r;
|
|
|
|
r = pa_stream_peek(pd->stream, &read_data, &read_length);
|
|
CHECK_SUCCESS_GOTO(ret, r == 0, unlock_and_fail);
|
|
|
|
if (read_length <= 0) {
|
|
pa_threaded_mainloop_wait(pd->mainloop);
|
|
CHECK_DEAD_GOTO(pd, ret, unlock_and_fail);
|
|
} else if (!read_data) {
|
|
/* There's a hole in the stream, skip it. We could generate
|
|
* silence, but that wouldn't work for compressed streams. */
|
|
r = pa_stream_drop(pd->stream);
|
|
CHECK_SUCCESS_GOTO(ret, r == 0, unlock_and_fail);
|
|
} else {
|
|
if (!pos) {
|
|
if (av_new_packet(pkt, pd->fragment_size) < 0) {
|
|
ret = AVERROR(ENOMEM);
|
|
goto unlock_and_fail;
|
|
}
|
|
|
|
dts = av_gettime();
|
|
pa_operation_unref(pa_stream_update_timing_info(pd->stream, NULL, NULL));
|
|
|
|
if (pa_stream_get_latency(pd->stream, &latency, &negative) >= 0) {
|
|
if (negative) {
|
|
dts += latency;
|
|
} else
|
|
dts -= latency;
|
|
} else {
|
|
av_log(s, AV_LOG_WARNING, "pa_stream_get_latency() failed\n");
|
|
}
|
|
}
|
|
if (pkt->size - pos < read_length) {
|
|
if (pos)
|
|
break;
|
|
pa_stream_drop(pd->stream);
|
|
/* Oversized fragment??? */
|
|
ret = AVERROR_EXTERNAL;
|
|
goto unlock_and_fail;
|
|
}
|
|
memcpy(pkt->data + pos, read_data, read_length);
|
|
pos += read_length;
|
|
pa_stream_drop(pd->stream);
|
|
}
|
|
}
|
|
|
|
pa_threaded_mainloop_unlock(pd->mainloop);
|
|
|
|
av_shrink_packet(pkt, pos);
|
|
|
|
if (pd->wallclock)
|
|
pkt->pts = ff_timefilter_update(pd->timefilter, dts, pd->last_period);
|
|
pd->last_period = pkt->size / pd->pa_frame_size;
|
|
|
|
return 0;
|
|
|
|
unlock_and_fail:
|
|
av_packet_unref(pkt);
|
|
pa_threaded_mainloop_unlock(pd->mainloop);
|
|
return ret;
|
|
}
|
|
|
|
static int pulse_get_device_list(AVFormatContext *h, AVDeviceInfoList *device_list)
|
|
{
|
|
PulseData *s = h->priv_data;
|
|
return ff_pulse_audio_get_devices(device_list, s->server, 0);
|
|
}
|
|
|
|
#define OFFSET(a) offsetof(PulseData, a)
|
|
#define D AV_OPT_FLAG_DECODING_PARAM
|
|
|
|
static const AVOption options[] = {
|
|
{ "server", "set PulseAudio server", OFFSET(server), AV_OPT_TYPE_STRING, {.str = NULL}, 0, 0, D },
|
|
{ "name", "set application name", OFFSET(name), AV_OPT_TYPE_STRING, {.str = LIBAVFORMAT_IDENT}, 0, 0, D },
|
|
{ "stream_name", "set stream description", OFFSET(stream_name), AV_OPT_TYPE_STRING, {.str = "record"}, 0, 0, D },
|
|
{ "sample_rate", "set sample rate in Hz", OFFSET(sample_rate), AV_OPT_TYPE_INT, {.i64 = 48000}, 1, INT_MAX, D },
|
|
{ "channels", "set number of audio channels", OFFSET(channels), AV_OPT_TYPE_INT, {.i64 = 2}, 1, INT_MAX, D },
|
|
{ "frame_size", "set number of bytes per frame", OFFSET(frame_size), AV_OPT_TYPE_INT, {.i64 = 1024}, 1, INT_MAX, D },
|
|
{ "fragment_size", "set buffering size, affects latency and cpu usage", OFFSET(fragment_size), AV_OPT_TYPE_INT, {.i64 = -1}, -1, INT_MAX, D },
|
|
{ "wallclock", "set the initial pts using the current time", OFFSET(wallclock), AV_OPT_TYPE_INT, {.i64 = 1}, -1, 1, D },
|
|
{ NULL },
|
|
};
|
|
|
|
static const AVClass pulse_demuxer_class = {
|
|
.class_name = "Pulse indev",
|
|
.item_name = av_default_item_name,
|
|
.option = options,
|
|
.version = LIBAVUTIL_VERSION_INT,
|
|
.category = AV_CLASS_CATEGORY_DEVICE_AUDIO_INPUT,
|
|
};
|
|
|
|
const AVInputFormat ff_pulse_demuxer = {
|
|
.name = "pulse",
|
|
.long_name = NULL_IF_CONFIG_SMALL("Pulse audio input"),
|
|
.priv_data_size = sizeof(PulseData),
|
|
.read_header = pulse_read_header,
|
|
.read_packet = pulse_read_packet,
|
|
.read_close = pulse_close,
|
|
.get_device_list = pulse_get_device_list,
|
|
.flags = AVFMT_NOFILE,
|
|
.priv_class = &pulse_demuxer_class,
|
|
};
|