All that remains in it are things that belong in avfilter_internal.h. Move them there and remove internal.h
		
			
				
	
	
		
			1191 lines
		
	
	
		
			38 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			1191 lines
		
	
	
		
			38 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * Copyright (c) 2012 Pavel Koshevoy <pkoshevoy at gmail dot com>
 | |
|  *
 | |
|  * This file is part of FFmpeg.
 | |
|  *
 | |
|  * FFmpeg is free software; you can redistribute it and/or
 | |
|  * modify it under the terms of the GNU Lesser General Public
 | |
|  * License as published by the Free Software Foundation; either
 | |
|  * version 2.1 of the License, or (at your option) any later version.
 | |
|  *
 | |
|  * FFmpeg is distributed in the hope that it will be useful,
 | |
|  * but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
|  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 | |
|  * Lesser General Public License for more details.
 | |
|  *
 | |
|  * You should have received a copy of the GNU Lesser General Public
 | |
|  * License along with FFmpeg; if not, write to the Free Software
 | |
|  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
 | |
|  */
 | |
| 
 | |
| /**
 | |
|  * @file
 | |
|  * tempo scaling audio filter -- an implementation of WSOLA algorithm
 | |
|  *
 | |
|  * Based on MIT licensed yaeAudioTempoFilter.h and yaeAudioFragment.h
 | |
|  * from Apprentice Video player by Pavel Koshevoy.
 | |
|  * https://sourceforge.net/projects/apprenticevideo/
 | |
|  *
 | |
|  * An explanation of SOLA algorithm is available at
 | |
|  * http://www.surina.net/article/time-and-pitch-scaling.html
 | |
|  *
 | |
|  * WSOLA is very similar to SOLA, only one major difference exists between
 | |
|  * these algorithms.  SOLA shifts audio fragments along the output stream,
 | |
|  * where as WSOLA shifts audio fragments along the input stream.
 | |
|  *
 | |
|  * The advantage of WSOLA algorithm is that the overlap region size is
 | |
|  * always the same, therefore the blending function is constant and
 | |
|  * can be precomputed.
 | |
|  */
 | |
| 
 | |
| #include <float.h>
 | |
| #include "libavutil/avassert.h"
 | |
| #include "libavutil/channel_layout.h"
 | |
| #include "libavutil/mem.h"
 | |
| #include "libavutil/opt.h"
 | |
| #include "libavutil/samplefmt.h"
 | |
| #include "libavutil/tx.h"
 | |
| #include "avfilter.h"
 | |
| #include "audio.h"
 | |
| #include "filters.h"
 | |
| 
 | |
| /**
 | |
|  * A fragment of audio waveform
 | |
|  */
 | |
| typedef struct AudioFragment {
 | |
|     // index of the first sample of this fragment in the overall waveform;
 | |
|     // 0: input sample position
 | |
|     // 1: output sample position
 | |
|     int64_t position[2];
 | |
| 
 | |
|     // original packed multi-channel samples:
 | |
|     uint8_t *data;
 | |
| 
 | |
|     // number of samples in this fragment:
 | |
|     int nsamples;
 | |
| 
 | |
|     // rDFT transform of the down-mixed mono fragment, used for
 | |
|     // fast waveform alignment via correlation in frequency domain:
 | |
|     float *xdat_in;
 | |
|     float *xdat;
 | |
| } AudioFragment;
 | |
| 
 | |
| /**
 | |
|  * Filter state machine states
 | |
|  */
 | |
| typedef enum {
 | |
|     YAE_LOAD_FRAGMENT,
 | |
|     YAE_ADJUST_POSITION,
 | |
|     YAE_RELOAD_FRAGMENT,
 | |
|     YAE_OUTPUT_OVERLAP_ADD,
 | |
|     YAE_FLUSH_OUTPUT,
 | |
| } FilterState;
 | |
| 
 | |
| /**
 | |
|  * Filter state machine
 | |
|  */
 | |
| typedef struct ATempoContext {
 | |
|     const AVClass *class;
 | |
| 
 | |
|     // ring-buffer of input samples, necessary because some times
 | |
|     // input fragment position may be adjusted backwards:
 | |
|     uint8_t *buffer;
 | |
| 
 | |
|     // ring-buffer maximum capacity, expressed in sample rate time base:
 | |
|     int ring;
 | |
| 
 | |
|     // ring-buffer house keeping:
 | |
|     int size;
 | |
|     int head;
 | |
|     int tail;
 | |
| 
 | |
|     // 0: input sample position corresponding to the ring buffer tail
 | |
|     // 1: output sample position
 | |
|     int64_t position[2];
 | |
| 
 | |
|     // first input timestamp, all other timestamps are offset by this one
 | |
|     int64_t start_pts;
 | |
| 
 | |
|     // sample format:
 | |
|     enum AVSampleFormat format;
 | |
| 
 | |
|     // number of channels:
 | |
|     int channels;
 | |
| 
 | |
|     // row of bytes to skip from one sample to next, across multple channels;
 | |
|     // stride = (number-of-channels * bits-per-sample-per-channel) / 8
 | |
|     int stride;
 | |
| 
 | |
|     // fragment window size, power-of-two integer:
 | |
|     int window;
 | |
| 
 | |
|     // Hann window coefficients, for feathering
 | |
|     // (blending) the overlapping fragment region:
 | |
|     float *hann;
 | |
| 
 | |
|     // tempo scaling factor:
 | |
|     double tempo;
 | |
| 
 | |
|     // a snapshot of previous fragment input and output position values
 | |
|     // captured when the tempo scale factor was set most recently:
 | |
|     int64_t origin[2];
 | |
| 
 | |
|     // current/previous fragment ring-buffer:
 | |
|     AudioFragment frag[2];
 | |
| 
 | |
|     // current fragment index:
 | |
|     uint64_t nfrag;
 | |
| 
 | |
|     // current state:
 | |
|     FilterState state;
 | |
| 
 | |
|     // for fast correlation calculation in frequency domain:
 | |
|     AVTXContext *real_to_complex;
 | |
|     AVTXContext *complex_to_real;
 | |
|     av_tx_fn r2c_fn, c2r_fn;
 | |
|     float *correlation_in;
 | |
|     float *correlation;
 | |
| 
 | |
|     // for managing AVFilterPad.request_frame and AVFilterPad.filter_frame
 | |
|     AVFrame *dst_buffer;
 | |
|     uint8_t *dst;
 | |
|     uint8_t *dst_end;
 | |
|     uint64_t nsamples_in;
 | |
|     uint64_t nsamples_out;
 | |
| } ATempoContext;
 | |
| 
 | |
| #define YAE_ATEMPO_MIN 0.5
 | |
| #define YAE_ATEMPO_MAX 100.0
 | |
| 
 | |
| #define OFFSET(x) offsetof(ATempoContext, x)
 | |
| 
 | |
| static const AVOption atempo_options[] = {
 | |
|     { "tempo", "set tempo scale factor",
 | |
|       OFFSET(tempo), AV_OPT_TYPE_DOUBLE, { .dbl = 1.0 },
 | |
|       YAE_ATEMPO_MIN,
 | |
|       YAE_ATEMPO_MAX,
 | |
|       AV_OPT_FLAG_AUDIO_PARAM | AV_OPT_FLAG_FILTERING_PARAM | AV_OPT_FLAG_RUNTIME_PARAM },
 | |
|     { NULL }
 | |
| };
 | |
| 
 | |
| AVFILTER_DEFINE_CLASS(atempo);
 | |
| 
 | |
| inline static AudioFragment *yae_curr_frag(ATempoContext *atempo)
 | |
| {
 | |
|     return &atempo->frag[atempo->nfrag % 2];
 | |
| }
 | |
| 
 | |
| inline static AudioFragment *yae_prev_frag(ATempoContext *atempo)
 | |
| {
 | |
|     return &atempo->frag[(atempo->nfrag + 1) % 2];
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * Reset filter to initial state, do not deallocate existing local buffers.
 | |
|  */
 | |
| static void yae_clear(ATempoContext *atempo)
 | |
| {
 | |
|     atempo->size = 0;
 | |
|     atempo->head = 0;
 | |
|     atempo->tail = 0;
 | |
| 
 | |
|     atempo->nfrag = 0;
 | |
|     atempo->state = YAE_LOAD_FRAGMENT;
 | |
|     atempo->start_pts = AV_NOPTS_VALUE;
 | |
| 
 | |
|     atempo->position[0] = 0;
 | |
|     atempo->position[1] = 0;
 | |
| 
 | |
|     atempo->origin[0] = 0;
 | |
|     atempo->origin[1] = 0;
 | |
| 
 | |
|     atempo->frag[0].position[0] = 0;
 | |
|     atempo->frag[0].position[1] = 0;
 | |
|     atempo->frag[0].nsamples    = 0;
 | |
| 
 | |
|     atempo->frag[1].position[0] = 0;
 | |
|     atempo->frag[1].position[1] = 0;
 | |
|     atempo->frag[1].nsamples    = 0;
 | |
| 
 | |
|     // shift left position of 1st fragment by half a window
 | |
|     // so that no re-normalization would be required for
 | |
|     // the left half of the 1st fragment:
 | |
|     atempo->frag[0].position[0] = -(int64_t)(atempo->window / 2);
 | |
|     atempo->frag[0].position[1] = -(int64_t)(atempo->window / 2);
 | |
| 
 | |
|     av_frame_free(&atempo->dst_buffer);
 | |
|     atempo->dst     = NULL;
 | |
|     atempo->dst_end = NULL;
 | |
| 
 | |
|     atempo->nsamples_in       = 0;
 | |
|     atempo->nsamples_out      = 0;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * Reset filter to initial state and deallocate all buffers.
 | |
|  */
 | |
| static void yae_release_buffers(ATempoContext *atempo)
 | |
| {
 | |
|     yae_clear(atempo);
 | |
| 
 | |
|     av_freep(&atempo->frag[0].data);
 | |
|     av_freep(&atempo->frag[1].data);
 | |
|     av_freep(&atempo->frag[0].xdat_in);
 | |
|     av_freep(&atempo->frag[1].xdat_in);
 | |
|     av_freep(&atempo->frag[0].xdat);
 | |
|     av_freep(&atempo->frag[1].xdat);
 | |
| 
 | |
|     av_freep(&atempo->buffer);
 | |
|     av_freep(&atempo->hann);
 | |
|     av_freep(&atempo->correlation_in);
 | |
|     av_freep(&atempo->correlation);
 | |
| 
 | |
|     av_tx_uninit(&atempo->real_to_complex);
 | |
|     av_tx_uninit(&atempo->complex_to_real);
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * Prepare filter for processing audio data of given format,
 | |
|  * sample rate and number of channels.
 | |
|  */
 | |
| static int yae_reset(ATempoContext *atempo,
 | |
|                      enum AVSampleFormat format,
 | |
|                      int sample_rate,
 | |
|                      int channels)
 | |
| {
 | |
|     const int sample_size = av_get_bytes_per_sample(format);
 | |
|     uint32_t nlevels  = 0;
 | |
|     float scale = 1.f, iscale = 1.f;
 | |
|     uint32_t pot;
 | |
|     int ret;
 | |
|     int i;
 | |
| 
 | |
|     atempo->format   = format;
 | |
|     atempo->channels = channels;
 | |
|     atempo->stride   = sample_size * channels;
 | |
| 
 | |
|     // pick a segment window size:
 | |
|     atempo->window = sample_rate / 24;
 | |
| 
 | |
|     // adjust window size to be a power-of-two integer:
 | |
|     nlevels = av_log2(atempo->window);
 | |
|     pot = 1 << nlevels;
 | |
|     av_assert0(pot <= atempo->window);
 | |
| 
 | |
|     if (pot < atempo->window) {
 | |
|         atempo->window = pot * 2;
 | |
|         nlevels++;
 | |
|     }
 | |
| 
 | |
|     /* av_realloc is not aligned enough, so simply discard all the old buffers
 | |
|      * (fortunately, their data does not need to be preserved) */
 | |
|     yae_release_buffers(atempo);
 | |
| 
 | |
|     // initialize audio fragment buffers:
 | |
|     if (!(atempo->frag[0].data    = av_calloc(atempo->window, atempo->stride)) ||
 | |
|         !(atempo->frag[1].data    = av_calloc(atempo->window, atempo->stride)) ||
 | |
|         !(atempo->frag[0].xdat_in = av_calloc(atempo->window + 1, sizeof(AVComplexFloat))) ||
 | |
|         !(atempo->frag[1].xdat_in = av_calloc(atempo->window + 1, sizeof(AVComplexFloat))) ||
 | |
|         !(atempo->frag[0].xdat    = av_calloc(atempo->window + 1, sizeof(AVComplexFloat))) ||
 | |
|         !(atempo->frag[1].xdat    = av_calloc(atempo->window + 1, sizeof(AVComplexFloat)))) {
 | |
|         ret = AVERROR(ENOMEM);
 | |
|         goto fail;
 | |
|     }
 | |
| 
 | |
|     // initialize rDFT contexts:
 | |
|     ret = av_tx_init(&atempo->real_to_complex, &atempo->r2c_fn,
 | |
|                      AV_TX_FLOAT_RDFT, 0, 1 << (nlevels + 1), &scale, 0);
 | |
|     if (ret < 0)
 | |
|         goto fail;
 | |
| 
 | |
|     ret = av_tx_init(&atempo->complex_to_real, &atempo->c2r_fn,
 | |
|                      AV_TX_FLOAT_RDFT, 1, 1 << (nlevels + 1), &iscale, 0);
 | |
|     if (ret < 0)
 | |
|         goto fail;
 | |
| 
 | |
|     if (!(atempo->correlation_in = av_calloc(atempo->window + 1, sizeof(AVComplexFloat))) ||
 | |
|         !(atempo->correlation    = av_calloc(atempo->window,     sizeof(AVComplexFloat)))) {
 | |
|         ret = AVERROR(ENOMEM);
 | |
|         goto fail;
 | |
|     }
 | |
| 
 | |
|     atempo->ring = atempo->window * 3;
 | |
|     atempo->buffer = av_calloc(atempo->ring, atempo->stride);
 | |
|     if (!atempo->buffer) {
 | |
|         ret = AVERROR(ENOMEM);
 | |
|         goto fail;
 | |
|     }
 | |
| 
 | |
|     // initialize the Hann window function:
 | |
|     atempo->hann = av_malloc_array(atempo->window, sizeof(float));
 | |
|     if (!atempo->hann) {
 | |
|         ret = AVERROR(ENOMEM);
 | |
|         goto fail;
 | |
|     }
 | |
| 
 | |
|     for (i = 0; i < atempo->window; i++) {
 | |
|         double t = (double)i / (double)(atempo->window - 1);
 | |
|         double h = 0.5 * (1.0 - cos(2.0 * M_PI * t));
 | |
|         atempo->hann[i] = (float)h;
 | |
|     }
 | |
| 
 | |
|     return 0;
 | |
| fail:
 | |
|     yae_release_buffers(atempo);
 | |
|     return ret;
 | |
| }
 | |
| 
 | |
| static int yae_update(AVFilterContext *ctx)
 | |
| {
 | |
|     const AudioFragment *prev;
 | |
|     ATempoContext *atempo = ctx->priv;
 | |
| 
 | |
|     prev = yae_prev_frag(atempo);
 | |
|     atempo->origin[0] = prev->position[0] + atempo->window / 2;
 | |
|     atempo->origin[1] = prev->position[1] + atempo->window / 2;
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * A helper macro for initializing complex data buffer with scalar data
 | |
|  * of a given type.
 | |
|  */
 | |
| #define yae_init_xdat(scalar_type, scalar_max)                          \
 | |
|     do {                                                                \
 | |
|         const uint8_t *src_end = src +                                  \
 | |
|             frag->nsamples * atempo->channels * sizeof(scalar_type);    \
 | |
|                                                                         \
 | |
|         float *xdat = frag->xdat_in;                                    \
 | |
|         scalar_type tmp;                                                \
 | |
|                                                                         \
 | |
|         if (atempo->channels == 1) {                                    \
 | |
|             for (; src < src_end; xdat++) {                             \
 | |
|                 tmp = *(const scalar_type *)src;                        \
 | |
|                 src += sizeof(scalar_type);                             \
 | |
|                                                                         \
 | |
|                 *xdat = (float)tmp;                                     \
 | |
|             }                                                           \
 | |
|         } else {                                                        \
 | |
|             float s, max, ti, si;                                       \
 | |
|             int i;                                                      \
 | |
|                                                                         \
 | |
|             for (; src < src_end; xdat++) {                             \
 | |
|                 tmp = *(const scalar_type *)src;                        \
 | |
|                 src += sizeof(scalar_type);                             \
 | |
|                                                                         \
 | |
|                 max = (float)tmp;                                       \
 | |
|                 s = FFMIN((float)scalar_max,                            \
 | |
|                           (float)fabsf(max));                           \
 | |
|                                                                         \
 | |
|                 for (i = 1; i < atempo->channels; i++) {                \
 | |
|                     tmp = *(const scalar_type *)src;                    \
 | |
|                     src += sizeof(scalar_type);                         \
 | |
|                                                                         \
 | |
|                     ti = (float)tmp;                                    \
 | |
|                     si = FFMIN((float)scalar_max,                       \
 | |
|                                (float)fabsf(ti));                       \
 | |
|                                                                         \
 | |
|                     if (s < si) {                                       \
 | |
|                         s   = si;                                       \
 | |
|                         max = ti;                                       \
 | |
|                     }                                                   \
 | |
|                 }                                                       \
 | |
|                                                                         \
 | |
|                 *xdat = max;                                            \
 | |
|             }                                                           \
 | |
|         }                                                               \
 | |
|     } while (0)
 | |
| 
 | |
| /**
 | |
|  * Initialize complex data buffer of a given audio fragment
 | |
|  * with down-mixed mono data of appropriate scalar type.
 | |
|  */
 | |
| static void yae_downmix(ATempoContext *atempo, AudioFragment *frag)
 | |
| {
 | |
|     // shortcuts:
 | |
|     const uint8_t *src = frag->data;
 | |
| 
 | |
|     // init complex data buffer used for FFT and Correlation:
 | |
|     memset(frag->xdat_in, 0, sizeof(AVComplexFloat) * (atempo->window + 1));
 | |
| 
 | |
|     if (atempo->format == AV_SAMPLE_FMT_U8) {
 | |
|         yae_init_xdat(uint8_t, 127);
 | |
|     } else if (atempo->format == AV_SAMPLE_FMT_S16) {
 | |
|         yae_init_xdat(int16_t, 32767);
 | |
|     } else if (atempo->format == AV_SAMPLE_FMT_S32) {
 | |
|         yae_init_xdat(int, 2147483647);
 | |
|     } else if (atempo->format == AV_SAMPLE_FMT_FLT) {
 | |
|         yae_init_xdat(float, 1);
 | |
|     } else if (atempo->format == AV_SAMPLE_FMT_DBL) {
 | |
|         yae_init_xdat(double, 1);
 | |
|     }
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * Populate the internal data buffer on as-needed basis.
 | |
|  *
 | |
|  * @return
 | |
|  *   0 if requested data was already available or was successfully loaded,
 | |
|  *   AVERROR(EAGAIN) if more input data is required.
 | |
|  */
 | |
| static int yae_load_data(ATempoContext *atempo,
 | |
|                          const uint8_t **src_ref,
 | |
|                          const uint8_t *src_end,
 | |
|                          int64_t stop_here)
 | |
| {
 | |
|     // shortcut:
 | |
|     const uint8_t *src = *src_ref;
 | |
|     const int read_size = stop_here - atempo->position[0];
 | |
| 
 | |
|     if (stop_here <= atempo->position[0]) {
 | |
|         return 0;
 | |
|     }
 | |
| 
 | |
|     // samples are not expected to be skipped, unless tempo is greater than 2:
 | |
|     av_assert0(read_size <= atempo->ring || atempo->tempo > 2.0);
 | |
| 
 | |
|     while (atempo->position[0] < stop_here && src < src_end) {
 | |
|         int src_samples = (src_end - src) / atempo->stride;
 | |
| 
 | |
|         // load data piece-wise, in order to avoid complicating the logic:
 | |
|         int nsamples = FFMIN(read_size, src_samples);
 | |
|         int na;
 | |
|         int nb;
 | |
| 
 | |
|         nsamples = FFMIN(nsamples, atempo->ring);
 | |
|         na = FFMIN(nsamples, atempo->ring - atempo->tail);
 | |
|         nb = FFMIN(nsamples - na, atempo->ring);
 | |
| 
 | |
|         if (na) {
 | |
|             uint8_t *a = atempo->buffer + atempo->tail * atempo->stride;
 | |
|             memcpy(a, src, na * atempo->stride);
 | |
| 
 | |
|             src += na * atempo->stride;
 | |
|             atempo->position[0] += na;
 | |
| 
 | |
|             atempo->size = FFMIN(atempo->size + na, atempo->ring);
 | |
|             atempo->tail = (atempo->tail + na) % atempo->ring;
 | |
|             atempo->head =
 | |
|                 atempo->size < atempo->ring ?
 | |
|                 atempo->tail - atempo->size :
 | |
|                 atempo->tail;
 | |
|         }
 | |
| 
 | |
|         if (nb) {
 | |
|             uint8_t *b = atempo->buffer;
 | |
|             memcpy(b, src, nb * atempo->stride);
 | |
| 
 | |
|             src += nb * atempo->stride;
 | |
|             atempo->position[0] += nb;
 | |
| 
 | |
|             atempo->size = FFMIN(atempo->size + nb, atempo->ring);
 | |
|             atempo->tail = (atempo->tail + nb) % atempo->ring;
 | |
|             atempo->head =
 | |
|                 atempo->size < atempo->ring ?
 | |
|                 atempo->tail - atempo->size :
 | |
|                 atempo->tail;
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     // pass back the updated source buffer pointer:
 | |
|     *src_ref = src;
 | |
| 
 | |
|     // sanity check:
 | |
|     av_assert0(atempo->position[0] <= stop_here);
 | |
| 
 | |
|     return atempo->position[0] == stop_here ? 0 : AVERROR(EAGAIN);
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * Populate current audio fragment data buffer.
 | |
|  *
 | |
|  * @return
 | |
|  *   0 when the fragment is ready,
 | |
|  *   AVERROR(EAGAIN) if more input data is required.
 | |
|  */
 | |
| static int yae_load_frag(ATempoContext *atempo,
 | |
|                          const uint8_t **src_ref,
 | |
|                          const uint8_t *src_end)
 | |
| {
 | |
|     // shortcuts:
 | |
|     AudioFragment *frag = yae_curr_frag(atempo);
 | |
|     uint8_t *dst;
 | |
|     int64_t missing, start, zeros;
 | |
|     uint32_t nsamples;
 | |
|     const uint8_t *a, *b;
 | |
|     int i0, i1, n0, n1, na, nb;
 | |
| 
 | |
|     int64_t stop_here = frag->position[0] + atempo->window;
 | |
|     if (src_ref && yae_load_data(atempo, src_ref, src_end, stop_here) != 0) {
 | |
|         return AVERROR(EAGAIN);
 | |
|     }
 | |
| 
 | |
|     // calculate the number of samples we don't have:
 | |
|     missing =
 | |
|         stop_here > atempo->position[0] ?
 | |
|         stop_here - atempo->position[0] : 0;
 | |
| 
 | |
|     nsamples =
 | |
|         missing < (int64_t)atempo->window ?
 | |
|         (uint32_t)(atempo->window - missing) : 0;
 | |
| 
 | |
|     // setup the output buffer:
 | |
|     frag->nsamples = nsamples;
 | |
|     dst = frag->data;
 | |
| 
 | |
|     start = atempo->position[0] - atempo->size;
 | |
| 
 | |
|     // what we don't have we substitute with zeros:
 | |
|     zeros =
 | |
|       frag->position[0] < start ?
 | |
|       FFMIN(start - frag->position[0], (int64_t)nsamples) : 0;
 | |
| 
 | |
|     if (zeros == nsamples) {
 | |
|         return 0;
 | |
|     }
 | |
| 
 | |
|     if (frag->position[0] < start) {
 | |
|         memset(dst, 0, zeros * atempo->stride);
 | |
|         dst += zeros * atempo->stride;
 | |
|     }
 | |
| 
 | |
|     // get the remaining data from the ring buffer:
 | |
|     na = (atempo->head < atempo->tail ?
 | |
|           atempo->tail - atempo->head :
 | |
|           atempo->ring - atempo->head);
 | |
| 
 | |
|     nb = atempo->head < atempo->tail ? 0 : atempo->tail;
 | |
| 
 | |
|     // sanity check:
 | |
|     av_assert0(nsamples <= zeros + na + nb);
 | |
| 
 | |
|     a = atempo->buffer + atempo->head * atempo->stride;
 | |
|     b = atempo->buffer;
 | |
| 
 | |
|     i0 = frag->position[0] + zeros - start;
 | |
|     i1 = i0 < na ? 0 : i0 - na;
 | |
| 
 | |
|     n0 = i0 < na ? FFMIN(na - i0, (int)(nsamples - zeros)) : 0;
 | |
|     n1 = nsamples - zeros - n0;
 | |
| 
 | |
|     if (n0) {
 | |
|         memcpy(dst, a + i0 * atempo->stride, n0 * atempo->stride);
 | |
|         dst += n0 * atempo->stride;
 | |
|     }
 | |
| 
 | |
|     if (n1) {
 | |
|         memcpy(dst, b + i1 * atempo->stride, n1 * atempo->stride);
 | |
|     }
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * Prepare for loading next audio fragment.
 | |
|  */
 | |
| static void yae_advance_to_next_frag(ATempoContext *atempo)
 | |
| {
 | |
|     const double fragment_step = atempo->tempo * (double)(atempo->window / 2);
 | |
| 
 | |
|     const AudioFragment *prev;
 | |
|     AudioFragment       *frag;
 | |
| 
 | |
|     atempo->nfrag++;
 | |
|     prev = yae_prev_frag(atempo);
 | |
|     frag = yae_curr_frag(atempo);
 | |
| 
 | |
|     frag->position[0] = prev->position[0] + (int64_t)fragment_step;
 | |
|     frag->position[1] = prev->position[1] + atempo->window / 2;
 | |
|     frag->nsamples    = 0;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * Calculate cross-correlation via rDFT.
 | |
|  *
 | |
|  * Multiply two vectors of complex numbers (result of real_to_complex rDFT)
 | |
|  * and transform back via complex_to_real rDFT.
 | |
|  */
 | |
| static void yae_xcorr_via_rdft(float *xcorr_in,
 | |
|                                float *xcorr,
 | |
|                                AVTXContext *complex_to_real,
 | |
|                                av_tx_fn c2r_fn,
 | |
|                                const AVComplexFloat *xa,
 | |
|                                const AVComplexFloat *xb,
 | |
|                                const int window)
 | |
| {
 | |
|     AVComplexFloat *xc = (AVComplexFloat *)xcorr_in;
 | |
|     int i;
 | |
| 
 | |
|     for (i = 0; i <= window; i++, xa++, xb++, xc++) {
 | |
|         xc->re = (xa->re * xb->re + xa->im * xb->im);
 | |
|         xc->im = (xa->im * xb->re - xa->re * xb->im);
 | |
|     }
 | |
| 
 | |
|     // apply inverse rDFT:
 | |
|     c2r_fn(complex_to_real, xcorr, xcorr_in, sizeof(*xc));
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * Calculate alignment offset for given fragment
 | |
|  * relative to the previous fragment.
 | |
|  *
 | |
|  * @return alignment offset of current fragment relative to previous.
 | |
|  */
 | |
| static int yae_align(AudioFragment *frag,
 | |
|                      const AudioFragment *prev,
 | |
|                      const int window,
 | |
|                      const int delta_max,
 | |
|                      const int drift,
 | |
|                      float *correlation_in,
 | |
|                      float *correlation,
 | |
|                      AVTXContext *complex_to_real,
 | |
|                      av_tx_fn c2r_fn)
 | |
| {
 | |
|     int       best_offset = -drift;
 | |
|     float     best_metric = -FLT_MAX;
 | |
|     float    *xcorr;
 | |
| 
 | |
|     int i0;
 | |
|     int i1;
 | |
|     int i;
 | |
| 
 | |
|     yae_xcorr_via_rdft(correlation_in,
 | |
|                        correlation,
 | |
|                        complex_to_real,
 | |
|                        c2r_fn,
 | |
|                        (const AVComplexFloat *)prev->xdat,
 | |
|                        (const AVComplexFloat *)frag->xdat,
 | |
|                        window);
 | |
| 
 | |
|     // identify search window boundaries:
 | |
|     i0 = FFMAX(window / 2 - delta_max - drift, 0);
 | |
|     i0 = FFMIN(i0, window);
 | |
| 
 | |
|     i1 = FFMIN(window / 2 + delta_max - drift, window - window / 16);
 | |
|     i1 = FFMAX(i1, 0);
 | |
| 
 | |
|     // identify cross-correlation peaks within search window:
 | |
|     xcorr = correlation + i0;
 | |
| 
 | |
|     for (i = i0; i < i1; i++, xcorr++) {
 | |
|         float metric = *xcorr;
 | |
| 
 | |
|         // normalize:
 | |
|         float drifti = (float)(drift + i);
 | |
|         metric *= drifti * (float)(i - i0) * (float)(i1 - i);
 | |
| 
 | |
|         if (metric > best_metric) {
 | |
|             best_metric = metric;
 | |
|             best_offset = i - window / 2;
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     return best_offset;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * Adjust current fragment position for better alignment
 | |
|  * with previous fragment.
 | |
|  *
 | |
|  * @return alignment correction.
 | |
|  */
 | |
| static int yae_adjust_position(ATempoContext *atempo)
 | |
| {
 | |
|     const AudioFragment *prev = yae_prev_frag(atempo);
 | |
|     AudioFragment       *frag = yae_curr_frag(atempo);
 | |
| 
 | |
|     const double prev_output_position =
 | |
|         (double)(prev->position[1] - atempo->origin[1] + atempo->window / 2) *
 | |
|         atempo->tempo;
 | |
| 
 | |
|     const double ideal_output_position =
 | |
|         (double)(prev->position[0] - atempo->origin[0] + atempo->window / 2);
 | |
| 
 | |
|     const int drift = (int)(prev_output_position - ideal_output_position);
 | |
| 
 | |
|     const int delta_max  = atempo->window / 2;
 | |
|     const int correction = yae_align(frag,
 | |
|                                      prev,
 | |
|                                      atempo->window,
 | |
|                                      delta_max,
 | |
|                                      drift,
 | |
|                                      atempo->correlation_in,
 | |
|                                      atempo->correlation,
 | |
|                                      atempo->complex_to_real,
 | |
|                                      atempo->c2r_fn);
 | |
| 
 | |
|     if (correction) {
 | |
|         // adjust fragment position:
 | |
|         frag->position[0] -= correction;
 | |
| 
 | |
|         // clear so that the fragment can be reloaded:
 | |
|         frag->nsamples = 0;
 | |
|     }
 | |
| 
 | |
|     return correction;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * A helper macro for blending the overlap region of previous
 | |
|  * and current audio fragment.
 | |
|  */
 | |
| #define yae_blend(scalar_type)                                          \
 | |
|     do {                                                                \
 | |
|         const scalar_type *aaa = (const scalar_type *)a;                \
 | |
|         const scalar_type *bbb = (const scalar_type *)b;                \
 | |
|                                                                         \
 | |
|         scalar_type *out     = (scalar_type *)dst;                      \
 | |
|         scalar_type *out_end = (scalar_type *)dst_end;                  \
 | |
|         int64_t i;                                                      \
 | |
|                                                                         \
 | |
|         for (i = 0; i < overlap && out < out_end;                       \
 | |
|              i++, atempo->position[1]++, wa++, wb++) {                  \
 | |
|             float w0 = *wa;                                             \
 | |
|             float w1 = *wb;                                             \
 | |
|             int j;                                                      \
 | |
|                                                                         \
 | |
|             for (j = 0; j < atempo->channels;                           \
 | |
|                  j++, aaa++, bbb++, out++) {                            \
 | |
|                 float t0 = (float)*aaa;                                 \
 | |
|                 float t1 = (float)*bbb;                                 \
 | |
|                                                                         \
 | |
|                 *out =                                                  \
 | |
|                     frag->position[0] + i < 0 ?                         \
 | |
|                     *aaa :                                              \
 | |
|                     (scalar_type)(t0 * w0 + t1 * w1);                   \
 | |
|             }                                                           \
 | |
|         }                                                               \
 | |
|         dst = (uint8_t *)out;                                           \
 | |
|     } while (0)
 | |
| 
 | |
| /**
 | |
|  * Blend the overlap region of previous and current audio fragment
 | |
|  * and output the results to the given destination buffer.
 | |
|  *
 | |
|  * @return
 | |
|  *   0 if the overlap region was completely stored in the dst buffer,
 | |
|  *   AVERROR(EAGAIN) if more destination buffer space is required.
 | |
|  */
 | |
| static int yae_overlap_add(ATempoContext *atempo,
 | |
|                            uint8_t **dst_ref,
 | |
|                            uint8_t *dst_end)
 | |
| {
 | |
|     // shortcuts:
 | |
|     const AudioFragment *prev = yae_prev_frag(atempo);
 | |
|     const AudioFragment *frag = yae_curr_frag(atempo);
 | |
| 
 | |
|     const int64_t start_here = FFMAX(atempo->position[1],
 | |
|                                      frag->position[1]);
 | |
| 
 | |
|     const int64_t stop_here = FFMIN(prev->position[1] + prev->nsamples,
 | |
|                                     frag->position[1] + frag->nsamples);
 | |
| 
 | |
|     const int64_t overlap = stop_here - start_here;
 | |
| 
 | |
|     const int64_t ia = start_here - prev->position[1];
 | |
|     const int64_t ib = start_here - frag->position[1];
 | |
| 
 | |
|     const float *wa = atempo->hann + ia;
 | |
|     const float *wb = atempo->hann + ib;
 | |
| 
 | |
|     const uint8_t *a = prev->data + ia * atempo->stride;
 | |
|     const uint8_t *b = frag->data + ib * atempo->stride;
 | |
| 
 | |
|     uint8_t *dst = *dst_ref;
 | |
| 
 | |
|     av_assert0(start_here <= stop_here &&
 | |
|                frag->position[1] <= start_here &&
 | |
|                overlap <= frag->nsamples);
 | |
| 
 | |
|     if (atempo->format == AV_SAMPLE_FMT_U8) {
 | |
|         yae_blend(uint8_t);
 | |
|     } else if (atempo->format == AV_SAMPLE_FMT_S16) {
 | |
|         yae_blend(int16_t);
 | |
|     } else if (atempo->format == AV_SAMPLE_FMT_S32) {
 | |
|         yae_blend(int);
 | |
|     } else if (atempo->format == AV_SAMPLE_FMT_FLT) {
 | |
|         yae_blend(float);
 | |
|     } else if (atempo->format == AV_SAMPLE_FMT_DBL) {
 | |
|         yae_blend(double);
 | |
|     }
 | |
| 
 | |
|     // pass-back the updated destination buffer pointer:
 | |
|     *dst_ref = dst;
 | |
| 
 | |
|     return atempo->position[1] == stop_here ? 0 : AVERROR(EAGAIN);
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * Feed as much data to the filter as it is able to consume
 | |
|  * and receive as much processed data in the destination buffer
 | |
|  * as it is able to produce or store.
 | |
|  */
 | |
| static void
 | |
| yae_apply(ATempoContext *atempo,
 | |
|           const uint8_t **src_ref,
 | |
|           const uint8_t *src_end,
 | |
|           uint8_t **dst_ref,
 | |
|           uint8_t *dst_end)
 | |
| {
 | |
|     while (1) {
 | |
|         if (atempo->state == YAE_LOAD_FRAGMENT) {
 | |
|             // load additional data for the current fragment:
 | |
|             if (yae_load_frag(atempo, src_ref, src_end) != 0) {
 | |
|                 break;
 | |
|             }
 | |
| 
 | |
|             // down-mix to mono:
 | |
|             yae_downmix(atempo, yae_curr_frag(atempo));
 | |
| 
 | |
|             // apply rDFT:
 | |
|             atempo->r2c_fn(atempo->real_to_complex, yae_curr_frag(atempo)->xdat, yae_curr_frag(atempo)->xdat_in, sizeof(float));
 | |
| 
 | |
|             // must load the second fragment before alignment can start:
 | |
|             if (!atempo->nfrag) {
 | |
|                 yae_advance_to_next_frag(atempo);
 | |
|                 continue;
 | |
|             }
 | |
| 
 | |
|             atempo->state = YAE_ADJUST_POSITION;
 | |
|         }
 | |
| 
 | |
|         if (atempo->state == YAE_ADJUST_POSITION) {
 | |
|             // adjust position for better alignment:
 | |
|             if (yae_adjust_position(atempo)) {
 | |
|                 // reload the fragment at the corrected position, so that the
 | |
|                 // Hann window blending would not require normalization:
 | |
|                 atempo->state = YAE_RELOAD_FRAGMENT;
 | |
|             } else {
 | |
|                 atempo->state = YAE_OUTPUT_OVERLAP_ADD;
 | |
|             }
 | |
|         }
 | |
| 
 | |
|         if (atempo->state == YAE_RELOAD_FRAGMENT) {
 | |
|             // load additional data if necessary due to position adjustment:
 | |
|             if (yae_load_frag(atempo, src_ref, src_end) != 0) {
 | |
|                 break;
 | |
|             }
 | |
| 
 | |
|             // down-mix to mono:
 | |
|             yae_downmix(atempo, yae_curr_frag(atempo));
 | |
| 
 | |
|             // apply rDFT:
 | |
|             atempo->r2c_fn(atempo->real_to_complex, yae_curr_frag(atempo)->xdat, yae_curr_frag(atempo)->xdat_in, sizeof(float));
 | |
| 
 | |
|             atempo->state = YAE_OUTPUT_OVERLAP_ADD;
 | |
|         }
 | |
| 
 | |
|         if (atempo->state == YAE_OUTPUT_OVERLAP_ADD) {
 | |
|             // overlap-add and output the result:
 | |
|             if (yae_overlap_add(atempo, dst_ref, dst_end) != 0) {
 | |
|                 break;
 | |
|             }
 | |
| 
 | |
|             // advance to the next fragment, repeat:
 | |
|             yae_advance_to_next_frag(atempo);
 | |
|             atempo->state = YAE_LOAD_FRAGMENT;
 | |
|         }
 | |
|     }
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * Flush any buffered data from the filter.
 | |
|  *
 | |
|  * @return
 | |
|  *   0 if all data was completely stored in the dst buffer,
 | |
|  *   AVERROR(EAGAIN) if more destination buffer space is required.
 | |
|  */
 | |
| static int yae_flush(ATempoContext *atempo,
 | |
|                      uint8_t **dst_ref,
 | |
|                      uint8_t *dst_end)
 | |
| {
 | |
|     AudioFragment *frag = yae_curr_frag(atempo);
 | |
|     int64_t overlap_end;
 | |
|     int64_t start_here;
 | |
|     int64_t stop_here;
 | |
|     int64_t offset;
 | |
| 
 | |
|     const uint8_t *src;
 | |
|     uint8_t *dst;
 | |
| 
 | |
|     int src_size;
 | |
|     int dst_size;
 | |
|     int nbytes;
 | |
| 
 | |
|     atempo->state = YAE_FLUSH_OUTPUT;
 | |
| 
 | |
|     if (!atempo->nfrag) {
 | |
|         // there is nothing to flush:
 | |
|         return 0;
 | |
|     }
 | |
| 
 | |
|     if (atempo->position[0] == frag->position[0] + frag->nsamples &&
 | |
|         atempo->position[1] == frag->position[1] + frag->nsamples) {
 | |
|         // the current fragment is already flushed:
 | |
|         return 0;
 | |
|     }
 | |
| 
 | |
|     if (frag->position[0] + frag->nsamples < atempo->position[0]) {
 | |
|         // finish loading the current (possibly partial) fragment:
 | |
|         yae_load_frag(atempo, NULL, NULL);
 | |
| 
 | |
|         if (atempo->nfrag) {
 | |
|             // down-mix to mono:
 | |
|             yae_downmix(atempo, frag);
 | |
| 
 | |
|             // apply rDFT:
 | |
|             atempo->r2c_fn(atempo->real_to_complex, frag->xdat, frag->xdat_in, sizeof(float));
 | |
| 
 | |
|             // align current fragment to previous fragment:
 | |
|             if (yae_adjust_position(atempo)) {
 | |
|                 // reload the current fragment due to adjusted position:
 | |
|                 yae_load_frag(atempo, NULL, NULL);
 | |
|             }
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     // flush the overlap region:
 | |
|     overlap_end = frag->position[1] + FFMIN(atempo->window / 2,
 | |
|                                             frag->nsamples);
 | |
| 
 | |
|     while (atempo->position[1] < overlap_end) {
 | |
|         if (yae_overlap_add(atempo, dst_ref, dst_end) != 0) {
 | |
|             return AVERROR(EAGAIN);
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     // check whether all of the input samples have been consumed:
 | |
|     if (frag->position[0] + frag->nsamples < atempo->position[0]) {
 | |
|         yae_advance_to_next_frag(atempo);
 | |
|         return AVERROR(EAGAIN);
 | |
|     }
 | |
| 
 | |
|     // flush the remainder of the current fragment:
 | |
|     start_here = FFMAX(atempo->position[1], overlap_end);
 | |
|     stop_here  = frag->position[1] + frag->nsamples;
 | |
|     offset     = start_here - frag->position[1];
 | |
|     av_assert0(start_here <= stop_here && frag->position[1] <= start_here);
 | |
| 
 | |
|     src = frag->data + offset * atempo->stride;
 | |
|     dst = (uint8_t *)*dst_ref;
 | |
| 
 | |
|     src_size = (int)(stop_here - start_here) * atempo->stride;
 | |
|     dst_size = dst_end - dst;
 | |
|     nbytes = FFMIN(src_size, dst_size);
 | |
| 
 | |
|     memcpy(dst, src, nbytes);
 | |
|     dst += nbytes;
 | |
| 
 | |
|     atempo->position[1] += (nbytes / atempo->stride);
 | |
| 
 | |
|     // pass-back the updated destination buffer pointer:
 | |
|     *dst_ref = (uint8_t *)dst;
 | |
| 
 | |
|     return atempo->position[1] == stop_here ? 0 : AVERROR(EAGAIN);
 | |
| }
 | |
| 
 | |
| static av_cold int init(AVFilterContext *ctx)
 | |
| {
 | |
|     ATempoContext *atempo = ctx->priv;
 | |
|     atempo->format = AV_SAMPLE_FMT_NONE;
 | |
|     atempo->state  = YAE_LOAD_FRAGMENT;
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static av_cold void uninit(AVFilterContext *ctx)
 | |
| {
 | |
|     ATempoContext *atempo = ctx->priv;
 | |
|     yae_release_buffers(atempo);
 | |
| }
 | |
| 
 | |
| // WSOLA necessitates an internal sliding window ring buffer
 | |
| // for incoming audio stream.
 | |
| //
 | |
| // Planar sample formats are too cumbersome to store in a ring buffer,
 | |
| // therefore planar sample formats are not supported.
 | |
| //
 | |
| static const enum AVSampleFormat sample_fmts[] = {
 | |
|     AV_SAMPLE_FMT_U8,
 | |
|     AV_SAMPLE_FMT_S16,
 | |
|     AV_SAMPLE_FMT_S32,
 | |
|     AV_SAMPLE_FMT_FLT,
 | |
|     AV_SAMPLE_FMT_DBL,
 | |
|     AV_SAMPLE_FMT_NONE
 | |
| };
 | |
| 
 | |
| static int config_props(AVFilterLink *inlink)
 | |
| {
 | |
|     AVFilterContext  *ctx = inlink->dst;
 | |
|     ATempoContext *atempo = ctx->priv;
 | |
| 
 | |
|     enum AVSampleFormat format = inlink->format;
 | |
|     int sample_rate = (int)inlink->sample_rate;
 | |
| 
 | |
|     return yae_reset(atempo, format, sample_rate, inlink->ch_layout.nb_channels);
 | |
| }
 | |
| 
 | |
| static int push_samples(ATempoContext *atempo,
 | |
|                         AVFilterLink *outlink,
 | |
|                         int n_out)
 | |
| {
 | |
|     int ret;
 | |
| 
 | |
|     atempo->dst_buffer->sample_rate = outlink->sample_rate;
 | |
|     atempo->dst_buffer->nb_samples  = n_out;
 | |
| 
 | |
|     // adjust the PTS:
 | |
|     atempo->dst_buffer->pts = atempo->start_pts +
 | |
|         av_rescale_q(atempo->nsamples_out,
 | |
|                      (AVRational){ 1, outlink->sample_rate },
 | |
|                      outlink->time_base);
 | |
| 
 | |
|     ret = ff_filter_frame(outlink, atempo->dst_buffer);
 | |
|     atempo->dst_buffer = NULL;
 | |
|     atempo->dst        = NULL;
 | |
|     atempo->dst_end    = NULL;
 | |
|     if (ret < 0)
 | |
|         return ret;
 | |
| 
 | |
|     atempo->nsamples_out += n_out;
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static int filter_frame(AVFilterLink *inlink, AVFrame *src_buffer)
 | |
| {
 | |
|     AVFilterContext  *ctx = inlink->dst;
 | |
|     ATempoContext *atempo = ctx->priv;
 | |
|     AVFilterLink *outlink = ctx->outputs[0];
 | |
| 
 | |
|     int ret = 0;
 | |
|     int n_in = src_buffer->nb_samples;
 | |
|     int n_out = (int)(0.5 + ((double)n_in) / atempo->tempo);
 | |
| 
 | |
|     const uint8_t *src = src_buffer->data[0];
 | |
|     const uint8_t *src_end = src + n_in * atempo->stride;
 | |
| 
 | |
|     if (atempo->start_pts == AV_NOPTS_VALUE)
 | |
|         atempo->start_pts = av_rescale_q(src_buffer->pts,
 | |
|                                          inlink->time_base,
 | |
|                                          outlink->time_base);
 | |
| 
 | |
|     while (src < src_end) {
 | |
|         if (!atempo->dst_buffer) {
 | |
|             atempo->dst_buffer = ff_get_audio_buffer(outlink, n_out);
 | |
|             if (!atempo->dst_buffer) {
 | |
|                 av_frame_free(&src_buffer);
 | |
|                 return AVERROR(ENOMEM);
 | |
|             }
 | |
|             av_frame_copy_props(atempo->dst_buffer, src_buffer);
 | |
| 
 | |
|             atempo->dst = atempo->dst_buffer->data[0];
 | |
|             atempo->dst_end = atempo->dst + n_out * atempo->stride;
 | |
|         }
 | |
| 
 | |
|         yae_apply(atempo, &src, src_end, &atempo->dst, atempo->dst_end);
 | |
| 
 | |
|         if (atempo->dst == atempo->dst_end) {
 | |
|             int n_samples = ((atempo->dst - atempo->dst_buffer->data[0]) /
 | |
|                              atempo->stride);
 | |
|             ret = push_samples(atempo, outlink, n_samples);
 | |
|             if (ret < 0)
 | |
|                 goto end;
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     atempo->nsamples_in += n_in;
 | |
| end:
 | |
|     av_frame_free(&src_buffer);
 | |
|     return ret;
 | |
| }
 | |
| 
 | |
| static int request_frame(AVFilterLink *outlink)
 | |
| {
 | |
|     AVFilterContext  *ctx = outlink->src;
 | |
|     ATempoContext *atempo = ctx->priv;
 | |
|     int ret;
 | |
| 
 | |
|     ret = ff_request_frame(ctx->inputs[0]);
 | |
| 
 | |
|     if (ret == AVERROR_EOF) {
 | |
|         // flush the filter:
 | |
|         int n_max = atempo->ring;
 | |
|         int n_out;
 | |
|         int err = AVERROR(EAGAIN);
 | |
| 
 | |
|         while (err == AVERROR(EAGAIN)) {
 | |
|             if (!atempo->dst_buffer) {
 | |
|                 atempo->dst_buffer = ff_get_audio_buffer(outlink, n_max);
 | |
|                 if (!atempo->dst_buffer)
 | |
|                     return AVERROR(ENOMEM);
 | |
| 
 | |
|                 atempo->dst = atempo->dst_buffer->data[0];
 | |
|                 atempo->dst_end = atempo->dst + n_max * atempo->stride;
 | |
|             }
 | |
| 
 | |
|             err = yae_flush(atempo, &atempo->dst, atempo->dst_end);
 | |
| 
 | |
|             n_out = ((atempo->dst - atempo->dst_buffer->data[0]) /
 | |
|                      atempo->stride);
 | |
| 
 | |
|             if (n_out) {
 | |
|                 ret = push_samples(atempo, outlink, n_out);
 | |
|                 if (ret < 0)
 | |
|                     return ret;
 | |
|             }
 | |
|         }
 | |
| 
 | |
|         av_frame_free(&atempo->dst_buffer);
 | |
|         atempo->dst     = NULL;
 | |
|         atempo->dst_end = NULL;
 | |
| 
 | |
|         return AVERROR_EOF;
 | |
|     }
 | |
| 
 | |
|     return ret;
 | |
| }
 | |
| 
 | |
| static int process_command(AVFilterContext *ctx,
 | |
|                            const char *cmd,
 | |
|                            const char *arg,
 | |
|                            char *res,
 | |
|                            int res_len,
 | |
|                            int flags)
 | |
| {
 | |
|     int ret = ff_filter_process_command(ctx, cmd, arg, res, res_len, flags);
 | |
| 
 | |
|     if (ret < 0)
 | |
|         return ret;
 | |
| 
 | |
|     return yae_update(ctx);
 | |
| }
 | |
| 
 | |
| static const AVFilterPad atempo_inputs[] = {
 | |
|     {
 | |
|         .name         = "default",
 | |
|         .type         = AVMEDIA_TYPE_AUDIO,
 | |
|         .filter_frame = filter_frame,
 | |
|         .config_props = config_props,
 | |
|     },
 | |
| };
 | |
| 
 | |
| static const AVFilterPad atempo_outputs[] = {
 | |
|     {
 | |
|         .name          = "default",
 | |
|         .request_frame = request_frame,
 | |
|         .type          = AVMEDIA_TYPE_AUDIO,
 | |
|     },
 | |
| };
 | |
| 
 | |
| const AVFilter ff_af_atempo = {
 | |
|     .name            = "atempo",
 | |
|     .description     = NULL_IF_CONFIG_SMALL("Adjust audio tempo."),
 | |
|     .init            = init,
 | |
|     .uninit          = uninit,
 | |
|     .process_command = process_command,
 | |
|     .priv_size       = sizeof(ATempoContext),
 | |
|     .priv_class      = &atempo_class,
 | |
|     FILTER_INPUTS(atempo_inputs),
 | |
|     FILTER_OUTPUTS(atempo_outputs),
 | |
|     FILTER_SAMPLEFMTS_ARRAY(sample_fmts),
 | |
| };
 |