1305 lines
40 KiB
C++
1305 lines
40 KiB
C++
/* -*- c-basic-offset: 4 indent-tabs-mode: nil -*- vi:set ts=8 sts=4 sw=4: */
|
|
|
|
/*
|
|
Rubber Band Library
|
|
An audio time-stretching and pitch-shifting library.
|
|
Copyright 2007-2024 Particular Programs Ltd.
|
|
|
|
This program is free software; you can redistribute it and/or
|
|
modify it under the terms of the GNU General Public License as
|
|
published by the Free Software Foundation; either version 2 of the
|
|
License, or (at your option) any later version. See the file
|
|
COPYING included with this distribution for more information.
|
|
|
|
Alternatively, if you have a valid commercial licence for the
|
|
Rubber Band Library obtained by agreement with the copyright
|
|
holders, you may redistribute and/or modify it under the terms
|
|
described in that licence.
|
|
|
|
If you wish to distribute code using the Rubber Band Library
|
|
under terms other than those of the GNU General Public License,
|
|
you must obtain a valid commercial licence before doing so.
|
|
*/
|
|
|
|
#include "R2Stretcher.h"
|
|
#include "StretcherChannelData.h"
|
|
|
|
#include "../common/StretchCalculator.h"
|
|
#include "../common/Resampler.h"
|
|
#include "../common/Profiler.h"
|
|
#include "../common/VectorOps.h"
|
|
#include "../common/sysutils.h"
|
|
#include "../common/mathmisc.h"
|
|
|
|
#include <cassert>
|
|
#include <cmath>
|
|
#include <set>
|
|
#include <map>
|
|
#include <deque>
|
|
#include <algorithm>
|
|
|
|
using namespace RubberBand;
|
|
|
|
namespace RubberBand {
|
|
|
|
#ifndef NO_THREADING
|
|
|
|
R2Stretcher::ProcessThread::ProcessThread(R2Stretcher *s, size_t c) :
|
|
m_s(s),
|
|
m_channel(c),
|
|
m_dataAvailable(std::string("data ") + char('A' + c)),
|
|
m_abandoning(false)
|
|
{ }
|
|
|
|
void
|
|
R2Stretcher::ProcessThread::run()
|
|
{
|
|
m_s->m_log.log(2, "thread getting going for channel", m_channel);
|
|
|
|
ChannelData &cd = *m_s->m_channelData[m_channel];
|
|
|
|
while (cd.inputSize == -1 ||
|
|
cd.inbuf->getReadSpace() > 0) {
|
|
|
|
bool any = false, last = false;
|
|
m_s->processChunks(m_channel, any, last);
|
|
|
|
if (last) break;
|
|
|
|
if (any) {
|
|
m_s->m_spaceAvailable.lock();
|
|
m_s->m_spaceAvailable.signal();
|
|
m_s->m_spaceAvailable.unlock();
|
|
}
|
|
|
|
m_dataAvailable.lock();
|
|
if (!m_s->testInbufReadSpace(m_channel) && !m_abandoning) {
|
|
m_dataAvailable.wait(50000); // bounded in case of abandonment
|
|
}
|
|
m_dataAvailable.unlock();
|
|
|
|
if (m_abandoning) {
|
|
m_s->m_log.log(2, "thread abandoning for channel", m_channel);
|
|
return;
|
|
}
|
|
}
|
|
|
|
bool any = false, last = false;
|
|
m_s->processChunks(m_channel, any, last);
|
|
m_s->m_spaceAvailable.lock();
|
|
m_s->m_spaceAvailable.signal();
|
|
m_s->m_spaceAvailable.unlock();
|
|
|
|
m_s->m_log.log(2, "thread done for channel", m_channel);
|
|
}
|
|
|
|
void
|
|
R2Stretcher::ProcessThread::signalDataAvailable()
|
|
{
|
|
m_dataAvailable.lock();
|
|
m_dataAvailable.signal();
|
|
m_dataAvailable.unlock();
|
|
}
|
|
|
|
void
|
|
R2Stretcher::ProcessThread::abandon()
|
|
{
|
|
m_abandoning = true;
|
|
}
|
|
|
|
#endif
|
|
|
|
bool
|
|
R2Stretcher::resampleBeforeStretching() const
|
|
{
|
|
// We can't resample before stretching in offline mode, because
|
|
// the stretch calculation is based on doing it the other way
|
|
// around. It would take more work (and testing) to enable this.
|
|
if (!m_realtime) return false;
|
|
|
|
if (m_options & RubberBandStretcher::OptionPitchHighQuality) {
|
|
return (m_pitchScale < 1.0); // better sound
|
|
} else if (m_options & RubberBandStretcher::OptionPitchHighConsistency) {
|
|
return false;
|
|
} else {
|
|
return (m_pitchScale > 1.0); // better performance
|
|
}
|
|
}
|
|
|
|
void
|
|
R2Stretcher::prepareChannelMS(size_t c,
|
|
const float *const *inputs,
|
|
size_t offset,
|
|
size_t samples,
|
|
float *prepared)
|
|
{
|
|
for (size_t i = 0; i < samples; ++i) {
|
|
float left = inputs[0][i + offset];
|
|
float right = inputs[1][i + offset];
|
|
float mid = (left + right) / 2;
|
|
float side = (left - right) / 2;
|
|
if (c == 0) {
|
|
prepared[i] = mid;
|
|
} else {
|
|
prepared[i] = side;
|
|
}
|
|
}
|
|
}
|
|
|
|
size_t
|
|
R2Stretcher::consumeChannel(size_t c,
|
|
const float *const *inputs,
|
|
size_t offset,
|
|
size_t samples,
|
|
bool final)
|
|
{
|
|
Profiler profiler("R2Stretcher::consumeChannel");
|
|
|
|
ChannelData &cd = *m_channelData[c];
|
|
RingBuffer<float> &inbuf = *cd.inbuf;
|
|
|
|
size_t toWrite = samples;
|
|
size_t writable = inbuf.getWriteSpace();
|
|
|
|
bool resampling = resampleBeforeStretching();
|
|
|
|
const float *input = 0;
|
|
|
|
bool useMidSide =
|
|
((m_options & RubberBandStretcher::OptionChannelsTogether) &&
|
|
(m_channels >= 2) &&
|
|
(c < 2));
|
|
|
|
if (resampling) {
|
|
|
|
Profiler profiler2("R2Stretcher::resample");
|
|
|
|
toWrite = int(ceil(samples / m_pitchScale));
|
|
bool shortened = false;
|
|
if (writable < toWrite) {
|
|
samples = int(floor(writable * m_pitchScale));
|
|
shortened = true;
|
|
if (samples == 0) return 0;
|
|
}
|
|
|
|
if (useMidSide) {
|
|
// Our pre-resample buffer size is also constrained by a
|
|
// buffer (cd.ms) of the same size as cd.inbuf
|
|
size_t ibsz = cd.inbuf->getSize();
|
|
if (ibsz < samples) {
|
|
samples = ibsz;
|
|
}
|
|
}
|
|
|
|
size_t reqSize = int(ceil(samples / m_pitchScale));
|
|
if (reqSize > cd.resamplebufSize) {
|
|
m_log.log(0, "WARNING: R2Stretcher::consumeChannel: resizing resampler buffer from and to", cd.resamplebufSize, reqSize);
|
|
cd.setResampleBufSize(reqSize);
|
|
}
|
|
|
|
#if defined(STRETCHER_IMPL_RESAMPLER_MUTEX_REQUIRED)
|
|
if (m_threaded) {
|
|
m_resamplerMutex.lock();
|
|
}
|
|
#endif
|
|
|
|
if (useMidSide) {
|
|
prepareChannelMS(c, inputs, offset, samples, cd.ms);
|
|
input = cd.ms;
|
|
} else {
|
|
input = inputs[c] + offset;
|
|
}
|
|
|
|
toWrite = cd.resampler->resample(&cd.resamplebuf,
|
|
cd.resamplebufSize,
|
|
&input,
|
|
samples,
|
|
1.0 / m_pitchScale,
|
|
final && !shortened);
|
|
|
|
#if defined(STRETCHER_IMPL_RESAMPLER_MUTEX_REQUIRED)
|
|
if (m_threaded) {
|
|
m_resamplerMutex.unlock();
|
|
}
|
|
#endif
|
|
}
|
|
|
|
if (writable < toWrite) {
|
|
if (resampling) {
|
|
m_log.log(1, "consumeChannel: resampler produced too much output, cannot use", toWrite, writable);
|
|
return 0;
|
|
}
|
|
toWrite = writable;
|
|
}
|
|
|
|
if (resampling) {
|
|
|
|
inbuf.write(cd.resamplebuf, toWrite);
|
|
cd.inCount += samples;
|
|
|
|
m_log.log(2, "consumeChannel: wrote to inbuf from resamplebuf, inCount now", toWrite, cd.inCount);
|
|
|
|
return samples;
|
|
|
|
} else {
|
|
|
|
if (useMidSide) {
|
|
prepareChannelMS(c, inputs, offset, toWrite, cd.ms);
|
|
input = cd.ms;
|
|
} else {
|
|
input = inputs[c] + offset;
|
|
}
|
|
|
|
inbuf.write(input, toWrite);
|
|
cd.inCount += toWrite;
|
|
|
|
m_log.log(2, "consumeChannel: wrote to inbuf from input, inCount now", toWrite, cd.inCount);
|
|
|
|
return toWrite;
|
|
}
|
|
}
|
|
|
|
void
|
|
R2Stretcher::processChunks(size_t c, bool &any, bool &last)
|
|
{
|
|
Profiler profiler("R2Stretcher::processChunks");
|
|
|
|
// Process as many chunks as there are available on the input
|
|
// buffer for channel c. This requires that the increments have
|
|
// already been calculated.
|
|
|
|
// This is the normal process method in offline mode.
|
|
|
|
ChannelData &cd = *m_channelData[c];
|
|
|
|
last = false;
|
|
any = false;
|
|
|
|
float *tmp = 0;
|
|
|
|
while (!last) {
|
|
|
|
if (!testInbufReadSpace(c)) {
|
|
m_log.log(2, "processChunks: out of input");
|
|
break;
|
|
}
|
|
|
|
any = true;
|
|
|
|
if (!cd.draining) {
|
|
size_t ready = cd.inbuf->getReadSpace();
|
|
assert(ready >= m_aWindowSize || cd.inputSize >= 0);
|
|
cd.inbuf->peek(cd.fltbuf, std::min(ready, m_aWindowSize));
|
|
cd.inbuf->skip(m_increment);
|
|
}
|
|
|
|
bool phaseReset = false;
|
|
size_t phaseIncrement, shiftIncrement;
|
|
getIncrements(c, phaseIncrement, shiftIncrement, phaseReset);
|
|
|
|
if (shiftIncrement <= m_aWindowSize) {
|
|
analyseChunk(c);
|
|
last = processChunkForChannel
|
|
(c, phaseIncrement, shiftIncrement, phaseReset);
|
|
} else {
|
|
size_t bit = m_aWindowSize/4;
|
|
m_log.log(2, "breaking down overlong increment into chunks from and to", shiftIncrement, bit);
|
|
if (!tmp) tmp = allocate<float>(m_aWindowSize);
|
|
analyseChunk(c);
|
|
v_copy(tmp, cd.fltbuf, m_aWindowSize);
|
|
for (size_t i = 0; i < shiftIncrement; i += bit) {
|
|
v_copy(cd.fltbuf, tmp, m_aWindowSize);
|
|
size_t thisIncrement = bit;
|
|
if (i + thisIncrement > shiftIncrement) {
|
|
thisIncrement = shiftIncrement - i;
|
|
}
|
|
last = processChunkForChannel
|
|
(c, phaseIncrement + i, thisIncrement, phaseReset);
|
|
phaseReset = false;
|
|
}
|
|
}
|
|
|
|
cd.chunkCount++;
|
|
m_log.log(3, "channel/last", c, last);
|
|
m_log.log(3, "channel/chunkCount", c, cd.chunkCount);
|
|
}
|
|
|
|
if (tmp) deallocate(tmp);
|
|
}
|
|
|
|
bool
|
|
R2Stretcher::processOneChunk()
|
|
{
|
|
Profiler profiler("R2Stretcher::processOneChunk");
|
|
|
|
m_log.log(3, "R2Stretcher::processOneChunk");
|
|
|
|
// Process a single chunk for all channels, provided there is
|
|
// enough data on each channel for at least one chunk. This is
|
|
// able to calculate increments as it goes along.
|
|
|
|
// This is the normal process method in RT mode.
|
|
|
|
for (size_t c = 0; c < m_channels; ++c) {
|
|
if (!testInbufReadSpace(c)) {
|
|
m_log.log(2, "processOneChunk: out of input");
|
|
return false;
|
|
}
|
|
ChannelData &cd = *m_channelData[c];
|
|
m_log.log(2, "read space and draining", cd.inbuf->getReadSpace(), cd.draining);
|
|
if (!cd.draining) {
|
|
size_t ready = cd.inbuf->getReadSpace();
|
|
assert(ready >= m_aWindowSize || cd.inputSize >= 0);
|
|
cd.inbuf->peek(cd.fltbuf, std::min(ready, m_aWindowSize));
|
|
cd.inbuf->skip(m_increment);
|
|
analyseChunk(c);
|
|
}
|
|
}
|
|
|
|
bool phaseReset = false;
|
|
size_t phaseIncrement, shiftIncrement;
|
|
if (!getIncrements(0, phaseIncrement, shiftIncrement, phaseReset)) {
|
|
calculateIncrements(phaseIncrement, shiftIncrement, phaseReset);
|
|
}
|
|
|
|
bool last = false;
|
|
for (size_t c = 0; c < m_channels; ++c) {
|
|
last = processChunkForChannel(c, phaseIncrement, shiftIncrement, phaseReset);
|
|
m_channelData[c]->chunkCount++;
|
|
}
|
|
|
|
m_log.log(3, "R2Stretcher::processOneChunk returning", last);
|
|
return last;
|
|
}
|
|
|
|
bool
|
|
R2Stretcher::testInbufReadSpace(size_t c)
|
|
{
|
|
Profiler profiler("R2Stretcher::testInbufReadSpace");
|
|
|
|
ChannelData &cd = *m_channelData[c];
|
|
RingBuffer<float> &inbuf = *cd.inbuf;
|
|
|
|
size_t rs = inbuf.getReadSpace();
|
|
|
|
if (rs < m_aWindowSize && !cd.draining) {
|
|
|
|
if (cd.inputSize == -1) {
|
|
|
|
// Not all the input data has been written to the inbuf
|
|
// (that's why the input size is not yet set). We can't
|
|
// process, because we don't have a full chunk of data, so
|
|
// our process chunk would contain some empty padding in
|
|
// its input -- and that would give incorrect output, as
|
|
// we know there is more input to come.
|
|
|
|
#ifndef NO_THREADING
|
|
if (!m_threaded) {
|
|
#endif
|
|
m_log.log(2, "Note: read space < chunk size when not all input written", inbuf.getReadSpace(), m_aWindowSize);
|
|
|
|
#ifndef NO_THREADING
|
|
}
|
|
#endif
|
|
return false;
|
|
}
|
|
|
|
if (rs == 0) {
|
|
m_log.log(2, "read space = 0, giving up");
|
|
return false;
|
|
} else if (rs < m_aWindowSize/2) {
|
|
m_log.log(2, "setting draining true with read space and window size", rs, m_aWindowSize);
|
|
m_log.log(2, "outbuf read space is", cd.outbuf->getReadSpace());
|
|
m_log.log(2, "accumulator fill is", cd.accumulatorFill);
|
|
cd.draining = true;
|
|
}
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool
|
|
R2Stretcher::processChunkForChannel(size_t c,
|
|
size_t phaseIncrement,
|
|
size_t shiftIncrement,
|
|
bool phaseReset)
|
|
{
|
|
Profiler profiler("R2Stretcher::processChunkForChannel");
|
|
|
|
// Process a single chunk on a single channel. This assumes
|
|
// enough input data is available; caller must have tested this
|
|
// using e.g. testInbufReadSpace first. Return true if this is
|
|
// the last chunk on the channel.
|
|
|
|
if (phaseReset) {
|
|
m_log.log(2, "processChunkForChannel: phase reset found, increments",
|
|
phaseIncrement, shiftIncrement);
|
|
}
|
|
|
|
ChannelData &cd = *m_channelData[c];
|
|
|
|
if (!cd.draining) {
|
|
|
|
// This is the normal processing case -- draining is only
|
|
// set when all the input has been used and we only need
|
|
// to write from the existing accumulator into the output.
|
|
|
|
// We know we have enough samples available in m_inbuf --
|
|
// this is usually m_aWindowSize, but we know that if fewer
|
|
// are available, it's OK to use zeroes for the rest
|
|
// (which the ring buffer will provide) because we've
|
|
// reached the true end of the data.
|
|
|
|
// We need to peek m_aWindowSize samples for processing, and
|
|
// then skip m_increment to advance the read pointer.
|
|
|
|
modifyChunk(c, phaseIncrement, phaseReset);
|
|
synthesiseChunk(c, shiftIncrement); // reads from cd.mag, cd.phase
|
|
|
|
if (m_log.getDebugLevel() > 2) {
|
|
if (phaseReset) {
|
|
for (int i = 0; i < 10; ++i) {
|
|
cd.accumulator[i] = 1.2f - (i % 3) * 1.2f;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
bool last = false;
|
|
|
|
if (cd.draining) {
|
|
m_log.log(2, "draining: accumulator fill and shift increment", cd.accumulatorFill, shiftIncrement);
|
|
m_log.log(2, "outbuf read space is", cd.outbuf->getReadSpace());
|
|
if (cd.accumulatorFill == 0) {
|
|
m_log.log(2, "draining: accumulator empty");
|
|
return true;
|
|
}
|
|
if (shiftIncrement == 0) {
|
|
m_log.log(0, "WARNING: draining: shiftIncrement == 0, can't handle that in this context: setting to", m_increment);
|
|
shiftIncrement = m_increment;
|
|
}
|
|
if (cd.accumulatorFill <= shiftIncrement) {
|
|
m_log.log(2, "draining: marking as last and reducing shift increment from and to", shiftIncrement, cd.accumulatorFill);
|
|
shiftIncrement = cd.accumulatorFill;
|
|
last = true;
|
|
}
|
|
}
|
|
|
|
int required = shiftIncrement;
|
|
|
|
if (m_pitchScale != 1.0) {
|
|
required = int(required / m_pitchScale) + 1;
|
|
}
|
|
|
|
int ws = cd.outbuf->getWriteSpace();
|
|
if (ws < required) {
|
|
m_log.log(1, "Buffer overrun on output for channel", c);
|
|
|
|
// The only correct thing we can do here is resize the buffer.
|
|
// We can't wait for the client thread to read some data out
|
|
// from the buffer so as to make more space, because the
|
|
// client thread (if we are threaded at all) is probably stuck
|
|
// in a process() call waiting for us to stow away enough
|
|
// input increments to allow the process() call to complete.
|
|
// This is an unhappy situation.
|
|
|
|
RingBuffer<float> *oldbuf = cd.outbuf;
|
|
cd.outbuf = oldbuf->resized(oldbuf->getSize() * 2);
|
|
|
|
m_log.log(2, "write space and space needed", ws, required);
|
|
m_log.log(2, "resized output buffer from and to", oldbuf->getSize(),
|
|
cd.outbuf->getSize());
|
|
|
|
m_emergencyScavenger.claim(oldbuf);
|
|
}
|
|
|
|
writeChunk(c, shiftIncrement, last);
|
|
m_log.log(3, "processChunkForChannel: accumulatorFill now; returning", cd.accumulatorFill, last);
|
|
return last;
|
|
}
|
|
|
|
void
|
|
R2Stretcher::calculateIncrements(size_t &phaseIncrementRtn,
|
|
size_t &shiftIncrementRtn,
|
|
bool &phaseReset)
|
|
{
|
|
Profiler profiler("R2Stretcher::calculateIncrements");
|
|
|
|
// Calculate the next upcoming phase and shift increment, on the
|
|
// basis that both channels are in sync. This is in contrast to
|
|
// getIncrements, which requires that all the increments have been
|
|
// calculated in advance but can then return increments
|
|
// corresponding to different chunks in different channels.
|
|
|
|
// Requires frequency domain representations of channel data in
|
|
// the mag and phase buffers in the channel.
|
|
|
|
// This function is only used in real-time mode.
|
|
|
|
phaseIncrementRtn = m_increment;
|
|
shiftIncrementRtn = m_increment;
|
|
phaseReset = false;
|
|
|
|
if (m_channels == 0) return;
|
|
|
|
ChannelData &cd = *m_channelData[0];
|
|
|
|
size_t bc = cd.chunkCount;
|
|
for (size_t c = 1; c < m_channels; ++c) {
|
|
if (m_channelData[c]->chunkCount != bc) {
|
|
m_log.log(0, "ERROR: R2Stretcher::calculateIncrements: Channels are not in sync");
|
|
return;
|
|
}
|
|
}
|
|
|
|
const int hs = m_fftSize/2 + 1;
|
|
|
|
// Normally we would mix down the time-domain signal and apply a
|
|
// single FFT, or else mix down the Cartesian form of the
|
|
// frequency-domain signal. Both of those would be inefficient
|
|
// from this position. Fortunately, the onset detectors should
|
|
// work reasonably well (maybe even better?) if we just sum the
|
|
// magnitudes of the frequency-domain channel signals and forget
|
|
// about phase entirely. Normally we don't expect the channel
|
|
// phases to cancel each other, and broadband effects will still
|
|
// be apparent.
|
|
|
|
float df = 0.f;
|
|
bool silent = false;
|
|
|
|
if (m_channels == 1) {
|
|
|
|
if (sizeof(process_t) == sizeof(double)) {
|
|
df = m_phaseResetAudioCurve->processDouble((double *)cd.mag, m_increment);
|
|
silent = (m_silentAudioCurve->processDouble((double *)cd.mag, m_increment) > 0.f);
|
|
} else {
|
|
df = m_phaseResetAudioCurve->processFloat((float *)cd.mag, m_increment);
|
|
silent = (m_silentAudioCurve->processFloat((float *)cd.mag, m_increment) > 0.f);
|
|
}
|
|
|
|
} else {
|
|
|
|
process_t *tmp = (process_t *)alloca(hs * sizeof(process_t));
|
|
|
|
v_zero(tmp, hs);
|
|
for (size_t c = 0; c < m_channels; ++c) {
|
|
v_add(tmp, m_channelData[c]->mag, hs);
|
|
}
|
|
|
|
if (sizeof(process_t) == sizeof(double)) {
|
|
df = m_phaseResetAudioCurve->processDouble((double *)tmp, m_increment);
|
|
silent = (m_silentAudioCurve->processDouble((double *)tmp, m_increment) > 0.f);
|
|
} else {
|
|
df = m_phaseResetAudioCurve->processFloat((float *)tmp, m_increment);
|
|
silent = (m_silentAudioCurve->processFloat((float *)tmp, m_increment) > 0.f);
|
|
}
|
|
}
|
|
|
|
double effectivePitchRatio = 1.0 / m_pitchScale;
|
|
if (cd.resampler) {
|
|
effectivePitchRatio = cd.resampler->getEffectiveRatio(effectivePitchRatio);
|
|
}
|
|
|
|
int incr = m_stretchCalculator->calculateSingle
|
|
(m_timeRatio, effectivePitchRatio, df, m_increment,
|
|
m_aWindowSize, m_sWindowSize, false);
|
|
|
|
if (m_lastProcessPhaseResetDf.getWriteSpace() > 0) {
|
|
m_lastProcessPhaseResetDf.write(&df, 1);
|
|
}
|
|
if (m_lastProcessOutputIncrements.getWriteSpace() > 0) {
|
|
m_lastProcessOutputIncrements.write(&incr, 1);
|
|
}
|
|
|
|
if (incr < 0) {
|
|
phaseReset = true;
|
|
incr = -incr;
|
|
}
|
|
|
|
// The returned increment is the phase increment. The shift
|
|
// increment for one chunk is the same as the phase increment for
|
|
// the following chunk (see comment below). This means we don't
|
|
// actually know the shift increment until we see the following
|
|
// phase increment... which is a bit of a problem.
|
|
|
|
// This implies we should use this increment for the shift
|
|
// increment, and make the following phase increment the same as
|
|
// it. This means in RT mode we'll be one chunk later with our
|
|
// phase reset than we would be in non-RT mode. The sensitivity
|
|
// of the broadband onset detector may mean that this isn't a
|
|
// problem -- test it and see.
|
|
|
|
shiftIncrementRtn = incr;
|
|
|
|
if (cd.prevIncrement == 0) {
|
|
phaseIncrementRtn = shiftIncrementRtn;
|
|
} else {
|
|
phaseIncrementRtn = cd.prevIncrement;
|
|
}
|
|
|
|
cd.prevIncrement = shiftIncrementRtn;
|
|
|
|
if (silent) ++m_silentHistory;
|
|
else m_silentHistory = 0;
|
|
|
|
if (m_silentHistory >= int(m_aWindowSize / m_increment) && !phaseReset) {
|
|
phaseReset = true;
|
|
m_log.log(2, "calculateIncrements: phase reset on silence: silent history", m_silentHistory);
|
|
}
|
|
}
|
|
|
|
bool
|
|
R2Stretcher::getIncrements(size_t channel,
|
|
size_t &phaseIncrementRtn,
|
|
size_t &shiftIncrementRtn,
|
|
bool &phaseReset)
|
|
{
|
|
Profiler profiler("R2Stretcher::getIncrements");
|
|
|
|
if (channel >= m_channels) {
|
|
phaseIncrementRtn = m_increment;
|
|
shiftIncrementRtn = m_increment;
|
|
phaseReset = false;
|
|
return false;
|
|
}
|
|
|
|
// There are two relevant output increments here. The first is
|
|
// the phase increment which we use when recalculating the phases
|
|
// for the current chunk; the second is the shift increment used
|
|
// to determine how far to shift the processing buffer after
|
|
// writing the chunk. The shift increment for one chunk is the
|
|
// same as the phase increment for the following chunk.
|
|
|
|
// When an onset occurs for which we need to reset phases, the
|
|
// increment given will be negative.
|
|
|
|
// When we reset phases, the previous shift increment (and so
|
|
// current phase increments) must have been m_increment to ensure
|
|
// consistency.
|
|
|
|
// m_outputIncrements stores phase increments.
|
|
|
|
ChannelData &cd = *m_channelData[channel];
|
|
bool gotData = true;
|
|
|
|
if (cd.chunkCount >= m_outputIncrements.size()) {
|
|
if (m_outputIncrements.size() == 0) {
|
|
phaseIncrementRtn = m_increment;
|
|
shiftIncrementRtn = m_increment;
|
|
phaseReset = false;
|
|
return false;
|
|
} else {
|
|
cd.chunkCount = m_outputIncrements.size()-1;
|
|
gotData = false;
|
|
}
|
|
}
|
|
|
|
int phaseIncrement = m_outputIncrements[cd.chunkCount];
|
|
|
|
int shiftIncrement = phaseIncrement;
|
|
if (cd.chunkCount + 1 < m_outputIncrements.size()) {
|
|
shiftIncrement = m_outputIncrements[cd.chunkCount + 1];
|
|
}
|
|
|
|
if (phaseIncrement < 0) {
|
|
phaseIncrement = -phaseIncrement;
|
|
phaseReset = true;
|
|
}
|
|
|
|
if (shiftIncrement < 0) {
|
|
shiftIncrement = -shiftIncrement;
|
|
}
|
|
|
|
if (shiftIncrement >= int(m_aWindowSize)) {
|
|
m_log.log(1, "WARNING: shiftIncrement >= analysis window size", shiftIncrement, m_aWindowSize);
|
|
m_log.log(1, "at chunk of total", cd.chunkCount, m_outputIncrements.size());
|
|
}
|
|
|
|
phaseIncrementRtn = phaseIncrement;
|
|
shiftIncrementRtn = shiftIncrement;
|
|
if (cd.chunkCount == 0) phaseReset = true; // don't mess with the first chunk
|
|
return gotData;
|
|
}
|
|
|
|
void
|
|
R2Stretcher::analyseChunk(size_t channel)
|
|
{
|
|
Profiler profiler("R2Stretcher::analyseChunk");
|
|
|
|
ChannelData &cd = *m_channelData[channel];
|
|
|
|
process_t *const R__ dblbuf = cd.dblbuf;
|
|
float *const R__ fltbuf = cd.fltbuf;
|
|
|
|
// cd.fltbuf is known to contain m_aWindowSize samples
|
|
|
|
if (m_aWindowSize > m_fftSize) {
|
|
m_afilter->cut(fltbuf);
|
|
}
|
|
|
|
cutShiftAndFold(dblbuf, m_fftSize, fltbuf, m_awindow);
|
|
|
|
cd.fft->forwardPolar(dblbuf, cd.mag, cd.phase);
|
|
}
|
|
|
|
void
|
|
R2Stretcher::modifyChunk(size_t channel,
|
|
size_t outputIncrement,
|
|
bool phaseReset)
|
|
{
|
|
Profiler profiler("R2Stretcher::modifyChunk");
|
|
|
|
ChannelData &cd = *m_channelData[channel];
|
|
|
|
if (phaseReset) {
|
|
m_log.log(2, "phase reset: leaving phases unmodified");
|
|
}
|
|
|
|
const process_t rate = process_t(m_sampleRate);
|
|
const int count = m_fftSize / 2;
|
|
|
|
bool unchanged = cd.unchanged && (outputIncrement == m_increment);
|
|
bool fullReset = phaseReset;
|
|
bool laminar = !(m_options & RubberBandStretcher::OptionPhaseIndependent);
|
|
bool bandlimited = (m_options & RubberBandStretcher::OptionTransientsMixed);
|
|
int bandlow = lrint((150 * m_fftSize) / rate);
|
|
int bandhigh = lrint((1000 * m_fftSize) / rate);
|
|
|
|
float r = getEffectiveRatio();
|
|
|
|
bool unity = (fabsf(r - 1.f) < 1.e-6f);
|
|
if (unity) {
|
|
if (!phaseReset) {
|
|
phaseReset = true;
|
|
bandlimited = true;
|
|
bandlow = lrint((cd.unityResetLow * m_fftSize) / rate);
|
|
bandhigh = count;
|
|
if (bandlow > 0) {
|
|
m_log.log(2, "unity: bandlow & high", bandlow, bandhigh);
|
|
}
|
|
}
|
|
cd.unityResetLow *= 0.9f;
|
|
} else {
|
|
cd.unityResetLow = 16000.f;
|
|
}
|
|
|
|
float freq0 = m_freq0;
|
|
float freq1 = m_freq1;
|
|
float freq2 = m_freq2;
|
|
|
|
if (laminar) {
|
|
if (r > 1) {
|
|
float rf0 = 600 + (600 * ((r-1)*(r-1)*(r-1)*2));
|
|
float f1ratio = freq1 / freq0;
|
|
float f2ratio = freq2 / freq0;
|
|
freq0 = std::max(freq0, rf0);
|
|
freq1 = freq0 * f1ratio;
|
|
freq2 = freq0 * f2ratio;
|
|
}
|
|
}
|
|
|
|
int limit0 = lrint((freq0 * m_fftSize) / rate);
|
|
int limit1 = lrint((freq1 * m_fftSize) / rate);
|
|
int limit2 = lrint((freq2 * m_fftSize) / rate);
|
|
|
|
if (limit1 < limit0) limit1 = limit0;
|
|
if (limit2 < limit1) limit2 = limit1;
|
|
|
|
process_t prevInstability = 0.0;
|
|
bool prevDirection = false;
|
|
|
|
process_t distance = 0.0;
|
|
const process_t maxdist = 8.0;
|
|
|
|
const int lookback = 1;
|
|
|
|
process_t distacc = 0.0;
|
|
|
|
for (int i = count; i >= 0; i -= lookback) {
|
|
|
|
bool resetThis = phaseReset;
|
|
|
|
if (bandlimited) {
|
|
if (resetThis) {
|
|
if (i > bandlow && i < bandhigh) {
|
|
resetThis = false;
|
|
fullReset = false;
|
|
}
|
|
}
|
|
}
|
|
|
|
process_t p = cd.phase[i];
|
|
process_t perr = 0.0;
|
|
process_t outphase = p;
|
|
|
|
process_t mi = maxdist;
|
|
if (i <= limit0) mi = 0.0;
|
|
else if (i <= limit1) mi = 1.0;
|
|
else if (i <= limit2) mi = 3.0;
|
|
|
|
if (!resetThis) {
|
|
|
|
process_t omega = (2 * M_PI * m_increment * i) / (m_fftSize);
|
|
|
|
process_t pp = cd.prevPhase[i];
|
|
process_t ep = pp + omega;
|
|
perr = princarg(p - ep);
|
|
|
|
process_t instability = fabs(perr - cd.prevError[i]);
|
|
bool direction = (perr > cd.prevError[i]);
|
|
|
|
bool inherit = false;
|
|
|
|
if (laminar) {
|
|
if (distance >= mi || i == count) {
|
|
inherit = false;
|
|
} else if (bandlimited && (i == bandhigh || i == bandlow)) {
|
|
inherit = false;
|
|
} else if (instability > prevInstability &&
|
|
direction == prevDirection) {
|
|
inherit = true;
|
|
}
|
|
}
|
|
|
|
process_t advance = outputIncrement * ((omega + perr) / m_increment);
|
|
|
|
if (inherit) {
|
|
process_t inherited =
|
|
cd.unwrappedPhase[i + lookback] - cd.prevPhase[i + lookback];
|
|
advance = ((advance * distance) +
|
|
(inherited * (maxdist - distance)))
|
|
/ maxdist;
|
|
outphase = p + advance;
|
|
distacc += distance;
|
|
distance += 1.0;
|
|
} else {
|
|
outphase = cd.unwrappedPhase[i] + advance;
|
|
distance = 0.0;
|
|
}
|
|
|
|
prevInstability = instability;
|
|
prevDirection = direction;
|
|
|
|
} else {
|
|
distance = 0.0;
|
|
}
|
|
|
|
cd.prevError[i] = perr;
|
|
cd.prevPhase[i] = p;
|
|
cd.phase[i] = outphase;
|
|
cd.unwrappedPhase[i] = outphase;
|
|
}
|
|
|
|
m_log.log(3, "mean inheritance distance", distacc / count);
|
|
|
|
if (fullReset) unchanged = true;
|
|
cd.unchanged = unchanged;
|
|
|
|
if (unchanged) {
|
|
m_log.log(2, "frame unchanged on channel", channel);
|
|
}
|
|
}
|
|
|
|
|
|
void
|
|
R2Stretcher::formantShiftChunk(size_t channel)
|
|
{
|
|
Profiler profiler("R2Stretcher::formantShiftChunk");
|
|
|
|
ChannelData &cd = *m_channelData[channel];
|
|
|
|
process_t *const R__ mag = cd.mag;
|
|
process_t *const R__ envelope = cd.envelope;
|
|
process_t *const R__ dblbuf = cd.dblbuf;
|
|
|
|
const int sz = m_fftSize;
|
|
const int hs = sz / 2;
|
|
const process_t factor = 1.0 / sz;
|
|
|
|
cd.fft->inverseCepstral(mag, dblbuf);
|
|
|
|
const int cutoff = m_sampleRate / 700;
|
|
|
|
dblbuf[0] /= 2;
|
|
dblbuf[cutoff-1] /= 2;
|
|
|
|
for (int i = cutoff; i < sz; ++i) {
|
|
dblbuf[i] = 0.0;
|
|
}
|
|
|
|
v_scale(dblbuf, factor, cutoff);
|
|
|
|
process_t *spare = (process_t *)alloca((hs + 1) * sizeof(process_t));
|
|
cd.fft->forward(dblbuf, envelope, spare);
|
|
|
|
v_exp(envelope, hs + 1);
|
|
v_divide(mag, envelope, hs + 1);
|
|
|
|
if (m_pitchScale > 1.0) {
|
|
// scaling up, we want a new envelope that is lower by the pitch factor
|
|
for (int target = 0; target <= hs; ++target) {
|
|
int source = lrint(target * m_pitchScale);
|
|
if (source > hs) {
|
|
envelope[target] = 0.0;
|
|
} else {
|
|
envelope[target] = envelope[source];
|
|
}
|
|
}
|
|
} else {
|
|
// scaling down, we want a new envelope that is higher by the pitch factor
|
|
for (int target = hs; target > 0; ) {
|
|
--target;
|
|
int source = lrint(target * m_pitchScale);
|
|
envelope[target] = envelope[source];
|
|
}
|
|
}
|
|
|
|
v_multiply(mag, envelope, hs+1);
|
|
|
|
cd.unchanged = false;
|
|
}
|
|
|
|
void
|
|
R2Stretcher::synthesiseChunk(size_t channel,
|
|
size_t shiftIncrement)
|
|
{
|
|
Profiler profiler("R2Stretcher::synthesiseChunk");
|
|
|
|
if ((m_options & RubberBandStretcher::OptionFormantPreserved) &&
|
|
(m_pitchScale != 1.0)) {
|
|
formantShiftChunk(channel);
|
|
}
|
|
|
|
ChannelData &cd = *m_channelData[channel];
|
|
|
|
process_t *const R__ dblbuf = cd.dblbuf;
|
|
float *const R__ fltbuf = cd.fltbuf;
|
|
float *const R__ accumulator = cd.accumulator;
|
|
float *const R__ windowAccumulator = cd.windowAccumulator;
|
|
|
|
const int fsz = m_fftSize;
|
|
const int hs = fsz / 2;
|
|
|
|
const int wsz = m_sWindowSize;
|
|
|
|
if (!cd.unchanged) {
|
|
|
|
// Our FFTs produced unscaled results. Scale before inverse
|
|
// transform rather than after, to avoid overflow if using a
|
|
// fixed-point FFT.
|
|
float factor = 1.f / fsz;
|
|
v_scale(cd.mag, factor, hs + 1);
|
|
|
|
cd.fft->inversePolar(cd.mag, cd.phase, cd.dblbuf);
|
|
|
|
if (wsz == fsz) {
|
|
v_convert(fltbuf, dblbuf + hs, hs);
|
|
v_convert(fltbuf + hs, dblbuf, hs);
|
|
} else {
|
|
v_zero(fltbuf, wsz);
|
|
int j = fsz - wsz/2;
|
|
while (j < 0) j += fsz;
|
|
for (int i = 0; i < wsz; ++i) {
|
|
fltbuf[i] += dblbuf[j];
|
|
if (++j == fsz) j = 0;
|
|
}
|
|
}
|
|
}
|
|
|
|
if (wsz > fsz) {
|
|
int p = shiftIncrement * 2;
|
|
if (cd.interpolatorScale != p) {
|
|
SincWindow<float>::write(cd.interpolator, wsz, p);
|
|
cd.interpolatorScale = p;
|
|
}
|
|
v_multiply(fltbuf, cd.interpolator, wsz);
|
|
}
|
|
|
|
m_swindow->cut(fltbuf);
|
|
v_add(accumulator, fltbuf, wsz);
|
|
cd.accumulatorFill = std::max(cd.accumulatorFill, size_t(wsz));
|
|
|
|
if (wsz > fsz) {
|
|
// reuse fltbuf to calculate interpolating window shape for
|
|
// window accumulator
|
|
v_copy(fltbuf, cd.interpolator, wsz);
|
|
m_swindow->cut(fltbuf);
|
|
v_add(windowAccumulator, fltbuf, wsz);
|
|
} else {
|
|
m_swindow->add(windowAccumulator, m_awindow->getArea() * 1.5f);
|
|
}
|
|
}
|
|
|
|
void
|
|
R2Stretcher::writeChunk(size_t channel, size_t shiftIncrement, bool last)
|
|
{
|
|
Profiler profiler("R2Stretcher::writeChunk");
|
|
|
|
ChannelData &cd = *m_channelData[channel];
|
|
|
|
float *const R__ accumulator = cd.accumulator;
|
|
float *const R__ windowAccumulator = cd.windowAccumulator;
|
|
|
|
const int sz = cd.accumulatorFill;
|
|
const int si = shiftIncrement;
|
|
|
|
m_log.log(3, "writeChunk: channel and shiftIncrement",
|
|
channel, shiftIncrement);
|
|
if (last) {
|
|
m_log.log(3, "writeChunk: last true");
|
|
}
|
|
|
|
v_divide(accumulator, windowAccumulator, si);
|
|
|
|
// for exact sample scaling (probably not meaningful if we
|
|
// were running in RT mode)
|
|
size_t theoreticalOut = 0;
|
|
if (cd.inputSize >= 0) {
|
|
theoreticalOut = lrint(cd.inputSize * m_timeRatio);
|
|
}
|
|
|
|
bool resampledAlready = resampleBeforeStretching();
|
|
|
|
if (!resampledAlready &&
|
|
(m_pitchScale != 1.0 ||
|
|
(m_options & RubberBandStretcher::OptionPitchHighConsistency)) &&
|
|
cd.resampler) {
|
|
|
|
Profiler profiler2("R2Stretcher::resample");
|
|
|
|
size_t reqSize = int(ceil(si / m_pitchScale));
|
|
if (reqSize > cd.resamplebufSize) {
|
|
// This shouldn't normally happen -- the buffer is
|
|
// supposed to be initialised with enough space in the
|
|
// first place. But we retain this check in case the
|
|
// pitch scale has changed since then, or the stretch
|
|
// calculator has gone mad, or something.
|
|
m_log.log(0, "WARNING: R2Stretcher::writeChunk: resizing resampler buffer from and to", cd.resamplebufSize, reqSize);
|
|
cd.setResampleBufSize(reqSize);
|
|
}
|
|
|
|
#if defined(STRETCHER_IMPL_RESAMPLER_MUTEX_REQUIRED)
|
|
if (m_threaded) {
|
|
m_resamplerMutex.lock();
|
|
}
|
|
#endif
|
|
|
|
size_t outframes = cd.resampler->resample(&cd.resamplebuf,
|
|
cd.resamplebufSize,
|
|
&cd.accumulator,
|
|
si,
|
|
1.0 / m_pitchScale,
|
|
last);
|
|
|
|
#if defined(STRETCHER_IMPL_RESAMPLER_MUTEX_REQUIRED)
|
|
if (m_threaded) {
|
|
m_resamplerMutex.unlock();
|
|
}
|
|
#endif
|
|
|
|
writeOutput(*cd.outbuf, cd.resamplebuf,
|
|
outframes, cd.outCount, theoreticalOut);
|
|
|
|
} else {
|
|
writeOutput(*cd.outbuf, accumulator,
|
|
si, cd.outCount, theoreticalOut);
|
|
}
|
|
|
|
v_move(accumulator, accumulator + si, sz - si);
|
|
v_zero(accumulator + sz - si, si);
|
|
|
|
v_move(windowAccumulator, windowAccumulator + si, sz - si);
|
|
v_zero(windowAccumulator + sz - si, si);
|
|
|
|
if (int(cd.accumulatorFill) > si) {
|
|
cd.accumulatorFill -= si;
|
|
} else {
|
|
cd.accumulatorFill = 0;
|
|
if (cd.draining) {
|
|
m_log.log(2, "writeChunk: setting outputComplete to true");
|
|
cd.outputComplete = true;
|
|
}
|
|
}
|
|
|
|
m_log.log(3, "writeChunk: accumulatorFill now", cd.accumulatorFill);
|
|
}
|
|
|
|
void
|
|
R2Stretcher::writeOutput(RingBuffer<float> &to,
|
|
float *from, size_t qty,
|
|
size_t &outCount, size_t theoreticalOut)
|
|
{
|
|
Profiler profiler("R2Stretcher::writeOutput");
|
|
|
|
// In non-RT mode, we don't want to write the first startSkip
|
|
// samples, because the first chunk is centred on the start of the
|
|
// output. In RT mode we didn't apply any pre-padding in
|
|
// configure(), so we don't want to remove any here.
|
|
|
|
size_t startSkip = 0;
|
|
if (!m_realtime) {
|
|
startSkip = lrintf((m_sWindowSize/2) / m_pitchScale);
|
|
}
|
|
|
|
if (outCount > startSkip) {
|
|
|
|
// this is the normal case
|
|
|
|
if (theoreticalOut > 0) {
|
|
m_log.log(2, "theoreticalOut and outCount",
|
|
theoreticalOut, outCount);
|
|
m_log.log(2, "startSkip and qty",
|
|
startSkip, qty);
|
|
if (outCount - startSkip <= theoreticalOut &&
|
|
outCount - startSkip + qty > theoreticalOut) {
|
|
qty = theoreticalOut - (outCount - startSkip);
|
|
m_log.log(2, "reducing qty to", qty);
|
|
}
|
|
}
|
|
|
|
m_log.log(3, "writing", qty);
|
|
|
|
size_t written = to.write(from, qty);
|
|
|
|
if (written < qty) {
|
|
m_log.log(0, "WARNING: writeOutput: buffer overrun: wanted to write and able to write", qty, written);
|
|
}
|
|
|
|
outCount += written;
|
|
|
|
m_log.log(3, "written and new outCount", written, outCount);
|
|
return;
|
|
}
|
|
|
|
// the rest of this is only used during the first startSkip samples
|
|
|
|
if (outCount + qty <= startSkip) {
|
|
m_log.log(2, "discarding with startSkip", startSkip);
|
|
m_log.log(2, "qty and outCount", qty, outCount);
|
|
outCount += qty;
|
|
return;
|
|
}
|
|
|
|
size_t off = startSkip - outCount;
|
|
m_log.log(2, "shortening with startSkip", startSkip);
|
|
m_log.log(2, "qty and outCount", qty, outCount);
|
|
m_log.log(2, "start offset and number written", off, qty - off);
|
|
to.write(from + off, qty - off);
|
|
outCount += qty;
|
|
}
|
|
|
|
int
|
|
R2Stretcher::available() const
|
|
{
|
|
Profiler profiler("R2Stretcher::available");
|
|
|
|
m_log.log(3, "R2Stretcher::available");
|
|
|
|
#ifndef NO_THREADING
|
|
if (m_threaded) {
|
|
MutexLocker locker(&m_threadSetMutex);
|
|
if (m_channelData.empty()) return 0;
|
|
} else {
|
|
if (m_channelData.empty()) return 0;
|
|
}
|
|
#endif
|
|
|
|
#ifndef NO_THREADING
|
|
if (!m_threaded) {
|
|
#endif
|
|
if (m_channelData[0]->inputSize >= 0) {
|
|
//!!! do we ever actually do this? if so, this method should not be const
|
|
// ^^^ yes, we do sometimes -- e.g. when fed a very short file
|
|
if (m_realtime) {
|
|
while (m_channelData[0]->inbuf->getReadSpace() > 0 ||
|
|
m_channelData[0]->draining) {
|
|
m_log.log(2, "calling processOneChunk from available");
|
|
if (((R2Stretcher *)this)->processOneChunk()) {
|
|
break;
|
|
}
|
|
}
|
|
} else {
|
|
for (size_t c = 0; c < m_channels; ++c) {
|
|
if (m_channelData[c]->inbuf->getReadSpace() > 0) {
|
|
m_log.log(2, "calling processChunks from available, channel" , c);
|
|
bool any = false, last = false;
|
|
((R2Stretcher *)this)->processChunks(c, any, last);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
#ifndef NO_THREADING
|
|
}
|
|
#endif
|
|
|
|
size_t min = 0;
|
|
bool consumed = true;
|
|
bool haveResamplers = false;
|
|
|
|
for (size_t i = 0; i < m_channels; ++i) {
|
|
size_t availIn = m_channelData[i]->inbuf->getReadSpace();
|
|
size_t availOut = m_channelData[i]->outbuf->getReadSpace();
|
|
m_log.log(3, "available in and out", availIn, availOut);
|
|
if (i == 0 || availOut < min) min = availOut;
|
|
if (!m_channelData[i]->outputComplete) consumed = false;
|
|
if (m_channelData[i]->resampler) haveResamplers = true;
|
|
}
|
|
|
|
if (min == 0 && consumed) {
|
|
m_log.log(2, "R2Stretcher::available: end of stream");
|
|
return -1;
|
|
}
|
|
if (m_pitchScale == 1.0) {
|
|
m_log.log(3, "R2Stretcher::available (not shifting): returning", min);
|
|
return min;
|
|
}
|
|
|
|
int rv;
|
|
if (haveResamplers) {
|
|
rv = min; // resampling has already happened
|
|
} else {
|
|
rv = int(floor(min / m_pitchScale));
|
|
}
|
|
m_log.log(3, "R2Stretcher::available (shifting): returning", rv);
|
|
return rv;
|
|
}
|
|
|
|
size_t
|
|
R2Stretcher::retrieve(float *const *output, size_t samples) const
|
|
{
|
|
Profiler profiler("R2Stretcher::retrieve");
|
|
|
|
m_log.log(3, "R2Stretcher::retrieve", samples);
|
|
|
|
size_t got = samples;
|
|
|
|
for (size_t c = 0; c < m_channels; ++c) {
|
|
size_t gotHere = m_channelData[c]->outbuf->read(output[c], got);
|
|
if (gotHere < got) {
|
|
if (c > 0) {
|
|
m_log.log(0, "R2Stretcher::retrieve: WARNING: channel imbalance detected");
|
|
}
|
|
got = gotHere;
|
|
}
|
|
}
|
|
|
|
if ((m_options & RubberBandStretcher::OptionChannelsTogether) &&
|
|
(m_channels >= 2)) {
|
|
for (size_t i = 0; i < got; ++i) {
|
|
float mid = output[0][i];
|
|
float side = output[1][i];
|
|
float left = mid + side;
|
|
float right = mid - side;
|
|
output[0][i] = left;
|
|
output[1][i] = right;
|
|
}
|
|
}
|
|
|
|
m_log.log(3, "R2Stretcher::retrieve returning", got);
|
|
|
|
return got;
|
|
}
|
|
|
|
}
|
|
|