Files
librubberband/src/StretcherProcess.cpp

1162 lines
36 KiB
C++
Raw Normal View History

2007-11-06 21:41:16 +00:00
/* -*- c-basic-offset: 4 indent-tabs-mode: nil -*- vi:set ts=8 sts=4 sw=4: */
/*
Rubber Band
An audio time-stretching and pitch-shifting library.
Copyright 2007-2011 Chris Cannam.
2007-11-06 21:41:16 +00:00
This program is free software; you can redistribute it and/or
modify it under the terms of the GNU General Public License as
published by the Free Software Foundation; either version 2 of the
License, or (at your option) any later version. See the file
COPYING included with this distribution for more information.
*/
#include "StretcherImpl.h"
#include "dsp/PercussiveAudioCurve.h"
#include "dsp/HighFrequencyAudioCurve.h"
#include "dsp/ConstantAudioCurve.h"
2007-11-06 21:41:16 +00:00
#include "StretchCalculator.h"
#include "StretcherChannelData.h"
#include "dsp/Resampler.h"
#include "base/Profiler.h"
#include "system/VectorOps.h"
#ifndef _WIN32
#include <alloca.h>
#endif
2007-11-06 21:41:16 +00:00
#include <cassert>
#include <cmath>
#include <set>
#include <map>
#include <deque>
using namespace RubberBand;
2007-11-06 21:41:16 +00:00
using std::cerr;
using std::endl;
namespace RubberBand {
RubberBandStretcher::Impl::ProcessThread::ProcessThread(Impl *s, size_t c) :
m_s(s),
m_channel(c),
m_dataAvailable(std::string("data ") + char('A' + c)),
m_abandoning(false)
{ }
2007-11-06 21:41:16 +00:00
void
RubberBandStretcher::Impl::ProcessThread::run()
2007-11-06 21:41:16 +00:00
{
if (m_s->m_debugLevel > 1) {
cerr << "thread " << m_channel << " getting going" << endl;
}
ChannelData &cd = *m_s->m_channelData[m_channel];
while (cd.inputSize == -1 ||
cd.inbuf->getReadSpace() > 0) {
2007-12-10 22:20:07 +00:00
// if (cd.inputSize != -1) {
// cerr << "inputSize == " << cd.inputSize
// << ", readSpace == " << cd.inbuf->getReadSpace() << endl;
// }
2007-11-06 21:41:16 +00:00
2007-11-26 11:00:47 +00:00
bool any = false, last = false;
m_s->processChunks(m_channel, any, last);
2007-11-06 21:41:16 +00:00
if (last) break;
if (any) {
m_s->m_spaceAvailable.lock();
m_s->m_spaceAvailable.signal();
m_s->m_spaceAvailable.unlock();
}
2007-11-26 11:00:47 +00:00
m_dataAvailable.lock();
if (!m_s->testInbufReadSpace(m_channel) && !m_abandoning) {
m_dataAvailable.wait(50000); // bounded in case of abandonment
2007-11-06 21:41:16 +00:00
}
m_dataAvailable.unlock();
if (m_abandoning) {
if (m_s->m_debugLevel > 1) {
cerr << "thread " << m_channel << " abandoning" << endl;
}
return;
}
2007-11-06 21:41:16 +00:00
}
2007-11-26 11:00:47 +00:00
bool any = false, last = false;
m_s->processChunks(m_channel, any, last);
m_s->m_spaceAvailable.lock();
2007-11-06 21:41:16 +00:00
m_s->m_spaceAvailable.signal();
m_s->m_spaceAvailable.unlock();
2007-11-06 21:41:16 +00:00
if (m_s->m_debugLevel > 1) {
cerr << "thread " << m_channel << " done" << endl;
}
}
void
RubberBandStretcher::Impl::ProcessThread::signalDataAvailable()
{
m_dataAvailable.lock();
m_dataAvailable.signal();
m_dataAvailable.unlock();
}
void
RubberBandStretcher::Impl::ProcessThread::abandon()
{
m_abandoning = true;
}
bool
RubberBandStretcher::Impl::resampleBeforeStretching() const
{
// We can't resample before stretching in offline mode, because
// the stretch calculation is based on doing it the other way
// around. It would take more work (and testing) to enable this.
if (!m_realtime) return false;
if (m_options & OptionPitchHighQuality) {
return (m_pitchScale < 1.0); // better sound
2008-07-01 20:49:24 +00:00
} else if (m_options & OptionPitchHighConsistency) {
return false;
} else {
return (m_pitchScale > 1.0); // better performance
}
}
size_t
RubberBandStretcher::Impl::consumeChannel(size_t c, const float *input,
size_t samples, bool final)
{
Profiler profiler("RubberBandStretcher::Impl::consumeChannel");
ChannelData &cd = *m_channelData[c];
RingBuffer<float> &inbuf = *cd.inbuf;
size_t toWrite = samples;
size_t writable = inbuf.getWriteSpace();
bool resampling = resampleBeforeStretching();
if (resampling) {
toWrite = int(ceil(samples / m_pitchScale));
if (writable < toWrite) {
samples = int(floor(writable * m_pitchScale));
if (samples == 0) return 0;
}
size_t reqSize = int(ceil(samples / m_pitchScale));
if (reqSize > cd.resamplebufSize) {
cerr << "WARNING: RubberBandStretcher::Impl::consumeChannel: resizing resampler buffer from "
<< cd.resamplebufSize << " to " << reqSize << endl;
cd.setResampleBufSize(reqSize);
}
toWrite = cd.resampler->resample(&input,
&cd.resamplebuf,
samples,
1.0 / m_pitchScale,
final);
}
if (writable < toWrite) {
if (resampling) {
return 0;
}
toWrite = writable;
}
if (resampling) {
inbuf.write(cd.resamplebuf, toWrite);
cd.inCount += samples;
return samples;
} else {
inbuf.write(input, toWrite);
cd.inCount += toWrite;
return toWrite;
}
}
2007-11-26 11:00:47 +00:00
void
RubberBandStretcher::Impl::processChunks(size_t c, bool &any, bool &last)
2007-11-06 21:41:16 +00:00
{
Profiler profiler("RubberBandStretcher::Impl::processChunks");
2007-11-06 21:41:16 +00:00
// Process as many chunks as there are available on the input
// buffer for channel c. This requires that the increments have
// already been calculated.
// This is the normal process method in offline mode.
2007-11-06 21:41:16 +00:00
ChannelData &cd = *m_channelData[c];
2007-11-26 11:00:47 +00:00
last = false;
any = false;
2007-11-06 21:41:16 +00:00
while (!last) {
2007-11-26 11:00:47 +00:00
if (!testInbufReadSpace(c)) {
// cerr << "not enough input" << endl;
break;
}
any = true;
2007-11-06 21:41:16 +00:00
if (!cd.draining) {
size_t got = cd.inbuf->peek(cd.fltbuf, m_aWindowSize);
assert(got == m_aWindowSize || cd.inputSize >= 0);
2007-11-06 21:41:16 +00:00
cd.inbuf->skip(m_increment);
}
bool phaseReset = false;
2007-11-06 21:41:16 +00:00
size_t phaseIncrement, shiftIncrement;
getIncrements(c, phaseIncrement, shiftIncrement, phaseReset);
2007-11-06 21:41:16 +00:00
if (shiftIncrement <= m_aWindowSize) {
analyseChunk(c);
last = processChunkForChannel
(c, phaseIncrement, shiftIncrement, phaseReset);
} else {
size_t bit = m_aWindowSize/4;
if (m_debugLevel > 1) {
cerr << "channel " << c << " breaking down overlong increment " << shiftIncrement << " into " << bit << "-size bits" << endl;
}
analyseChunk(c);
float *tmp = (float *)alloca(m_aWindowSize * sizeof(float));
v_copy(tmp, cd.fltbuf, m_aWindowSize);
for (size_t i = 0; i < shiftIncrement; i += bit) {
v_copy(cd.fltbuf, tmp, m_aWindowSize);
size_t thisIncrement = bit;
if (i + thisIncrement > shiftIncrement) {
thisIncrement = shiftIncrement - i;
}
last = processChunkForChannel
(c, phaseIncrement + i, thisIncrement, phaseReset);
phaseReset = false;
}
}
cd.chunkCount++;
2007-11-06 21:41:16 +00:00
if (m_debugLevel > 2) {
cerr << "channel " << c << ": last = " << last << ", chunkCount = " << cd.chunkCount << endl;
2007-11-06 21:41:16 +00:00
}
}
}
bool
RubberBandStretcher::Impl::processOneChunk()
2007-11-06 21:41:16 +00:00
{
Profiler profiler("RubberBandStretcher::Impl::processOneChunk");
2007-11-06 21:41:16 +00:00
// Process a single chunk for all channels, provided there is
// enough data on each channel for at least one chunk. This is
// able to calculate increments as it goes along.
// This is the normal process method in RT mode.
2007-11-06 21:41:16 +00:00
for (size_t c = 0; c < m_channels; ++c) {
if (!testInbufReadSpace(c)) return false;
ChannelData &cd = *m_channelData[c];
if (!cd.draining) {
size_t got = cd.inbuf->peek(cd.fltbuf, m_aWindowSize);
assert(got == m_aWindowSize || cd.inputSize >= 0);
2007-11-06 21:41:16 +00:00
cd.inbuf->skip(m_increment);
analyseChunk(c);
}
}
bool phaseReset = false;
2007-11-06 21:41:16 +00:00
size_t phaseIncrement, shiftIncrement;
if (!getIncrements(0, phaseIncrement, shiftIncrement, phaseReset)) {
calculateIncrements(phaseIncrement, shiftIncrement, phaseReset);
2007-11-06 21:41:16 +00:00
}
bool last = false;
for (size_t c = 0; c < m_channels; ++c) {
last = processChunkForChannel(c, phaseIncrement, shiftIncrement, phaseReset);
m_channelData[c]->chunkCount++;
2007-11-06 21:41:16 +00:00
}
return last;
}
bool
RubberBandStretcher::Impl::testInbufReadSpace(size_t c)
2007-11-06 21:41:16 +00:00
{
Profiler profiler("RubberBandStretcher::Impl::testInbufReadSpace");
2007-11-06 21:41:16 +00:00
ChannelData &cd = *m_channelData[c];
RingBuffer<float> &inbuf = *cd.inbuf;
size_t rs = inbuf.getReadSpace();
if (rs < m_aWindowSize && !cd.draining) {
2007-11-06 21:41:16 +00:00
if (cd.inputSize == -1) {
// Not all the input data has been written to the inbuf
// (that's why the input size is not yet set). We can't
// process, because we don't have a full chunk of data, so
// our process chunk would contain some empty padding in
2007-11-06 21:41:16 +00:00
// its input -- and that would give incorrect output, as
// we know there is more input to come.
if (!m_threaded) {
// cerr << "WARNING: RubberBandStretcher: read space < chunk size ("
// << inbuf.getReadSpace() << " < " << m_windowSize
2007-11-06 21:41:16 +00:00
// << ") when not all input written, on processChunks for channel " << c << endl;
}
return false;
}
if (rs == 0) {
if (m_debugLevel > 1) {
cerr << "read space = 0, giving up" << endl;
}
return false;
} else if (rs < m_aWindowSize/2) {
2007-11-06 21:41:16 +00:00
if (m_debugLevel > 1) {
cerr << "read space = " << rs << ", setting draining true" << endl;
}
cd.draining = true;
}
}
return true;
}
bool
RubberBandStretcher::Impl::processChunkForChannel(size_t c,
2007-11-06 21:41:16 +00:00
size_t phaseIncrement,
size_t shiftIncrement,
bool phaseReset)
2007-11-06 21:41:16 +00:00
{
Profiler profiler("RubberBandStretcher::Impl::processChunkForChannel");
2007-11-06 21:41:16 +00:00
// Process a single chunk on a single channel. This assumes
// enough input data is available; caller must have tested this
// using e.g. testInbufReadSpace first. Return true if this is
// the last chunk on the channel.
if (phaseReset && (m_debugLevel > 1)) {
cerr << "processChunkForChannel: phase reset found, incrs "
<< phaseIncrement << ":" << shiftIncrement << endl;
2007-11-06 21:41:16 +00:00
}
ChannelData &cd = *m_channelData[c];
if (!cd.draining) {
// This is the normal processing case -- draining is only
// set when all the input has been used and we only need
// to write from the existing accumulator into the output.
// We know we have enough samples available in m_inbuf --
// this is usually m_aWindowSize, but we know that if fewer
2007-11-06 21:41:16 +00:00
// are available, it's OK to use zeroes for the rest
// (which the ring buffer will provide) because we've
// reached the true end of the data.
// We need to peek m_aWindowSize samples for processing, and
2007-11-06 21:41:16 +00:00
// then skip m_increment to advance the read pointer.
modifyChunk(c, phaseIncrement, phaseReset);
synthesiseChunk(c, shiftIncrement); // reads from cd.mag, cd.phase
2007-11-06 21:41:16 +00:00
if (m_debugLevel > 2) {
if (phaseReset) {
2007-11-06 21:41:16 +00:00
for (int i = 0; i < 10; ++i) {
2007-11-28 11:50:47 +00:00
cd.accumulator[i] = 1.2f - (i % 3) * 1.2f;
2007-11-06 21:41:16 +00:00
}
}
}
}
bool last = false;
if (cd.draining) {
if (m_debugLevel > 1) {
cerr << "draining: accumulator fill = " << cd.accumulatorFill << " (shiftIncrement = " << shiftIncrement << ")" << endl;
}
if (shiftIncrement == 0) {
cerr << "WARNING: draining: shiftIncrement == 0, can't handle that in this context: setting to " << m_increment << endl;
shiftIncrement = m_increment;
}
if (cd.accumulatorFill <= shiftIncrement) {
if (m_debugLevel > 1) {
cerr << "reducing shift increment from " << shiftIncrement
<< " to " << cd.accumulatorFill
<< " and marking as last" << endl;
}
shiftIncrement = cd.accumulatorFill;
last = true;
}
}
int required = shiftIncrement;
if (m_pitchScale != 1.0) {
required = int(required / m_pitchScale) + 1;
}
int ws = cd.outbuf->getWriteSpace();
if (ws < required) {
if (m_debugLevel > 0) {
cerr << "Buffer overrun on output for channel " << c << endl;
2007-11-06 21:41:16 +00:00
}
// The only correct thing we can do here is resize the buffer.
// We can't wait for the client thread to read some data out
// from the buffer so as to make more space, because the
// client thread (if we are threaded at all) is probably stuck
// in a process() call waiting for us to stow away enough
// input increments to allow the process() call to complete.
// This is an unhappy situation.
2007-11-06 21:41:16 +00:00
RingBuffer<float> *oldbuf = cd.outbuf;
cd.outbuf = oldbuf->resized(oldbuf->getSize() + (required - ws));
m_emergencyScavenger.claim(oldbuf);
2007-11-06 21:41:16 +00:00
}
2007-11-06 21:41:16 +00:00
writeChunk(c, shiftIncrement, last);
return last;
}
void
RubberBandStretcher::Impl::calculateIncrements(size_t &phaseIncrementRtn,
2007-11-06 21:41:16 +00:00
size_t &shiftIncrementRtn,
bool &phaseReset)
2007-11-06 21:41:16 +00:00
{
Profiler profiler("RubberBandStretcher::Impl::calculateIncrements");
2007-11-06 21:41:16 +00:00
// cerr << "calculateIncrements" << endl;
// Calculate the next upcoming phase and shift increment, on the
// basis that both channels are in sync. This is in contrast to
// getIncrements, which requires that all the increments have been
// calculated in advance but can then return increments
// corresponding to different chunks in different channels.
2007-11-06 21:41:16 +00:00
// Requires frequency domain representations of channel data in
// the mag and phase buffers in the channel.
// This function is only used in real-time mode.
phaseIncrementRtn = m_increment;
shiftIncrementRtn = m_increment;
phaseReset = false;
2007-11-06 21:41:16 +00:00
if (m_channels == 0) return;
ChannelData &cd = *m_channelData[0];
size_t bc = cd.chunkCount;
2007-11-06 21:41:16 +00:00
for (size_t c = 1; c < m_channels; ++c) {
if (m_channelData[c]->chunkCount != bc) {
cerr << "ERROR: RubberBandStretcher::Impl::calculateIncrements: Channels are not in sync" << endl;
2007-11-06 21:41:16 +00:00
return;
}
}
const int hs = m_fftSize/2 + 1;
2007-11-06 21:41:16 +00:00
// Normally we would mix down the time-domain signal and apply a
// single FFT, or else mix down the Cartesian form of the
// frequency-domain signal. Both of those would be inefficient
// from this position. Fortunately, the onset detectors should
// work reasonably well (maybe even better?) if we just sum the
// magnitudes of the frequency-domain channel signals and forget
// about phase entirely. Normally we don't expect the channel
// phases to cancel each other, and broadband effects will still
// be apparent.
float df = 0.f;
bool silent = false;
2007-11-06 21:41:16 +00:00
if (m_channels == 1) {
if (sizeof(process_t) == sizeof(double)) {
df = m_phaseResetAudioCurve->processDouble((double *)cd.mag, m_increment);
silent = (m_silentAudioCurve->processDouble((double *)cd.mag, m_increment) > 0.f);
} else {
df = m_phaseResetAudioCurve->processFloat((float *)cd.mag, m_increment);
silent = (m_silentAudioCurve->processFloat((float *)cd.mag, m_increment) > 0.f);
}
} else {
process_t *tmp = (process_t *)alloca(hs * sizeof(process_t));
v_zero(tmp, hs);
for (size_t c = 0; c < m_channels; ++c) {
v_add(tmp, m_channelData[c]->mag, hs);
2007-11-06 21:41:16 +00:00
}
if (sizeof(process_t) == sizeof(double)) {
df = m_phaseResetAudioCurve->processDouble((double *)tmp, m_increment);
silent = (m_silentAudioCurve->processDouble((double *)tmp, m_increment) > 0.f);
} else {
df = m_phaseResetAudioCurve->processFloat((float *)tmp, m_increment);
silent = (m_silentAudioCurve->processFloat((float *)tmp, m_increment) > 0.f);
}
}
2007-11-06 21:41:16 +00:00
int incr = m_stretchCalculator->calculateSingle
(getEffectiveRatio(), df, m_increment);
2007-11-06 21:41:16 +00:00
m_lastProcessPhaseResetDf.write(&df, 1);
2007-11-06 21:41:16 +00:00
m_lastProcessOutputIncrements.write(&incr, 1);
if (incr < 0) {
phaseReset = true;
2007-11-06 21:41:16 +00:00
incr = -incr;
}
// The returned increment is the phase increment. The shift
// increment for one chunk is the same as the phase increment for
// the following chunk (see comment below). This means we don't
// actually know the shift increment until we see the following
// phase increment... which is a bit of a problem.
// This implies we should use this increment for the shift
// increment, and make the following phase increment the same as
// it. This means in RT mode we'll be one chunk later with our
// phase reset than we would be in non-RT mode. The sensitivity
// of the broadband onset detector may mean that this isn't a
// problem -- test it and see.
2007-11-06 21:41:16 +00:00
shiftIncrementRtn = incr;
if (cd.prevIncrement == 0) {
phaseIncrementRtn = shiftIncrementRtn;
} else {
phaseIncrementRtn = cd.prevIncrement;
}
cd.prevIncrement = shiftIncrementRtn;
if (silent) ++m_silentHistory;
else m_silentHistory = 0;
if (m_silentHistory >= int(m_aWindowSize / m_increment) && !phaseReset) {
phaseReset = true;
if (m_debugLevel > 1) {
cerr << "calculateIncrements: phase reset on silence (silent history == "
<< m_silentHistory << ")" << endl;
}
}
2007-11-06 21:41:16 +00:00
}
bool
RubberBandStretcher::Impl::getIncrements(size_t channel,
2007-11-06 21:41:16 +00:00
size_t &phaseIncrementRtn,
size_t &shiftIncrementRtn,
bool &phaseReset)
2007-11-06 21:41:16 +00:00
{
Profiler profiler("RubberBandStretcher::Impl::getIncrements");
2007-11-06 21:41:16 +00:00
if (channel >= m_channels) {
phaseIncrementRtn = m_increment;
shiftIncrementRtn = m_increment;
phaseReset = false;
2007-11-06 21:41:16 +00:00
return false;
}
// There are two relevant output increments here. The first is
// the phase increment which we use when recalculating the phases
// for the current chunk; the second is the shift increment used
// to determine how far to shift the processing buffer after
// writing the chunk. The shift increment for one chunk is the
// same as the phase increment for the following chunk.
// When an onset occurs for which we need to reset phases, the
2007-11-06 21:41:16 +00:00
// increment given will be negative.
// When we reset phases, the previous shift increment (and so
2007-11-06 21:41:16 +00:00
// current phase increments) must have been m_increment to ensure
// consistency.
// m_outputIncrements stores phase increments.
ChannelData &cd = *m_channelData[channel];
bool gotData = true;
if (cd.chunkCount >= m_outputIncrements.size()) {
// cerr << "WARNING: RubberBandStretcher::Impl::getIncrements:"
// << " chunk count " << cd.chunkCount << " >= "
// << m_outputIncrements.size() << endl;
2007-11-06 21:41:16 +00:00
if (m_outputIncrements.size() == 0) {
phaseIncrementRtn = m_increment;
shiftIncrementRtn = m_increment;
phaseReset = false;
2007-11-06 21:41:16 +00:00
return false;
} else {
cd.chunkCount = m_outputIncrements.size()-1;
2007-11-06 21:41:16 +00:00
gotData = false;
}
}
int phaseIncrement = m_outputIncrements[cd.chunkCount];
2007-11-06 21:41:16 +00:00
int shiftIncrement = phaseIncrement;
if (cd.chunkCount + 1 < m_outputIncrements.size()) {
shiftIncrement = m_outputIncrements[cd.chunkCount + 1];
2007-11-06 21:41:16 +00:00
}
if (phaseIncrement < 0) {
phaseIncrement = -phaseIncrement;
phaseReset = true;
2007-11-06 21:41:16 +00:00
}
if (shiftIncrement < 0) {
shiftIncrement = -shiftIncrement;
}
/*
if (shiftIncrement >= int(m_windowSize)) {
cerr << "*** ERROR: RubberBandStretcher::Impl::processChunks: shiftIncrement " << shiftIncrement << " >= windowSize " << m_windowSize << " at " << cd.chunkCount << " (of " << m_outputIncrements.size() << ")" << endl;
shiftIncrement = m_windowSize;
2007-11-06 21:41:16 +00:00
}
*/
2007-11-06 21:41:16 +00:00
phaseIncrementRtn = phaseIncrement;
shiftIncrementRtn = shiftIncrement;
if (cd.chunkCount == 0) phaseReset = true; // don't mess with the first chunk
2007-11-06 21:41:16 +00:00
return gotData;
}
void
RubberBandStretcher::Impl::analyseChunk(size_t channel)
2007-11-06 21:41:16 +00:00
{
Profiler profiler("RubberBandStretcher::Impl::analyseChunk");
2007-11-06 21:41:16 +00:00
ChannelData &cd = *m_channelData[channel];
process_t *const R__ dblbuf = cd.dblbuf;
float *const R__ fltbuf = cd.fltbuf;
// cd.fltbuf is known to contain m_aWindowSize samples
2007-11-06 21:41:16 +00:00
if (m_aWindowSize > m_fftSize) {
m_afilter->cut(fltbuf);
2007-11-06 21:41:16 +00:00
}
cutShiftAndFold(dblbuf, m_fftSize, fltbuf, m_awindow);
cd.fft->forwardPolar(dblbuf, cd.mag, cd.phase);
2007-11-06 21:41:16 +00:00
}
void
RubberBandStretcher::Impl::modifyChunk(size_t channel,
size_t outputIncrement,
bool phaseReset)
2007-11-06 21:41:16 +00:00
{
Profiler profiler("RubberBandStretcher::Impl::modifyChunk");
2007-11-06 21:41:16 +00:00
ChannelData &cd = *m_channelData[channel];
if (phaseReset && m_debugLevel > 1) {
cerr << "phase reset: leaving phases unmodified" << endl;
2007-11-06 21:41:16 +00:00
}
const process_t rate = m_sampleRate;
const int count = m_fftSize / 2;
bool unchanged = cd.unchanged && (outputIncrement == m_increment);
bool fullReset = phaseReset;
bool laminar = !(m_options & OptionPhaseIndependent);
bool bandlimited = (m_options & OptionTransientsMixed);
int bandlow = lrint((150 * m_fftSize) / rate);
int bandhigh = lrint((1000 * m_fftSize) / rate);
float freq0 = m_freq0;
float freq1 = m_freq1;
float freq2 = m_freq2;
2007-11-06 21:41:16 +00:00
if (laminar) {
float r = getEffectiveRatio();
if (r > 1) {
float rf0 = 600 + (600 * ((r-1)*(r-1)*(r-1)*2));
float f1ratio = freq1 / freq0;
float f2ratio = freq2 / freq0;
freq0 = std::max(freq0, rf0);
freq1 = freq0 * f1ratio;
freq2 = freq0 * f2ratio;
2007-11-06 21:41:16 +00:00
}
}
int limit0 = lrint((freq0 * m_fftSize) / rate);
int limit1 = lrint((freq1 * m_fftSize) / rate);
int limit2 = lrint((freq2 * m_fftSize) / rate);
if (limit1 < limit0) limit1 = limit0;
if (limit2 < limit1) limit2 = limit1;
process_t prevInstability = 0.0;
2008-07-08 15:32:52 +00:00
bool prevDirection = false;
2007-11-06 21:41:16 +00:00
process_t distance = 0.0;
const process_t maxdist = 8.0;
2008-07-08 15:41:21 +00:00
const int lookback = 1;
process_t distacc = 0.0;
2008-07-08 15:41:21 +00:00
for (int i = count; i >= 0; i -= lookback) {
2007-11-06 21:41:16 +00:00
bool resetThis = phaseReset;
if (bandlimited) {
if (resetThis) {
if (i > bandlow && i < bandhigh) {
resetThis = false;
fullReset = false;
}
}
2007-11-08 15:02:47 +00:00
}
process_t p = cd.phase[i];
process_t perr = 0.0;
process_t outphase = p;
2007-11-06 21:41:16 +00:00
process_t mi = maxdist;
if (i <= limit0) mi = 0.0;
else if (i <= limit1) mi = 1.0;
else if (i <= limit2) mi = 3.0;
if (!resetThis) {
2007-11-06 21:41:16 +00:00
process_t omega = (2 * M_PI * m_increment * i) / (m_fftSize);
2007-11-06 21:41:16 +00:00
process_t pp = cd.prevPhase[i];
process_t ep = pp + omega;
perr = princarg(p - ep);
process_t instability = fabs(perr - cd.prevError[i]);
bool direction = (perr > cd.prevError[i]);
2007-11-06 21:41:16 +00:00
bool inherit = false;
2007-11-06 21:41:16 +00:00
if (laminar) {
if (distance >= mi || i == count) {
inherit = false;
} else if (bandlimited && (i == bandhigh || i == bandlow)) {
inherit = false;
} else if (instability > prevInstability &&
direction == prevDirection) {
inherit = true;
}
}
process_t advance = outputIncrement * ((omega + perr) / m_increment);
if (inherit) {
process_t inherited =
cd.unwrappedPhase[i + lookback] - cd.prevPhase[i + lookback];
advance = ((advance * distance) +
(inherited * (maxdist - distance)))
/ maxdist;
outphase = p + advance;
distacc += distance;
distance += 1.0;
} else {
outphase = cd.unwrappedPhase[i] + advance;
distance = 0.0;
2007-11-06 21:41:16 +00:00
}
prevInstability = instability;
prevDirection = direction;
2007-11-06 21:41:16 +00:00
} else {
distance = 0.0;
2007-11-06 21:41:16 +00:00
}
cd.prevError[i] = perr;
cd.prevPhase[i] = p;
cd.phase[i] = outphase;
cd.unwrappedPhase[i] = outphase;
}
if (m_debugLevel > 2) {
cerr << "mean inheritance distance = " << distacc / count << endl;
}
if (fullReset) unchanged = true;
cd.unchanged = unchanged;
if (unchanged && m_debugLevel > 1) {
cerr << "frame unchanged on channel " << channel << endl;
}
}
2007-11-06 21:41:16 +00:00
2007-11-06 21:41:16 +00:00
void
RubberBandStretcher::Impl::formantShiftChunk(size_t channel)
2007-11-06 21:41:16 +00:00
{
Profiler profiler("RubberBandStretcher::Impl::formantShiftChunk");
2007-11-06 21:41:16 +00:00
ChannelData &cd = *m_channelData[channel];
process_t *const R__ mag = cd.mag;
process_t *const R__ envelope = cd.envelope;
process_t *const R__ dblbuf = cd.dblbuf;
const int sz = m_fftSize;
const int hs = sz / 2;
const process_t factor = 1.0 / sz;
cd.fft->inverseCepstral(mag, dblbuf);
2008-06-09 20:46:37 +00:00
const int cutoff = m_sampleRate / 700;
// cerr <<"cutoff = "<< cutoff << ", m_sampleRate/cutoff = " << m_sampleRate/cutoff << endl;
dblbuf[0] /= 2;
dblbuf[cutoff-1] /= 2;
2007-11-06 21:41:16 +00:00
for (int i = cutoff; i < sz; ++i) {
dblbuf[i] = 0.0;
2007-11-06 21:41:16 +00:00
}
v_scale(dblbuf, factor, cutoff);
cd.fft->forward(dblbuf, envelope, 0);
v_exp(envelope, hs + 1);
v_divide(mag, envelope, hs + 1);
2007-11-06 21:41:16 +00:00
if (m_pitchScale > 1.0) {
// scaling up, we want a new envelope that is lower by the pitch factor
for (int target = 0; target <= hs; ++target) {
int source = lrint(target * m_pitchScale);
if (source > hs) {
envelope[target] = 0.0;
} else {
envelope[target] = envelope[source];
}
}
} else {
// scaling down, we want a new envelope that is higher by the pitch factor
for (int target = hs; target > 0; ) {
--target;
int source = lrint(target * m_pitchScale);
envelope[target] = envelope[source];
}
}
2007-11-06 21:41:16 +00:00
v_multiply(mag, envelope, hs+1);
cd.unchanged = false;
}
void
RubberBandStretcher::Impl::synthesiseChunk(size_t channel,
size_t shiftIncrement)
{
Profiler profiler("RubberBandStretcher::Impl::synthesiseChunk");
if ((m_options & OptionFormantPreserved) &&
(m_pitchScale != 1.0)) {
formantShiftChunk(channel);
}
ChannelData &cd = *m_channelData[channel];
process_t *const R__ dblbuf = cd.dblbuf;
float *const R__ fltbuf = cd.fltbuf;
float *const R__ accumulator = cd.accumulator;
float *const R__ windowAccumulator = cd.windowAccumulator;
const int fsz = m_fftSize;
const int hs = fsz / 2;
const int wsz = m_sWindowSize;
if (!cd.unchanged) {
cd.fft->inversePolar(cd.mag, cd.phase, cd.dblbuf);
// our ffts produced unscaled results
float factor = 1.f / fsz;
v_scale(dblbuf, factor, fsz);
if (wsz == fsz) {
v_convert(fltbuf, dblbuf + hs, hs);
v_convert(fltbuf + hs, dblbuf, hs);
} else {
v_zero(fltbuf, wsz);
int j = fsz - wsz/2;
while (j < 0) j += fsz;
for (int i = 0; i < wsz; ++i) {
fltbuf[i] += dblbuf[j];
if (++j == fsz) j = 0;
}
}
}
if (wsz > fsz) {
int p = shiftIncrement * 2;
if (cd.interpolatorScale != p) {
SincWindow<float>::write(cd.interpolator, wsz, p);
cd.interpolatorScale = p;
}
v_multiply(fltbuf, cd.interpolator, wsz);
}
m_swindow->cut(fltbuf);
v_add(accumulator, fltbuf, wsz);
cd.accumulatorFill = wsz;
if (wsz > fsz) {
// reuse fltbuf to calculate interpolating window shape for
// window accumulator
v_copy(fltbuf, cd.interpolator, wsz);
m_swindow->cut(fltbuf);
v_add(windowAccumulator, fltbuf, wsz);
} else {
m_swindow->add(windowAccumulator, m_awindow->getArea() * 1.5f);
}
2007-11-06 21:41:16 +00:00
}
void
RubberBandStretcher::Impl::writeChunk(size_t channel, size_t shiftIncrement, bool last)
2007-11-06 21:41:16 +00:00
{
Profiler profiler("RubberBandStretcher::Impl::writeChunk");
2007-11-06 21:41:16 +00:00
ChannelData &cd = *m_channelData[channel];
float *const R__ accumulator = cd.accumulator;
float *const R__ windowAccumulator = cd.windowAccumulator;
2007-11-06 21:41:16 +00:00
const int sz = m_sWindowSize;
const int si = shiftIncrement;
2007-11-06 21:41:16 +00:00
if (m_debugLevel > 2) {
cerr << "writeChunk(" << channel << ", " << shiftIncrement << ", " << last << ")" << endl;
}
v_divide(accumulator, windowAccumulator, si);
2007-11-06 21:41:16 +00:00
// for exact sample scaling (probably not meaningful if we
// were running in RT mode)
size_t theoreticalOut = 0;
if (cd.inputSize >= 0) {
theoreticalOut = lrint(cd.inputSize * m_timeRatio);
}
bool resampledAlready = resampleBeforeStretching();
2007-11-06 21:41:16 +00:00
2008-07-01 20:49:24 +00:00
if (!resampledAlready &&
(m_pitchScale != 1.0 || m_options & OptionPitchHighConsistency) &&
cd.resampler) {
size_t reqSize = int(ceil(si / m_pitchScale));
2007-11-06 21:41:16 +00:00
if (reqSize > cd.resamplebufSize) {
// This shouldn't normally happen -- the buffer is
// supposed to be initialised with enough space in the
// first place. But we retain this check in case the
// pitch scale has changed since then, or the stretch
// calculator has gone mad, or something.
cerr << "WARNING: RubberBandStretcher::Impl::writeChunk: resizing resampler buffer from "
2007-11-06 21:41:16 +00:00
<< cd.resamplebufSize << " to " << reqSize << endl;
cd.setResampleBufSize(reqSize);
2007-11-06 21:41:16 +00:00
}
size_t outframes = cd.resampler->resample(&cd.accumulator,
&cd.resamplebuf,
si,
2007-11-06 21:41:16 +00:00
1.0 / m_pitchScale,
last);
2008-07-04 19:31:23 +00:00
2007-11-06 21:41:16 +00:00
writeOutput(*cd.outbuf, cd.resamplebuf,
outframes, cd.outCount, theoreticalOut);
} else {
writeOutput(*cd.outbuf, accumulator,
si, cd.outCount, theoreticalOut);
2007-11-06 21:41:16 +00:00
}
v_move(accumulator, accumulator + si, sz - si);
v_zero(accumulator + sz - si, si);
2007-11-06 21:41:16 +00:00
v_move(windowAccumulator, windowAccumulator + si, sz - si);
v_zero(windowAccumulator + sz - si, si);
2007-11-06 21:41:16 +00:00
if (int(cd.accumulatorFill) > si) {
cd.accumulatorFill -= si;
2007-11-06 21:41:16 +00:00
} else {
cd.accumulatorFill = 0;
if (cd.draining) {
if (m_debugLevel > 1) {
cerr << "RubberBandStretcher::Impl::processChunks: setting outputComplete to true" << endl;
2007-11-06 21:41:16 +00:00
}
cd.outputComplete = true;
}
}
}
void
RubberBandStretcher::Impl::writeOutput(RingBuffer<float> &to, float *from, size_t qty, size_t &outCount, size_t theoreticalOut)
2007-11-06 21:41:16 +00:00
{
Profiler profiler("RubberBandStretcher::Impl::writeOutput");
2007-11-06 21:41:16 +00:00
// In non-RT mode, we don't want to write the first startSkip
// samples, because the first chunk is centred on the start of the
2007-11-06 21:41:16 +00:00
// output. In RT mode we didn't apply any pre-padding in
// configure(), so we don't want to remove any here.
size_t startSkip = 0;
if (!m_realtime) {
startSkip = lrintf((m_sWindowSize/2) / m_pitchScale);
2007-11-06 21:41:16 +00:00
}
if (outCount > startSkip) {
// this is the normal case
2007-11-06 21:41:16 +00:00
if (theoreticalOut > 0) {
if (m_debugLevel > 1) {
cerr << "theoreticalOut = " << theoreticalOut
<< ", outCount = " << outCount
<< ", startSkip = " << startSkip
<< ", qty = " << qty << endl;
2007-11-06 21:41:16 +00:00
}
if (outCount - startSkip <= theoreticalOut &&
outCount - startSkip + qty > theoreticalOut) {
qty = theoreticalOut - (outCount - startSkip);
if (m_debugLevel > 1) {
cerr << "reduce qty to " << qty << endl;
}
}
}
if (m_debugLevel > 2) {
cerr << "writing " << qty << endl;
}
size_t written = to.write(from, qty);
if (written < qty) {
cerr << "WARNING: RubberBandStretcher::Impl::writeOutput: "
<< "Buffer overrun on output: wrote " << written
<< " of " << qty << " samples" << endl;
}
outCount += written;
2007-11-06 21:41:16 +00:00
return;
}
// the rest of this is only used during the first startSkip samples
2007-11-06 21:41:16 +00:00
if (outCount + qty <= startSkip) {
if (m_debugLevel > 1) {
cerr << "qty = " << qty << ", startSkip = "
<< startSkip << ", outCount = " << outCount
<< ", discarding" << endl;
2007-11-06 21:41:16 +00:00
}
outCount += qty;
return;
}
size_t off = startSkip - outCount;
if (m_debugLevel > 1) {
cerr << "qty = " << qty << ", startSkip = "
<< startSkip << ", outCount = " << outCount
<< ", writing " << qty - off
<< " from start offset " << off << endl;
2007-11-06 21:41:16 +00:00
}
to.write(from + off, qty - off);
outCount += qty;
}
int
RubberBandStretcher::Impl::available() const
2007-11-06 21:41:16 +00:00
{
Profiler profiler("RubberBandStretcher::Impl::available");
2007-11-06 21:41:16 +00:00
if (m_threaded) {
MutexLocker locker(&m_threadSetMutex);
if (m_channelData.empty()) return 0;
} else {
if (m_channelData.empty()) return 0;
}
if (!m_threaded) {
for (size_t c = 0; c < m_channels; ++c) {
if (m_channelData[c]->inputSize >= 0) {
// cerr << "available: m_done true" << endl;
2007-11-06 21:41:16 +00:00
if (m_channelData[c]->inbuf->getReadSpace() > 0) {
// cerr << "calling processChunks(" << c << ") from available" << endl;
2007-11-06 21:41:16 +00:00
//!!! do we ever actually do this? if so, this method should not be const
// ^^^ yes, we do sometimes -- e.g. when fed a very short file
2007-11-26 11:00:47 +00:00
bool any = false, last = false;
((RubberBandStretcher::Impl *)this)->processChunks(c, any, last);
2007-11-06 21:41:16 +00:00
}
}
}
}
size_t min = 0;
bool consumed = true;
bool haveResamplers = false;
for (size_t i = 0; i < m_channels; ++i) {
size_t availIn = m_channelData[i]->inbuf->getReadSpace();
size_t availOut = m_channelData[i]->outbuf->getReadSpace();
if (m_debugLevel > 2) {
cerr << "available on channel " << i << ": " << availOut << " (waiting: " << availIn << ")" << endl;
}
if (i == 0 || availOut < min) min = availOut;
if (!m_channelData[i]->outputComplete) consumed = false;
if (m_channelData[i]->resampler) haveResamplers = true;
}
if (min == 0 && consumed) return -1;
if (m_pitchScale == 1.0) return min;
if (haveResamplers) return min; // resampling has already happened
return int(floor(min / m_pitchScale));
}
size_t
RubberBandStretcher::Impl::retrieve(float *const *output, size_t samples) const
2007-11-06 21:41:16 +00:00
{
Profiler profiler("RubberBandStretcher::Impl::retrieve");
2007-11-06 21:41:16 +00:00
size_t got = samples;
for (size_t c = 0; c < m_channels; ++c) {
size_t gotHere = m_channelData[c]->outbuf->read(output[c], got);
if (gotHere < got) {
if (c > 0) {
if (m_debugLevel > 0) {
cerr << "RubberBandStretcher::Impl::retrieve: WARNING: channel imbalance detected" << endl;
2007-11-06 21:41:16 +00:00
}
}
got = gotHere;
}
}
return got;
}
}