mirror of
https://github.com/xiaoyifang/goldendict-ng.git
synced 2024-11-24 04:24:09 +00:00
578 lines
14 KiB
C++
578 lines
14 KiB
C++
#ifdef MAKE_FFMPEG_PLAYER
|
|
|
|
#include "ffmpegaudio.hh"
|
|
|
|
#include <math.h>
|
|
#include <errno.h>
|
|
|
|
#ifndef INT64_C
|
|
#define INT64_C(c) (c ## LL)
|
|
#endif
|
|
|
|
#ifndef UINT64_C
|
|
#define UINT64_C(c) (c ## ULL)
|
|
#endif
|
|
|
|
#include <ao/ao.h>
|
|
|
|
extern "C" {
|
|
#include <libavcodec/avcodec.h>
|
|
#include <libavformat/avformat.h>
|
|
#include <libavutil/avutil.h>
|
|
#include "libswresample/swresample.h"
|
|
}
|
|
|
|
#include <QString>
|
|
#include <QDataStream>
|
|
#include <QDebug>
|
|
|
|
#include <vector>
|
|
|
|
#include "gddebug.hh"
|
|
#include "utils.hh"
|
|
|
|
using std::vector;
|
|
|
|
namespace Ffmpeg
|
|
{
|
|
|
|
QMutex DecoderThread::deviceMutex_;
|
|
|
|
static inline QString avErrorString( int errnum )
|
|
{
|
|
char buf[64];
|
|
av_strerror( errnum, buf, 64 );
|
|
return QString::fromLatin1( buf );
|
|
}
|
|
|
|
AudioService & AudioService::instance()
|
|
{
|
|
static AudioService a;
|
|
return a;
|
|
}
|
|
|
|
AudioService::AudioService()
|
|
{
|
|
ao_initialize();
|
|
}
|
|
|
|
AudioService::~AudioService()
|
|
{
|
|
emit cancelPlaying( true );
|
|
ao_shutdown();
|
|
}
|
|
|
|
void AudioService::playMemory( const char * ptr, int size )
|
|
{
|
|
emit cancelPlaying( false );
|
|
QByteArray audioData( ptr, size );
|
|
DecoderThread * thread = new DecoderThread( audioData, this );
|
|
|
|
connect( thread, SIGNAL( error( QString ) ), this, SIGNAL( error( QString ) ) );
|
|
connect( this, SIGNAL( cancelPlaying( bool ) ), thread, SLOT( cancel( bool ) ), Qt::DirectConnection );
|
|
connect( thread, SIGNAL( finished() ), thread, SLOT( deleteLater() ) );
|
|
|
|
thread->start();
|
|
}
|
|
|
|
void AudioService::stop()
|
|
{
|
|
emit cancelPlaying( false );
|
|
}
|
|
|
|
struct DecoderContext
|
|
{
|
|
enum
|
|
{
|
|
kBufferSize = 32768
|
|
};
|
|
|
|
static QMutex deviceMutex_;
|
|
QAtomicInt & isCancelled_;
|
|
QByteArray audioData_;
|
|
QDataStream audioDataStream_;
|
|
AVFormatContext * formatContext_;
|
|
#if LIBAVCODEC_VERSION_MAJOR < 59
|
|
AVCodec * codec_;
|
|
#else
|
|
const AVCodec * codec_;
|
|
#endif
|
|
AVCodecContext * codecContext_;
|
|
AVIOContext * avioContext_;
|
|
AVStream * audioStream_;
|
|
ao_device * aoDevice_;
|
|
bool avformatOpened_;
|
|
|
|
SwrContext *swr_;
|
|
|
|
DecoderContext( QByteArray const & audioData, QAtomicInt & isCancelled );
|
|
~DecoderContext();
|
|
|
|
bool openCodec( QString & errorString );
|
|
void closeCodec();
|
|
bool openOutputDevice( QString & errorString );
|
|
void closeOutputDevice();
|
|
bool play( QString & errorString );
|
|
bool normalizeAudio( AVFrame * frame, vector<char> & samples );
|
|
void playFrame( AVFrame * frame );
|
|
};
|
|
|
|
DecoderContext::DecoderContext( QByteArray const & audioData, QAtomicInt & isCancelled ):
|
|
isCancelled_( isCancelled ),
|
|
audioData_( audioData ),
|
|
audioDataStream_( audioData_ ),
|
|
formatContext_( NULL ),
|
|
codec_( NULL ),
|
|
codecContext_( NULL ),
|
|
avioContext_( NULL ),
|
|
audioStream_( NULL ),
|
|
aoDevice_( NULL ),
|
|
avformatOpened_( false ),
|
|
swr_( NULL )
|
|
{
|
|
}
|
|
|
|
DecoderContext::~DecoderContext()
|
|
{
|
|
closeOutputDevice();
|
|
closeCodec();
|
|
}
|
|
|
|
static int readAudioData( void * opaque, unsigned char * buffer, int bufferSize )
|
|
{
|
|
QDataStream * pStream = ( QDataStream * )opaque;
|
|
// This function is passed as the read_packet callback into avio_alloc_context().
|
|
// The documentation for this callback parameter states:
|
|
// For stream protocols, must never return 0 but rather a proper AVERROR code.
|
|
if( pStream->atEnd() )
|
|
return AVERROR_EOF;
|
|
const int bytesRead = pStream->readRawData( ( char * )buffer, bufferSize );
|
|
// QDataStream::readRawData() returns 0 at EOF => return AVERROR_EOF in this case.
|
|
// An error is unlikely here, so just print a warning and return AVERROR_EOF too.
|
|
if( bytesRead < 0 )
|
|
gdWarning( "readAudioData: error while reading raw data." );
|
|
return bytesRead > 0 ? bytesRead : AVERROR_EOF;
|
|
}
|
|
|
|
bool DecoderContext::openCodec( QString & errorString )
|
|
{
|
|
formatContext_ = avformat_alloc_context();
|
|
if ( !formatContext_ )
|
|
{
|
|
errorString = QObject::tr( "avformat_alloc_context() failed." );
|
|
return false;
|
|
}
|
|
|
|
unsigned char * avioBuffer = ( unsigned char * )av_malloc( kBufferSize + AV_INPUT_BUFFER_PADDING_SIZE );
|
|
|
|
if ( !avioBuffer )
|
|
{
|
|
errorString = QObject::tr( "av_malloc() failed." );
|
|
return false;
|
|
}
|
|
|
|
// Don't free buffer allocated here (if succeeded), it will be cleaned up automatically.
|
|
avioContext_ = avio_alloc_context( avioBuffer, kBufferSize, 0, &audioDataStream_, readAudioData, NULL, NULL );
|
|
if ( !avioContext_ )
|
|
{
|
|
av_free( avioBuffer );
|
|
errorString = QObject::tr( "avio_alloc_context() failed." );
|
|
return false;
|
|
}
|
|
|
|
avioContext_->seekable = 0;
|
|
avioContext_->write_flag = 0;
|
|
|
|
// If pb not set, avformat_open_input() simply crash.
|
|
formatContext_->pb = avioContext_;
|
|
formatContext_->flags |= AVFMT_FLAG_CUSTOM_IO;
|
|
|
|
int ret = 0;
|
|
avformatOpened_ = true;
|
|
|
|
ret = avformat_open_input( &formatContext_, "_STREAM_", NULL, NULL );
|
|
if ( ret < 0 )
|
|
{
|
|
errorString = QObject::tr( "avformat_open_input() failed: %1." ).arg( avErrorString( ret ) );
|
|
return false;
|
|
}
|
|
|
|
ret = avformat_find_stream_info( formatContext_, NULL );
|
|
if ( ret < 0 )
|
|
{
|
|
errorString = QObject::tr( "avformat_find_stream_info() failed: %1." ).arg( avErrorString( ret ) );
|
|
return false;
|
|
}
|
|
|
|
// Find audio stream, use the first audio stream if available
|
|
for ( unsigned i = 0; i < formatContext_->nb_streams; i++ )
|
|
{
|
|
if ( formatContext_->streams[i]->codecpar->codec_type == AVMEDIA_TYPE_AUDIO )
|
|
{
|
|
audioStream_ = formatContext_->streams[i];
|
|
break;
|
|
}
|
|
}
|
|
if ( !audioStream_ )
|
|
{
|
|
errorString = QObject::tr( "Could not find audio stream." );
|
|
return false;
|
|
}
|
|
|
|
codec_ = avcodec_find_decoder( audioStream_->codecpar->codec_id );
|
|
if ( !codec_ )
|
|
{
|
|
errorString = QObject::tr( "Codec [id: %1] not found." ).arg( audioStream_->codecpar->codec_id );
|
|
return false;
|
|
}
|
|
codecContext_ = avcodec_alloc_context3( codec_ );
|
|
if ( !codecContext_ )
|
|
{
|
|
errorString = QObject::tr( "avcodec_alloc_context3() failed." );
|
|
return false;
|
|
}
|
|
avcodec_parameters_to_context( codecContext_, audioStream_->codecpar );
|
|
|
|
ret = avcodec_open2( codecContext_, codec_, NULL );
|
|
if ( ret < 0 )
|
|
{
|
|
errorString = QObject::tr( "avcodec_open2() failed: %1." ).arg( avErrorString( ret ) );
|
|
return false;
|
|
}
|
|
|
|
av_log( NULL, AV_LOG_INFO, "Codec open: %s: channels: %d, rate: %d, format: %s\n", codec_->long_name,
|
|
codecContext_->channels, codecContext_->sample_rate, av_get_sample_fmt_name( codecContext_->sample_fmt ) );
|
|
|
|
if ( codecContext_->sample_fmt == AV_SAMPLE_FMT_S32 ||
|
|
codecContext_->sample_fmt == AV_SAMPLE_FMT_S32P ||
|
|
codecContext_->sample_fmt == AV_SAMPLE_FMT_FLT ||
|
|
codecContext_->sample_fmt == AV_SAMPLE_FMT_FLTP ||
|
|
codecContext_->sample_fmt == AV_SAMPLE_FMT_DBL ||
|
|
codecContext_->sample_fmt == AV_SAMPLE_FMT_DBLP )
|
|
{
|
|
swr_ = swr_alloc_set_opts( NULL,
|
|
codecContext_->channel_layout,
|
|
AV_SAMPLE_FMT_S16,
|
|
codecContext_->sample_rate,
|
|
codecContext_->channel_layout,
|
|
codecContext_->sample_fmt,
|
|
codecContext_->sample_rate,
|
|
0,
|
|
NULL );
|
|
swr_init( swr_ );
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
void DecoderContext::closeCodec()
|
|
{
|
|
if ( swr_ )
|
|
{
|
|
swr_free( &swr_ );
|
|
}
|
|
|
|
if ( !formatContext_ )
|
|
{
|
|
if ( avioContext_ )
|
|
{
|
|
av_free( avioContext_->buffer );
|
|
avioContext_ = NULL;
|
|
}
|
|
return;
|
|
}
|
|
|
|
// avformat_open_input() is not called, just free the buffer associated with
|
|
// the AVIOContext, and the AVFormatContext
|
|
if ( !avformatOpened_ )
|
|
{
|
|
if ( formatContext_ )
|
|
{
|
|
avformat_free_context( formatContext_ );
|
|
formatContext_ = NULL;
|
|
}
|
|
|
|
if ( avioContext_ )
|
|
{
|
|
av_free( avioContext_->buffer );
|
|
avioContext_ = NULL;
|
|
}
|
|
return;
|
|
}
|
|
|
|
avformatOpened_ = false;
|
|
|
|
// Closing a codec context without prior avcodec_open2() will result in
|
|
// a crash in ffmpeg
|
|
if ( audioStream_ && codecContext_ && codec_ )
|
|
{
|
|
audioStream_->discard = AVDISCARD_ALL;
|
|
avcodec_close( codecContext_ );
|
|
avcodec_free_context( &codecContext_ );
|
|
}
|
|
|
|
avformat_close_input( &formatContext_ );
|
|
av_free( avioContext_->buffer );
|
|
}
|
|
|
|
bool DecoderContext::openOutputDevice( QString & errorString )
|
|
{
|
|
// Prepare for audio output
|
|
int aoDriverId = ao_default_driver_id();
|
|
ao_info * aoDrvInfo = ao_driver_info( aoDriverId );
|
|
|
|
if ( aoDriverId < 0 || !aoDrvInfo )
|
|
{
|
|
errorString = QObject::tr( "Cannot find usable audio output device." );
|
|
return false;
|
|
}
|
|
|
|
ao_sample_format aoSampleFormat;
|
|
memset (&aoSampleFormat, 0, sizeof(aoSampleFormat) );
|
|
aoSampleFormat.channels = codecContext_->channels;
|
|
aoSampleFormat.rate = codecContext_->sample_rate;
|
|
aoSampleFormat.byte_format = AO_FMT_NATIVE;
|
|
aoSampleFormat.matrix = 0;
|
|
aoSampleFormat.bits = qMin( 16, av_get_bytes_per_sample( codecContext_->sample_fmt ) << 3 );
|
|
|
|
if ( aoSampleFormat.bits == 0 )
|
|
{
|
|
errorString = QObject::tr( "Unsupported sample format." );
|
|
return false;
|
|
}
|
|
|
|
av_log( NULL, AV_LOG_INFO, "ao_open_live(): %s: channels: %d, rate: %d, bits: %d\n",
|
|
aoDrvInfo->name, aoSampleFormat.channels, aoSampleFormat.rate, aoSampleFormat.bits );
|
|
|
|
aoDevice_ = ao_open_live( aoDriverId, &aoSampleFormat, NULL );
|
|
if ( !aoDevice_ )
|
|
{
|
|
errorString = QObject::tr( "ao_open_live() failed: " );
|
|
|
|
switch ( errno )
|
|
{
|
|
case AO_ENODRIVER:
|
|
errorString += QObject::tr( "No driver." );
|
|
break;
|
|
case AO_ENOTLIVE:
|
|
errorString += QObject::tr( "This driver is not a live output device." );
|
|
break;
|
|
case AO_EBADOPTION:
|
|
errorString += QObject::tr( "A valid option key has an invalid value." );
|
|
break;
|
|
case AO_EOPENDEVICE:
|
|
errorString += QObject::tr( "Cannot open the device: %1, channels: %2, rate: %3, bits: %4." )
|
|
.arg( aoDrvInfo->short_name )
|
|
.arg( aoSampleFormat.channels )
|
|
.arg( aoSampleFormat.rate )
|
|
.arg( aoSampleFormat.bits );
|
|
break;
|
|
default:
|
|
errorString += QObject::tr( "Unknown error." );
|
|
break;
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
void DecoderContext::closeOutputDevice()
|
|
{
|
|
// ao_close() is synchronous, it will wait until all audio streams flushed
|
|
if ( aoDevice_ )
|
|
{
|
|
ao_close( aoDevice_ );
|
|
aoDevice_ = NULL;
|
|
}
|
|
}
|
|
|
|
bool DecoderContext::play( QString & errorString )
|
|
{
|
|
AVFrame * frame = av_frame_alloc();
|
|
if ( !frame )
|
|
{
|
|
errorString = QObject::tr( "avcodec_alloc_frame() failed." );
|
|
return false;
|
|
}
|
|
|
|
AVPacket packet;
|
|
av_init_packet( &packet );
|
|
|
|
while ( !Utils::AtomicInt::loadAcquire( isCancelled_ ) &&
|
|
av_read_frame( formatContext_, &packet ) >= 0 )
|
|
{
|
|
if ( packet.stream_index == audioStream_->index )
|
|
{
|
|
AVPacket pack = packet;
|
|
int ret = avcodec_send_packet( codecContext_, &pack );
|
|
/* read all the output frames (in general there may be any number of them) */
|
|
while( ret >= 0 )
|
|
{
|
|
ret = avcodec_receive_frame( codecContext_, frame);
|
|
|
|
if ( Utils::AtomicInt::loadAcquire( isCancelled_ ) || ret < 0 )
|
|
break;
|
|
|
|
playFrame( frame );
|
|
}
|
|
|
|
}
|
|
|
|
av_packet_unref( &packet );
|
|
|
|
}
|
|
|
|
/* flush the decoder */
|
|
av_init_packet( &packet );
|
|
packet.data = NULL;
|
|
packet.size = 0;
|
|
int ret = avcodec_send_packet(codecContext_, &packet );
|
|
while( ret >= 0 )
|
|
{
|
|
ret = avcodec_receive_frame(codecContext_, frame);
|
|
if ( Utils::AtomicInt::loadAcquire( isCancelled_ ) || ret < 0 )
|
|
break;
|
|
playFrame( frame );
|
|
}
|
|
av_frame_free( &frame );
|
|
return true;
|
|
}
|
|
|
|
bool DecoderContext::normalizeAudio( AVFrame * frame, vector<char> & samples )
|
|
{
|
|
int lineSize = 0;
|
|
int dataSize = av_samples_get_buffer_size( &lineSize, codecContext_->channels,
|
|
frame->nb_samples, codecContext_->sample_fmt, 1 );
|
|
|
|
// Portions from: https://code.google.com/p/lavfilters/source/browse/decoder/LAVAudio/LAVAudio.cpp
|
|
// But this one use 8, 16, 32 bits integer, respectively.
|
|
switch ( codecContext_->sample_fmt )
|
|
{
|
|
case AV_SAMPLE_FMT_U8:
|
|
case AV_SAMPLE_FMT_S16:
|
|
{
|
|
samples.resize( dataSize );
|
|
memcpy( &samples.front(), frame->data[0], lineSize );
|
|
}
|
|
break;
|
|
// Planar
|
|
case AV_SAMPLE_FMT_U8P:
|
|
{
|
|
samples.resize( dataSize );
|
|
|
|
uint8_t * out = ( uint8_t * )&samples.front();
|
|
for ( int i = 0; i < frame->nb_samples; i++ )
|
|
{
|
|
for ( int ch = 0; ch < codecContext_->channels; ch++ )
|
|
{
|
|
*out++ = ( ( uint8_t * )frame->extended_data[ch] )[i];
|
|
}
|
|
}
|
|
}
|
|
break;
|
|
case AV_SAMPLE_FMT_S16P:
|
|
{
|
|
samples.resize( dataSize );
|
|
|
|
int16_t * out = ( int16_t * )&samples.front();
|
|
for ( int i = 0; i < frame->nb_samples; i++ )
|
|
{
|
|
for ( int ch = 0; ch < codecContext_->channels; ch++ )
|
|
{
|
|
*out++ = ( ( int16_t * )frame->extended_data[ch] )[i];
|
|
}
|
|
}
|
|
}
|
|
break;
|
|
case AV_SAMPLE_FMT_S32:
|
|
/* Pass through */
|
|
case AV_SAMPLE_FMT_S32P:
|
|
/* Pass through */
|
|
case AV_SAMPLE_FMT_FLT:
|
|
/* Pass through */
|
|
case AV_SAMPLE_FMT_FLTP:
|
|
/* Pass through */
|
|
{
|
|
samples.resize( dataSize / 2 );
|
|
|
|
uint8_t *out = ( uint8_t * )&samples.front();
|
|
swr_convert( swr_, &out, frame->nb_samples, (const uint8_t**)frame->extended_data, frame->nb_samples );
|
|
}
|
|
break;
|
|
case AV_SAMPLE_FMT_DBL:
|
|
case AV_SAMPLE_FMT_DBLP:
|
|
{
|
|
samples.resize( dataSize / 4 );
|
|
|
|
uint8_t *out = ( uint8_t * )&samples.front();
|
|
swr_convert( swr_, &out, frame->nb_samples, (const uint8_t**)frame->extended_data, frame->nb_samples );
|
|
}
|
|
break;
|
|
default:
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
void DecoderContext::playFrame( AVFrame * frame )
|
|
{
|
|
if ( !frame )
|
|
return;
|
|
|
|
vector<char> samples;
|
|
if ( normalizeAudio( frame, samples ) )
|
|
ao_play( aoDevice_, &samples.front(), samples.size() );
|
|
}
|
|
|
|
DecoderThread::DecoderThread( QByteArray const & audioData, QObject * parent ) :
|
|
QThread( parent ),
|
|
isCancelled_( 0 ),
|
|
audioData_( audioData )
|
|
{
|
|
}
|
|
|
|
DecoderThread::~DecoderThread()
|
|
{
|
|
isCancelled_.ref();
|
|
}
|
|
|
|
void DecoderThread::run()
|
|
{
|
|
QString errorString;
|
|
DecoderContext d( audioData_, isCancelled_ );
|
|
|
|
if ( !d.openCodec( errorString ) )
|
|
{
|
|
emit error( errorString );
|
|
return;
|
|
}
|
|
|
|
while ( !deviceMutex_.tryLock( 100 ) )
|
|
{
|
|
if ( Utils::AtomicInt::loadAcquire( isCancelled_ ) )
|
|
return;
|
|
}
|
|
|
|
if ( !d.openOutputDevice( errorString ) )
|
|
emit error( errorString );
|
|
else if ( !d.play( errorString ) )
|
|
emit error( errorString );
|
|
|
|
d.closeOutputDevice();
|
|
deviceMutex_.unlock();
|
|
}
|
|
|
|
void DecoderThread::cancel( bool waitUntilFinished )
|
|
{
|
|
isCancelled_.ref();
|
|
if ( waitUntilFinished )
|
|
this->wait();
|
|
}
|
|
|
|
}
|
|
|
|
#endif // MAKE_FFMPEG_PLAYER
|