mirror of
https://github.com/xiaoyifang/goldendict-ng.git
synced 2024-11-23 20:14:05 +00:00
Add support for Aard dictionaries (aar-html)
This commit is contained in:
parent
79d24b80f6
commit
81ee55aff7
910
aard.cc
Normal file
910
aard.cc
Normal file
|
@ -0,0 +1,910 @@
|
|||
/* This file is (c) 2008-2011 Konstantin Isakov <ikm@goldendict.org>
|
||||
* Part of GoldenDict. Licensed under GPLv3 or later, see the LICENSE file */
|
||||
|
||||
#include "aard.hh"
|
||||
#include "btreeidx.hh"
|
||||
#include "folding.hh"
|
||||
#include "utf8.hh"
|
||||
#include "chunkedstorage.hh"
|
||||
#include "langcoder.hh"
|
||||
#include "dprintf.hh"
|
||||
#include "fsencoding.hh"
|
||||
#include "decompress.hh"
|
||||
|
||||
#include <map>
|
||||
#include <set>
|
||||
#include <string>
|
||||
|
||||
#ifdef _MSC_VER
|
||||
#include <stub_msvc.h>
|
||||
#endif
|
||||
|
||||
#include <QString>
|
||||
#include <QSemaphore>
|
||||
#include <QThreadPool>
|
||||
#include <QAtomicInt>
|
||||
#include <QDomDocument>
|
||||
|
||||
#include "ufile.hh"
|
||||
#include "wstring_qt.hh"
|
||||
|
||||
namespace Aard {
|
||||
|
||||
using std::map;
|
||||
using std::multimap;
|
||||
using std::pair;
|
||||
using std::set;
|
||||
using std::string;
|
||||
using gd::wstring;
|
||||
|
||||
using BtreeIndexing::WordArticleLink;
|
||||
using BtreeIndexing::IndexedWords;
|
||||
using BtreeIndexing::IndexInfo;
|
||||
|
||||
namespace {
|
||||
|
||||
DEF_EX_STR( exNotDctFile, "Not an Sdictionary file", Dictionary::Ex )
|
||||
DEF_EX_STR( exCantReadFile, "Can't read file", Dictionary::Ex )
|
||||
DEF_EX_STR( exWordIsTooLarge, "Enountered a word that is too large:", Dictionary::Ex )
|
||||
DEF_EX_STR( exSuddenEndOfFile, "Sudden end of file", Dictionary::Ex )
|
||||
|
||||
#ifdef _MSC_VER
|
||||
#pragma pack( push, 1 )
|
||||
#endif
|
||||
|
||||
// Big-Endian template
|
||||
// http://habrahabr.ru/blogs/cpp/121811/
|
||||
|
||||
template<typename T>
|
||||
struct BigEndian
|
||||
{
|
||||
union
|
||||
{
|
||||
unsigned char bytes[sizeof(T)];
|
||||
T raw_value;
|
||||
};
|
||||
|
||||
BigEndian(T t = T())
|
||||
{
|
||||
operator =(t);
|
||||
}
|
||||
|
||||
BigEndian(const BigEndian<T> & t)
|
||||
{
|
||||
raw_value = t.raw_value;
|
||||
}
|
||||
|
||||
operator const T() const
|
||||
{
|
||||
T t = T();
|
||||
for (unsigned i = 0; i < sizeof(T); i++)
|
||||
t |= T(bytes[sizeof(T) - 1 - i]) << (i << 3);
|
||||
return t;
|
||||
}
|
||||
|
||||
const T operator = (const T t)
|
||||
{
|
||||
for (unsigned i = 0; i < sizeof(T); i++)
|
||||
bytes[sizeof(T) - 1 - i] = (unsigned char)( t >> (i << 3) );
|
||||
return t;
|
||||
}
|
||||
|
||||
}
|
||||
#ifndef _MSC_VER
|
||||
__attribute__((packed))
|
||||
#endif
|
||||
;
|
||||
|
||||
typedef BigEndian< uint16_t > uint16_be;
|
||||
typedef BigEndian< uint32_t > uint32_be;
|
||||
|
||||
/// AAR file header
|
||||
struct AAR_header
|
||||
{
|
||||
char signature[4];
|
||||
char checksum[40];
|
||||
uint16_be version;
|
||||
char uuid[16];
|
||||
uint16_be volume;
|
||||
uint16_be totalVolumes;
|
||||
uint32_be metaLength;
|
||||
uint32_be wordsCount;
|
||||
uint32_be articleOffset;
|
||||
char indexItemFormat[4];
|
||||
char keyLengthFormat[2];
|
||||
char articleLengthFormat[2];
|
||||
}
|
||||
#ifndef _MSC_VER
|
||||
__attribute__((packed))
|
||||
#endif
|
||||
;
|
||||
|
||||
struct IndexElement
|
||||
{
|
||||
uint32_be wordOffset;
|
||||
uint32_be articleOffset;
|
||||
}
|
||||
#ifndef _MSC_VER
|
||||
__attribute__((packed))
|
||||
#endif
|
||||
;
|
||||
|
||||
enum
|
||||
{
|
||||
Signature = 0x58524141, // AARX on little-endian, XRAA on big-endian
|
||||
CurrentFormatVersion = 1 + BtreeIndexing::FormatVersion + Folding::Version
|
||||
};
|
||||
|
||||
struct IdxHeader
|
||||
{
|
||||
uint32_t signature; // First comes the signature, AARX
|
||||
uint32_t formatVersion; // File format version (CurrentFormatVersion)
|
||||
uint32_t chunksOffset; // The offset to chunks' storage
|
||||
uint32_t indexBtreeMaxElements; // Two fields from IndexInfo
|
||||
uint32_t indexRootOffset;
|
||||
uint32_t wordCount;
|
||||
uint32_t articleCount;
|
||||
uint32_t langFrom; // Source language
|
||||
uint32_t langTo; // Target language
|
||||
}
|
||||
#ifndef _MSC_VER
|
||||
__attribute__((packed))
|
||||
#endif
|
||||
;
|
||||
|
||||
#ifdef _MSC_VER
|
||||
#pragma pack( pop, 1 )
|
||||
#endif
|
||||
|
||||
bool indexIsOldOrBad( string const & indexFile )
|
||||
{
|
||||
File::Class idx( indexFile, "rb" );
|
||||
|
||||
IdxHeader header;
|
||||
|
||||
return idx.readRecords( &header, sizeof( header ), 1 ) != 1 ||
|
||||
header.signature != Signature ||
|
||||
header.formatVersion != CurrentFormatVersion;
|
||||
}
|
||||
|
||||
void readJSONValue( string const & source, string & str, uint32_t & pos)
|
||||
{
|
||||
int level = 1;
|
||||
char endChar;
|
||||
str.push_back( source[pos] );
|
||||
if( source[pos] == '{')
|
||||
endChar = '}';
|
||||
else if( source[pos] == '[' )
|
||||
endChar = ']';
|
||||
else if( source[pos] == '\"' )
|
||||
{
|
||||
str.clear();
|
||||
endChar = '\"';
|
||||
}
|
||||
else
|
||||
endChar = ',';
|
||||
|
||||
pos++;
|
||||
char ch = 0;
|
||||
char lastCh = 0;
|
||||
while( !( ch == endChar && lastCh != '\\' && level == 0 )
|
||||
&& pos < source.size() )
|
||||
{
|
||||
lastCh = ch;
|
||||
ch = source[ pos++ ];
|
||||
if( ( ch == '{' || ch == '[' ) && lastCh != '\\' )
|
||||
level++;
|
||||
if( ( ch == '}' || ch == ']' ) && lastCh != '\\' )
|
||||
level--;
|
||||
|
||||
if( ch == endChar &&
|
||||
( ( ch == '\"' && lastCh != '\\' ) || ch == ',' )
|
||||
&& level == 1)
|
||||
break;
|
||||
str.push_back( ch );
|
||||
}
|
||||
}
|
||||
|
||||
class AardDictionary: public BtreeIndexing::BtreeDictionary
|
||||
{
|
||||
Mutex idxMutex;
|
||||
File::Class idx;
|
||||
IdxHeader idxHeader;
|
||||
ChunkedStorage::Reader chunks;
|
||||
string dictionaryName;
|
||||
File::Class df;
|
||||
QIcon dictionaryIcon, dictionaryNativeIcon;
|
||||
bool dictionaryIconLoaded;
|
||||
|
||||
public:
|
||||
|
||||
AardDictionary( string const & id, string const & indexFile,
|
||||
vector< string > const & dictionaryFiles );
|
||||
|
||||
~AardDictionary();
|
||||
|
||||
virtual string getName() throw()
|
||||
{ return dictionaryName; }
|
||||
|
||||
virtual map< Dictionary::Property, string > getProperties() throw()
|
||||
{ return map< Dictionary::Property, string >(); }
|
||||
|
||||
virtual unsigned long getArticleCount() throw()
|
||||
{ return idxHeader.articleCount; }
|
||||
|
||||
virtual unsigned long getWordCount() throw()
|
||||
{ return idxHeader.wordCount; }
|
||||
|
||||
virtual QIcon getIcon() throw();
|
||||
|
||||
virtual QIcon getNativeIcon() throw();
|
||||
|
||||
inline virtual quint32 getLangFrom() const
|
||||
{ return idxHeader.langFrom; }
|
||||
|
||||
inline virtual quint32 getLangTo() const
|
||||
{ return idxHeader.langTo; }
|
||||
|
||||
virtual sptr< Dictionary::DataRequest > getArticle( wstring const &,
|
||||
vector< wstring > const & alts,
|
||||
wstring const & )
|
||||
throw( std::exception );
|
||||
|
||||
private:
|
||||
|
||||
void loadIcon();
|
||||
|
||||
/// Loads the article.
|
||||
void loadArticle( uint32_t address,
|
||||
string & articleText );
|
||||
string convert( string const & in_data );
|
||||
|
||||
friend class AardArticleRequest;
|
||||
};
|
||||
|
||||
AardDictionary::AardDictionary( string const & id,
|
||||
string const & indexFile,
|
||||
vector< string > const & dictionaryFiles ):
|
||||
BtreeDictionary( id, dictionaryFiles ),
|
||||
idx( indexFile, "rb" ),
|
||||
idxHeader( idx.read< IdxHeader >() ),
|
||||
chunks( idx, idxHeader.chunksOffset ),
|
||||
df( dictionaryFiles[ 0 ], "rb" ),
|
||||
dictionaryIconLoaded( false )
|
||||
{
|
||||
// Read dictionary name
|
||||
|
||||
idx.seek( sizeof( idxHeader ) );
|
||||
vector< char > dName( idx.read< uint32_t >() );
|
||||
if( dName.size() )
|
||||
{
|
||||
idx.read( &dName.front(), dName.size() );
|
||||
dictionaryName = string( &dName.front(), dName.size() );
|
||||
}
|
||||
|
||||
// Initialize the index
|
||||
|
||||
openIndex( IndexInfo( idxHeader.indexBtreeMaxElements,
|
||||
idxHeader.indexRootOffset ),
|
||||
idx, idxMutex );
|
||||
}
|
||||
|
||||
AardDictionary::~AardDictionary()
|
||||
{
|
||||
df.close();
|
||||
}
|
||||
|
||||
QIcon AardDictionary::getNativeIcon() throw()
|
||||
{
|
||||
loadIcon();
|
||||
return dictionaryNativeIcon;
|
||||
}
|
||||
|
||||
QIcon AardDictionary::getIcon() throw()
|
||||
{
|
||||
loadIcon();
|
||||
return dictionaryIcon;
|
||||
}
|
||||
|
||||
void AardDictionary::loadIcon()
|
||||
{
|
||||
if ( dictionaryIconLoaded )
|
||||
return;
|
||||
|
||||
QString fileName =
|
||||
QDir::fromNativeSeparators( FsEncoding::decode( getDictionaryFilenames()[ 0 ].c_str() ) );
|
||||
|
||||
// Remove the extension
|
||||
|
||||
fileName.chop( 3 );
|
||||
fileName += "bmp";
|
||||
QFileInfo info( fileName );
|
||||
|
||||
if ( !info.exists() )
|
||||
{
|
||||
fileName.chop( 3 );
|
||||
fileName += "png";
|
||||
info = QFileInfo( fileName );
|
||||
}
|
||||
|
||||
if ( info.exists() )
|
||||
{
|
||||
QImage img( fileName );
|
||||
|
||||
if ( !img.isNull() )
|
||||
{
|
||||
// Load successful
|
||||
|
||||
// Apply the color key
|
||||
|
||||
img.setAlphaChannel( img.createMaskFromColor( QColor( 192, 192, 192 ).rgb(),
|
||||
Qt::MaskOutColor ) );
|
||||
|
||||
dictionaryNativeIcon = QIcon( QPixmap::fromImage( img ) );
|
||||
|
||||
// Transform it to be square
|
||||
int max = img.width() > img.height() ? img.width() : img.height();
|
||||
|
||||
QImage result( max, max, QImage::Format_ARGB32 );
|
||||
result.fill( 0 ); // Black transparent
|
||||
|
||||
QPainter painter( &result );
|
||||
|
||||
painter.drawImage( QPoint( img.width() == max ? 0 : ( max - img.width() ) / 2,
|
||||
img.height() == max ? 0 : ( max - img.height() ) / 2 ),
|
||||
img );
|
||||
|
||||
painter.end();
|
||||
|
||||
dictionaryIcon = QIcon( QPixmap::fromImage( result ) );
|
||||
}
|
||||
}
|
||||
|
||||
if ( dictionaryIcon.isNull() )
|
||||
{
|
||||
// Load failed -- use default icons
|
||||
dictionaryNativeIcon = dictionaryIcon = QIcon(":/icons/icon32_aard.png");
|
||||
}
|
||||
|
||||
dictionaryIconLoaded = true;
|
||||
}
|
||||
|
||||
string AardDictionary::convert( const string & in )
|
||||
{
|
||||
string inConverted;
|
||||
char inCh, lastCh = 0;
|
||||
bool afterEol = false;
|
||||
|
||||
for( string::const_iterator i = in.begin(), j = in.end(); i != j; ++i )
|
||||
{
|
||||
inCh = *i;
|
||||
if( lastCh == '\\' )
|
||||
{
|
||||
inConverted.erase( inConverted.size() - 1 );
|
||||
lastCh = 0;
|
||||
if( inCh == 'n' )
|
||||
{
|
||||
inConverted.append( "<br/>");
|
||||
afterEol = true;
|
||||
continue;
|
||||
}
|
||||
else if( inCh == 'r')
|
||||
continue;
|
||||
}
|
||||
else if( inCh == ' ' && afterEol )
|
||||
{
|
||||
inConverted.append( " " );
|
||||
continue;
|
||||
} else
|
||||
lastCh = inCh;
|
||||
afterEol = false;
|
||||
inConverted.push_back( inCh );
|
||||
}
|
||||
|
||||
QDomDocument dd;
|
||||
QString errorStr;
|
||||
int errorLine, errorColumn;
|
||||
|
||||
if( !dd.setContent( QByteArray( inConverted.c_str() ), false, &errorStr, &errorLine, &errorColumn ) )
|
||||
{
|
||||
FDPRINTF( stderr, "Aard article parse failed: %s at %d,%d\n", errorStr.toLocal8Bit().constData(), errorLine, errorColumn );
|
||||
FDPRINTF( stderr, "The input was: %s\n", in.c_str() );
|
||||
return inConverted;
|
||||
}
|
||||
|
||||
QDomNodeList nodes = dd.elementsByTagName( "a" ); // References
|
||||
for( int i = 0; i < nodes.count(); i++ )
|
||||
{
|
||||
QDomElement el = nodes.at( i ).toElement();
|
||||
QString ref = el.attribute( "href", "" );
|
||||
if( ref.size() == 0 || ref.indexOf( "http://") != -1 || ref[0] == '#' )
|
||||
continue;
|
||||
if( ref.indexOf( "w:") == 0 || ref.indexOf( "s:") == 0 )
|
||||
ref.replace( 0, 2, "bword:" );
|
||||
else
|
||||
ref.insert( 0, "bword:" );
|
||||
el.setAttribute( "href", ref );
|
||||
}
|
||||
|
||||
return dd.toByteArray().data();
|
||||
}
|
||||
|
||||
void AardDictionary::loadArticle( uint32_t address,
|
||||
string & articleText )
|
||||
{
|
||||
uint32_t articleOffset = address;
|
||||
uint32_t articleSize;
|
||||
uint32_be size;
|
||||
|
||||
vector< char > articleBody;
|
||||
|
||||
articleText.clear();
|
||||
|
||||
df.seek( articleOffset );
|
||||
df.read( &size, sizeof(size) );
|
||||
articleSize = size;
|
||||
articleBody.resize( articleSize );
|
||||
df.read( &articleBody.front(), articleSize );
|
||||
|
||||
if ( articleBody.empty() )
|
||||
throw exCantReadFile( getDictionaryFilenames()[ 0 ] );
|
||||
|
||||
string text = decompressBzip2( articleBody.data(), articleSize );
|
||||
if( text.empty() )
|
||||
text = decompressZlib( articleBody.data(), articleSize );
|
||||
if( text.empty() )
|
||||
text = string( articleBody.data(), articleSize );
|
||||
|
||||
uint32_t n = 0;
|
||||
while( n < text.size() && text[n] != '\"' )
|
||||
n++;
|
||||
|
||||
if( n >= text.size() )
|
||||
return;
|
||||
|
||||
readJSONValue( text, articleText, n );
|
||||
|
||||
if( articleText.empty() )
|
||||
{
|
||||
n = text.find( "\"r\"" );
|
||||
if( n != string::npos )
|
||||
{
|
||||
n += 3;
|
||||
while( n < text.size() && text[n] != '\"' )
|
||||
n++;
|
||||
|
||||
if( n >= text.size() )
|
||||
return;
|
||||
|
||||
string link;
|
||||
readJSONValue( text, link, n );
|
||||
if( !link.empty() )
|
||||
articleText = "<a href=\"" + link + "\">" + link + "</a>";
|
||||
}
|
||||
}
|
||||
|
||||
if( !articleText.empty() )
|
||||
articleText = convert( articleText );
|
||||
|
||||
articleText = "<div class=\"sdict\">" + articleText + "</div>";
|
||||
}
|
||||
|
||||
/// AardDictionary::getArticle()
|
||||
|
||||
class AardArticleRequest;
|
||||
|
||||
class AardArticleRequestRunnable: public QRunnable
|
||||
{
|
||||
AardArticleRequest & r;
|
||||
QSemaphore & hasExited;
|
||||
|
||||
public:
|
||||
|
||||
AardArticleRequestRunnable( AardArticleRequest & r_,
|
||||
QSemaphore & hasExited_ ): r( r_ ),
|
||||
hasExited( hasExited_ )
|
||||
{}
|
||||
|
||||
~AardArticleRequestRunnable()
|
||||
{
|
||||
hasExited.release();
|
||||
}
|
||||
|
||||
virtual void run();
|
||||
};
|
||||
|
||||
class AardArticleRequest: public Dictionary::DataRequest
|
||||
{
|
||||
friend class AardArticleRequestRunnable;
|
||||
|
||||
wstring word;
|
||||
vector< wstring > alts;
|
||||
AardDictionary & dict;
|
||||
|
||||
QAtomicInt isCancelled;
|
||||
QSemaphore hasExited;
|
||||
|
||||
public:
|
||||
|
||||
AardArticleRequest( wstring const & word_,
|
||||
vector< wstring > const & alts_,
|
||||
AardDictionary & dict_ ):
|
||||
word( word_ ), alts( alts_ ), dict( dict_ )
|
||||
{
|
||||
QThreadPool::globalInstance()->start(
|
||||
new AardArticleRequestRunnable( *this, hasExited ) );
|
||||
}
|
||||
|
||||
void run(); // Run from another thread by DslArticleRequestRunnable
|
||||
|
||||
virtual void cancel()
|
||||
{
|
||||
isCancelled.ref();
|
||||
}
|
||||
|
||||
~AardArticleRequest()
|
||||
{
|
||||
isCancelled.ref();
|
||||
hasExited.acquire();
|
||||
}
|
||||
};
|
||||
|
||||
void AardArticleRequestRunnable::run()
|
||||
{
|
||||
r.run();
|
||||
}
|
||||
|
||||
void AardArticleRequest::run()
|
||||
{
|
||||
if ( isCancelled )
|
||||
{
|
||||
finish();
|
||||
return;
|
||||
}
|
||||
|
||||
vector< WordArticleLink > chain = dict.findArticles( word );
|
||||
|
||||
for( unsigned x = 0; x < alts.size(); ++x )
|
||||
{
|
||||
/// Make an additional query for each alt
|
||||
|
||||
vector< WordArticleLink > altChain = dict.findArticles( alts[ x ] );
|
||||
|
||||
chain.insert( chain.end(), altChain.begin(), altChain.end() );
|
||||
}
|
||||
|
||||
multimap< wstring, pair< string, string > > mainArticles, alternateArticles;
|
||||
|
||||
set< uint32_t > articlesIncluded; // Some synonims make it that the articles
|
||||
// appear several times. We combat this
|
||||
// by only allowing them to appear once.
|
||||
|
||||
wstring wordCaseFolded = Folding::applySimpleCaseOnly( word );
|
||||
|
||||
for( unsigned x = 0; x < chain.size(); ++x )
|
||||
{
|
||||
if ( isCancelled )
|
||||
{
|
||||
finish();
|
||||
return;
|
||||
}
|
||||
|
||||
if ( articlesIncluded.find( chain[ x ].articleOffset ) != articlesIncluded.end() )
|
||||
continue; // We already have this article in the body.
|
||||
|
||||
// Now grab that article
|
||||
|
||||
string headword, articleText;
|
||||
|
||||
headword = chain[ x ].word;
|
||||
dict.loadArticle( chain[ x ].articleOffset, articleText );
|
||||
|
||||
// Ok. Now, does it go to main articles, or to alternate ones? We list
|
||||
// main ones first, and alternates after.
|
||||
|
||||
// We do the case-folded comparison here.
|
||||
|
||||
wstring headwordStripped =
|
||||
Folding::applySimpleCaseOnly( Utf8::decode( headword ) );
|
||||
|
||||
multimap< wstring, pair< string, string > > & mapToUse =
|
||||
( wordCaseFolded == headwordStripped ) ?
|
||||
mainArticles : alternateArticles;
|
||||
|
||||
mapToUse.insert( pair< wstring, pair< string, string > >(
|
||||
Folding::applySimpleCaseOnly( Utf8::decode( headword ) ),
|
||||
pair< string, string >( headword, articleText ) ) );
|
||||
|
||||
articlesIncluded.insert( chain[ x ].articleOffset );
|
||||
}
|
||||
|
||||
if ( mainArticles.empty() && alternateArticles.empty() )
|
||||
{
|
||||
// No such word
|
||||
finish();
|
||||
return;
|
||||
}
|
||||
|
||||
string result;
|
||||
|
||||
multimap< wstring, pair< string, string > >::const_iterator i;
|
||||
|
||||
for( i = mainArticles.begin(); i != mainArticles.end(); ++i )
|
||||
{
|
||||
result += "<h3>";
|
||||
result += i->second.first;
|
||||
result += "</h3>";
|
||||
result += i->second.second;
|
||||
}
|
||||
|
||||
for( i = alternateArticles.begin(); i != alternateArticles.end(); ++i )
|
||||
{
|
||||
result += "<h3>";
|
||||
result += i->second.first;
|
||||
result += "</h3>";
|
||||
result += i->second.second;
|
||||
}
|
||||
|
||||
Mutex::Lock _( dataMutex );
|
||||
|
||||
data.resize( result.size() );
|
||||
|
||||
memcpy( &data.front(), result.data(), result.size() );
|
||||
|
||||
hasAnyData = true;
|
||||
|
||||
finish();
|
||||
}
|
||||
|
||||
map< string, string > parseMetaData( string const & metaData )
|
||||
{
|
||||
// Parsing JSON string
|
||||
map< string, string > data;
|
||||
string name, value;
|
||||
uint32_t n = 0;
|
||||
|
||||
while( metaData[n] != '{' && n < metaData.length() )
|
||||
n++;
|
||||
while( n < metaData.length() )
|
||||
{
|
||||
// Skip to '"'
|
||||
while( metaData[n] != '\"' && n < metaData.length() )
|
||||
n++;
|
||||
if( ++n >= metaData.length() )
|
||||
break;
|
||||
|
||||
// Read name
|
||||
while( !( ( metaData[n] == '\"' || metaData[n] == '{' ) && metaData[n-1] != '\\' )
|
||||
&& n < metaData.length() )
|
||||
name.push_back( metaData[n++]);
|
||||
|
||||
// Skip to ':'
|
||||
if( ++n >= metaData.length() )
|
||||
break;
|
||||
while( metaData[n] != ':' && n < metaData.length() )
|
||||
n++;
|
||||
if( ++n >= metaData.length() )
|
||||
break;
|
||||
|
||||
// Find value start after ':'
|
||||
while( !( ( metaData[n] == '\"'
|
||||
|| metaData[n] == '{'
|
||||
|| metaData[n] == '['
|
||||
|| ( metaData[n] >= '0' && metaData[n] <= '9' ) )
|
||||
&& metaData[n-1] != '\\' )
|
||||
&& n < metaData.length() )
|
||||
n++;
|
||||
if( n >= metaData.length() )
|
||||
break;
|
||||
|
||||
readJSONValue( metaData, value, n);
|
||||
|
||||
data[name] = value;
|
||||
|
||||
name.clear();
|
||||
value.clear();
|
||||
if( ++n >= metaData.length() )
|
||||
break;
|
||||
}
|
||||
return data;
|
||||
}
|
||||
|
||||
sptr< Dictionary::DataRequest > AardDictionary::getArticle( wstring const & word,
|
||||
vector< wstring > const & alts,
|
||||
wstring const & )
|
||||
throw( std::exception )
|
||||
{
|
||||
return new AardArticleRequest( word, alts, *this );
|
||||
}
|
||||
|
||||
} // anonymous namespace
|
||||
|
||||
vector< sptr< Dictionary::Class > > makeDictionaries(
|
||||
vector< string > const & fileNames,
|
||||
string const & indicesDir,
|
||||
Dictionary::Initializing & initializing )
|
||||
throw( std::exception )
|
||||
{
|
||||
vector< sptr< Dictionary::Class > > dictionaries;
|
||||
|
||||
for( vector< string >::const_iterator i = fileNames.begin(); i != fileNames.end();
|
||||
++i )
|
||||
{
|
||||
// Skip files with the extensions different to .aar to speed up the
|
||||
// scanning
|
||||
if ( i->size() < 4 ||
|
||||
strcasecmp( i->c_str() + ( i->size() - 4 ), ".aar" ) != 0 )
|
||||
continue;
|
||||
|
||||
// Got the file -- check if we need to rebuid the index
|
||||
|
||||
vector< string > dictFiles( 1, *i );
|
||||
|
||||
string dictId = Dictionary::makeDictionaryId( dictFiles );
|
||||
|
||||
string indexFile = indicesDir + dictId;
|
||||
|
||||
if ( Dictionary::needToRebuildIndex( dictFiles, indexFile ) ||
|
||||
indexIsOldOrBad( indexFile ) )
|
||||
{
|
||||
try
|
||||
{
|
||||
File::Class df( *i, "rb" );
|
||||
|
||||
AAR_header dictHeader;
|
||||
|
||||
df.read( &dictHeader, sizeof(dictHeader) );
|
||||
if( strncmp( dictHeader.signature, "aard", 4 )
|
||||
|| strncmp( dictHeader.indexItemFormat, ">LL", 4 )
|
||||
|| strncmp( dictHeader.keyLengthFormat, ">H", 2 )
|
||||
|| strncmp( dictHeader.articleLengthFormat, ">L", 2) )
|
||||
{
|
||||
DPRINTF( "File %s is not in supported aard format", i->c_str() );
|
||||
continue;
|
||||
}
|
||||
|
||||
vector< char > data;
|
||||
uint32_t size = dictHeader.metaLength;
|
||||
|
||||
data.resize( size );
|
||||
df.read( &data.front(), size );
|
||||
string metaStr = decompressBzip2( data.data(), size );
|
||||
if( metaStr.empty() )
|
||||
metaStr = decompressZlib( data.data(), size );
|
||||
|
||||
map< string, string > meta = parseMetaData( metaStr );
|
||||
|
||||
if( meta.empty() )
|
||||
{
|
||||
DPRINTF( "File %s has invalid metadata", i->c_str() );
|
||||
continue;
|
||||
}
|
||||
|
||||
string dictName;
|
||||
map< string, string >::const_iterator iter = meta.find( "title" );
|
||||
if( iter != meta.end() )
|
||||
dictName = iter->second;
|
||||
|
||||
uint16_t volumes = dictHeader.totalVolumes;
|
||||
if( volumes > 1 )
|
||||
{
|
||||
QString ss;
|
||||
ss.sprintf( " (%i/%i)", (uint16_t)(dictHeader.volume), volumes );
|
||||
dictName += ss.toLocal8Bit().data();
|
||||
}
|
||||
|
||||
string langFrom;
|
||||
iter = meta.find( "index_language" );
|
||||
if( iter != meta.end() )
|
||||
langFrom = iter->second;
|
||||
|
||||
string langTo;
|
||||
iter = meta.find( "article_language" );
|
||||
if( iter != meta.end() )
|
||||
langTo = iter->second;
|
||||
|
||||
initializing.indexingDictionary( dictName );
|
||||
|
||||
File::Class idx( indexFile, "wb" );
|
||||
IdxHeader idxHeader;
|
||||
memset( &idxHeader, 0, sizeof( idxHeader ) );
|
||||
|
||||
// We write a dummy header first. At the end of the process the header
|
||||
// will be rewritten with the right values.
|
||||
|
||||
idx.write( idxHeader );
|
||||
|
||||
idx.write( (uint32_t) dictName.size() );
|
||||
if( !dictName.empty() )
|
||||
idx.write( dictName.data(), dictName.size() );
|
||||
|
||||
IndexedWords indexedWords;
|
||||
|
||||
ChunkedStorage::Writer chunks( idx );
|
||||
|
||||
uint32_t wordCount = dictHeader.wordsCount;
|
||||
set< uint32_t > articleOffsets;
|
||||
uint32_t pos = df.tell();
|
||||
uint32_t wordsBase = pos + wordCount * sizeof( IndexElement );
|
||||
uint32_t articlesBase = dictHeader.articleOffset;
|
||||
|
||||
for( uint32_t j = 0; j < wordCount; j++ )
|
||||
{
|
||||
IndexElement el;
|
||||
|
||||
df.seek( pos );
|
||||
df.read( &el, sizeof(el) );
|
||||
uint32_t articleOffset = articlesBase + el.articleOffset;
|
||||
uint32_t wordOffset = wordsBase + el.wordOffset;
|
||||
|
||||
df.seek( wordOffset );
|
||||
|
||||
uint16_be sizeBE;
|
||||
df.read( &sizeBE, sizeof(sizeBE) );
|
||||
uint16_t wordSize = sizeBE;
|
||||
data.resize( wordSize );
|
||||
df.read( &data.front(), wordSize );
|
||||
|
||||
if( articleOffsets.find( articleOffset ) == articleOffsets.end() )
|
||||
articleOffsets.insert( articleOffset );
|
||||
|
||||
// Insert new entry
|
||||
indexedWords.addWord( Utf8::decode( string( data.data(), wordSize ) ), articleOffset);
|
||||
|
||||
pos += sizeof(el);
|
||||
}
|
||||
// Finish with the chunks
|
||||
|
||||
idxHeader.chunksOffset = chunks.finish();
|
||||
|
||||
// Build index
|
||||
|
||||
IndexInfo idxInfo = BtreeIndexing::buildIndex( indexedWords, idx );
|
||||
|
||||
idxHeader.indexBtreeMaxElements = idxInfo.btreeMaxElements;
|
||||
idxHeader.indexRootOffset = idxInfo.rootOffset;
|
||||
|
||||
indexedWords.clear(); // Release memory -- no need for this data
|
||||
|
||||
// That concludes it. Update the header.
|
||||
|
||||
idxHeader.signature = Signature;
|
||||
idxHeader.formatVersion = CurrentFormatVersion;
|
||||
|
||||
idxHeader.articleCount = articleOffsets.size();
|
||||
idxHeader.wordCount = wordCount;
|
||||
|
||||
if( langFrom.size() == 3)
|
||||
idxHeader.langFrom = LangCoder::code3toInt( langFrom.c_str() );
|
||||
else if( langFrom.size() == 2 )
|
||||
idxHeader.langFrom = LangCoder::code2toInt( langFrom.c_str() );
|
||||
|
||||
if( langTo.size() == 3)
|
||||
idxHeader.langTo = LangCoder::code3toInt( langTo.c_str() );
|
||||
else if( langTo.size() == 2 )
|
||||
idxHeader.langTo = LangCoder::code2toInt( langTo.c_str() );
|
||||
|
||||
idx.rewind();
|
||||
|
||||
idx.write( &idxHeader, sizeof( idxHeader ) );
|
||||
}
|
||||
catch( std::exception & e )
|
||||
{
|
||||
FDPRINTF( stderr, "Aard dictionary indexing failed: %s, error: %s\n",
|
||||
i->c_str(), e.what() );
|
||||
continue;
|
||||
}
|
||||
catch( ... )
|
||||
{
|
||||
FDPRINTF( stderr, "Aard dictionary indexing failed\n" );
|
||||
continue;
|
||||
}
|
||||
} // if need to rebuild
|
||||
dictionaries.push_back( new AardDictionary( dictId,
|
||||
indexFile,
|
||||
dictFiles ) );
|
||||
}
|
||||
return dictionaries;
|
||||
}
|
||||
|
||||
}
|
23
aard.hh
Normal file
23
aard.hh
Normal file
|
@ -0,0 +1,23 @@
|
|||
/* This file is (c) 2008-2012 Konstantin Isakov <ikm@goldendict.org>
|
||||
* Part of GoldenDict. Licensed under GPLv3 or later, see the LICENSE file */
|
||||
|
||||
#ifndef __AARD_HH_INCLUDED__
|
||||
#define __AARD_HH_INCLUDED__
|
||||
|
||||
#include "dictionary.hh"
|
||||
|
||||
/// Support for the aard dictionaries.
|
||||
namespace Aard {
|
||||
|
||||
using std::vector;
|
||||
using std::string;
|
||||
|
||||
vector< sptr< Dictionary::Class > > makeDictionaries(
|
||||
vector< string > const & fileNames,
|
||||
string const & indicesDir,
|
||||
Dictionary::Initializing & )
|
||||
throw( std::exception );
|
||||
|
||||
}
|
||||
|
||||
#endif
|
72
decompress.cc
Normal file
72
decompress.cc
Normal file
|
@ -0,0 +1,72 @@
|
|||
#include <stdlib.h>
|
||||
|
||||
#include "decompress.hh"
|
||||
#include "zlib.h"
|
||||
#include "bzlib.h"
|
||||
|
||||
string decompressZlib( char * bufptr, unsigned length )
|
||||
{
|
||||
z_stream zs;
|
||||
char buf[2048];
|
||||
string str;
|
||||
int res;
|
||||
memset( &zs, 0, sizeof(zs) );
|
||||
zs.next_in = (Bytef *)bufptr;
|
||||
zs.avail_in = length;
|
||||
while( 1 )
|
||||
{
|
||||
res = inflateInit( &zs );
|
||||
if( res != Z_OK )
|
||||
break;
|
||||
while( res != Z_STREAM_END )
|
||||
{
|
||||
memset( buf, 0, sizeof(buf) );
|
||||
zs.next_out = (Bytef *)buf;
|
||||
zs.avail_out = 2047;
|
||||
res = inflate( &zs, Z_SYNC_FLUSH );
|
||||
str += buf;
|
||||
if( res != Z_OK && res != Z_STREAM_END )
|
||||
break;
|
||||
}
|
||||
break;
|
||||
}
|
||||
inflateEnd( &zs );
|
||||
if( res != Z_STREAM_END )
|
||||
str.clear();
|
||||
return str;
|
||||
}
|
||||
|
||||
string decompressBzip2( char * bufptr, unsigned length )
|
||||
{
|
||||
bz_stream zs;
|
||||
char buf[2048];
|
||||
string str;
|
||||
int res;
|
||||
memset( &zs, 0, sizeof(zs) );
|
||||
zs.next_in = bufptr;
|
||||
zs.avail_in = length;
|
||||
zs.total_in_lo32 = length;
|
||||
while( 1 )
|
||||
{
|
||||
res = BZ2_bzDecompressInit( &zs, 0, 0 );
|
||||
if( res != BZ_OK )
|
||||
break;
|
||||
while( res != BZ_STREAM_END )
|
||||
{
|
||||
memset( buf, 0, sizeof(buf) );
|
||||
zs.next_out = buf;
|
||||
zs.avail_out = 2047;
|
||||
zs.total_out_lo32 = length;
|
||||
res = BZ2_bzDecompress( &zs );
|
||||
str += buf;
|
||||
if( res != BZ_OK && res != BZ_STREAM_END )
|
||||
break;
|
||||
}
|
||||
break;
|
||||
}
|
||||
BZ2_bzDecompressEnd( &zs );
|
||||
if( res != BZ_STREAM_END )
|
||||
str.clear();
|
||||
return str;
|
||||
}
|
||||
|
12
decompress.hh
Normal file
12
decompress.hh
Normal file
|
@ -0,0 +1,12 @@
|
|||
#ifndef __DECOMPRESS_HH_INCLUDED__
|
||||
#define __DECOMPRESS_HH_INCLUDED__
|
||||
|
||||
#include <string>
|
||||
|
||||
using std::string;
|
||||
|
||||
string decompressZlib( char * bufptr, unsigned length );
|
||||
|
||||
string decompressBzip2( char * bufptr, unsigned length );
|
||||
|
||||
#endif // DECOMPRESS_HH
|
|
@ -192,7 +192,9 @@ HEADERS += folding.hh \
|
|||
gdappstyle.hh \
|
||||
ufile.hh \
|
||||
xdxf.hh \
|
||||
sdict.hh
|
||||
sdict.hh \
|
||||
decompress.hh \
|
||||
aard.hh
|
||||
FORMS += groups.ui \
|
||||
dictgroupwidget.ui \
|
||||
mainwindow.ui \
|
||||
|
@ -282,7 +284,9 @@ SOURCES += folding.cc \
|
|||
gdappstyle.cc \
|
||||
ufile.cc \
|
||||
xdxf.cc \
|
||||
sdict.cc
|
||||
sdict.cc \
|
||||
decompress.cc \
|
||||
aard.cc
|
||||
win32 {
|
||||
SOURCES += mouseover_win32/ThTypes.c \
|
||||
wordbyauto.cc \
|
||||
|
|
BIN
icons/icon32_aard.png
Normal file
BIN
icons/icon32_aard.png
Normal file
Binary file not shown.
After Width: | Height: | Size: 3.5 KiB |
|
@ -22,6 +22,7 @@
|
|||
#include "fsencoding.hh"
|
||||
#include "xdxf.hh"
|
||||
#include "sdict.hh"
|
||||
#include "aard.hh"
|
||||
|
||||
#include <QMessageBox>
|
||||
#include <QDir>
|
||||
|
@ -42,7 +43,7 @@ LoadDictionaries::LoadDictionaries( Config::Class const & cfg ):
|
|||
|
||||
nameFilters << "*.bgl" << "*.ifo" << "*.lsa" << "*.dat"
|
||||
<< "*.dsl" << "*.dsl.dz" << "*.index" << "*.xdxf"
|
||||
<< "*.xdxf.dz" << "*.dct";
|
||||
<< "*.xdxf.dz" << "*.dct" << "*.aar";
|
||||
}
|
||||
|
||||
void LoadDictionaries::run()
|
||||
|
@ -155,6 +156,13 @@ void LoadDictionaries::handlePath( Config::Path const & path )
|
|||
dictionaries.insert( dictionaries.end(), sdictDictionaries.begin(),
|
||||
sdictDictionaries.end() );
|
||||
}
|
||||
{
|
||||
vector< sptr< Dictionary::Class > > aardDictionaries =
|
||||
Aard::makeDictionaries( allFiles, FsEncoding::encode( Config::getIndexDir() ), *this );
|
||||
|
||||
dictionaries.insert( dictionaries.end(), aardDictionaries.begin(),
|
||||
aardDictionaries.end() );
|
||||
}
|
||||
}
|
||||
|
||||
void LoadDictionaries::indexingDictionary( string const & dictionaryName ) throw()
|
||||
|
|
|
@ -48,5 +48,6 @@
|
|||
<file>icons/error.png</file>
|
||||
<file>icons/macicon.png</file>
|
||||
<file>icons/icon32_sdict.png</file>
|
||||
<file>icons/icon32_aard.png</file>
|
||||
</qresource>
|
||||
</RCC>
|
||||
|
|
70
sdict.cc
70
sdict.cc
|
@ -9,13 +9,11 @@
|
|||
#include "langcoder.hh"
|
||||
#include "dprintf.hh"
|
||||
#include "fsencoding.hh"
|
||||
#include "decompress.hh"
|
||||
|
||||
#include <zlib.h>
|
||||
#include <bzlib.h>
|
||||
#include <map>
|
||||
#include <set>
|
||||
#include <string>
|
||||
#include <stdlib.h>
|
||||
|
||||
#ifdef _MSC_VER
|
||||
#include <stub_msvc.h>
|
||||
|
@ -124,72 +122,6 @@ bool indexIsOldOrBad( string const & indexFile )
|
|||
header.formatVersion != CurrentFormatVersion;
|
||||
}
|
||||
|
||||
string decompressZlib( char * bufptr, unsigned length )
|
||||
{
|
||||
z_stream zs;
|
||||
char buf[2048];
|
||||
string str;
|
||||
int res;
|
||||
memset( &zs, 0, sizeof(zs) );
|
||||
zs.next_in = (Bytef *)bufptr;
|
||||
zs.avail_in = length;
|
||||
while( 1 )
|
||||
{
|
||||
res = inflateInit( &zs );
|
||||
if( res != Z_OK )
|
||||
break;
|
||||
while( res != Z_STREAM_END )
|
||||
{
|
||||
memset( buf, 0, sizeof(buf) );
|
||||
zs.next_out = (Bytef *)buf;
|
||||
zs.avail_out = 2047;
|
||||
res = inflate( &zs, Z_SYNC_FLUSH );
|
||||
str += buf;
|
||||
if( res != Z_OK && res != Z_STREAM_END )
|
||||
break;
|
||||
}
|
||||
break;
|
||||
}
|
||||
inflateEnd( &zs );
|
||||
if( res != Z_STREAM_END )
|
||||
str.clear();
|
||||
return str;
|
||||
}
|
||||
|
||||
string decompressBzip2( char * bufptr, unsigned length )
|
||||
{
|
||||
bz_stream zs;
|
||||
char buf[2048];
|
||||
string str;
|
||||
int res;
|
||||
memset( &zs, 0, sizeof(zs) );
|
||||
zs.next_in = bufptr;
|
||||
zs.avail_in = length;
|
||||
zs.total_in_lo32 = length;
|
||||
while( 1 )
|
||||
{
|
||||
res = BZ2_bzDecompressInit( &zs, 0, 0 );
|
||||
if( res != BZ_OK )
|
||||
break;
|
||||
while( res != BZ_STREAM_END )
|
||||
{
|
||||
memset( buf, 0, sizeof(buf) );
|
||||
zs.next_out = buf;
|
||||
zs.avail_out = 2047;
|
||||
zs.total_out_lo32 = length;
|
||||
res = BZ2_bzDecompress( &zs );
|
||||
str += buf;
|
||||
if( res != BZ_OK && res != BZ_STREAM_END )
|
||||
break;
|
||||
}
|
||||
break;
|
||||
}
|
||||
BZ2_bzDecompressEnd( &zs );
|
||||
if( res != BZ_STREAM_END )
|
||||
str.clear();
|
||||
return str;
|
||||
}
|
||||
|
||||
class SdictDictionary: public BtreeIndexing::BtreeDictionary
|
||||
{
|
||||
Mutex idxMutex;
|
||||
|
|
Loading…
Reference in a new issue