mirror of
https://github.com/xiaoyifang/goldendict-ng.git
synced 2024-11-23 20:14:05 +00:00
501 lines
15 KiB
C++
501 lines
15 KiB
C++
/* This file is (c) 2008-2012 Konstantin Isakov <ikm@goldendict.org>
|
|
* Part of GoldenDict. Licensed under GPLv3 or later, see the LICENSE file */
|
|
|
|
#include "wordfinder.hh"
|
|
#include "folding.hh"
|
|
#include "wstring_qt.hh"
|
|
#include <map>
|
|
#include "gddebug.hh"
|
|
|
|
using std::vector;
|
|
using std::list;
|
|
using gd::wstring;
|
|
using gd::wchar;
|
|
using std::map;
|
|
using std::pair;
|
|
|
|
WordFinder::WordFinder( QObject * parent ):
|
|
QObject( parent ),
|
|
searchInProgress( false ),
|
|
updateResultsTimer( this ),
|
|
searchQueued( false )
|
|
{
|
|
updateResultsTimer.setInterval( 1000 ); // We use a one second update timer
|
|
updateResultsTimer.setSingleShot( true );
|
|
|
|
connect( &updateResultsTimer, &QTimer::timeout, this, &WordFinder::updateResults, Qt::QueuedConnection );
|
|
}
|
|
|
|
WordFinder::~WordFinder()
|
|
{
|
|
clear();
|
|
}
|
|
|
|
void WordFinder::prefixMatch( QString const & str,
|
|
std::vector< sptr< Dictionary::Class > > const & dicts,
|
|
unsigned long maxResults,
|
|
Dictionary::Features features )
|
|
{
|
|
cancel();
|
|
|
|
searchQueued = true;
|
|
searchType = PrefixMatch;
|
|
inputWord = str;
|
|
inputDicts = &dicts;
|
|
requestedMaxResults = maxResults;
|
|
requestedFeatures = features;
|
|
|
|
resultsArray.clear();
|
|
resultsIndex.clear();
|
|
searchResults.clear();
|
|
|
|
if ( queuedRequests.empty() ) {
|
|
// No requests are queued, no need to wait for them to finish.
|
|
startSearch();
|
|
}
|
|
|
|
// Else some requests are still queued, last one to finish would trigger
|
|
// new search. This shouldn't take a lot of time, since they were all
|
|
// cancelled, but still it could take some time.
|
|
}
|
|
void WordFinder::stemmedMatch( QString const & str,
|
|
std::vector< sptr< Dictionary::Class > > const & dicts,
|
|
unsigned minLength,
|
|
unsigned maxSuffixVariation,
|
|
unsigned long maxResults,
|
|
Dictionary::Features features )
|
|
{
|
|
cancel();
|
|
|
|
searchQueued = true;
|
|
searchType = StemmedMatch;
|
|
inputWord = str;
|
|
inputDicts = &dicts;
|
|
requestedMaxResults = maxResults;
|
|
requestedFeatures = features;
|
|
stemmedMinLength = minLength;
|
|
stemmedMaxSuffixVariation = maxSuffixVariation;
|
|
|
|
resultsArray.clear();
|
|
resultsIndex.clear();
|
|
searchResults.clear();
|
|
|
|
if ( queuedRequests.empty() ) {
|
|
startSearch();
|
|
}
|
|
}
|
|
|
|
void WordFinder::expressionMatch( QString const & str,
|
|
std::vector< sptr< Dictionary::Class > > const & dicts,
|
|
unsigned long maxResults,
|
|
Dictionary::Features features )
|
|
{
|
|
cancel();
|
|
|
|
searchQueued = true;
|
|
searchType = ExpressionMatch;
|
|
inputWord = str;
|
|
inputDicts = &dicts;
|
|
requestedMaxResults = maxResults;
|
|
requestedFeatures = features;
|
|
|
|
resultsArray.clear();
|
|
resultsIndex.clear();
|
|
searchResults.clear();
|
|
|
|
if ( queuedRequests.empty() ) {
|
|
// No requests are queued, no need to wait for them to finish.
|
|
startSearch();
|
|
}
|
|
}
|
|
|
|
void WordFinder::startSearch()
|
|
{
|
|
if ( !searchQueued ) {
|
|
return; // Search was probably cancelled
|
|
}
|
|
|
|
// Clear the requests just in case
|
|
queuedRequests.clear();
|
|
finishedRequests.clear();
|
|
|
|
searchErrorString.clear();
|
|
searchResultsUncertain = false;
|
|
|
|
searchQueued = false;
|
|
searchInProgress = true;
|
|
|
|
// Gather all writings of the word
|
|
|
|
if ( allWordWritings.size() != 1 ) {
|
|
allWordWritings.resize( 1 );
|
|
}
|
|
|
|
allWordWritings[ 0 ] = gd::toWString( inputWord );
|
|
|
|
for ( const auto & inputDict : *inputDicts ) {
|
|
vector< wstring > writings = inputDict->getAlternateWritings( allWordWritings[ 0 ] );
|
|
|
|
allWordWritings.insert( allWordWritings.end(), writings.begin(), writings.end() );
|
|
}
|
|
|
|
// Query each dictionary for all word writings
|
|
|
|
for ( const auto & inputDict : *inputDicts ) {
|
|
if ( ( inputDict->getFeatures() & requestedFeatures ) != requestedFeatures ) {
|
|
continue;
|
|
}
|
|
|
|
for ( const auto & allWordWriting : allWordWritings ) {
|
|
try {
|
|
sptr< Dictionary::WordSearchRequest > sr = ( searchType == PrefixMatch || searchType == ExpressionMatch ) ?
|
|
inputDict->prefixMatch( allWordWriting, requestedMaxResults ) :
|
|
inputDict->stemmedMatch( allWordWriting, stemmedMinLength, stemmedMaxSuffixVariation, requestedMaxResults );
|
|
|
|
connect( sr.get(), &Dictionary::Request::finished, this, &WordFinder::requestFinished, Qt::QueuedConnection );
|
|
|
|
queuedRequests.push_back( sr );
|
|
}
|
|
catch ( std::exception & e ) {
|
|
gdWarning( "Word \"%s\" search error (%s) in \"%s\"\n",
|
|
inputWord.toUtf8().data(),
|
|
e.what(),
|
|
inputDict->getName().c_str() );
|
|
}
|
|
}
|
|
}
|
|
|
|
// Handle any requests finished already
|
|
|
|
requestFinished();
|
|
}
|
|
|
|
void WordFinder::cancel()
|
|
{
|
|
searchQueued = false;
|
|
searchInProgress = false;
|
|
|
|
cancelSearches();
|
|
}
|
|
|
|
void WordFinder::clear()
|
|
{
|
|
cancel();
|
|
queuedRequests.clear();
|
|
finishedRequests.clear();
|
|
}
|
|
|
|
void WordFinder::requestFinished()
|
|
{
|
|
bool newResults = false;
|
|
|
|
// See how many new requests have finished, and if we have any new results
|
|
for ( auto i = queuedRequests.begin(); i != queuedRequests.end(); ) {
|
|
if ( ( *i )->isFinished() ) {
|
|
if ( searchInProgress && !( *i )->getErrorString().isEmpty() ) {
|
|
searchErrorString = tr( "Failed to query some dictionaries." );
|
|
}
|
|
|
|
if ( ( *i )->isUncertain() ) {
|
|
searchResultsUncertain = true;
|
|
}
|
|
|
|
if ( ( *i )->matchesCount() ) {
|
|
newResults = true;
|
|
|
|
// This list is handled by updateResults()
|
|
finishedRequests.splice( finishedRequests.end(), queuedRequests, i++ );
|
|
}
|
|
else { // We won't do anything with it anymore, so we erase it
|
|
queuedRequests.erase( i++ );
|
|
}
|
|
}
|
|
else {
|
|
++i;
|
|
}
|
|
}
|
|
|
|
if ( !searchInProgress ) {
|
|
// There is no search in progress, so we just wait until there's
|
|
// no requests left
|
|
|
|
if ( queuedRequests.empty() ) {
|
|
// We got rid of all queries, queued search can now start
|
|
finishedRequests.clear();
|
|
|
|
if ( searchQueued ) {
|
|
startSearch();
|
|
}
|
|
}
|
|
|
|
return;
|
|
}
|
|
|
|
if ( newResults && !queuedRequests.empty() && !updateResultsTimer.isActive() ) {
|
|
// If we have got some new results, but not all of them, we would start a
|
|
// timer to update a user some time in the future
|
|
updateResultsTimer.start();
|
|
}
|
|
|
|
if ( queuedRequests.empty() ) {
|
|
// Search is finished.
|
|
updateResults();
|
|
}
|
|
}
|
|
|
|
namespace {
|
|
|
|
|
|
unsigned saturated( unsigned x )
|
|
{
|
|
return x < 255 ? x : 255;
|
|
}
|
|
|
|
/// Checks whether the first string has the second one inside, surrounded from
|
|
/// both sides by either whitespace, punctuation or begin/end of string.
|
|
/// If true is returned, pos holds the offset in the haystack. If the offset
|
|
/// is larger than 255, it is set to 255.
|
|
bool hasSurroundedWithWs( wstring const & haystack, wstring const & needle, wstring::size_type & pos )
|
|
{
|
|
if ( haystack.size() < needle.size() ) {
|
|
return false; // Needle won't even fit into a haystack
|
|
}
|
|
|
|
for ( pos = 0;; ++pos ) {
|
|
pos = haystack.find( needle, pos );
|
|
|
|
if ( pos == wstring::npos ) {
|
|
return false; // Not found
|
|
}
|
|
|
|
if ( ( !pos || Folding::isWhitespace( haystack[ pos - 1 ] ) || Folding::isPunct( haystack[ pos - 1 ] ) )
|
|
&& ( ( pos + needle.size() == haystack.size() ) || Folding::isWhitespace( haystack[ pos + needle.size() ] )
|
|
|| Folding::isPunct( haystack[ pos + needle.size() ] ) ) ) {
|
|
pos = saturated( pos );
|
|
|
|
return true;
|
|
}
|
|
}
|
|
}
|
|
|
|
} // namespace
|
|
|
|
void WordFinder::updateResults()
|
|
{
|
|
if ( !searchInProgress ) {
|
|
return; // Old queued signal
|
|
}
|
|
|
|
if ( updateResultsTimer.isActive() ) {
|
|
updateResultsTimer.stop(); // Can happen when we were done before it'd expire
|
|
}
|
|
|
|
wstring original = Folding::applySimpleCaseOnly( allWordWritings[ 0 ] );
|
|
|
|
for ( auto i = finishedRequests.begin(); i != finishedRequests.end(); ) {
|
|
for ( size_t count = ( *i )->matchesCount(), x = 0; x < count; ++x ) {
|
|
wstring match = ( **i )[ x ].word;
|
|
int weight = ( **i )[ x ].weight;
|
|
wstring lowerCased = Folding::applySimpleCaseOnly( match );
|
|
|
|
if ( searchType == ExpressionMatch ) {
|
|
unsigned ws;
|
|
|
|
for ( ws = 0; ws < allWordWritings.size(); ws++ ) {
|
|
if ( ws == 0 ) {
|
|
// Check for prefix match with original expression
|
|
if ( lowerCased.compare( 0, original.size(), original ) == 0 ) {
|
|
break;
|
|
}
|
|
}
|
|
else if ( lowerCased == Folding::applySimpleCaseOnly( allWordWritings[ ws ] ) ) {
|
|
break;
|
|
}
|
|
}
|
|
|
|
if ( ws >= allWordWritings.size() ) {
|
|
// No exact matches found
|
|
continue;
|
|
}
|
|
weight = ws;
|
|
}
|
|
auto insertResult =
|
|
resultsIndex.insert( pair< wstring, ResultsArray::iterator >( lowerCased, resultsArray.end() ) );
|
|
|
|
if ( !insertResult.second ) {
|
|
// Wasn't inserted since there was already an item -- check the case
|
|
if ( insertResult.first->second->word != match ) {
|
|
// The case is different -- agree on a lowercase version
|
|
insertResult.first->second->word = lowerCased;
|
|
}
|
|
if ( !weight && insertResult.first->second->wasSuggested ) {
|
|
insertResult.first->second->wasSuggested = false;
|
|
}
|
|
}
|
|
else {
|
|
resultsArray.emplace_back();
|
|
|
|
resultsArray.back().word = match;
|
|
resultsArray.back().rank = INT_MAX;
|
|
resultsArray.back().wasSuggested = ( weight != 0 );
|
|
|
|
insertResult.first->second = --resultsArray.end();
|
|
}
|
|
}
|
|
finishedRequests.erase( i++ );
|
|
}
|
|
|
|
size_t maxSearchResults = 500;
|
|
|
|
if ( !resultsArray.empty() ) {
|
|
if ( searchType == PrefixMatch ) {
|
|
/// Assign each result a category, storing it in the rank's field
|
|
|
|
enum Category {
|
|
ExactMatch,
|
|
ExactNoFullCaseMatch,
|
|
ExactNoDiaMatch,
|
|
ExactNoPunctMatch,
|
|
ExactNoWsMatch,
|
|
ExactInsideMatch,
|
|
ExactNoDiaInsideMatch,
|
|
ExactNoPunctInsideMatch,
|
|
PrefixMatch,
|
|
PrefixNoDiaMatch,
|
|
PrefixNoPunctMatch,
|
|
PrefixNoWsMatch,
|
|
WorstMatch,
|
|
Multiplier = 256 // Categories should be multiplied by Multiplier
|
|
};
|
|
|
|
for ( const auto & allWordWriting : allWordWritings ) {
|
|
wstring target = Folding::applySimpleCaseOnly( allWordWriting );
|
|
wstring targetNoFullCase = Folding::applyFullCaseOnly( target );
|
|
wstring targetNoDia = Folding::applyDiacriticsOnly( targetNoFullCase );
|
|
wstring targetNoPunct = Folding::applyPunctOnly( targetNoDia );
|
|
wstring targetNoWs = Folding::applyWhitespaceOnly( targetNoPunct );
|
|
|
|
wstring::size_type matchPos = 0;
|
|
|
|
for ( const auto & i : resultsIndex ) {
|
|
wstring resultNoFullCase, resultNoDia, resultNoPunct, resultNoWs;
|
|
|
|
int rank;
|
|
|
|
if ( i.first == target ) {
|
|
rank = ExactMatch * Multiplier;
|
|
}
|
|
else if ( ( resultNoFullCase = Folding::applyFullCaseOnly( i.first ) ) == targetNoFullCase ) {
|
|
rank = ExactNoFullCaseMatch * Multiplier;
|
|
}
|
|
else if ( ( resultNoDia = Folding::applyDiacriticsOnly( resultNoFullCase ) ) == targetNoDia ) {
|
|
rank = ExactNoDiaMatch * Multiplier;
|
|
}
|
|
else if ( ( resultNoPunct = Folding::applyPunctOnly( resultNoDia ) ) == targetNoPunct ) {
|
|
rank = ExactNoPunctMatch * Multiplier;
|
|
}
|
|
else if ( ( resultNoWs = Folding::applyWhitespaceOnly( resultNoPunct ) ) == targetNoWs ) {
|
|
rank = ExactNoWsMatch * Multiplier;
|
|
}
|
|
else if ( hasSurroundedWithWs( i.first, target, matchPos ) ) {
|
|
rank = ExactInsideMatch * Multiplier + matchPos;
|
|
}
|
|
else if ( hasSurroundedWithWs( resultNoDia, targetNoDia, matchPos ) ) {
|
|
rank = ExactNoDiaInsideMatch * Multiplier + matchPos;
|
|
}
|
|
else if ( hasSurroundedWithWs( resultNoPunct, targetNoPunct, matchPos ) ) {
|
|
rank = ExactNoPunctInsideMatch * Multiplier + matchPos;
|
|
}
|
|
else if ( i.first.size() > target.size() && i.first.compare( 0, target.size(), target ) == 0 ) {
|
|
rank = PrefixMatch * Multiplier + saturated( i.first.size() );
|
|
}
|
|
else if ( resultNoDia.size() > targetNoDia.size()
|
|
&& resultNoDia.compare( 0, targetNoDia.size(), targetNoDia ) == 0 ) {
|
|
rank = PrefixNoDiaMatch * Multiplier + saturated( i.first.size() );
|
|
}
|
|
else if ( resultNoPunct.size() > targetNoPunct.size()
|
|
&& resultNoPunct.compare( 0, targetNoPunct.size(), targetNoPunct ) == 0 ) {
|
|
rank = PrefixNoPunctMatch * Multiplier + saturated( i.first.size() );
|
|
}
|
|
else if ( resultNoWs.size() > targetNoWs.size()
|
|
&& resultNoWs.compare( 0, targetNoWs.size(), targetNoWs ) == 0 ) {
|
|
rank = PrefixNoWsMatch * Multiplier + saturated( i.first.size() );
|
|
}
|
|
else {
|
|
rank = WorstMatch * Multiplier;
|
|
}
|
|
|
|
if ( i.second->rank > rank ) {
|
|
i.second->rank = rank; // We store the best rank of any writing
|
|
}
|
|
}
|
|
}
|
|
|
|
resultsArray.sort( SortByRank() );
|
|
}
|
|
else if ( searchType == StemmedMatch ) {
|
|
// Handling stemmed matches
|
|
|
|
// We use two factors -- first is the number of characters strings share
|
|
// in their beginnings, and second, the length of the strings. Here we assign
|
|
// only the first one, storing it in rank. Then we sort the results using
|
|
// SortByRankAndLength.
|
|
for ( const auto & allWordWriting : allWordWritings ) {
|
|
wstring target = Folding::apply( allWordWriting );
|
|
|
|
for ( const auto & i : resultsIndex ) {
|
|
wstring resultFolded = Folding::apply( i.first );
|
|
|
|
int charsInCommon = 0;
|
|
|
|
for ( wchar const *t = target.c_str(), *r = resultFolded.c_str(); *t && *t == *r;
|
|
++t, ++r, ++charsInCommon ) {
|
|
;
|
|
}
|
|
|
|
int rank = -charsInCommon; // Negated so the lesser-than
|
|
// comparison would yield right
|
|
// results.
|
|
|
|
if ( i.second->rank > rank ) {
|
|
i.second->rank = rank; // We store the best rank of any writing
|
|
}
|
|
}
|
|
}
|
|
|
|
resultsArray.sort( SortByRankAndLength() );
|
|
|
|
maxSearchResults = 15;
|
|
}
|
|
}
|
|
|
|
searchResults.clear();
|
|
searchResults.reserve( resultsArray.size() < maxSearchResults ? resultsArray.size() : maxSearchResults );
|
|
|
|
for ( const auto & i : resultsArray ) {
|
|
if ( searchResults.size() < maxSearchResults ) {
|
|
searchResults.emplace_back( QString::fromStdU32String( i.word ), i.wasSuggested );
|
|
}
|
|
else {
|
|
break;
|
|
}
|
|
}
|
|
|
|
if ( !queuedRequests.empty() ) {
|
|
// There are still some unhandled results.
|
|
emit updated();
|
|
}
|
|
else {
|
|
// That were all of them.
|
|
searchInProgress = false;
|
|
emit finished();
|
|
}
|
|
}
|
|
|
|
void WordFinder::cancelSearches()
|
|
{
|
|
for ( auto & queuedRequest : queuedRequests ) {
|
|
queuedRequest->cancel();
|
|
}
|
|
}
|