mirror of
https://github.com/yanyiwu/cppjieba.git
synced 2025-07-18 00:00:12 +08:00
403 lines
7.9 KiB
C++
403 lines
7.9 KiB
C++
/************************************
|
||
* file enc : utf8
|
||
* author : wuyanyi09@gmail.com
|
||
************************************/
|
||
#include "KeyWordExt.h"
|
||
|
||
|
||
namespace CppJieba
|
||
{
|
||
|
||
KeyWordExt::KeyWordExt()
|
||
{
|
||
}
|
||
|
||
KeyWordExt::~KeyWordExt()
|
||
{
|
||
}
|
||
|
||
bool KeyWordExt::init(const char * const filePath)
|
||
{
|
||
LogInfo(string_format("init(%s) start", filePath));
|
||
if(!checkFileExist(filePath))
|
||
{
|
||
LogError(string_format("cann't find file[%s].",filePath));
|
||
return false;
|
||
}
|
||
bool retFlag = _segment.init();
|
||
return retFlag;
|
||
|
||
}
|
||
|
||
bool KeyWordExt::loadPriorSubWords(const char * const filePath)
|
||
{
|
||
LogInfo(string_format("loadPriorSubWords(%s) start", filePath));
|
||
if(!checkFileExist(filePath))
|
||
{
|
||
LogError(string_format("cann't find file[%s].",filePath));
|
||
return false;
|
||
}
|
||
if(!_priorSubWords.empty())
|
||
{
|
||
LogError("_priorSubWords has been initted before");
|
||
return false;
|
||
}
|
||
ifstream infile(filePath);
|
||
string subword;
|
||
while(getline(infile, subword))
|
||
{
|
||
_priorSubWords.push_back(subword);
|
||
}
|
||
LogInfo(string_format("loadPriorSubWords(%s) end", filePath));
|
||
infile.close();
|
||
return true;
|
||
}
|
||
|
||
bool KeyWordExt::loadStopWords(const char * const filePath)
|
||
{
|
||
|
||
LogInfo(string_format("loadStopWords(%s) start", filePath));
|
||
if(!_stopWords.empty())
|
||
{
|
||
LogError("_stopWords has been loaded before! ");
|
||
return false;
|
||
}
|
||
if(!checkFileExist(filePath))
|
||
{
|
||
LogError(string_format("cann't find file[%s].",filePath));
|
||
return false;
|
||
}
|
||
|
||
ifstream ifile(filePath);
|
||
string line;
|
||
while(getline(ifile, line))
|
||
{
|
||
_stopWords.insert(line);
|
||
}
|
||
LogInfo(string_format("load stopwords[%d] finished.", _stopWords.size()));
|
||
|
||
return true;
|
||
}
|
||
|
||
bool KeyWordExt::dispose()
|
||
{
|
||
_segment.dispose();
|
||
return true;
|
||
}
|
||
|
||
bool KeyWordExt::_wordInfoCompare(const WordInfo& a, const WordInfo& b)
|
||
{
|
||
return a.weight > b.weight;
|
||
}
|
||
|
||
bool KeyWordExt::_sortWLIDF(vector<WordInfo>& wordInfos)
|
||
{
|
||
//size_t wLenSum = 0;
|
||
for(uint i = 0; i < wordInfos.size(); i++)
|
||
{
|
||
wordInfos[i].wLen = gEncoding.getWordLength(wordInfos[i].word);
|
||
if(0 == wordInfos[i].wLen)
|
||
{
|
||
LogFatal("wLen is 0");
|
||
return false;
|
||
}
|
||
//wLenSum += wordInfos[i].wLen;
|
||
}
|
||
|
||
/*
|
||
if(0 == wLenSum)
|
||
{
|
||
LogFatal("wLenSum == 0.");
|
||
return false;
|
||
}*/
|
||
|
||
for(uint i = 0; i < wordInfos.size(); i++)
|
||
{
|
||
WordInfo& wInfo = wordInfos[i];
|
||
double logWordFreq = _segment.getWordWeight(wInfo.word);
|
||
wInfo.idf = -logWordFreq;
|
||
size_t wLen = gEncoding.getWordLength(wInfo.word);
|
||
if(0 == wLen)
|
||
{
|
||
LogFatal("getUtf8WordLen(%s) return 0");
|
||
}
|
||
wInfo.weight = log(double(wLen + 1)) * wInfo.idf;
|
||
}
|
||
sort(wordInfos.begin(), wordInfos.end(), _wordInfoCompare);
|
||
return true;
|
||
}
|
||
|
||
bool KeyWordExt::_extractTopN(const vector<string>& words, vector<string>& keywords, uint topN)
|
||
{
|
||
keywords.clear();
|
||
vector<WordInfo> wordInfos;
|
||
for(uint i = 0; i < words.size(); i++)
|
||
{
|
||
WordInfo wInfo;
|
||
wInfo.word = words[i];
|
||
wordInfos.push_back(wInfo);
|
||
}
|
||
|
||
_sortWLIDF(wordInfos);
|
||
LogDebug(string_format("calc weight & sorted:%s",joinWordInfos(wordInfos).c_str()));
|
||
|
||
_prioritizeSubWords(wordInfos);
|
||
//LogDebug(string_format("_prioritizeSubWords res:%s", joinWordInfos(wordInfos).c_str()));
|
||
//extract TopN
|
||
for(uint i = 0; i < topN && i < wordInfos.size(); i++)
|
||
{
|
||
keywords.push_back(wordInfos[i].word);
|
||
}
|
||
return true;
|
||
}
|
||
|
||
bool KeyWordExt::extract(const string& utf8Str, vector<string>& keywords, uint topN)
|
||
{
|
||
LogDebug(string_format("title:[%s]",utf8Str.c_str()));
|
||
|
||
bool retFlag;
|
||
vector<string> words;
|
||
retFlag = _segment.cutDAG(utf8Str, words);
|
||
if(!retFlag)
|
||
{
|
||
LogError(string_format("cutDAG(%s) failed.", utf8Str.c_str()));
|
||
return false;
|
||
}
|
||
LogDebug(string_format("cutDAG result:[%s]", joinStr(words, ",").c_str()));
|
||
|
||
retFlag = _filter(words);
|
||
if(!retFlag)
|
||
{
|
||
LogError("_filter failed.");
|
||
return false;
|
||
}
|
||
LogDebug(string_format("_filter res:[%s]", joinStr(words, ",").c_str()));
|
||
|
||
retFlag = _extractTopN(words, keywords, topN);
|
||
if(!retFlag)
|
||
{
|
||
LogError("_extractTopN failed.");
|
||
return false;
|
||
}
|
||
//LogDebug("_extractTopN finished.");
|
||
|
||
LogDebug(string_format("ext res:[%s]", joinStr(keywords, ",").c_str()));
|
||
return true;
|
||
}
|
||
|
||
bool KeyWordExt::_filter(vector<string>& strs)
|
||
{
|
||
bool retFlag;
|
||
retFlag = _filterDuplicate(strs);
|
||
if(!retFlag)
|
||
{
|
||
LogError("_filterDuplicate failed.");
|
||
return false;
|
||
}
|
||
//LogDebug(string_format("_filterDuplicate res:[%s]", joinStr(strs, ",").c_str()));
|
||
|
||
retFlag = _filterSingleWord(strs);
|
||
if(!retFlag)
|
||
{
|
||
LogError("_filterSingleWord failed.");
|
||
return false;
|
||
}
|
||
//LogDebug(string_format("_filterSingleWord res:[%s]", joinStr(strs, ",").c_str()));
|
||
|
||
retFlag = _filterStopWords(strs);
|
||
if(!retFlag)
|
||
{
|
||
LogError("_filterStopWords failed.");
|
||
return false;
|
||
}
|
||
//LogDebug(string_format("_filterStopWords res:[%s]", joinStr(strs, ",").c_str()));
|
||
|
||
retFlag = _filterSubstr(strs);
|
||
if(!retFlag)
|
||
{
|
||
LogError("_filterSubstr failed.");
|
||
return false;
|
||
}
|
||
//LogDebug(string_format("_filterSubstr res:[%s]", joinStr(strs, ",").c_str()));
|
||
|
||
return true;
|
||
}
|
||
|
||
bool KeyWordExt::_filterStopWords(vector<string>& strs)
|
||
{
|
||
if(_stopWords.empty())
|
||
{
|
||
return true;
|
||
}
|
||
for(VSI it = strs.begin(); it != strs.end();)
|
||
{
|
||
if(_stopWords.find(*it) != _stopWords.end())
|
||
{
|
||
it = strs.erase(it);
|
||
}
|
||
else
|
||
{
|
||
it ++;
|
||
}
|
||
}
|
||
return true;
|
||
}
|
||
|
||
|
||
bool KeyWordExt::_filterDuplicate(vector<string>& strs)
|
||
{
|
||
set<string> st;
|
||
for(VSI it = strs.begin(); it != strs.end(); )
|
||
{
|
||
if(st.find(*it) != st.end())
|
||
{
|
||
it = strs.erase(it);
|
||
}
|
||
else
|
||
{
|
||
st.insert(*it);
|
||
it++;
|
||
}
|
||
}
|
||
return true;
|
||
}
|
||
|
||
bool KeyWordExt::_filterSingleWord(vector<string>& strs)
|
||
{
|
||
for(vector<string>::iterator it = strs.begin(); it != strs.end();)
|
||
{
|
||
|
||
// filter single word
|
||
if(1 == gEncoding.getWordLength(*it))
|
||
{
|
||
it = strs.erase(it);
|
||
}
|
||
else
|
||
{
|
||
it++;
|
||
}
|
||
}
|
||
return true;
|
||
}
|
||
|
||
bool KeyWordExt::_filterSubstr(vector<string>& strs)
|
||
{
|
||
vector<string> tmp = strs;
|
||
set<string> subs;
|
||
for(VSI it = strs.begin(); it != strs.end(); it ++)
|
||
{
|
||
for(uint j = 0; j < tmp.size(); j++)
|
||
{
|
||
if(*it != tmp[j] && string::npos != tmp[j].find(*it, 0))
|
||
{
|
||
subs.insert(*it);
|
||
}
|
||
}
|
||
}
|
||
|
||
//erase subs from strs
|
||
for(VSI it = strs.begin(); it != strs.end(); )
|
||
{
|
||
if(subs.end() != subs.find(*it))
|
||
{
|
||
LogDebug(string_format("_filterSubstr filter [%s].", it->c_str()));
|
||
it = strs.erase(it);
|
||
}
|
||
else
|
||
{
|
||
it ++;
|
||
}
|
||
}
|
||
return true;
|
||
}
|
||
|
||
bool KeyWordExt::_isContainSubWords(const string& word)
|
||
{
|
||
for(uint i = 0; i < _priorSubWords.size(); i++)
|
||
{
|
||
if(string::npos != word.find(_priorSubWords[i]))
|
||
{
|
||
return true;
|
||
}
|
||
}
|
||
return false;
|
||
}
|
||
|
||
bool KeyWordExt::_prioritizeSubWords(vector<WordInfo>& wordInfos)
|
||
{
|
||
if(2 > wordInfos.size())
|
||
{
|
||
return true;
|
||
}
|
||
|
||
WordInfo prior;
|
||
bool flag = false;
|
||
for(vector<WordInfo>::iterator it = wordInfos.begin(); it != wordInfos.end(); )
|
||
{
|
||
if(_isContainSubWords(it->word))
|
||
{
|
||
prior = *it;
|
||
it = wordInfos.erase(it);
|
||
flag = true;
|
||
break;
|
||
}
|
||
else
|
||
{
|
||
it ++;
|
||
}
|
||
}
|
||
if(flag)
|
||
{
|
||
wordInfos.insert(wordInfos.begin(), prior);
|
||
}
|
||
return true;
|
||
}
|
||
}
|
||
|
||
|
||
#ifdef KEYWORDEXT_UT
|
||
|
||
using namespace CppJieba;
|
||
|
||
int main()
|
||
{
|
||
KeyWordExt ext;
|
||
if(!ext.init("../dicts/segdict.utf8.v2.1"))
|
||
{
|
||
return 1;
|
||
}
|
||
ext.loadStopWords("stopwords.tmp");
|
||
|
||
if(!ext.loadPriorSubWords("prior.utf8"))
|
||
{
|
||
cerr<<"err"<<endl;
|
||
return 1;
|
||
}
|
||
//segment.init("dicts/jieba.dict.utf8");
|
||
|
||
vector<string> res;
|
||
string title;
|
||
/*title = "我来到北京清华大学";
|
||
res.clear();
|
||
ext.extract(title, res, 5);
|
||
PRINT_VECTOR(res);
|
||
|
||
|
||
title = "特价!camel骆驼 柔软舒适头层牛皮平底凉鞋女 休闲平跟妈妈鞋夏";
|
||
res.clear();
|
||
ext.extract(title, res, 5);
|
||
PRINT_VECTOR(res);
|
||
*/
|
||
|
||
|
||
title = "韩国MAGGIC 爆花石水晶飞机 锁骨链短款项链 女 进口韩国饰品";
|
||
res.clear();
|
||
ext.extract(title, res, 5);
|
||
PRINT_VECTOR(res);
|
||
|
||
ext.dispose();
|
||
return 0;
|
||
}
|
||
|
||
#endif
|