adjust code for limonp v0.5.5 to solve macro name conflicts

This commit is contained in:
yanyiwu 2016-02-19 16:15:23 +08:00
parent fc04cf750a
commit 3e28b4bcb1
4 changed files with 26 additions and 26 deletions

View File

@ -97,7 +97,7 @@ class DictTrie {
size_t lineno = 0; size_t lineno = 0;
for (size_t i = 0; i < files.size(); i++) { for (size_t i = 0; i < files.size(); i++) {
ifstream ifs(files[i].c_str()); ifstream ifs(files[i].c_str());
CHECK(ifs.is_open()) << "open " << files[i] << " failed"; XCHECK(ifs.is_open()) << "open " << files[i] << " failed";
string line; string line;
DictUnit node_info; DictUnit node_info;
vector<string> buf; vector<string> buf;
@ -118,7 +118,7 @@ class DictTrie {
} }
} }
} }
LOG(INFO) << "load userdicts " << filePaths << ", lines: " << lineno; XLOG(INFO) << "load userdicts " << filePaths << ", lines: " << lineno;
} }
bool MakeNodeInfo(DictUnit& node_info, bool MakeNodeInfo(DictUnit& node_info,
@ -126,7 +126,7 @@ class DictTrie {
double weight, double weight,
const string& tag) { const string& tag) {
if (!TransCode::Decode(word, node_info.word)) { if (!TransCode::Decode(word, node_info.word)) {
LOG(ERROR) << "Decode " << word << " failed."; XLOG(ERROR) << "Decode " << word << " failed.";
return false; return false;
} }
node_info.weight = weight; node_info.weight = weight;
@ -136,14 +136,14 @@ class DictTrie {
void LoadDict(const string& filePath) { void LoadDict(const string& filePath) {
ifstream ifs(filePath.c_str()); ifstream ifs(filePath.c_str());
CHECK(ifs.is_open()) << "open " << filePath << " failed."; XCHECK(ifs.is_open()) << "open " << filePath << " failed.";
string line; string line;
vector<string> buf; vector<string> buf;
DictUnit node_info; DictUnit node_info;
for (size_t lineno = 0; getline(ifs, line); lineno++) { for (size_t lineno = 0; getline(ifs, line); lineno++) {
Split(line, buf, " "); Split(line, buf, " ");
CHECK(buf.size() == DICT_COLUMN_NUM) << "split result illegal, line:" << line; XCHECK(buf.size() == DICT_COLUMN_NUM) << "split result illegal, line:" << line;
MakeNodeInfo(node_info, MakeNodeInfo(node_info,
buf[0], buf[0],
atof(buf[1].c_str()), atof(buf[1].c_str()),
@ -157,7 +157,7 @@ class DictTrie {
} }
void SetStaticWordWeights(UserWordWeightOption option) { void SetStaticWordWeights(UserWordWeightOption option) {
CHECK(!static_node_infos_.empty()); XCHECK(!static_node_infos_.empty());
vector<DictUnit> x = static_node_infos_; vector<DictUnit> x = static_node_infos_;
sort(x.begin(), x.end(), WeightCompare); sort(x.begin(), x.end(), WeightCompare);
min_weight_ = x[0].weight; min_weight_ = x[0].weight;

View File

@ -33,43 +33,43 @@ struct HMMModel {
} }
void LoadModel(const string& filePath) { void LoadModel(const string& filePath) {
ifstream ifile(filePath.c_str()); ifstream ifile(filePath.c_str());
CHECK(ifile.is_open()) << "open " << filePath << " failed"; XCHECK(ifile.is_open()) << "open " << filePath << " failed";
string line; string line;
vector<string> tmp; vector<string> tmp;
vector<string> tmp2; vector<string> tmp2;
//Load startProb //Load startProb
CHECK(GetLine(ifile, line)); XCHECK(GetLine(ifile, line));
Split(line, tmp, " "); Split(line, tmp, " ");
CHECK(tmp.size() == STATUS_SUM); XCHECK(tmp.size() == STATUS_SUM);
for (size_t j = 0; j< tmp.size(); j++) { for (size_t j = 0; j< tmp.size(); j++) {
startProb[j] = atof(tmp[j].c_str()); startProb[j] = atof(tmp[j].c_str());
} }
//Load transProb //Load transProb
for (size_t i = 0; i < STATUS_SUM; i++) { for (size_t i = 0; i < STATUS_SUM; i++) {
CHECK(GetLine(ifile, line)); XCHECK(GetLine(ifile, line));
Split(line, tmp, " "); Split(line, tmp, " ");
CHECK(tmp.size() == STATUS_SUM); XCHECK(tmp.size() == STATUS_SUM);
for (size_t j =0; j < STATUS_SUM; j++) { for (size_t j =0; j < STATUS_SUM; j++) {
transProb[i][j] = atof(tmp[j].c_str()); transProb[i][j] = atof(tmp[j].c_str());
} }
} }
//Load emitProbB //Load emitProbB
CHECK(GetLine(ifile, line)); XCHECK(GetLine(ifile, line));
CHECK(LoadEmitProb(line, emitProbB)); XCHECK(LoadEmitProb(line, emitProbB));
//Load emitProbE //Load emitProbE
CHECK(GetLine(ifile, line)); XCHECK(GetLine(ifile, line));
CHECK(LoadEmitProb(line, emitProbE)); XCHECK(LoadEmitProb(line, emitProbE));
//Load emitProbM //Load emitProbM
CHECK(GetLine(ifile, line)); XCHECK(GetLine(ifile, line));
CHECK(LoadEmitProb(line, emitProbM)); XCHECK(LoadEmitProb(line, emitProbM));
//Load emitProbS //Load emitProbS
CHECK(GetLine(ifile, line)); XCHECK(GetLine(ifile, line));
CHECK(LoadEmitProb(line, emitProbS)); XCHECK(LoadEmitProb(line, emitProbS));
} }
double GetEmitProb(const EmitProbMap* ptMp, Rune key, double GetEmitProb(const EmitProbMap* ptMp, Rune key,
double defVal)const { double defVal)const {
@ -102,11 +102,11 @@ struct HMMModel {
for (size_t i = 0; i < tmp.size(); i++) { for (size_t i = 0; i < tmp.size(); i++) {
Split(tmp[i], tmp2, ":"); Split(tmp[i], tmp2, ":");
if (2 != tmp2.size()) { if (2 != tmp2.size()) {
LOG(ERROR) << "emitProb illegal."; XLOG(ERROR) << "emitProb illegal.";
return false; return false;
} }
if (!TransCode::Decode(tmp2[0], unicode) || unicode.size() != 1) { if (!TransCode::Decode(tmp2[0], unicode) || unicode.size() != 1) {
LOG(ERROR) << "TransCode failed."; XLOG(ERROR) << "TransCode failed.";
return false; return false;
} }
mp[unicode[0]] = atof(tmp2[1].c_str()); mp[unicode[0]] = atof(tmp2[1].c_str());

View File

@ -86,7 +86,7 @@ class KeywordExtractor {
private: private:
void LoadIdfDict(const string& idfPath) { void LoadIdfDict(const string& idfPath) {
ifstream ifs(idfPath.c_str()); ifstream ifs(idfPath.c_str());
CHECK(ifs.is_open()) << "open " << idfPath << " failed"; XCHECK(ifs.is_open()) << "open " << idfPath << " failed";
string line ; string line ;
vector<string> buf; vector<string> buf;
double idf = 0.0; double idf = 0.0;
@ -95,12 +95,12 @@ class KeywordExtractor {
for (; getline(ifs, line); lineno++) { for (; getline(ifs, line); lineno++) {
buf.clear(); buf.clear();
if (line.empty()) { if (line.empty()) {
LOG(ERROR) << "lineno: " << lineno << " empty. skipped."; XLOG(ERROR) << "lineno: " << lineno << " empty. skipped.";
continue; continue;
} }
Split(line, buf, " "); Split(line, buf, " ");
if (buf.size() != 2) { if (buf.size() != 2) {
LOG(ERROR) << "line: " << line << ", lineno: " << lineno << " empty. skipped."; XLOG(ERROR) << "line: " << line << ", lineno: " << lineno << " empty. skipped.";
continue; continue;
} }
idf = atof(buf[1].c_str()); idf = atof(buf[1].c_str());
@ -115,7 +115,7 @@ class KeywordExtractor {
} }
void LoadStopWordDict(const string& filePath) { void LoadStopWordDict(const string& filePath) {
ifstream ifs(filePath.c_str()); ifstream ifs(filePath.c_str());
CHECK(ifs.is_open()) << "open " << filePath << " failed"; XCHECK(ifs.is_open()) << "open " << filePath << " failed";
string line ; string line ;
while (getline(ifs, line)) { while (getline(ifs, line)) {
stopWords_.insert(line); stopWords_.insert(line);

View File

@ -35,7 +35,7 @@ class PosTagger {
assert(dict != NULL); assert(dict != NULL);
for (vector<string>::iterator itr = CutRes.begin(); itr != CutRes.end(); ++itr) { for (vector<string>::iterator itr = CutRes.begin(); itr != CutRes.end(); ++itr) {
if (!TransCode::Decode(*itr, unico)) { if (!TransCode::Decode(*itr, unico)) {
LOG(ERROR) << "Decode failed."; XLOG(ERROR) << "Decode failed.";
return false; return false;
} }
tmp = dict->Find(unico.begin(), unico.end()); tmp = dict->Find(unico.begin(), unico.end());