Merge pull request #582 from hosiet/pr-fix-typo-codespell

Fix typos found by codespell
This commit is contained in:
Sun Junyi 2018-09-20 10:44:47 +08:00 committed by GitHub
commit 843cdc2b7c
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
2 changed files with 3 additions and 3 deletions

View File

@ -164,7 +164,7 @@ class Tokenizer(object):
self.initialized = True self.initialized = True
default_logger.debug( default_logger.debug(
"Loading model cost %.3f seconds." % (time.time() - t1)) "Loading model cost %.3f seconds." % (time.time() - t1))
default_logger.debug("Prefix dict has been built succesfully.") default_logger.debug("Prefix dict has been built successfully.")
def check_initialized(self): def check_initialized(self):
if not self.initialized: if not self.initialized:
@ -275,7 +275,7 @@ class Tokenizer(object):
def cut(self, sentence, cut_all=False, HMM=True): def cut(self, sentence, cut_all=False, HMM=True):
''' '''
The main function that segments an entire sentence that contains The main function that segments an entire sentence that contains
Chinese characters into seperated words. Chinese characters into separated words.
Parameter: Parameter:
- sentence: The str(unicode) to be segmented. - sentence: The str(unicode) to be segmented.

View File

@ -44,7 +44,7 @@ GitHub: https://github.com/fxsjy/jieba
setup(name='jieba', setup(name='jieba',
version='0.39', version='0.39',
description='Chinese Words Segementation Utilities', description='Chinese Words Segmentation Utilities',
long_description=LONGDOC, long_description=LONGDOC,
author='Sun, Junyi', author='Sun, Junyi',
author_email='ccnusjy@gmail.com', author_email='ccnusjy@gmail.com',