mirror of
https://github.com/fxsjy/jieba.git
synced 2025-07-10 00:01:33 +08:00
Fix typos found by codespell
This commit is contained in:
parent
cb0de2973b
commit
17ef8abba3
@ -161,7 +161,7 @@ class Tokenizer(object):
|
|||||||
self.initialized = True
|
self.initialized = True
|
||||||
default_logger.debug(
|
default_logger.debug(
|
||||||
"Loading model cost %.3f seconds." % (time.time() - t1))
|
"Loading model cost %.3f seconds." % (time.time() - t1))
|
||||||
default_logger.debug("Prefix dict has been built succesfully.")
|
default_logger.debug("Prefix dict has been built successfully.")
|
||||||
|
|
||||||
def check_initialized(self):
|
def check_initialized(self):
|
||||||
if not self.initialized:
|
if not self.initialized:
|
||||||
@ -272,7 +272,7 @@ class Tokenizer(object):
|
|||||||
def cut(self, sentence, cut_all=False, HMM=True):
|
def cut(self, sentence, cut_all=False, HMM=True):
|
||||||
'''
|
'''
|
||||||
The main function that segments an entire sentence that contains
|
The main function that segments an entire sentence that contains
|
||||||
Chinese characters into seperated words.
|
Chinese characters into separated words.
|
||||||
|
|
||||||
Parameter:
|
Parameter:
|
||||||
- sentence: The str(unicode) to be segmented.
|
- sentence: The str(unicode) to be segmented.
|
||||||
|
2
setup.py
2
setup.py
@ -44,7 +44,7 @@ GitHub: https://github.com/fxsjy/jieba
|
|||||||
|
|
||||||
setup(name='jieba',
|
setup(name='jieba',
|
||||||
version='0.39',
|
version='0.39',
|
||||||
description='Chinese Words Segementation Utilities',
|
description='Chinese Words Segmentation Utilities',
|
||||||
long_description=LONGDOC,
|
long_description=LONGDOC,
|
||||||
author='Sun, Junyi',
|
author='Sun, Junyi',
|
||||||
author_email='ccnusjy@gmail.com',
|
author_email='ccnusjy@gmail.com',
|
||||||
|
Loading…
x
Reference in New Issue
Block a user