Merge pull request #582 from hosiet/pr-fix-typo-codespell

Fix typos found by codespell
pull/463/merge
Sun Junyi 6 years ago committed by GitHub
commit 843cdc2b7c
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23

@ -164,7 +164,7 @@ class Tokenizer(object):
self.initialized = True
default_logger.debug(
"Loading model cost %.3f seconds." % (time.time() - t1))
default_logger.debug("Prefix dict has been built succesfully.")
default_logger.debug("Prefix dict has been built successfully.")
def check_initialized(self):
if not self.initialized:
@ -275,7 +275,7 @@ class Tokenizer(object):
def cut(self, sentence, cut_all=False, HMM=True):
'''
The main function that segments an entire sentence that contains
Chinese characters into seperated words.
Chinese characters into separated words.
Parameter:
- sentence: The str(unicode) to be segmented.

@ -44,7 +44,7 @@ GitHub: https://github.com/fxsjy/jieba
setup(name='jieba',
version='0.39',
description='Chinese Words Segementation Utilities',
description='Chinese Words Segmentation Utilities',
long_description=LONGDOC,
author='Sun, Junyi',
author_email='ccnusjy@gmail.com',

Loading…
Cancel
Save