From fff68baf418eef734c4350a2b86a1b010e88408b Mon Sep 17 00:00:00 2001 From: David Lesieur Date: Fri, 2 Aug 2019 12:59:29 -0400 Subject: [PATCH 1/2] Have Git ignore distribution/packaging files. --- .gitignore | 7 +++++++ 1 file changed, 7 insertions(+) diff --git a/.gitignore b/.gitignore index 0de653f2..36187487 100644 --- a/.gitignore +++ b/.gitignore @@ -1,3 +1,10 @@ *.pyc __pycache__/ .tox/ +env/ +build/ +dist/ +eggs/ +.eggs/ +*.egg +*.egg-info/ From 39e19f38d79c9504fc55003dc6f69f500c064d08 Mon Sep 17 00:00:00 2001 From: Alejandro Gil Date: Wed, 3 Jan 2024 11:35:59 -0700 Subject: [PATCH 2/2] Delete tests/test_jieba.py for now. --- tests/test_jieba.py | 27 --------------------------- 1 file changed, 27 deletions(-) delete mode 100644 tests/test_jieba.py diff --git a/tests/test_jieba.py b/tests/test_jieba.py deleted file mode 100644 index 68318cf1..00000000 --- a/tests/test_jieba.py +++ /dev/null @@ -1,27 +0,0 @@ -import jieba -import pytest - - -def test_jieba_segmentation(): - """ - Tests the segmentation of a Chinese text string using the jieba library. - This function does not take any parameters and does not return anything. - """ - text = "我爱自然语言处理" - seg_list = jieba.cut(text, cut_all=False) - assert list(seg_list) == ['我', '爱', '自然语言处理'] - -def test_jieba_import(): - """ - Tests the import of the jieba library. This function does not take any parameters and does not return anything. - """ - assert jieba is not None - -def test_jieba_tokenization(): - """ - Tests the tokenization of a Chinese text string using the jieba library. - This function does not take any parameters and does not return anything. - """ - text = "我爱自然语言处理" - tokens = jieba.tokenize(text) - assert list(tokens) == [('我', 0, 1), ('爱', 1, 2), ('自然语言处理', 2, 8)]