re-common 10.0.21__py3-none-any.whl → 10.0.22__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- re_common/baselibrary/__init__.py +4 -4
- re_common/baselibrary/baseabs/__init__.py +6 -6
- re_common/baselibrary/baseabs/baseabs.py +26 -26
- re_common/baselibrary/database/mbuilder.py +132 -132
- re_common/baselibrary/database/moudle.py +93 -93
- re_common/baselibrary/database/msqlite3.py +194 -194
- re_common/baselibrary/database/mysql.py +169 -169
- re_common/baselibrary/database/sql_factory.py +26 -26
- re_common/baselibrary/mthread/MThreadingRun.py +486 -486
- re_common/baselibrary/mthread/MThreadingRunEvent.py +349 -349
- re_common/baselibrary/mthread/__init__.py +2 -2
- re_common/baselibrary/mthread/mythreading.py +695 -695
- re_common/baselibrary/pakge_other/socks.py +404 -404
- re_common/baselibrary/readconfig/config_factory.py +18 -18
- re_common/baselibrary/readconfig/ini_config.py +317 -317
- re_common/baselibrary/readconfig/toml_config.py +49 -49
- re_common/baselibrary/temporary/envdata.py +36 -36
- re_common/baselibrary/tools/all_requests/aiohttp_request.py +118 -118
- re_common/baselibrary/tools/all_requests/httpx_requet.py +102 -102
- re_common/baselibrary/tools/all_requests/mrequest.py +412 -412
- re_common/baselibrary/tools/all_requests/requests_request.py +81 -81
- re_common/baselibrary/tools/batch_compre/bijiao_batch.py +31 -31
- re_common/baselibrary/tools/contrast_db3.py +123 -123
- re_common/baselibrary/tools/copy_file.py +39 -39
- re_common/baselibrary/tools/db3_2_sizedb3.py +102 -102
- re_common/baselibrary/tools/foreachgz.py +39 -39
- re_common/baselibrary/tools/get_attr.py +10 -10
- re_common/baselibrary/tools/image_to_pdf.py +61 -61
- re_common/baselibrary/tools/java_code_deal.py +139 -139
- re_common/baselibrary/tools/javacode.py +79 -79
- re_common/baselibrary/tools/mdb_db3.py +48 -48
- re_common/baselibrary/tools/merge_file.py +171 -171
- re_common/baselibrary/tools/merge_gz_file.py +165 -165
- re_common/baselibrary/tools/mhdfstools/down_hdfs_files.py +42 -42
- re_common/baselibrary/tools/mhdfstools/hdfst.py +42 -42
- re_common/baselibrary/tools/mhdfstools/up_hdfs_files.py +38 -38
- re_common/baselibrary/tools/mongo_tools.py +50 -50
- re_common/baselibrary/tools/move_file.py +170 -170
- re_common/baselibrary/tools/move_mongo/mongo_table_to_file.py +63 -63
- re_common/baselibrary/tools/move_mongo/move_mongo_table.py +354 -354
- re_common/baselibrary/tools/move_mongo/use_mttf.py +18 -18
- re_common/baselibrary/tools/move_mongo/use_mv.py +93 -93
- re_common/baselibrary/tools/mpandas/mpandasreadexcel.py +125 -125
- re_common/baselibrary/tools/mpandas/pandas_visualization.py +7 -7
- re_common/baselibrary/tools/myparsel.py +104 -104
- re_common/baselibrary/tools/rename_dir_file.py +37 -37
- re_common/baselibrary/tools/sequoiadb_utils.py +398 -398
- re_common/baselibrary/tools/split_line_to_many.py +25 -25
- re_common/baselibrary/tools/stringtodicts.py +33 -33
- re_common/baselibrary/tools/workwechant_bot.py +84 -84
- re_common/baselibrary/utils/baseaiohttp.py +296 -296
- re_common/baselibrary/utils/baseaiomysql.py +87 -87
- re_common/baselibrary/utils/baseallstep.py +191 -191
- re_common/baselibrary/utils/baseavro.py +19 -19
- re_common/baselibrary/utils/baseboto3.py +291 -291
- re_common/baselibrary/utils/basecsv.py +32 -32
- re_common/baselibrary/utils/basedict.py +133 -133
- re_common/baselibrary/utils/basedir.py +241 -241
- re_common/baselibrary/utils/baseencode.py +351 -351
- re_common/baselibrary/utils/baseencoding.py +28 -28
- re_common/baselibrary/utils/baseesdsl.py +86 -86
- re_common/baselibrary/utils/baseexcel.py +264 -264
- re_common/baselibrary/utils/baseexcept.py +109 -109
- re_common/baselibrary/utils/basefile.py +654 -654
- re_common/baselibrary/utils/baseftp.py +214 -214
- re_common/baselibrary/utils/basegzip.py +60 -60
- re_common/baselibrary/utils/basehdfs.py +135 -135
- re_common/baselibrary/utils/basehttpx.py +268 -268
- re_common/baselibrary/utils/baseip.py +87 -87
- re_common/baselibrary/utils/basejson.py +2 -2
- re_common/baselibrary/utils/baselist.py +32 -32
- re_common/baselibrary/utils/basemotor.py +190 -190
- re_common/baselibrary/utils/basemssql.py +98 -98
- re_common/baselibrary/utils/baseodbc.py +113 -113
- re_common/baselibrary/utils/basepandas.py +302 -302
- re_common/baselibrary/utils/basepeewee.py +11 -11
- re_common/baselibrary/utils/basepika.py +180 -180
- re_common/baselibrary/utils/basepydash.py +143 -143
- re_common/baselibrary/utils/basepymongo.py +230 -230
- re_common/baselibrary/utils/basequeue.py +22 -22
- re_common/baselibrary/utils/baserar.py +57 -57
- re_common/baselibrary/utils/baserequest.py +279 -279
- re_common/baselibrary/utils/baseset.py +8 -8
- re_common/baselibrary/utils/basesmb.py +403 -403
- re_common/baselibrary/utils/basestring.py +382 -382
- re_common/baselibrary/utils/basetime.py +320 -320
- re_common/baselibrary/utils/baseurl.py +121 -121
- re_common/baselibrary/utils/basezip.py +57 -57
- re_common/baselibrary/utils/core/__init__.py +7 -7
- re_common/baselibrary/utils/core/bottomutils.py +18 -18
- re_common/baselibrary/utils/core/mdeprecated.py +327 -327
- re_common/baselibrary/utils/core/mlamada.py +16 -16
- re_common/baselibrary/utils/core/msginfo.py +25 -25
- re_common/baselibrary/utils/core/requests_core.py +103 -103
- re_common/baselibrary/utils/fateadm.py +429 -429
- re_common/baselibrary/utils/importfun.py +123 -123
- re_common/baselibrary/utils/mfaker.py +57 -57
- re_common/baselibrary/utils/my_abc/__init__.py +3 -3
- re_common/baselibrary/utils/my_abc/better_abc.py +32 -32
- re_common/baselibrary/utils/mylogger.py +414 -414
- re_common/baselibrary/utils/myredisclient.py +861 -861
- re_common/baselibrary/utils/pipupgrade.py +21 -21
- re_common/baselibrary/utils/ringlist.py +85 -85
- re_common/baselibrary/utils/version_compare.py +36 -36
- re_common/baselibrary/utils/ydmhttp.py +126 -126
- re_common/facade/lazy_import.py +11 -11
- re_common/facade/loggerfacade.py +25 -25
- re_common/facade/mysqlfacade.py +467 -467
- re_common/facade/now.py +31 -31
- re_common/facade/sqlite3facade.py +257 -257
- re_common/facade/use/mq_use_facade.py +83 -83
- re_common/facade/use/proxy_use_facade.py +19 -19
- re_common/libtest/base_dict_test.py +19 -19
- re_common/libtest/baseavro_test.py +13 -13
- re_common/libtest/basefile_test.py +14 -14
- re_common/libtest/basemssql_test.py +77 -77
- re_common/libtest/baseodbc_test.py +7 -7
- re_common/libtest/basepandas_test.py +38 -38
- re_common/libtest/get_attr_test/get_attr_test_settings.py +14 -14
- re_common/libtest/get_attr_test/settings.py +54 -54
- re_common/libtest/idencode_test.py +53 -53
- re_common/libtest/iniconfig_test.py +35 -35
- re_common/libtest/ip_test.py +34 -34
- re_common/libtest/merge_file_test.py +20 -20
- re_common/libtest/mfaker_test.py +8 -8
- re_common/libtest/mm3_test.py +31 -31
- re_common/libtest/mylogger_test.py +88 -88
- re_common/libtest/myparsel_test.py +27 -27
- re_common/libtest/mysql_test.py +151 -151
- re_common/libtest/pymongo_test.py +21 -21
- re_common/libtest/split_test.py +11 -11
- re_common/libtest/sqlite3_merge_test.py +5 -5
- re_common/libtest/sqlite3_test.py +34 -34
- re_common/libtest/tomlconfig_test.py +30 -30
- re_common/libtest/use_tools_test/__init__.py +2 -2
- re_common/libtest/user/__init__.py +4 -4
- re_common/studio/__init__.py +4 -4
- re_common/studio/assignment_expressions.py +36 -36
- re_common/studio/mydash/test1.py +18 -18
- re_common/studio/pydashstudio/first.py +9 -9
- re_common/studio/streamlitstudio/first_app.py +65 -65
- re_common/studio/streamlitstudio/uber_pickups.py +23 -23
- re_common/studio/test.py +18 -18
- re_common/v2/baselibrary/decorators/utils.py +59 -59
- re_common/v2/baselibrary/s3object/baseboto3.py +230 -230
- re_common/v2/baselibrary/tools/WeChatRobot.py +79 -79
- re_common/v2/baselibrary/tools/ac_ahocorasick.py +75 -75
- re_common/v2/baselibrary/tools/dict_tools.py +37 -37
- re_common/v2/baselibrary/tools/dolphinscheduler.py +187 -187
- re_common/v2/baselibrary/tools/hdfs_data_processer.py +338 -338
- re_common/v2/baselibrary/tools/list_tools.py +65 -65
- re_common/v2/baselibrary/tools/search_hash_tools.py +54 -54
- re_common/v2/baselibrary/tools/text_matcher.py +326 -326
- re_common/v2/baselibrary/tools/unionfind_tools.py +60 -60
- re_common/v2/baselibrary/utils/BusinessStringUtil.py +196 -196
- re_common/v2/baselibrary/utils/author_smi.py +360 -360
- re_common/v2/baselibrary/utils/base_string_similarity.py +158 -158
- re_common/v2/baselibrary/utils/basedict.py +37 -37
- re_common/v2/baselibrary/utils/basehdfs.py +161 -161
- re_common/v2/baselibrary/utils/basepika.py +180 -180
- re_common/v2/baselibrary/utils/basetime.py +77 -77
- re_common/v2/baselibrary/utils/db.py +38 -38
- re_common/v2/baselibrary/utils/json_cls.py +16 -16
- re_common/v2/baselibrary/utils/mq.py +83 -83
- re_common/v2/baselibrary/utils/n_ary_expression_tree.py +243 -243
- re_common/v2/baselibrary/utils/string_bool.py +149 -149
- re_common/v2/baselibrary/utils/string_clear.py +204 -202
- re_common/v2/baselibrary/utils/string_smi.py +18 -18
- re_common/v2/baselibrary/utils/stringutils.py +213 -213
- re_common/vip/base_step_process.py +11 -11
- re_common/vip/baseencodeid.py +90 -90
- re_common/vip/changetaskname.py +28 -28
- re_common/vip/core_var.py +24 -24
- re_common/vip/mmh3Hash.py +89 -89
- re_common/vip/proxy/allproxys.py +127 -127
- re_common/vip/proxy/allproxys_thread.py +159 -159
- re_common/vip/proxy/cnki_proxy.py +153 -153
- re_common/vip/proxy/kuaidaili.py +87 -87
- re_common/vip/proxy/proxy_all.py +113 -113
- re_common/vip/proxy/update_kuaidaili_0.py +42 -42
- re_common/vip/proxy/wanfang_proxy.py +152 -152
- re_common/vip/proxy/wp_proxy_all.py +181 -181
- re_common/vip/read_rawid_to_txt.py +91 -91
- re_common/vip/title/__init__.py +5 -5
- re_common/vip/title/transform/TransformBookTitleToZt.py +125 -125
- re_common/vip/title/transform/TransformConferenceTitleToZt.py +139 -139
- re_common/vip/title/transform/TransformCstadTitleToZt.py +195 -195
- re_common/vip/title/transform/TransformJournalTitleToZt.py +203 -203
- re_common/vip/title/transform/TransformPatentTitleToZt.py +132 -132
- re_common/vip/title/transform/TransformRegulationTitleToZt.py +114 -114
- re_common/vip/title/transform/TransformStandardTitleToZt.py +135 -135
- re_common/vip/title/transform/TransformThesisTitleToZt.py +135 -135
- re_common/vip/title/transform/__init__.py +10 -10
- {re_common-10.0.21.dist-info → re_common-10.0.22.dist-info}/LICENSE +201 -201
- {re_common-10.0.21.dist-info → re_common-10.0.22.dist-info}/METADATA +16 -16
- re_common-10.0.22.dist-info/RECORD +227 -0
- {re_common-10.0.21.dist-info → re_common-10.0.22.dist-info}/WHEEL +1 -1
- re_common-10.0.21.dist-info/RECORD +0 -227
- {re_common-10.0.21.dist-info → re_common-10.0.22.dist-info}/top_level.txt +0 -0
|
@@ -1,66 +1,66 @@
|
|
|
1
|
-
import itertools
|
|
2
|
-
from typing import List, Any, Tuple
|
|
3
|
-
|
|
4
|
-
|
|
5
|
-
def check_no_duplicates_2d(lst_2d):
|
|
6
|
-
"""
|
|
7
|
-
检查二维列表的每一行是否无重复
|
|
8
|
-
如果有重复值 返回 False
|
|
9
|
-
如果没有重复 返回True
|
|
10
|
-
"""
|
|
11
|
-
for row in lst_2d:
|
|
12
|
-
# 将行转为集合,比较长度
|
|
13
|
-
if len(row) != len(set(row)):
|
|
14
|
-
return False
|
|
15
|
-
return True
|
|
16
|
-
|
|
17
|
-
|
|
18
|
-
def generate_cross_list_combinations(lists: List[List[Any]]) -> List[Tuple[Any, Any]]:
|
|
19
|
-
"""
|
|
20
|
-
生成不同列表间的所有两两组合(元组长度为2)
|
|
21
|
-
|
|
22
|
-
参数:
|
|
23
|
-
lists: 包含多个列表的列表,例如 [[1,2], ['a','b'], ['x','y']]
|
|
24
|
-
|
|
25
|
-
返回:
|
|
26
|
-
包含所有跨列表两两组合的列表,每个组合是一个元组
|
|
27
|
-
例如 [(1,'a'), (1,'b'), (2,'a'), ..., ('a','x'), ('a','y'), ...]
|
|
28
|
-
"""
|
|
29
|
-
combinations = []
|
|
30
|
-
for i in range(len(lists)):
|
|
31
|
-
for j in range(i + 1, len(lists)):
|
|
32
|
-
combinations.extend(itertools.product(lists[i], lists[j]))
|
|
33
|
-
return combinations
|
|
34
|
-
|
|
35
|
-
|
|
36
|
-
def filter_and_sort_by_smi(all_list, top_n=1000):
|
|
37
|
-
|
|
38
|
-
"""
|
|
39
|
-
要求 list 里面第一个是比较大小的数据 第二个是实际数据
|
|
40
|
-
"""
|
|
41
|
-
|
|
42
|
-
# 1. 去重:按 doc_id 去重,保留 smi 最大的记录
|
|
43
|
-
unique_dict = {}
|
|
44
|
-
for smi, doc_id in all_list:
|
|
45
|
-
if doc_id not in unique_dict or smi > unique_dict[doc_id][0]:
|
|
46
|
-
unique_dict[doc_id] = (smi, doc_id)
|
|
47
|
-
|
|
48
|
-
# 2. 转换为列表并排序
|
|
49
|
-
unique_list = sorted(unique_dict.values(), key=lambda x: x[0], reverse=True)
|
|
50
|
-
|
|
51
|
-
# 3. 取前 top_n 个
|
|
52
|
-
return unique_list[:top_n]
|
|
53
|
-
|
|
54
|
-
|
|
55
|
-
def list_to_dict(list_data,key_name):
|
|
56
|
-
# 使用 defaultdict 来处理重复 id
|
|
57
|
-
from collections import defaultdict
|
|
58
|
-
|
|
59
|
-
dict_data = defaultdict(list)
|
|
60
|
-
|
|
61
|
-
for item in list_data:
|
|
62
|
-
dict_data[item[key_name]].append(item)
|
|
63
|
-
|
|
64
|
-
# 将 defaultdict 转换成普通字典
|
|
65
|
-
dict_data = dict(dict_data)
|
|
1
|
+
import itertools
|
|
2
|
+
from typing import List, Any, Tuple
|
|
3
|
+
|
|
4
|
+
|
|
5
|
+
def check_no_duplicates_2d(lst_2d):
|
|
6
|
+
"""
|
|
7
|
+
检查二维列表的每一行是否无重复
|
|
8
|
+
如果有重复值 返回 False
|
|
9
|
+
如果没有重复 返回True
|
|
10
|
+
"""
|
|
11
|
+
for row in lst_2d:
|
|
12
|
+
# 将行转为集合,比较长度
|
|
13
|
+
if len(row) != len(set(row)):
|
|
14
|
+
return False
|
|
15
|
+
return True
|
|
16
|
+
|
|
17
|
+
|
|
18
|
+
def generate_cross_list_combinations(lists: List[List[Any]]) -> List[Tuple[Any, Any]]:
|
|
19
|
+
"""
|
|
20
|
+
生成不同列表间的所有两两组合(元组长度为2)
|
|
21
|
+
|
|
22
|
+
参数:
|
|
23
|
+
lists: 包含多个列表的列表,例如 [[1,2], ['a','b'], ['x','y']]
|
|
24
|
+
|
|
25
|
+
返回:
|
|
26
|
+
包含所有跨列表两两组合的列表,每个组合是一个元组
|
|
27
|
+
例如 [(1,'a'), (1,'b'), (2,'a'), ..., ('a','x'), ('a','y'), ...]
|
|
28
|
+
"""
|
|
29
|
+
combinations = []
|
|
30
|
+
for i in range(len(lists)):
|
|
31
|
+
for j in range(i + 1, len(lists)):
|
|
32
|
+
combinations.extend(itertools.product(lists[i], lists[j]))
|
|
33
|
+
return combinations
|
|
34
|
+
|
|
35
|
+
|
|
36
|
+
def filter_and_sort_by_smi(all_list, top_n=1000):
|
|
37
|
+
|
|
38
|
+
"""
|
|
39
|
+
要求 list 里面第一个是比较大小的数据 第二个是实际数据
|
|
40
|
+
"""
|
|
41
|
+
|
|
42
|
+
# 1. 去重:按 doc_id 去重,保留 smi 最大的记录
|
|
43
|
+
unique_dict = {}
|
|
44
|
+
for smi, doc_id in all_list:
|
|
45
|
+
if doc_id not in unique_dict or smi > unique_dict[doc_id][0]:
|
|
46
|
+
unique_dict[doc_id] = (smi, doc_id)
|
|
47
|
+
|
|
48
|
+
# 2. 转换为列表并排序
|
|
49
|
+
unique_list = sorted(unique_dict.values(), key=lambda x: x[0], reverse=True)
|
|
50
|
+
|
|
51
|
+
# 3. 取前 top_n 个
|
|
52
|
+
return unique_list[:top_n]
|
|
53
|
+
|
|
54
|
+
|
|
55
|
+
def list_to_dict(list_data,key_name):
|
|
56
|
+
# 使用 defaultdict 来处理重复 id
|
|
57
|
+
from collections import defaultdict
|
|
58
|
+
|
|
59
|
+
dict_data = defaultdict(list)
|
|
60
|
+
|
|
61
|
+
for item in list_data:
|
|
62
|
+
dict_data[item[key_name]].append(item)
|
|
63
|
+
|
|
64
|
+
# 将 defaultdict 转换成普通字典
|
|
65
|
+
dict_data = dict(dict_data)
|
|
66
66
|
return dict_data
|
|
@@ -1,54 +1,54 @@
|
|
|
1
|
-
from typing import List
|
|
2
|
-
|
|
3
|
-
import jieba
|
|
4
|
-
from datasketch import MinHash, minhash
|
|
5
|
-
|
|
6
|
-
from re_common.v2.baselibrary.decorators.utils import deprecated
|
|
7
|
-
from re_common.v2.baselibrary.utils.string_bool import is_single_cjk_char
|
|
8
|
-
|
|
9
|
-
@deprecated("请使用 TextMatcherV2 中的方法代替。")
|
|
10
|
-
def tokenize(text: str, stopwords=None) -> List[str]:
|
|
11
|
-
"""
|
|
12
|
-
分词并移除停用词
|
|
13
|
-
"""
|
|
14
|
-
if stopwords is None:
|
|
15
|
-
stopwords = []
|
|
16
|
-
words = jieba.lcut(text)
|
|
17
|
-
|
|
18
|
-
# 统计单字符数据 长度,防止结巴分词分不了的单词 将数据分为单个字符
|
|
19
|
-
|
|
20
|
-
# 这里为什么使用函数 而不是在推导式中兼容,主要是在一些 spark中 推导式的if 条件不遵循最短路径原则会将表达式当做一个整体算子
|
|
21
|
-
def is_singel_en(i):
|
|
22
|
-
if len(i) == 1 and not is_single_cjk_char(i):
|
|
23
|
-
return True
|
|
24
|
-
return False
|
|
25
|
-
|
|
26
|
-
one_char_size = len([i for i in words if is_singel_en(i)])
|
|
27
|
-
all_size = len(words)
|
|
28
|
-
# 如果单字符个数超过一定比例 就直接用空格分词
|
|
29
|
-
if all_size != 0 and one_char_size / all_size > 0.6:
|
|
30
|
-
words = [i for i in text.split() if i.strip()]
|
|
31
|
-
|
|
32
|
-
# 过滤停用词和空字符
|
|
33
|
-
words = [w for w in words if w not in stopwords and w.strip()]
|
|
34
|
-
return words
|
|
35
|
-
|
|
36
|
-
@deprecated("请使用 TextMatcherV2 中的方法代替。")
|
|
37
|
-
def create_minhash(words: List[str], num_perm=128) -> MinHash:
|
|
38
|
-
"""
|
|
39
|
-
为分词结果创建 MinHash
|
|
40
|
-
"""
|
|
41
|
-
minhash = MinHash(num_perm=num_perm)
|
|
42
|
-
for word in words:
|
|
43
|
-
minhash.update(word.encode("utf-8"))
|
|
44
|
-
return minhash
|
|
45
|
-
|
|
46
|
-
@deprecated("请使用 TextMatcherV2 中的方法代替。")
|
|
47
|
-
def get_str_minhash(title):
|
|
48
|
-
from re_common.v2.baselibrary.utils.string_clear import rel_clear
|
|
49
|
-
rel_title = rel_clear(title)
|
|
50
|
-
if not rel_title:
|
|
51
|
-
return ""
|
|
52
|
-
words = tokenize(rel_title)
|
|
53
|
-
minhash = create_minhash(words)
|
|
54
|
-
return minhash
|
|
1
|
+
from typing import List
|
|
2
|
+
|
|
3
|
+
import jieba
|
|
4
|
+
from datasketch import MinHash, minhash
|
|
5
|
+
|
|
6
|
+
from re_common.v2.baselibrary.decorators.utils import deprecated
|
|
7
|
+
from re_common.v2.baselibrary.utils.string_bool import is_single_cjk_char
|
|
8
|
+
|
|
9
|
+
@deprecated("请使用 TextMatcherV2 中的方法代替。")
|
|
10
|
+
def tokenize(text: str, stopwords=None) -> List[str]:
|
|
11
|
+
"""
|
|
12
|
+
分词并移除停用词
|
|
13
|
+
"""
|
|
14
|
+
if stopwords is None:
|
|
15
|
+
stopwords = []
|
|
16
|
+
words = jieba.lcut(text)
|
|
17
|
+
|
|
18
|
+
# 统计单字符数据 长度,防止结巴分词分不了的单词 将数据分为单个字符
|
|
19
|
+
|
|
20
|
+
# 这里为什么使用函数 而不是在推导式中兼容,主要是在一些 spark中 推导式的if 条件不遵循最短路径原则会将表达式当做一个整体算子
|
|
21
|
+
def is_singel_en(i):
|
|
22
|
+
if len(i) == 1 and not is_single_cjk_char(i):
|
|
23
|
+
return True
|
|
24
|
+
return False
|
|
25
|
+
|
|
26
|
+
one_char_size = len([i for i in words if is_singel_en(i)])
|
|
27
|
+
all_size = len(words)
|
|
28
|
+
# 如果单字符个数超过一定比例 就直接用空格分词
|
|
29
|
+
if all_size != 0 and one_char_size / all_size > 0.6:
|
|
30
|
+
words = [i for i in text.split() if i.strip()]
|
|
31
|
+
|
|
32
|
+
# 过滤停用词和空字符
|
|
33
|
+
words = [w for w in words if w not in stopwords and w.strip()]
|
|
34
|
+
return words
|
|
35
|
+
|
|
36
|
+
@deprecated("请使用 TextMatcherV2 中的方法代替。")
|
|
37
|
+
def create_minhash(words: List[str], num_perm=128) -> MinHash:
|
|
38
|
+
"""
|
|
39
|
+
为分词结果创建 MinHash
|
|
40
|
+
"""
|
|
41
|
+
minhash = MinHash(num_perm=num_perm)
|
|
42
|
+
for word in words:
|
|
43
|
+
minhash.update(word.encode("utf-8"))
|
|
44
|
+
return minhash
|
|
45
|
+
|
|
46
|
+
@deprecated("请使用 TextMatcherV2 中的方法代替。")
|
|
47
|
+
def get_str_minhash(title):
|
|
48
|
+
from re_common.v2.baselibrary.utils.string_clear import rel_clear
|
|
49
|
+
rel_title = rel_clear(title)
|
|
50
|
+
if not rel_title:
|
|
51
|
+
return ""
|
|
52
|
+
words = tokenize(rel_title)
|
|
53
|
+
minhash = create_minhash(words)
|
|
54
|
+
return minhash
|