re-common 10.0.24__tar.gz → 10.0.25__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {re_common-10.0.24/re_common.egg-info → re_common-10.0.25}/PKG-INFO +1 -1
- re_common-10.0.25/re_common/v2/baselibrary/tools/data_processer/base.py +53 -0
- re_common-10.0.25/re_common/v2/baselibrary/tools/data_processer/data_processer.py +508 -0
- re_common-10.0.25/re_common/v2/baselibrary/tools/data_processer/data_reader.py +187 -0
- re_common-10.0.25/re_common/v2/baselibrary/tools/data_processer/data_writer.py +38 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/baselibrary/tools/dict_tools.py +7 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/baselibrary/tools/list_tools.py +5 -1
- {re_common-10.0.24/re_common/v2/baselibrary/business_utils → re_common-10.0.25/re_common/v2/baselibrary/utils}/BusinessStringUtil.py +1 -0
- re_common-10.0.25/re_common/v2/baselibrary/utils/api_net_utils.py +270 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/baselibrary/utils/db.py +21 -2
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/baselibrary/utils/string_clear.py +6 -1
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/baselibrary/utils/stringutils.py +15 -0
- re_common-10.0.25/re_common/vip/proxy/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25/re_common.egg-info}/PKG-INFO +1 -1
- {re_common-10.0.24 → re_common-10.0.25}/re_common.egg-info/SOURCES.txt +6 -0
- {re_common-10.0.24 → re_common-10.0.25}/setup.py +1 -1
- {re_common-10.0.24 → re_common-10.0.25}/LICENSE +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/README.md +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/pyproject.toml +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/baseabs/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/baseabs/baseabs.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/database/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/database/mbuilder.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/database/moudle.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/database/msqlite3.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/database/mysql.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/database/sql_factory.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/mthread/MThreadingRun.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/mthread/MThreadingRunEvent.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/mthread/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/mthread/mythreading.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/pakge_other/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/pakge_other/socks.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/readconfig/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/readconfig/config_factory.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/readconfig/ini_config.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/readconfig/toml_config.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/temporary/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/temporary/envdata.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/all_requests/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/all_requests/aiohttp_request.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/all_requests/httpx_requet.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/all_requests/mrequest.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/all_requests/requests_request.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/batch_compre/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/batch_compre/bijiao_batch.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/contrast_db3.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/copy_file.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/db3_2_sizedb3.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/foreachgz.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/get_attr.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/image_to_pdf.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/java_code_deal.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/javacode.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/mdb_db3.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/merge_file.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/merge_gz_file.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/mhdfstools/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/mhdfstools/down_hdfs_files.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/mhdfstools/hdfst.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/mhdfstools/up_hdfs_files.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/mongo_tools.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/move_file.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/move_mongo/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/move_mongo/mongo_table_to_file.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/move_mongo/move_mongo_table.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/move_mongo/use_mttf.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/move_mongo/use_mv.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/mpandas/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/mpandas/mpandasreadexcel.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/mpandas/pandas_visualization.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/myparsel.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/rename_dir_file.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/sequoiadb_utils.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/split_line_to_many.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/stringtodicts.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/tools/workwechant_bot.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/baseaiohttp.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/baseaiomysql.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/baseallstep.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/baseavro.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/baseboto3.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/basecsv.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/basedict.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/basedir.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/baseencode.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/baseencoding.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/baseesdsl.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/baseexcel.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/baseexcept.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/basefile.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/baseftp.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/basegzip.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/basehdfs.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/basehttpx.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/baseip.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/basejson.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/baselist.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/basemotor.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/basemssql.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/baseodbc.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/basepandas.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/basepeewee.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/basepika.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/basepydash.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/basepymongo.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/basequeue.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/baserar.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/baserequest.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/baseset.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/basesmb.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/basestring.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/basetime.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/basetuple.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/baseurl.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/basezip.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/core/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/core/bottomutils.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/core/mdeprecated.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/core/mlamada.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/core/msginfo.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/core/requests_core.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/fateadm.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/importfun.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/mfaker.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/my_abc/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/my_abc/better_abc.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/mylogger.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/myredisclient.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/pipupgrade.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/ringlist.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/version_compare.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/baselibrary/utils/ydmhttp.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/facade/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/facade/lazy_import.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/facade/loggerfacade.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/facade/mysqlfacade.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/facade/now.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/facade/sqlite3facade.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/facade/use/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/facade/use/mq_use_facade.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/facade/use/proxy_use_facade.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/libtest/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/libtest/base_dict_test.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/libtest/baseavro_test.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/libtest/basefile_test.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/libtest/basemssql_test.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/libtest/baseodbc_test.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/libtest/basepandas_test.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/libtest/get_attr_test/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/libtest/get_attr_test/get_attr_test_settings.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/libtest/get_attr_test/settings.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/libtest/idencode_test.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/libtest/iniconfig_test.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/libtest/ip_test.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/libtest/merge_file_test.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/libtest/mfaker_test.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/libtest/mm3_test.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/libtest/mylogger_test.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/libtest/myparsel_test.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/libtest/mysql_test.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/libtest/pymongo_test.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/libtest/split_test.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/libtest/sqlite3_merge_test.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/libtest/sqlite3_test.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/libtest/tomlconfig_test.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/libtest/use_tools_test/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/libtest/user/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/studio/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/studio/assignment_expressions.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/studio/mydash/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/studio/mydash/test1.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/studio/pydashstudio/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/studio/pydashstudio/first.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/studio/streamlitstudio/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/studio/streamlitstudio/first_app.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/studio/streamlitstudio/uber_pickups.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/studio/test.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/baselibrary/__init__.py +0 -0
- {re_common-10.0.24/re_common/v2/baselibrary/utils → re_common-10.0.25/re_common/v2/baselibrary/business_utils}/BusinessStringUtil.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/baselibrary/business_utils/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/baselibrary/business_utils/rel_tools.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/baselibrary/decorators/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/baselibrary/decorators/utils.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/baselibrary/helpers/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/baselibrary/s3object/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/baselibrary/s3object/baseboto3.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/baselibrary/tools/WeChatRobot.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/baselibrary/tools/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/baselibrary/tools/ac_ahocorasick.py +0 -0
- {re_common-10.0.24/re_common/v2/baselibrary/utils → re_common-10.0.25/re_common/v2/baselibrary/tools/data_processer}/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/baselibrary/tools/dolphinscheduler.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/baselibrary/tools/hdfs_data_processer.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/baselibrary/tools/search_hash_tools.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/baselibrary/tools/text_matcher.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/baselibrary/tools/unionfind_tools.py +0 -0
- {re_common-10.0.24/re_common/vip → re_common-10.0.25/re_common/v2/baselibrary/utils}/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/baselibrary/utils/author_smi.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/baselibrary/utils/base_string_similarity.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/baselibrary/utils/basedict.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/baselibrary/utils/basehdfs.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/baselibrary/utils/basepika.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/baselibrary/utils/basetime.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/baselibrary/utils/json_cls.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/baselibrary/utils/mq.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/baselibrary/utils/n_ary_expression_tree.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/baselibrary/utils/string_bool.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/v2/baselibrary/utils/string_smi.py +0 -0
- {re_common-10.0.24/re_common/vip/proxy → re_common-10.0.25/re_common/vip}/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/vip/base_step_process.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/vip/baseencodeid.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/vip/changetaskname.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/vip/core_var.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/vip/mmh3Hash.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/vip/proxy/allproxys.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/vip/proxy/allproxys_thread.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/vip/proxy/cnki_proxy.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/vip/proxy/kuaidaili.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/vip/proxy/proxy_all.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/vip/proxy/update_kuaidaili_0.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/vip/proxy/wanfang_proxy.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/vip/proxy/wp_proxy_all.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/vip/read_rawid_to_txt.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/vip/title/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/vip/title/transform/TransformBookTitleToZt.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/vip/title/transform/TransformConferenceTitleToZt.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/vip/title/transform/TransformCstadTitleToZt.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/vip/title/transform/TransformJournalTitleToZt.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/vip/title/transform/TransformPatentTitleToZt.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/vip/title/transform/TransformRegulationTitleToZt.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/vip/title/transform/TransformStandardTitleToZt.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/vip/title/transform/TransformThesisTitleToZt.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common/vip/title/transform/__init__.py +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common.egg-info/dependency_links.txt +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/re_common.egg-info/top_level.txt +0 -0
- {re_common-10.0.24 → re_common-10.0.25}/setup.cfg +0 -0
|
@@ -0,0 +1,53 @@
|
|
|
1
|
+
from abc import ABC, abstractmethod
|
|
2
|
+
from typing import List, Generator
|
|
3
|
+
|
|
4
|
+
|
|
5
|
+
class BaseFileReader(ABC):
|
|
6
|
+
|
|
7
|
+
def __init__(self, batch_size: int = 10000):
|
|
8
|
+
self.batch_size = batch_size
|
|
9
|
+
self.read_model = 1
|
|
10
|
+
|
|
11
|
+
@abstractmethod
|
|
12
|
+
def list_files(self, path: str) -> List[str]:
|
|
13
|
+
"""列出路径下所有目标文件"""
|
|
14
|
+
pass
|
|
15
|
+
|
|
16
|
+
@abstractmethod
|
|
17
|
+
def count_lines(self, file_path: str) -> int:
|
|
18
|
+
"""统计文件行数"""
|
|
19
|
+
pass
|
|
20
|
+
|
|
21
|
+
@abstractmethod
|
|
22
|
+
def read_lines(self, file_path: str) -> Generator[List[str], None, None]:
|
|
23
|
+
"""读取文件内容,返回批量数据"""
|
|
24
|
+
pass
|
|
25
|
+
|
|
26
|
+
@abstractmethod
|
|
27
|
+
def read_all(self, file_path: str) -> List[List[str]]:
|
|
28
|
+
"""读取整个文件,默认按1000行分批"""
|
|
29
|
+
return [line for line in self.read_lines(file_path)]
|
|
30
|
+
|
|
31
|
+
def read_select(self, file_path: str) -> Generator[List[str], None, None]:
|
|
32
|
+
if self.read_model == 1:
|
|
33
|
+
for batch_data in self.read_lines(file_path):
|
|
34
|
+
yield batch_data
|
|
35
|
+
elif self.read_model == 2:
|
|
36
|
+
for batch_data in self.read_all(file_path):
|
|
37
|
+
yield batch_data
|
|
38
|
+
else:
|
|
39
|
+
raise Exception("模式选择错误")
|
|
40
|
+
|
|
41
|
+
|
|
42
|
+
class BaseFileWriter(ABC):
|
|
43
|
+
|
|
44
|
+
def __init__(self, file_path: str, compress: bool = True, overwrite: bool = True, encoding: str = "utf-8"):
|
|
45
|
+
self.file_path = file_path
|
|
46
|
+
self.compress = compress
|
|
47
|
+
self.encoding = encoding
|
|
48
|
+
self.overwrite = overwrite
|
|
49
|
+
|
|
50
|
+
@abstractmethod
|
|
51
|
+
def write_lines(self, lines: List[str], file_path: str):
|
|
52
|
+
"""写入多行文本到文件,支持压缩"""
|
|
53
|
+
pass
|
|
@@ -0,0 +1,508 @@
|
|
|
1
|
+
import asyncio
|
|
2
|
+
import os
|
|
3
|
+
import re
|
|
4
|
+
import sqlite3
|
|
5
|
+
import time
|
|
6
|
+
import traceback
|
|
7
|
+
from concurrent.futures import ProcessPoolExecutor
|
|
8
|
+
from pathlib import Path
|
|
9
|
+
from typing import List, Callable, Any
|
|
10
|
+
|
|
11
|
+
from filelock import FileLock
|
|
12
|
+
from tenacity import retry, stop_after_attempt, wait_exponential, wait_random, retry_if_result
|
|
13
|
+
|
|
14
|
+
from re_common.v2.baselibrary.tools.data_processer.base import BaseFileReader, BaseFileWriter
|
|
15
|
+
|
|
16
|
+
|
|
17
|
+
class DatabaseHandler:
|
|
18
|
+
def __init__(self, db_file="processed_files.db"):
|
|
19
|
+
self.db_file = db_file
|
|
20
|
+
self.lock_file = f"{self.db_file}.lock"
|
|
21
|
+
self._init_db()
|
|
22
|
+
|
|
23
|
+
def _init_db(self):
|
|
24
|
+
with FileLock(self.lock_file):
|
|
25
|
+
"""初始化 SQLite 数据库"""
|
|
26
|
+
with sqlite3.connect(self.db_file) as conn:
|
|
27
|
+
cursor = conn.cursor()
|
|
28
|
+
cursor.execute("""
|
|
29
|
+
CREATE TABLE IF NOT EXISTS processed_files (
|
|
30
|
+
file_path TEXT PRIMARY KEY
|
|
31
|
+
)
|
|
32
|
+
""")
|
|
33
|
+
conn.commit()
|
|
34
|
+
|
|
35
|
+
def save_processed_file(self, file_path):
|
|
36
|
+
"""保存处理过的文件"""
|
|
37
|
+
with FileLock(self.lock_file):
|
|
38
|
+
with sqlite3.connect(self.db_file) as conn:
|
|
39
|
+
cursor = conn.cursor()
|
|
40
|
+
cursor.execute(
|
|
41
|
+
"INSERT OR IGNORE INTO processed_files (file_path) VALUES (?)",
|
|
42
|
+
(file_path,)
|
|
43
|
+
)
|
|
44
|
+
conn.commit()
|
|
45
|
+
|
|
46
|
+
def save_processed_files_many(self, file_paths):
|
|
47
|
+
"""批量保存处理过的文件路径"""
|
|
48
|
+
if not file_paths:
|
|
49
|
+
return
|
|
50
|
+
with FileLock(self.lock_file):
|
|
51
|
+
with sqlite3.connect(self.db_file) as conn:
|
|
52
|
+
cursor = conn.cursor()
|
|
53
|
+
cursor.executemany(
|
|
54
|
+
"INSERT OR IGNORE INTO processed_files (file_path) VALUES (?)",
|
|
55
|
+
((fp,) for fp in file_paths)
|
|
56
|
+
)
|
|
57
|
+
conn.commit()
|
|
58
|
+
|
|
59
|
+
def is_file_processed(self, file_path):
|
|
60
|
+
"""检查文件是否已处理"""
|
|
61
|
+
with FileLock(self.lock_file):
|
|
62
|
+
with sqlite3.connect(self.db_file) as conn:
|
|
63
|
+
cursor = conn.cursor()
|
|
64
|
+
cursor.execute(
|
|
65
|
+
"SELECT file_path FROM processed_files WHERE file_path = ?",
|
|
66
|
+
(file_path,)
|
|
67
|
+
)
|
|
68
|
+
result = cursor.fetchone()
|
|
69
|
+
return result is not None
|
|
70
|
+
|
|
71
|
+
def fake_processed_files(self, start_index, end_index, file_list):
|
|
72
|
+
try:
|
|
73
|
+
# 将字符串序号转换为整数
|
|
74
|
+
start = int(start_index)
|
|
75
|
+
end = int(end_index)
|
|
76
|
+
|
|
77
|
+
# 验证序号范围
|
|
78
|
+
if start >= end:
|
|
79
|
+
raise ValueError(f"起始序号 {start_index} 必须小于结束序号 {end_index}")
|
|
80
|
+
|
|
81
|
+
list_formatted_num = []
|
|
82
|
+
# 为范围内的每个序号生成文件名
|
|
83
|
+
for num in range(start, end):
|
|
84
|
+
# 将序号格式化为5位字符串 (00120, 00121,...)
|
|
85
|
+
formatted_num = f"{num:05d}"
|
|
86
|
+
list_formatted_num.append(formatted_num)
|
|
87
|
+
|
|
88
|
+
skip_path_list = []
|
|
89
|
+
skip_formatted_num = []
|
|
90
|
+
for file_path in file_list:
|
|
91
|
+
re_f_num = re.findall(r'(?<!\d)\d{5}(?!\d)', str(Path(file_path).stem))
|
|
92
|
+
if re_f_num:
|
|
93
|
+
if re_f_num[0] in list_formatted_num:
|
|
94
|
+
skip_path_list.append(file_path)
|
|
95
|
+
skip_formatted_num.append(re_f_num[0])
|
|
96
|
+
|
|
97
|
+
for item_list in [skip_path_list[i:i + 2000] for i in range(0, len(skip_path_list), 2000)]:
|
|
98
|
+
self.save_processed_files_many(item_list)
|
|
99
|
+
for file_path in item_list:
|
|
100
|
+
print(f"伪造处理记录: {file_path}")
|
|
101
|
+
|
|
102
|
+
no_fil_num = set(list_formatted_num) - set(skip_formatted_num)
|
|
103
|
+
if len(no_fil_num) > 0:
|
|
104
|
+
print(f"没有对应num的文件,伪造失败数量为{len(no_fil_num)},样例:{list(no_fil_num)[:10]}")
|
|
105
|
+
print(f"成功伪造处理记录:序号 {start_index} 到 {end_index}(不含)的文件")
|
|
106
|
+
|
|
107
|
+
except ValueError as e:
|
|
108
|
+
print(f"错误: 序号格式无效 - {str(e)}")
|
|
109
|
+
except Exception as e:
|
|
110
|
+
print(f"伪造处理记录时出错: {str(e)}")
|
|
111
|
+
|
|
112
|
+
|
|
113
|
+
class DataProcessor:
|
|
114
|
+
def __init__(
|
|
115
|
+
self,
|
|
116
|
+
reader: BaseFileReader,
|
|
117
|
+
writer: BaseFileWriter = None,
|
|
118
|
+
db_handler: DatabaseHandler = None,
|
|
119
|
+
db_file="processed_files.db",
|
|
120
|
+
batch_size=50,
|
|
121
|
+
retry_limit=3,
|
|
122
|
+
):
|
|
123
|
+
self.reader = reader
|
|
124
|
+
self.writer = writer
|
|
125
|
+
self.db_file = db_file
|
|
126
|
+
self.batch_size = batch_size
|
|
127
|
+
self.retry_limit = retry_limit
|
|
128
|
+
self.db_handler = db_handler if db_handler else DatabaseHandler(db_file=db_file)
|
|
129
|
+
|
|
130
|
+
async def retry_process_data(self, data, process_func):
|
|
131
|
+
"""处理数据并执行处理函数"""
|
|
132
|
+
|
|
133
|
+
def on_retry(retry_state):
|
|
134
|
+
# 每次抛错进入该函数打印消息
|
|
135
|
+
print(
|
|
136
|
+
f"重试次数: {retry_state.attempt_number}/{self.retry_limit},数据内容: {retry_state.args[0]}\n"
|
|
137
|
+
f"异常信息: {retry_state.outcome.exception()}"
|
|
138
|
+
)
|
|
139
|
+
|
|
140
|
+
def on_retry_error(retry_state):
|
|
141
|
+
# 最后抛错后调用
|
|
142
|
+
original_exc = retry_state.outcome.exception()
|
|
143
|
+
raise RuntimeError(
|
|
144
|
+
f"处理数据失败,达到重试上限。data: {retry_state.args[0]}") from original_exc # 抛出的自定义异常中 保留 __process_func() 里的原始错误堆栈信息(traceback)
|
|
145
|
+
|
|
146
|
+
@retry(stop=stop_after_attempt(3),
|
|
147
|
+
wait=wait_exponential(multiplier=1, min=2, max=20),
|
|
148
|
+
before_sleep=on_retry, # 每次抛错后使用
|
|
149
|
+
retry_error_callback=on_retry_error, # 如果到最后都没有成功 抛错
|
|
150
|
+
reraise=True)
|
|
151
|
+
async def __process_func(_data):
|
|
152
|
+
return await process_func(_data)
|
|
153
|
+
|
|
154
|
+
return await __process_func(data)
|
|
155
|
+
|
|
156
|
+
async def process_file(self, hdfs_file_path, process_func, write_dir):
|
|
157
|
+
"""处理单个 gz 文件"""
|
|
158
|
+
total_lines = self.reader.count_lines(hdfs_file_path)
|
|
159
|
+
processed_lines = 0
|
|
160
|
+
start_time = time.time()
|
|
161
|
+
results = []
|
|
162
|
+
# # 这里根据不同的配置选用不同的读取文件的方法
|
|
163
|
+
for lines in self.reader.read_select(hdfs_file_path):
|
|
164
|
+
processing_start_time = time.time() # 记录本批处理开始时间
|
|
165
|
+
|
|
166
|
+
tasks = []
|
|
167
|
+
for line in lines:
|
|
168
|
+
# try:
|
|
169
|
+
# data = json.loads(line)
|
|
170
|
+
# tasks.append(self.retry_process_data(data, process_func))
|
|
171
|
+
# except json.JSONDecodeError as e:
|
|
172
|
+
# raise Exception(f"解析JSON失败: {e}, 行内容: {line.strip()}")
|
|
173
|
+
tasks.append(self.retry_process_data(line, process_func))
|
|
174
|
+
|
|
175
|
+
# await AsyncTaskPool(self.batch_size).run(tasks) # AsyncTaskPool 适用于一次提交所有任务, 限制并发数执行
|
|
176
|
+
results.extend(await asyncio.gather(*tasks))
|
|
177
|
+
|
|
178
|
+
processed_lines += len(lines)
|
|
179
|
+
|
|
180
|
+
elapsed_time = time.time() - start_time # 已用时间
|
|
181
|
+
processing_time = time.time() - processing_start_time # 本次处理时间
|
|
182
|
+
avg_processing_time = (
|
|
183
|
+
(elapsed_time * 1000) / processed_lines if processed_lines > 0 else float("inf")
|
|
184
|
+
) # 平均每条数据的处理时间(毫秒)
|
|
185
|
+
|
|
186
|
+
# 估算剩余时间
|
|
187
|
+
remaining_time = (
|
|
188
|
+
((avg_processing_time / 1000) * (total_lines - processed_lines))
|
|
189
|
+
if processed_lines > 0
|
|
190
|
+
else float("inf")
|
|
191
|
+
)
|
|
192
|
+
|
|
193
|
+
# 显示总进度信息
|
|
194
|
+
print(
|
|
195
|
+
f"文件: {hdfs_file_path} 总进度: {processed_lines}/{total_lines} 行 | "
|
|
196
|
+
f"已用时间: {elapsed_time:.2f}秒 | 本次处理时间: {processing_time:.2f}秒 | "
|
|
197
|
+
f"预估剩余时间: {remaining_time:.2f}秒 | 平均每条处理时间: {avg_processing_time:.2f}毫秒"
|
|
198
|
+
)
|
|
199
|
+
|
|
200
|
+
if write_dir is not None:
|
|
201
|
+
if not self.writer:
|
|
202
|
+
raise Exception("没有配置写数据的对象")
|
|
203
|
+
write_path = write_dir.rstrip("/") + f"/{Path(hdfs_file_path).stem}"
|
|
204
|
+
self.writer.write_lines([str(item) for item in results], write_path)
|
|
205
|
+
|
|
206
|
+
# 最终进度显示
|
|
207
|
+
final_elapsed_time = time.time() - start_time # 最终已用时间
|
|
208
|
+
print(
|
|
209
|
+
f"文件: {hdfs_file_path} 处理完成 | 总进度: {processed_lines}/{total_lines} 行 | "
|
|
210
|
+
f"总已用时间: {final_elapsed_time:.2f}秒 | "
|
|
211
|
+
f"平均每条处理时间: {(final_elapsed_time * 1000) / processed_lines:.2f}毫秒"
|
|
212
|
+
if processed_lines > 0
|
|
213
|
+
else "处理无数据"
|
|
214
|
+
)
|
|
215
|
+
|
|
216
|
+
self.db_handler.save_processed_file(hdfs_file_path) # 保存处理过的文件
|
|
217
|
+
|
|
218
|
+
async def retry_process_file(self, hdfs_file_path, process_func, write_dir):
|
|
219
|
+
"""带重试机制的文件处理"""
|
|
220
|
+
|
|
221
|
+
def on_retry(retry_state):
|
|
222
|
+
# 每次抛错进入该函数打印消息
|
|
223
|
+
exc = retry_state.outcome.exception()
|
|
224
|
+
tb = ''.join(traceback.format_exception(type(exc), exc, exc.__traceback__))
|
|
225
|
+
print(tb)
|
|
226
|
+
|
|
227
|
+
print(
|
|
228
|
+
f"处理文件 {retry_state.args[0]} 时发生错误: {exc},正在重试 {retry_state.attempt_number}/{self.retry_limit}")
|
|
229
|
+
|
|
230
|
+
def on_retry_error(retry_state):
|
|
231
|
+
# 最后抛错后调用
|
|
232
|
+
print(f"处理文件 {retry_state.args[0]} 失败,达到重试上限")
|
|
233
|
+
return False
|
|
234
|
+
|
|
235
|
+
@retry(stop=stop_after_attempt(3),
|
|
236
|
+
wait=wait_exponential(multiplier=1, min=2, max=20),
|
|
237
|
+
before_sleep=on_retry, # 每次抛错后使用
|
|
238
|
+
retry_error_callback=on_retry_error, # 如果最后没成功 返回 False
|
|
239
|
+
reraise=True)
|
|
240
|
+
async def __process_func(_hdfs_file_path, _process_func, _write_dir):
|
|
241
|
+
await self.process_file(_hdfs_file_path, _process_func, _write_dir)
|
|
242
|
+
return True # 成功处理后退出
|
|
243
|
+
|
|
244
|
+
return await __process_func(hdfs_file_path, process_func, write_dir)
|
|
245
|
+
|
|
246
|
+
def get_file_list(self, hdfs_dir):
|
|
247
|
+
# 获取所有任务文件
|
|
248
|
+
all_files = self.reader.list_files(hdfs_dir)
|
|
249
|
+
for file_path in all_files:
|
|
250
|
+
yield file_path
|
|
251
|
+
|
|
252
|
+
@retry(stop=stop_after_attempt(3),
|
|
253
|
+
wait=wait_random(min=10, max=30),
|
|
254
|
+
retry=retry_if_result(lambda result: not result), # 如果返回值是 False(失败),则重试 最后会抛出一个默认错误tenacity.RetryError:
|
|
255
|
+
reraise=True)
|
|
256
|
+
async def _batch_process_file(self, hdfs_file_path: str, process_func: Callable[[str], Any],
|
|
257
|
+
write_dir: str = None):
|
|
258
|
+
"""批量更新所有 gz 文件"""
|
|
259
|
+
# all_succeed = True
|
|
260
|
+
# for hdfs_file_path in self.get_file_list(hdfs_dir):
|
|
261
|
+
# if self.db_handler.is_file_processed(hdfs_file_path):
|
|
262
|
+
# print(f"跳过已处理文件: {hdfs_file_path}")
|
|
263
|
+
# continue # 如果文件已处理,跳过
|
|
264
|
+
# succeed = await self.retry_process_file(hdfs_file_path, process_func, write_dir) # 处理文件
|
|
265
|
+
# if succeed is False:
|
|
266
|
+
# all_succeed = False
|
|
267
|
+
#
|
|
268
|
+
# if all_succeed:
|
|
269
|
+
# # 处理完成后删除数据库文件
|
|
270
|
+
# try:
|
|
271
|
+
# if os.path.exists(self.db_file):
|
|
272
|
+
# os.remove(self.db_file)
|
|
273
|
+
# print(f"已删除断点重试文件: {self.db_file}")
|
|
274
|
+
# return True
|
|
275
|
+
# except Exception as e:
|
|
276
|
+
# print(f"删除断点重试文件失败: {e}")
|
|
277
|
+
# return False
|
|
278
|
+
if self.db_handler.is_file_processed(hdfs_file_path):
|
|
279
|
+
print(f"跳过已处理文件: {hdfs_file_path}")
|
|
280
|
+
return True # 如果文件已处理,跳过
|
|
281
|
+
succeed = await self.retry_process_file(hdfs_file_path, process_func, write_dir) # 处理文件
|
|
282
|
+
return succeed
|
|
283
|
+
|
|
284
|
+
async def process_file_bulk(self, hdfs_file_path, process_func, write_dir):
|
|
285
|
+
"""按批次处理单个文件,批量数据传递给处理函数"""
|
|
286
|
+
# 获取文件的数据总量
|
|
287
|
+
total_lines = self.reader.count_lines(hdfs_file_path)
|
|
288
|
+
processed_lines = 0
|
|
289
|
+
start_time = time.time()
|
|
290
|
+
|
|
291
|
+
results = []
|
|
292
|
+
tasks = []
|
|
293
|
+
# 这里根据不同的配置选用不同的读取文件的方法
|
|
294
|
+
for lines in self.reader.read_select(hdfs_file_path):
|
|
295
|
+
processing_start_time = time.time() # 记录本批处理开始时间
|
|
296
|
+
|
|
297
|
+
# batch_data = []
|
|
298
|
+
# for line in lines:
|
|
299
|
+
# try:
|
|
300
|
+
# data = json.loads(line)
|
|
301
|
+
# batch_data.append(data)
|
|
302
|
+
# except json.JSONDecodeError as e:
|
|
303
|
+
# raise Exception(f"解析JSON失败: {e}, 行内容: {line.strip()}")
|
|
304
|
+
|
|
305
|
+
# 处理读取到的批次数据
|
|
306
|
+
if lines:
|
|
307
|
+
tasks.append(process_func(lines)) # 将批次数据传递给处理函数并收集任务
|
|
308
|
+
processed_lines += len(lines) # 更新已处理行数
|
|
309
|
+
|
|
310
|
+
# 当积累的任务数量达到 batch_size 时并发处理所有任务
|
|
311
|
+
if len(tasks) >= self.batch_size:
|
|
312
|
+
results.extend(await asyncio.gather(*tasks))
|
|
313
|
+
elapsed_time = time.time() - start_time # 已用时间
|
|
314
|
+
processing_time = time.time() - processing_start_time # 本次处理时间
|
|
315
|
+
avg_processing_time = (
|
|
316
|
+
(elapsed_time * 1000) / processed_lines if processed_lines > 0 else float("inf")
|
|
317
|
+
) # 平均每条数据的处理时间(毫秒)
|
|
318
|
+
|
|
319
|
+
# 估算剩余时间
|
|
320
|
+
remaining_time = (
|
|
321
|
+
((avg_processing_time / 1000) * (total_lines - processed_lines))
|
|
322
|
+
if processed_lines > 0
|
|
323
|
+
else float("inf")
|
|
324
|
+
)
|
|
325
|
+
|
|
326
|
+
# 显示总进度信息
|
|
327
|
+
print(
|
|
328
|
+
f"文件: {hdfs_file_path} 总进度: {processed_lines}/{total_lines} 行 | "
|
|
329
|
+
f"已用时间: {elapsed_time:.2f}秒 | 本次处理时间: {processing_time:.2f}秒 | "
|
|
330
|
+
f"预估剩余时间: {remaining_time:.2f}秒 | 平均每条处理时间: {avg_processing_time:.2f}毫秒"
|
|
331
|
+
)
|
|
332
|
+
|
|
333
|
+
# 清空任务列表,准备下一批处理
|
|
334
|
+
tasks.clear()
|
|
335
|
+
# 处理剩余的任务
|
|
336
|
+
if tasks:
|
|
337
|
+
results.extend(await asyncio.gather(*tasks)) # 处理未达到 batch_size 的剩余任务
|
|
338
|
+
|
|
339
|
+
if write_dir is not None:
|
|
340
|
+
if not self.writer:
|
|
341
|
+
raise Exception("没有配置写数据的对象")
|
|
342
|
+
write_path = write_dir.rstrip("/") + f"/{Path(hdfs_file_path).stem}"
|
|
343
|
+
self.writer.write_lines([str(item) for items in results for item in items], write_path)
|
|
344
|
+
|
|
345
|
+
# 最终进度显示
|
|
346
|
+
final_elapsed_time = time.time() - start_time # 最终已用时间
|
|
347
|
+
print(
|
|
348
|
+
f"文件: {hdfs_file_path} 处理完成 | 总进度: {processed_lines}/{total_lines} 行 | "
|
|
349
|
+
f"总已用时间: {final_elapsed_time:.2f}秒 | "
|
|
350
|
+
f"平均每条处理时间: {(final_elapsed_time * 1000) / processed_lines:.2f}毫秒"
|
|
351
|
+
if processed_lines > 0
|
|
352
|
+
else "处理无数据"
|
|
353
|
+
)
|
|
354
|
+
|
|
355
|
+
self.db_handler.save_processed_file(hdfs_file_path)
|
|
356
|
+
|
|
357
|
+
async def retry_process_file_bulk(self, hdfs_file_path, process_func, write_dir):
|
|
358
|
+
"""带重试机制的批量文件处理"""
|
|
359
|
+
|
|
360
|
+
def on_retry(retry_state):
|
|
361
|
+
# 每次抛错进入该函数打印消息
|
|
362
|
+
exc = retry_state.outcome.exception()
|
|
363
|
+
tb = ''.join(traceback.format_exception(type(exc), exc, exc.__traceback__))
|
|
364
|
+
print(tb)
|
|
365
|
+
print(
|
|
366
|
+
f"处理文件 {retry_state.args[0]} 时发生错误: {exc},正在重试 {retry_state.attempt_number}/{self.retry_limit}")
|
|
367
|
+
|
|
368
|
+
def on_retry_error(retry_state):
|
|
369
|
+
# 最后抛错后调用
|
|
370
|
+
print(f"处理文件 {retry_state.args[0]} 失败,达到重试上限")
|
|
371
|
+
return False
|
|
372
|
+
|
|
373
|
+
@retry(stop=stop_after_attempt(3),
|
|
374
|
+
wait=wait_exponential(multiplier=1, min=2, max=20),
|
|
375
|
+
before_sleep=on_retry, # 每次抛错后使用
|
|
376
|
+
retry_error_callback=on_retry_error, # 如果最后没成功 返回 False
|
|
377
|
+
reraise=True)
|
|
378
|
+
async def __process_func(_hdfs_file_path, _process_func, write_dir):
|
|
379
|
+
await self.process_file_bulk(_hdfs_file_path, _process_func, write_dir)
|
|
380
|
+
return True # 成功处理后退出
|
|
381
|
+
|
|
382
|
+
return await __process_func(hdfs_file_path, process_func, write_dir)
|
|
383
|
+
|
|
384
|
+
async def batch_process_file(self, hdfs_dir: str, process_func: Callable[[List[str]], Any] | Callable[[str], Any],
|
|
385
|
+
write_dir: str = None, is_bulk: bool = False):
|
|
386
|
+
all_succeed = True
|
|
387
|
+
for hdfs_file_path in self.get_file_list(hdfs_dir):
|
|
388
|
+
if is_bulk:
|
|
389
|
+
succeed = await self._batch_process_file_bulk(hdfs_file_path, process_func, write_dir)
|
|
390
|
+
else:
|
|
391
|
+
succeed = await self._batch_process_file(hdfs_file_path, process_func, write_dir)
|
|
392
|
+
if succeed is False:
|
|
393
|
+
all_succeed = False
|
|
394
|
+
if all_succeed:
|
|
395
|
+
# 处理完成后删除数据库文件
|
|
396
|
+
try:
|
|
397
|
+
if os.path.exists(self.db_file):
|
|
398
|
+
os.remove(self.db_file)
|
|
399
|
+
print(f"已删除断点重试文件: {self.db_file}")
|
|
400
|
+
return True
|
|
401
|
+
except Exception as e:
|
|
402
|
+
print(f"删除断点重试文件失败: {e}")
|
|
403
|
+
return False
|
|
404
|
+
|
|
405
|
+
@retry(stop=stop_after_attempt(3),
|
|
406
|
+
wait=wait_random(min=10, max=30),
|
|
407
|
+
retry=retry_if_result(lambda result: not result), # 如果返回值是 False(失败),则重试 最后会抛出一个默认错误tenacity.RetryError:
|
|
408
|
+
reraise=True)
|
|
409
|
+
async def _batch_process_file_bulk(self, hdfs_file_path: str, process_func: Callable[[List[str]], Any],
|
|
410
|
+
write_dir: str = None):
|
|
411
|
+
"""批量处理 gz 文件中的数据"""
|
|
412
|
+
# 获取所有文件
|
|
413
|
+
# all_succeed = True
|
|
414
|
+
# for hdfs_file_path in self.get_file_list(hdfs_dir):
|
|
415
|
+
# # 查看是否跳过文件
|
|
416
|
+
# if self.db_handler.is_file_processed(hdfs_file_path):
|
|
417
|
+
# print(f"跳过已处理文件: {hdfs_file_path}")
|
|
418
|
+
# continue # 跳过已处理文件
|
|
419
|
+
# # 开始批量处理文件
|
|
420
|
+
# succeed = await self.retry_process_file_bulk(hdfs_file_path, process_func, write_dir)
|
|
421
|
+
# if succeed is False:
|
|
422
|
+
# all_succeed = False
|
|
423
|
+
#
|
|
424
|
+
# if all_succeed:
|
|
425
|
+
# # 处理完成后删除数据库文件
|
|
426
|
+
# try:
|
|
427
|
+
# if os.path.exists(self.db_file):
|
|
428
|
+
# os.remove(self.db_file)
|
|
429
|
+
# print(f"已删除断点重试文件: {self.db_file}")
|
|
430
|
+
# return True
|
|
431
|
+
# except Exception as e:
|
|
432
|
+
# print(f"删除断点重试文件失败: {e}")
|
|
433
|
+
# return False
|
|
434
|
+
# 查看是否跳过文件
|
|
435
|
+
if self.db_handler.is_file_processed(hdfs_file_path):
|
|
436
|
+
print(f"跳过已处理文件: {hdfs_file_path}")
|
|
437
|
+
return True # 跳过已处理文件
|
|
438
|
+
# 开始批量处理文件
|
|
439
|
+
succeed = await self.retry_process_file_bulk(hdfs_file_path, process_func, write_dir)
|
|
440
|
+
return succeed
|
|
441
|
+
|
|
442
|
+
|
|
443
|
+
# 全局变量,每个进程独立持有
|
|
444
|
+
_processor: DataProcessor | None = None
|
|
445
|
+
_process_func: Callable[[List[str]], Any] | Callable[[str], Any] | None = None
|
|
446
|
+
_process_args: dict
|
|
447
|
+
|
|
448
|
+
|
|
449
|
+
def get_data_processor_func(process_args):
|
|
450
|
+
_func_reader = process_args["reader_func"]
|
|
451
|
+
_reader_args = process_args["reader_kwargs"]
|
|
452
|
+
reader = _func_reader(**_reader_args)
|
|
453
|
+
writer = None
|
|
454
|
+
if process_args["is_writer"]:
|
|
455
|
+
_func_writer = process_args["writer_func"]
|
|
456
|
+
_writer_args = process_args["writer_kwargs"]
|
|
457
|
+
writer = _func_writer(**_writer_args)
|
|
458
|
+
|
|
459
|
+
data_kwargs = {
|
|
460
|
+
"reader": reader,
|
|
461
|
+
"writer": writer,
|
|
462
|
+
"db_file": process_args["db_file"]
|
|
463
|
+
}
|
|
464
|
+
if process_args.get("batch_size"):
|
|
465
|
+
data_kwargs["batch_size"] = process_args["batch_size"]
|
|
466
|
+
if process_args.get("retry_limit"):
|
|
467
|
+
data_kwargs["retry_limit"] = process_args["retry_limit"]
|
|
468
|
+
|
|
469
|
+
return DataProcessor(**data_kwargs)
|
|
470
|
+
|
|
471
|
+
|
|
472
|
+
def init_worker(process_func, process_args):
|
|
473
|
+
global _processor, _process_func, _process_args
|
|
474
|
+
_processor = get_data_processor_func(process_args)
|
|
475
|
+
_process_func = process_func
|
|
476
|
+
_process_args = process_args
|
|
477
|
+
|
|
478
|
+
_init_func = _process_args.get("init_work", None)
|
|
479
|
+
if _init_func:
|
|
480
|
+
_init_func()
|
|
481
|
+
|
|
482
|
+
_async_init_work = _process_args.get("async_init_work", None)
|
|
483
|
+
if _init_func:
|
|
484
|
+
asyncio.run(_async_init_work())
|
|
485
|
+
|
|
486
|
+
|
|
487
|
+
def worker(path_file):
|
|
488
|
+
if _process_args["is_bulk"]:
|
|
489
|
+
return asyncio.run(_processor._batch_process_file_bulk(path_file, _process_func, _process_args["write_dir"]))
|
|
490
|
+
else:
|
|
491
|
+
return asyncio.run(_processor._batch_process_file(path_file, _process_func, _process_args["write_dir"]))
|
|
492
|
+
|
|
493
|
+
|
|
494
|
+
def run_worker_many(hdfs_dir: str, process_func: Callable[[List[str]], Any] | Callable[[str], Any],
|
|
495
|
+
data_process_args: dict, max_workers=4):
|
|
496
|
+
processor = get_data_processor_func(data_process_args)
|
|
497
|
+
all_file = list(processor.get_file_list(hdfs_dir))
|
|
498
|
+
with ProcessPoolExecutor(
|
|
499
|
+
max_workers=max_workers,
|
|
500
|
+
initializer=init_worker,
|
|
501
|
+
initargs=(process_func, data_process_args)
|
|
502
|
+
) as executor:
|
|
503
|
+
# 提交任务并等待结果
|
|
504
|
+
results = executor.map(worker, all_file)
|
|
505
|
+
# 输出结果
|
|
506
|
+
for result in results:
|
|
507
|
+
if result:
|
|
508
|
+
print(result)
|