mdbq 3.6.3__tar.gz → 3.6.4__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (48) hide show
  1. {mdbq-3.6.3 → mdbq-3.6.4}/PKG-INFO +1 -1
  2. mdbq-3.6.4/mdbq/redis/getredis.py +291 -0
  3. mdbq-3.6.3/mdbq/redis/getredis.py → mdbq-3.6.4/mdbq/redis/getredis_bak20250131.py +114 -31
  4. mdbq-3.6.4/mdbq/redis/getredis_deepseek.py +235 -0
  5. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq.egg-info/PKG-INFO +1 -1
  6. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq.egg-info/SOURCES.txt +2 -0
  7. {mdbq-3.6.3 → mdbq-3.6.4}/setup.py +1 -1
  8. {mdbq-3.6.3 → mdbq-3.6.4}/README.txt +0 -0
  9. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/__init__.py +0 -0
  10. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/__version__.py +0 -0
  11. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/aggregation/__init__.py +0 -0
  12. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/aggregation/aggregation.py +0 -0
  13. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/aggregation/datashow.py +0 -0
  14. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/aggregation/optimize_data.py +0 -0
  15. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/aggregation/query_data.py +0 -0
  16. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/bdup/__init__.py +0 -0
  17. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/bdup/bdup.py +0 -0
  18. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/config/__init__.py +0 -0
  19. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/config/myconfig.py +0 -0
  20. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/config/products.py +0 -0
  21. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/config/set_support.py +0 -0
  22. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/dataframe/__init__.py +0 -0
  23. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/dataframe/converter.py +0 -0
  24. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/log/__init__.py +0 -0
  25. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/log/mylogger.py +0 -0
  26. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/mongo/__init__.py +0 -0
  27. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/mongo/mongo.py +0 -0
  28. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/mysql/__init__.py +0 -0
  29. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/mysql/mysql.py +0 -0
  30. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/mysql/recheck_mysql.py +0 -0
  31. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/mysql/s_query.py +0 -0
  32. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/mysql/year_month_day.py +0 -0
  33. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/other/__init__.py +0 -0
  34. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/other/download_sku_picture.py +0 -0
  35. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/other/porxy.py +0 -0
  36. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/other/pov_city.py +0 -0
  37. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/other/sku_picture.py +0 -0
  38. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/other/ua_sj.py +0 -0
  39. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/pbix/__init__.py +0 -0
  40. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/pbix/pbix_refresh.py +0 -0
  41. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/pbix/refresh_all.py +0 -0
  42. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/pbix/refresh_all_old.py +0 -0
  43. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/redis/__init__.py +0 -0
  44. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/spider/__init__.py +0 -0
  45. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq/spider/aikucun.py +0 -0
  46. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq.egg-info/dependency_links.txt +0 -0
  47. {mdbq-3.6.3 → mdbq-3.6.4}/mdbq.egg-info/top_level.txt +0 -0
  48. {mdbq-3.6.3 → mdbq-3.6.4}/setup.cfg +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: mdbq
3
- Version: 3.6.3
3
+ Version: 3.6.4
4
4
  Home-page: https://pypi.org/project/mdbq
5
5
  Author: xigua,
6
6
  Author-email: 2587125111@qq.com
@@ -0,0 +1,291 @@
1
+ # -*- coding: UTF-8 –*-
2
+ import os.path
3
+ import redis
4
+ import socket
5
+ from mdbq.mysql import s_query
6
+ from mdbq.config import myconfig
7
+ import pandas as pd
8
+ import json
9
+ import datetime
10
+ import threading
11
+ import logging
12
+
13
+
14
+ if socket.gethostname() == 'company' or socket.gethostname() == 'Mac2.local':
15
+ conf = myconfig.main()
16
+ conf_data = conf['Windows']['company']['mysql']['local']
17
+ username, password, host, port = conf_data['username'], conf_data['password'], conf_data['host'], conf_data['port']
18
+ redis_password = conf['Windows']['company']['redis']['local']['password']
19
+ elif socket.gethostname() == 'MacBookPro':
20
+ conf = myconfig.main()
21
+ conf_data = conf['Windows']['xigua_lx']['mysql']['local']
22
+ username, password, host, port = conf_data['username'], conf_data['password'], conf_data['host'], conf_data['port']
23
+ redis_password = conf['Windows']['company']['redis']['local']['password']
24
+ else:
25
+ conf = myconfig.main()
26
+ conf_data = conf['Windows']['xigua_lx']['mysql']['local']
27
+ username, password, host, port = conf_data['username'], conf_data['password'], conf_data['host'], conf_data['port']
28
+ redis_password = conf['Windows']['company']['redis']['local']['password'] # redis 使用本地数据,全部机子相同
29
+
30
+ logging.basicConfig(level=logging.INFO, format='%(asctime)s | %(levelname)s | %(message)s')
31
+
32
+ # 获取当前模块的日志记录器
33
+ logger = logging.getLogger(__name__)
34
+
35
+ # 创建一个文件处理器,用于将日志写入文件
36
+ file_handler = logging.FileHandler(os.path.join('/Users/xigua/Downloads', 'redis.log'))
37
+ file_handler.setLevel(logging.INFO) # 设置文件处理器的日志级别
38
+
39
+ # 创建一个日志格式器,并设置给文件处理器
40
+ formatter = logging.Formatter('[%(asctime)s] %(levelname)s %(message)s', datefmt='%Y-%m-%d %H:%M:%S')
41
+ file_handler.setFormatter(formatter)
42
+
43
+ # 将文件处理器添加到日志记录器
44
+ logger.addHandler(file_handler)
45
+
46
+
47
+ class RedisData(object):
48
+ def __init__(self, redis_engine, download, cache_ttl: int):
49
+ self.redis_engine = redis_engine # Redis 数据处理引擎
50
+ self.download = download # MySQL 数据处理引擎
51
+ self.cache_ttl = cache_ttl * 60 # 缓存过期时间(秒)
52
+
53
+ def get_from_mysql(
54
+ self,
55
+ db_name: str,
56
+ table_name: str,
57
+ set_year: bool,
58
+ start_date,
59
+ end_date
60
+ ) -> pd.DataFrame:
61
+ """
62
+ 从 MySQL 读取数据并返回 DataFrame
63
+
64
+ Args:
65
+ set_year: 表名是否包含年份后缀
66
+ """
67
+ dfs = []
68
+ if set_year:
69
+ current_year = datetime.datetime.today().year
70
+ for year in range(2024, current_year + 1):
71
+ df = self._fetch_table_data(
72
+ db_name, f"{table_name}_{year}", start_date, end_date
73
+ )
74
+ if df is not None:
75
+ dfs.append(df)
76
+ else:
77
+ df = self._fetch_table_data(db_name, table_name, start_date, end_date)
78
+ if df is not None:
79
+ dfs.append(df)
80
+
81
+ combined_df = pd.concat(dfs, ignore_index=True) if dfs else pd.DataFrame()
82
+ if combined_df.empty:
83
+ logger.info(f"警告: {db_name}.{table_name} 未读取到数据")
84
+ else:
85
+ combined_df = self._convert_date_columns(combined_df)
86
+ return combined_df
87
+
88
+ def get_from_redis(
89
+ self,
90
+ db_name: str,
91
+ table_name: str,
92
+ set_year: bool,
93
+ start_date,
94
+ end_date
95
+ ) -> pd.DataFrame:
96
+ """
97
+ 从 Redis 获取数据,若缓存过期/不完整则触发异步更新
98
+ """
99
+ start_dt = pd.to_datetime(start_date)
100
+ end_dt = pd.to_datetime(end_date)
101
+ cache_key = self._generate_cache_key(db_name, table_name, set_year)
102
+
103
+ # 尝试获取缓存元数据
104
+ try:
105
+ ttl = self.redis_engine.ttl(cache_key)
106
+ cache_data = self._fetch_redis_data(cache_key)
107
+ except Exception as e:
108
+ logger.info(f"Redis 连接异常: {e},直接访问 MySQL")
109
+ return self.get_from_mysql(db_name, table_name, set_year, start_date, end_date)
110
+
111
+ # 缓存失效处理逻辑
112
+ if ttl < 60 or cache_data.empty:
113
+ self._trigger_async_cache_update(
114
+ cache_key, db_name, table_name, set_year, start_date, end_date, cache_data
115
+ )
116
+ return self.get_from_mysql(db_name, table_name, set_year, start_date, end_date)
117
+
118
+ # 处理有效缓存数据
119
+ filtered_df = self._filter_by_date_range(cache_data, start_dt, end_dt)
120
+ if not filtered_df.empty:
121
+ return filtered_df
122
+
123
+ # 缓存数据不满足查询范围要求
124
+ self._trigger_async_cache_update(
125
+ cache_key, db_name, table_name, set_year, start_date, end_date, cache_data
126
+ )
127
+ return self.get_from_mysql(db_name, table_name, set_year, start_date, end_date)
128
+
129
+ def set_redis(
130
+ self,
131
+ cache_key: str,
132
+ db_name: str,
133
+ table_name: str,
134
+ set_year: bool,
135
+ start_date,
136
+ end_date,
137
+ existing_data: pd.DataFrame
138
+ ) -> pd.DataFrame:
139
+ """
140
+ 异步更新 Redis 缓存,合并新旧数据
141
+ """
142
+ try:
143
+ # 从 MySQL 获取新数据
144
+ new_data = self.get_from_mysql(db_name, table_name, set_year, start_date, end_date)
145
+ if new_data.empty:
146
+ return pd.DataFrame()
147
+
148
+ # 合并历史数据
149
+ combined_data = self._merge_data(new_data, existing_data)
150
+
151
+ # 序列化并存储到 Redis
152
+ serialized_data = self._serialize_data(combined_data)
153
+ self.redis_engine.set(cache_key, serialized_data)
154
+ self.redis_engine.expire(cache_key, self.cache_ttl)
155
+
156
+ logger.info(f"缓存更新 {cache_key} | 数据量: {len(combined_data)}")
157
+ return combined_data
158
+
159
+ except Exception as e:
160
+ logger.info(f"缓存更新失败: {cache_key} - {str(e)}")
161
+ return pd.DataFrame()
162
+
163
+ # Helper Methods ------------------------------------------------
164
+
165
+ def _fetch_table_data(
166
+ self,
167
+ db_name: str,
168
+ table_name: str,
169
+ start_date,
170
+ end_date
171
+ ) -> pd.DataFrame:
172
+ """封装 MySQL 数据获取逻辑"""
173
+ try:
174
+ return self.download.data_to_df(
175
+ db_name=db_name,
176
+ table_name=table_name,
177
+ start_date=start_date,
178
+ end_date=end_date,
179
+ projection={}
180
+ )
181
+ except Exception as e:
182
+ logger.info(f"MySQL 查询异常 {db_name}.{table_name}: {e}")
183
+ return pd.DataFrame()
184
+
185
+ def _fetch_redis_data(self, cache_key: str) -> pd.DataFrame:
186
+ """从 Redis 获取并解析数据(自动转换日期列)"""
187
+ try:
188
+ data = self.redis_engine.get(cache_key)
189
+ if not data:
190
+ return pd.DataFrame()
191
+ # 反序列化数据
192
+ df = pd.DataFrame(json.loads(data.decode("utf-8")))
193
+ return self._convert_date_columns(df)
194
+ except Exception as e:
195
+ logger.info(f"Redis 数据解析失败 {cache_key}: {e}")
196
+ return pd.DataFrame()
197
+
198
+ def _convert_date_columns(self, df: pd.DataFrame) -> pd.DataFrame:
199
+ """统一处理日期列转换"""
200
+ if "日期" in df.columns:
201
+ df["日期"] = pd.to_datetime(df["日期"], format="%Y-%m-%d", errors="coerce")
202
+ return df
203
+
204
+ def _generate_cache_key(self, db_name: str, table_name: str, set_year: bool) -> str:
205
+ """生成标准化的缓存键"""
206
+ return f"{db_name}:{table_name}_haveyear" if set_year else f"{db_name}:{table_name}"
207
+
208
+ def _filter_by_date_range(
209
+ self,
210
+ df: pd.DataFrame,
211
+ start_dt: datetime.datetime,
212
+ end_dt: datetime.datetime
213
+ ) -> pd.DataFrame:
214
+ """按日期范围筛选数据"""
215
+ if "日期" not in df.columns:
216
+ return df
217
+ date_mask = (df["日期"] >= start_dt) & (df["日期"] <= end_dt)
218
+ return df[date_mask].copy()
219
+
220
+ def _trigger_async_cache_update(
221
+ self,
222
+ cache_key: str,
223
+ db_name: str,
224
+ table_name: str,
225
+ set_year: bool,
226
+ start_date: str,
227
+ end_date: str,
228
+ existing_data: pd.DataFrame
229
+ ):
230
+ """启动异步缓存更新线程"""
231
+ thread = threading.Thread(
232
+ target=self.set_redis,
233
+ args=(cache_key, db_name, table_name, set_year, start_date, end_date, existing_data),
234
+ daemon=True
235
+ )
236
+ thread.start()
237
+
238
+ def _merge_data(self, new_data: pd.DataFrame, existing_data: pd.DataFrame) -> pd.DataFrame:
239
+ """合并新旧数据集"""
240
+ if existing_data.empty or "日期" not in existing_data.columns:
241
+ return new_data
242
+
243
+ new_min = new_data["日期"].min()
244
+ new_max = new_data["日期"].max()
245
+ valid_historical = existing_data[
246
+ (existing_data["日期"] < new_min) | (existing_data["日期"] > new_max)
247
+ ]
248
+ return pd.concat([new_data, valid_historical], ignore_index=True).drop_duplicates(subset=["日期"])
249
+
250
+ def _serialize_data(self, df: pd.DataFrame) -> str:
251
+ """序列化 DataFrame 并处理日期类型"""
252
+ temp_df = df.copy()
253
+ date_cols = temp_df.select_dtypes(include=["datetime64[ns]"]).columns
254
+ for col in date_cols:
255
+ temp_df[col] = temp_df[col].dt.strftime("%Y-%m-%d")
256
+ return temp_df.to_json(orient="records", force_ascii=False)
257
+
258
+
259
+ if __name__ == '__main__':
260
+ # # ****************************************************
261
+ # # 这一部分在外部定义,只需要定义一次,开始
262
+ # redis_config = {
263
+ # 'host': '127.0.0.1',
264
+ # 'port': 6379, # 默认Redis端口
265
+ # 'db': 0, # 默认Redis数据库索引
266
+ # # 'username': 'default',
267
+ # 'password': redis_password,
268
+ # }
269
+ # # redis 实例化
270
+ # r = redis.Redis(**redis_config)
271
+ # # mysql 实例化
272
+ # d = s_query.QueryDatas(username=username, password=password, host=host, port=port)
273
+ # # 将两个库的实例化对象传给 RedisData 类,并实例化数据处理引擎
274
+ # m = RedisData(redis_engin=r, download=d)
275
+ # # ****************************************************
276
+ #
277
+ # # 以下为动态获取数据库数据
278
+ # db_name = '聚合数据'
279
+ # table_name = '多店推广场景_按日聚合'
280
+ # set_year = False
281
+ # df = m.get_from_redis(
282
+ # db_name=db_name,
283
+ # table_name=table_name,
284
+ # set_year=set_year,
285
+ # start_date='2025-01-01',
286
+ # end_date='2025-01-31'
287
+ # )
288
+ # logger.info(df)
289
+ #
290
+
291
+ logger.info(socket.gethostname())
@@ -14,6 +14,11 @@ if socket.gethostname() == 'company' or socket.gethostname() == 'Mac2.local':
14
14
  conf_data = conf['Windows']['company']['mysql']['local']
15
15
  username, password, host, port = conf_data['username'], conf_data['password'], conf_data['host'], conf_data['port']
16
16
  redis_password = conf['Windows']['company']['redis']['local']['password']
17
+ # elif socket.gethostname() == 'MacBook-Pro.local':
18
+ # conf = myconfig.main()
19
+ # conf_data = conf['Windows']['xigua_lx']['mysql']['local']
20
+ # username, password, host, port = conf_data['username'], conf_data['password'], conf_data['host'], conf_data['port']
21
+ # redis_password = conf['Windows']['company']['redis']['local']['password']
17
22
  else:
18
23
  conf = myconfig.main()
19
24
  conf_data = conf['Windows']['xigua_lx']['mysql']['local']
@@ -101,7 +106,7 @@ class RedisData(object):
101
106
  # Bug: 如果外部请求日期小于 table 最小日期,每次都要从 mysql 获取数据,即使 redis 缓存了数据
102
107
  if start_date < min_date: # 外部请求日期小于 redis 数据最小日期
103
108
  # 3. redis 有数据但数据不完整时
104
- print(f'{start_date} -- {min_date} 数据日期不对劲需要更新')
109
+ print(f'{start_date} -- {min_date} 数据日期需要更新')
105
110
  thread = threading.Thread(target=self.set_redis, args=(my_key, _db_name, _table_name, _set_year, start_date, end_date, _df))
106
111
  thread.start()
107
112
  # _df = self.set_redis(my_key=my_key, _db_name=_db_name, _table_name=_table_name, _set_year=_set_year, start_date=start_date, end_date=end_date)
@@ -112,6 +117,82 @@ class RedisData(object):
112
117
  return _df
113
118
 
114
119
  def set_redis(self, my_key, _db_name, _table_name, _set_year, start_date, end_date, before_df):
120
+ """
121
+ 从MySQL读取数据并存储到Redis(异步执行)
122
+
123
+ Args:
124
+ my_key: Redis存储键名
125
+ _db_name: 数据库名称
126
+ _table_name: 数据表名称
127
+ _set_year: 数据集年份
128
+ start_date: 查询开始日期
129
+ end_date: 查询结束日期
130
+ before_df: 合并用的历史数据
131
+
132
+ Returns:
133
+ pd.DataFrame: 处理后的数据集(含历史数据合并)
134
+ """
135
+ # 异常处理容器
136
+ datetime_cols = []
137
+
138
+ try:
139
+ # 从MySQL获取数据
140
+ _df = self.get_from_mysql(
141
+ _db_name=_db_name,
142
+ _table_name=_table_name,
143
+ start_date=start_date,
144
+ end_date=end_date,
145
+ _set_year=_set_year
146
+ )
147
+
148
+ # 日期列处理(当新旧数据都存在日期列时)
149
+ if '日期' in _df.columns and '日期' in before_df.columns:
150
+ # 获取当前数据时间范围
151
+ _min_date, _max_date = _df['日期'].min(), _df['日期'].max()
152
+
153
+ # 筛选需要保留的历史数据
154
+ mask = (before_df['日期'] < _min_date) | (before_df['日期'] > _max_date)
155
+ valid_history = before_df[mask]
156
+
157
+ # 合并数据
158
+ _df = pd.concat([_df, valid_history], ignore_index=True, axis=0)
159
+ _df.drop_duplicates(subset='日期', keep='first', inplace=True) # 可选去重
160
+
161
+ # 预处理时间类型转换
162
+ datetime_cols = _df.select_dtypes(include=['datetime64[ns]']).columns.tolist()
163
+ if datetime_cols:
164
+ _df[datetime_cols] = _df[datetime_cols].astype(str)
165
+
166
+ # 空数据检查
167
+ if _df.empty:
168
+ print(f'Warning: {_table_name} 空数据集,跳过Redis存储')
169
+ return pd.DataFrame()
170
+
171
+ # Redis存储操作
172
+ self.redis_engin.set(my_key, _df.to_json(orient='records', force_ascii=False))
173
+ self.redis_engin.expire(my_key, self.minute * 60)
174
+
175
+ # 恢复时间类型(返回用)
176
+ if datetime_cols:
177
+ _df[datetime_cols] = _df[datetime_cols].apply(pd.to_datetime, errors='coerce')
178
+
179
+ # 记录操作日志
180
+ print(f'{datetime.datetime.now().strftime("%Y-%m-%d %H:%M:%S")}'
181
+ f' | 刷新Redis {_db_name}:{_table_name}'
182
+ f' | 数据量:{len(_df)}行')
183
+
184
+ except Exception as e:
185
+ print(f'Error: {_table_name} 数据处理失败 - {str(e)}')
186
+ _df = pd.DataFrame()
187
+
188
+ finally:
189
+ # 确保返回前恢复时间类型
190
+ if datetime_cols and not _df.empty:
191
+ _df[datetime_cols] = _df[datetime_cols].apply(pd.to_datetime, errors='ignore')
192
+
193
+ return _df
194
+
195
+ def set_redis_bak(self, my_key, _db_name, _table_name, _set_year, start_date, end_date, before_df):
115
196
  """
116
197
  从 mysql 读取数据并存储 redis
117
198
  由于这个函数是异步执行的,从页面段首次加载数据时,可能返回空,等待异步执行结束后会正常返回数据
@@ -149,34 +230,36 @@ class RedisData(object):
149
230
  print(f'{now}: 刷新 redis -> {_db_name}:{_table_name}')
150
231
  return _df
151
232
 
152
-
153
233
  if __name__ == '__main__':
154
- # ****************************************************
155
- # 这一部分在外部定义,只需要定义一次,开始
156
- redis_config = {
157
- 'host': '127.0.0.1',
158
- 'port': 6379, # 默认Redis端口
159
- 'db': 0, # 默认Redis数据库索引
160
- # 'username': 'default',
161
- 'password': redis_password,
162
- }
163
- # redis 实例化
164
- r = redis.Redis(**redis_config)
165
- # mysql 实例化
166
- d = s_query.QueryDatas(username=username, password=password, host=host, port=port)
167
- # 将两个库的实例化对象传给 RedisData 类,并实例化数据处理引擎
168
- m = RedisData(redis_engin=r, download=d)
169
- # ****************************************************
170
-
171
- # 以下为动态获取数据库数据
172
- db_name = '聚合数据'
173
- table_name = '多店推广场景_按日聚合'
174
- set_year = False
175
- df = m.get_from_redis(
176
- _db_name=db_name,
177
- _table_name=table_name,
178
- _set_year=set_year,
179
- start_date='2025-01-01',
180
- end_date='2025-01-31'
181
- )
182
- print(df)
234
+ # # ****************************************************
235
+ # # 这一部分在外部定义,只需要定义一次,开始
236
+ # redis_config = {
237
+ # 'host': '127.0.0.1',
238
+ # 'port': 6379, # 默认Redis端口
239
+ # 'db': 0, # 默认Redis数据库索引
240
+ # # 'username': 'default',
241
+ # 'password': redis_password,
242
+ # }
243
+ # # redis 实例化
244
+ # r = redis.Redis(**redis_config)
245
+ # # mysql 实例化
246
+ # d = s_query.QueryDatas(username=username, password=password, host=host, port=port)
247
+ # # 将两个库的实例化对象传给 RedisData 类,并实例化数据处理引擎
248
+ # m = RedisData(redis_engin=r, download=d)
249
+ # # ****************************************************
250
+ #
251
+ # # 以下为动态获取数据库数据
252
+ # db_name = '聚合数据'
253
+ # table_name = '多店推广场景_按日聚合'
254
+ # set_year = False
255
+ # df = m.get_from_redis(
256
+ # _db_name=db_name,
257
+ # _table_name=table_name,
258
+ # _set_year=set_year,
259
+ # start_date='2025-01-01',
260
+ # end_date='2025-01-31'
261
+ # )
262
+ # print(df)
263
+ #
264
+
265
+ print(socket.gethostname())
@@ -0,0 +1,235 @@
1
+ # -*- coding: UTF-8 –*-
2
+ import redis
3
+ import socket
4
+ from mdbq.mysql import s_query
5
+ from mdbq.config import myconfig
6
+ import pandas as pd
7
+ import json
8
+ import datetime
9
+ import threading
10
+
11
+
12
+ if socket.gethostname() == 'company' or socket.gethostname() == 'Mac2.local':
13
+ conf = myconfig.main()
14
+ conf_data = conf['Windows']['company']['mysql']['local']
15
+ username, password, host, port = conf_data['username'], conf_data['password'], conf_data['host'], conf_data['port']
16
+ redis_password = conf['Windows']['company']['redis']['local']['password']
17
+ # elif socket.gethostname() == 'MacBook-Pro.local':
18
+ # conf = myconfig.main()
19
+ # conf_data = conf['Windows']['xigua_lx']['mysql']['local']
20
+ # username, password, host, port = conf_data['username'], conf_data['password'], conf_data['host'], conf_data['port']
21
+ # redis_password = conf['Windows']['company']['redis']['local']['password']
22
+ else:
23
+ conf = myconfig.main()
24
+ conf_data = conf['Windows']['xigua_lx']['mysql']['local']
25
+ username, password, host, port = conf_data['username'], conf_data['password'], conf_data['host'], conf_data['port']
26
+ redis_password = conf['Windows']['company']['redis']['local']['password'] # redis 使用本地数据,全部机子相同
27
+
28
+
29
+ class RedisData:
30
+ def __init__(self, redis_engine, download, cache_ttl: int = 60):
31
+ """
32
+ 初始化Redis数据处理对象
33
+
34
+ :param redis_engine: Redis连接引擎
35
+ :param download: 数据库下载处理器
36
+ :param cache_ttl: 缓存过期时间(分钟)
37
+ """
38
+ self.redis_engine = redis_engine
39
+ self.download = download
40
+ self.cache_ttl = cache_ttl * 60 # 转换为秒
41
+ self.lock = threading.Lock() # 线程锁
42
+
43
+ def _handle_datetime_columns(self, df: pd.DataFrame) -> pd.DataFrame:
44
+ """统一处理日期列转换"""
45
+ if '日期' in df.columns:
46
+ df['日期'] = pd.to_datetime(df['日期'], errors='coerce', format='%Y-%m-%d')
47
+ return df
48
+
49
+ def get_from_mysql(self, db_name: str, table_name: str,
50
+ start_date, end_date,
51
+ set_year: bool) -> pd.DataFrame:
52
+ """
53
+ 从MySQL获取数据
54
+
55
+ :param set_year: 是否按年份分表
56
+ """
57
+ try:
58
+ if set_year:
59
+ current_year = datetime.datetime.now().year
60
+ dfs = []
61
+ # 动态获取需要查询的年份范围
62
+ min_year = min(2024, pd.to_datetime(start_date).year) # 根据实际需求调整
63
+ for year in range(min_year, current_year + 1):
64
+ table = f"{table_name}_{year}"
65
+ df = self.download.data_to_df(
66
+ db_name=db_name,
67
+ table_name=table,
68
+ start_date=start_date,
69
+ end_date=end_date,
70
+ projection={}
71
+ )
72
+ if not df.empty:
73
+ dfs.append(df)
74
+ _df = pd.concat(dfs, ignore_index=True) if dfs else pd.DataFrame()
75
+ else:
76
+ _df = self.download.data_to_df(
77
+ db_name=db_name,
78
+ table_name=table_name,
79
+ start_date=start_date,
80
+ end_date=end_date,
81
+ projection={}
82
+ )
83
+
84
+ if _df.empty:
85
+ print(f"空数据 - {db_name}.{table_name}")
86
+ return pd.DataFrame()
87
+
88
+ return self._handle_datetime_columns(_df)
89
+
90
+ except Exception as e:
91
+ print(f"MySQL查询失败: {str(e)}")
92
+ return pd.DataFrame()
93
+
94
+ def get_from_redis(self, db_name: str, table_name: str,
95
+ start_date, end_date,
96
+ set_year: bool) -> pd.DataFrame:
97
+ """
98
+ 从Redis获取数据(带自动缓存更新)
99
+ """
100
+ start_dt = pd.to_datetime(start_date)
101
+ end_dt = pd.to_datetime(end_date)
102
+
103
+ # 生成统一缓存键
104
+ cache_key = f"{db_name}:{table_name}{'_year' if set_year else ''}"
105
+
106
+ try:
107
+ # 检查缓存状态
108
+ with self.lock: # 保证线程安全
109
+ ttl = self.redis_engine.ttl(cache_key)
110
+
111
+ if ttl < 300: # 剩余时间小于5分钟时触发更新
112
+ print(f"异步缓存更新: {cache_key}")
113
+ threading.Thread(
114
+ target=self._update_cache,
115
+ args=(cache_key, db_name, table_name,
116
+ start_date, end_date, set_year),
117
+ daemon=True
118
+ ).start()
119
+
120
+ # 获取缓存数据
121
+ cached_data = self.redis_engine.get(cache_key)
122
+ if not cached_data:
123
+ return self._fallback_to_mysql(db_name, table_name,
124
+ start_date, end_date, set_year)
125
+ json_str = cached_data.decode('utf-8')
126
+ _df = pd.read_json(json_str, orient='records')
127
+ _df = self._handle_datetime_columns(_df)
128
+
129
+ # 数据范围校验
130
+ if '日期' in _df.columns:
131
+ cache_min = _df['日期'].min()
132
+ cache_max = _df['日期'].max()
133
+
134
+ # 请求范围超出缓存范围时需要更新
135
+ if start_dt < cache_min or end_dt > cache_max:
136
+ print(f"请求范围超出缓存 {start_dt.strftime('%Y-%m-%d ')} - {end_dt.strftime('%Y-%m-%d ')}")
137
+ self._update_cache(cache_key, db_name, table_name,
138
+ start_date, end_date, set_year, _df)
139
+ return self._fallback_to_mysql(db_name, table_name,
140
+ start_date, end_date, set_year)
141
+
142
+ return _df[(start_dt <= _df['日期']) & (_df['日期'] <= end_dt)]
143
+ return _df
144
+
145
+ except Exception as e:
146
+ print(f"Redis操作失败: {str(e)}")
147
+ return self._fallback_to_mysql(db_name, table_name,
148
+ start_date, end_date, set_year)
149
+
150
+ def _update_cache(self, cache_key: str, db_name: str, table_name: str,
151
+ start_date: str, end_date: str, set_year: bool,
152
+ existing_df: pd.DataFrame = None) -> None:
153
+ """缓存更新核心逻辑"""
154
+ try:
155
+ # 获取最新数据
156
+ new_data = self.get_from_mysql(
157
+ db_name=db_name,
158
+ table_name=table_name,
159
+ start_date=start_date,
160
+ end_date=end_date,
161
+ set_year=set_year
162
+ )
163
+
164
+ # 合并历史数据
165
+ if existing_df is not None and not new_data.empty:
166
+ combined = pd.concat([existing_df, new_data], ignore_index=True)
167
+ combined = combined.drop_duplicates(subset='日期', keep='last')
168
+ else:
169
+ combined = new_data
170
+
171
+ if not combined.empty:
172
+ # 转换日期类型为字符串
173
+ temp_df = combined.copy()
174
+ datetime_cols = temp_df.select_dtypes(include=['datetime64[ns]']).columns
175
+ temp_df[datetime_cols] = temp_df[datetime_cols].astype(str)
176
+
177
+ # 存储到Redis
178
+ with self.lock:
179
+ self.redis_engine.set(
180
+ cache_key,
181
+ temp_df.to_json(orient='records', force_ascii=False),
182
+ ex=self.cache_ttl
183
+ )
184
+ print(f"缓存更新成功: {cache_key} | 记录数: {len(combined)}")
185
+
186
+ except Exception as e:
187
+ print(f"缓存更新失败: {str(e)}")
188
+
189
+ def _fallback_to_mysql(self, db_name: str, table_name: str,
190
+ start_date: str, end_date: str,
191
+ set_year: bool) -> pd.DataFrame:
192
+ """降级到直接MySQL查询"""
193
+ print(f"降级到MySQL查询: {db_name}.{table_name}")
194
+ return self.get_from_mysql(
195
+ db_name=db_name,
196
+ table_name=table_name,
197
+ start_date=start_date,
198
+ end_date=end_date,
199
+ set_year=set_year
200
+ )
201
+
202
+
203
+ if __name__ == '__main__':
204
+ # # ****************************************************
205
+ # # 这一部分在外部定义,只需要定义一次,开始
206
+ # redis_config = {
207
+ # 'host': '127.0.0.1',
208
+ # 'port': 6379, # 默认Redis端口
209
+ # 'db': 0, # 默认Redis数据库索引
210
+ # # 'username': 'default',
211
+ # 'password': redis_password,
212
+ # }
213
+ # # redis 实例化
214
+ # r = redis.Redis(**redis_config)
215
+ # # mysql 实例化
216
+ # d = s_query.QueryDatas(username=username, password=password, host=host, port=port)
217
+ # # 将两个库的实例化对象传给 RedisData 类,并实例化数据处理引擎
218
+ # m = RedisData(redis_engin=r, download=d)
219
+ # # ****************************************************
220
+ #
221
+ # # 以下为动态获取数据库数据
222
+ # db_name = '聚合数据'
223
+ # table_name = '多店推广场景_按日聚合'
224
+ # set_year = False
225
+ # df = m.get_from_redis(
226
+ # _db_name=db_name,
227
+ # _table_name=table_name,
228
+ # _set_year=set_year,
229
+ # start_date='2025-01-01',
230
+ # end_date='2025-01-31'
231
+ # )
232
+ # print(df)
233
+ #
234
+
235
+ print(socket.gethostname())
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: mdbq
3
- Version: 3.6.3
3
+ Version: 3.6.4
4
4
  Home-page: https://pypi.org/project/mdbq
5
5
  Author: xigua,
6
6
  Author-email: 2587125111@qq.com
@@ -40,5 +40,7 @@ mdbq/pbix/refresh_all.py
40
40
  mdbq/pbix/refresh_all_old.py
41
41
  mdbq/redis/__init__.py
42
42
  mdbq/redis/getredis.py
43
+ mdbq/redis/getredis_bak20250131.py
44
+ mdbq/redis/getredis_deepseek.py
43
45
  mdbq/spider/__init__.py
44
46
  mdbq/spider/aikucun.py
@@ -3,7 +3,7 @@
3
3
  from setuptools import setup, find_packages
4
4
 
5
5
  setup(name='mdbq',
6
- version='3.6.3',
6
+ version='3.6.4',
7
7
  author='xigua, ',
8
8
  author_email="2587125111@qq.com",
9
9
  url='https://pypi.org/project/mdbq',
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes