tf_lstm3_pre.py 7.0 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148
  1. #!/usr/bin/env python
  2. # -*- coding:utf-8 -*-
  3. # @FileName :tf_lstm_pre.py
  4. # @Time :2025/2/13 10:52
  5. # @Author :David
  6. # @Company: shenyang JY
  7. import json, copy
  8. import numpy as np
  9. from flask import Flask, request, g
  10. import logging, argparse, traceback
  11. from common.database_dml_koi import *
  12. from common.processing_data_common import missing_features, str_to_list
  13. from data_processing.data_operation.data_handler import DataHandler
  14. from threading import Lock
  15. import time, yaml
  16. from copy import deepcopy
  17. model_lock = Lock()
  18. from itertools import chain
  19. from common.logs import Log
  20. from common.data_utils import deep_update
  21. logger = Log('tf_ts3').logger
  22. np.random.seed(42) # NumPy随机种子
  23. # tf.set_random_seed(42) # TensorFlow随机种子
  24. app = Flask('tf_lstm3_pre——service')
  25. current_dir = os.path.dirname(os.path.abspath(__file__))
  26. with open(os.path.join(current_dir, 'lstm.yaml'), 'r', encoding='utf-8') as f:
  27. global_config = yaml.safe_load(f) # 只读的全局配置
  28. @app.before_request
  29. def update_config():
  30. # ------------ 整理参数,整合请求参数 ------------
  31. # 深拷贝全局配置 + 合并请求参数
  32. current_config = deepcopy(global_config)
  33. request_args = request.values.to_dict()
  34. # features参数规则:1.有传入,解析,覆盖 2. 无传入,不覆盖,原始值
  35. request_args['features'] = request_args['features'].split(',') if 'features' in request_args else current_config['features']
  36. request_args['time_series'] = request_args.get('time_series', 3)
  37. request_args['lstm_type'] = request_args.get('lstm_type', 1)
  38. current_config = deep_update(current_config, request_args)
  39. # 存储到请求上下文
  40. g.opt = argparse.Namespace(**current_config)
  41. g.dh = DataHandler(logger, current_config) # 每个请求独立实例
  42. if int(request_args['lstm_type']) == 1:
  43. from models_processing.model_tf.tf_bilstm import TSHandler
  44. g.ts = TSHandler(logger, current_config)
  45. else:
  46. from models_processing.model_tf.tf_bilstm_2 import TSHandler2
  47. g.ts = TSHandler2(logger, current_config)
  48. @app.route('/tf_lstm3_predict', methods=['POST'])
  49. def model_prediction_lstm3():
  50. # 获取程序开始时间
  51. start_time = time.time()
  52. result = {}
  53. success = 0
  54. dh = g.dh
  55. ts = g.ts
  56. args = deepcopy(g.opt.__dict__)
  57. logger.info("Program starts execution!")
  58. try:
  59. pre_data = get_data_from_mongo(args)
  60. if args.get('algorithm_test', 0):
  61. field_mapping = {'clearsky_ghi': 'clearskyGhi', 'dni_calcd': 'dniCalcd','surface_pressure': 'surfacePressure'}
  62. pre_data = pre_data.rename(columns=field_mapping)
  63. feature_scaler, target_scaler = get_scaler_model_from_mongo(args)
  64. ts.opt.cap = round(target_scaler.transform(np.array([[float(args['cap'])]]))[0, 0], 2)
  65. ts.get_model(args)
  66. dh.opt.features = json.loads(ts.model_params)['Model']['features'].split(',')
  67. scaled_pre_x, pre_data = dh.pre_data_handler(pre_data, feature_scaler, time_series=args['time_series'], lstm_type=3)
  68. res = list(chain.from_iterable(target_scaler.inverse_transform(ts.predict(scaled_pre_x))))
  69. pre_data['farm_id'] = args.get('farm_id', 'null')
  70. if int(args.get('algorithm_test', 0)):
  71. pre_data = pre_data.iloc[dh.opt.Model["time_step"]:]
  72. pre_data[args['model_name']] = res[:len(pre_data)]
  73. pre_data.rename(columns={args['col_time']: 'dateTime'}, inplace=True)
  74. pre_data = pre_data[['dateTime', 'farm_id', args['target'], args['model_name'], 'dq']]
  75. pre_data = pre_data.melt(id_vars=['dateTime', 'farm_id', args['target']], var_name='model', value_name='power_forecast')
  76. res_cols = ['dateTime', 'power_forecast', 'farm_id', args['target'], 'model']
  77. if 'howLongAgo' in args:
  78. pre_data['howLongAgo'] = int(args['howLongAgo'])
  79. res_cols += ['howLongAgo']
  80. else:
  81. pre_data = pre_data.iloc[dh.opt.Model["time_step"]:]
  82. pre_data['power_forecast'] = res[:len(pre_data)]
  83. pre_data.rename(columns={args['col_time']: 'date_time'}, inplace=True)
  84. res_cols = ['date_time', 'power_forecast', 'farm_id']
  85. pre_data = pre_data[res_cols]
  86. pre_data.loc[:, 'power_forecast'] = pre_data['power_forecast'].round(2)
  87. pre_data.loc[pre_data['power_forecast'] > float(args['cap']), 'power_forecast'] = float(args['cap'])
  88. pre_data.loc[pre_data['power_forecast'] < 0, 'power_forecast'] = 0
  89. insert_data_into_mongo(pre_data, args)
  90. success = 1
  91. except Exception as e:
  92. my_exception = traceback.format_exc()
  93. my_exception.replace("\n", "\t")
  94. result['msg'] = my_exception
  95. end_time = time.time()
  96. result['success'] = success
  97. result['args'] = args
  98. result['start_time'] = time.strftime('%Y-%m-%d %H:%M:%S', time.localtime(start_time))
  99. result['end_time'] = time.strftime('%Y-%m-%d %H:%M:%S', time.localtime(end_time))
  100. print("Program execution ends!")
  101. return result
  102. if __name__ == "__main__":
  103. print("Program starts execution!")
  104. from waitress import serve
  105. serve(app, host="0.0.0.0", port=10122,
  106. threads=8, # 指定线程数(默认4,根据硬件调整)
  107. channel_timeout=600 # 连接超时时间(秒)
  108. )
  109. print("server start!")
  110. # ------------------------测试代码------------------------
  111. # args_dict = {"mongodb_database": 'david_test', 'scaler_table': 'j00083_scaler', 'model_name': 'bp1.0.test',
  112. # 'model_table': 'j00083_model', 'mongodb_read_table': 'j00083_test', 'col_time': 'date_time', 'mongodb_write_table': 'j00083_rs',
  113. # 'features': 'speed10,direction10,speed30,direction30,speed50,direction50,speed70,direction70,speed90,direction90,speed110,direction110,speed150,direction150,speed170,direction170'}
  114. # args_dict['features'] = args_dict['features'].split(',')
  115. # arguments.update(args_dict)
  116. # dh = DataHandler(logger, arguments)
  117. # ts = TSHandler(logger)
  118. # opt = argparse.Namespace(**arguments)
  119. #
  120. # opt.Model['input_size'] = len(opt.features)
  121. # pre_data = get_data_from_mongo(args_dict)
  122. # feature_scaler, target_scaler = get_scaler_model_from_mongo(arguments)
  123. # pre_x = dh.pre_data_handler(pre_data, feature_scaler, opt)
  124. # ts.get_model(arguments)
  125. # result = ts.predict(pre_x)
  126. # result1 = list(chain.from_iterable(target_scaler.inverse_transform([result.flatten()])))
  127. # pre_data['power_forecast'] = result1[:len(pre_data)]
  128. # pre_data['farm_id'] = 'J00083'
  129. # pre_data['cdq'] = 1
  130. # pre_data['dq'] = 1
  131. # pre_data['zq'] = 1
  132. # pre_data.rename(columns={arguments['col_time']: 'date_time'}, inplace=True)
  133. # pre_data = pre_data[['date_time', 'power_forecast', 'farm_id', 'cdq', 'dq', 'zq']]
  134. #
  135. # pre_data['power_forecast'] = pre_data['power_forecast'].round(2)
  136. # pre_data.loc[pre_data['power_forecast'] > opt.cap, 'power_forecast'] = opt.cap
  137. # pre_data.loc[pre_data['power_forecast'] < 0, 'power_forecast'] = 0
  138. #
  139. # insert_data_into_mongo(pre_data, arguments)