aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorLadue, David (dl3158) <dl3158@att.com>2018-03-19 11:19:34 -0400
committerLadue, David (dl3158) <dl3158@att.com>2018-03-19 15:39:21 -0400
commit623442e3e9deab63f9d5c4bf88fd7077bc7f64b5 (patch)
tree9e80533141b3c03274dafd7d00e29777cd60d6b6
parentfb0aff81014b052a938a60b0815d6052404be162 (diff)
codestyle updates and docker
Change-Id: I68dafc17a4e9734bd86f925a720b22a951fe8d5f Issue-ID: DCAEGEN2-271 Signed-off-by: Ladue, David (dl3158) <dl3158@att.com>
-rw-r--r--bin/mod/trapd_file_utils.py41
-rw-r--r--bin/mod/trapd_get_cbs_config.py22
-rw-r--r--bin/mod/trapd_http_session.py3
-rw-r--r--bin/mod/trapd_logging.py44
-rw-r--r--bin/mod/trapd_settings.py19
-rw-r--r--bin/snmptrapd.py192
-rw-r--r--etc/snmptrapd.json4
-rw-r--r--requirements.txt1
8 files changed, 185 insertions, 141 deletions
diff --git a/bin/mod/trapd_file_utils.py b/bin/mod/trapd_file_utils.py
index 2da099b..e62b528 100644
--- a/bin/mod/trapd_file_utils.py
+++ b/bin/mod/trapd_file_utils.py
@@ -60,8 +60,8 @@ def roll_all_logs():
# NOTE: this will go away when onap logging is standardized/available
try:
# open various ecomp logs - if any fails, exit
- for fd in [tds.eelf_error_fd, tds.eelf_debug_fd, tds.eelf_audit_fd, \
- tds.eelf_metrics_fd, tds.arriving_traps_fd, tds.json_traps_fd]:
+ for fd in [tds.eelf_error_fd, tds.eelf_debug_fd, tds.eelf_audit_fd,
+ tds.eelf_metrics_fd, tds.arriving_traps_fd, tds.json_traps_fd]:
fd.close()
roll_file(tds.eelf_error_file_name)
@@ -86,7 +86,8 @@ def roll_all_logs():
try:
tds.json_traps_fd = open_file(tds.json_traps_filename)
except Exception as e:
- msg = ("Error opening json_log %s : %s" % (json_traps_filename, str(e)))
+ msg = ("Error opening json_log %s : %s" %
+ (json_traps_filename, str(e)))
stdout_logger(msg)
cleanup_and_exit(1, tds.pid_file_name)
@@ -96,11 +97,11 @@ def roll_all_logs():
try:
tds.arriving_traps_fd = open_file(tds.arriving_traps_filename)
except Exception as e:
- msg = ("Error opening arriving traps %s : %s" % (arriving_traps_filename, str(e)))
+ msg = ("Error opening arriving traps %s : %s" %
+ (arriving_traps_filename, str(e)))
stdout_logger(msg)
cleanup_and_exit(1, tds.pid_file_name)
-
# # # # # # # # # # # # # # # # # # #
# fx: setup_ecomp_logs -> log in eelf format until standard
@@ -116,7 +117,8 @@ def open_eelf_logs():
try:
# open various ecomp logs - if any fails, exit
- tds.eelf_error_file_name = (tds.c_config['files.eelf_base_dir'] + "/" + tds.c_config['files.eelf_error'])
+ tds.eelf_error_file_name = (
+ tds.c_config['files.eelf_base_dir'] + "/" + tds.c_config['files.eelf_error'])
tds.eelf_error_fd = open_file(tds.eelf_error_file_name)
except Exception as e:
@@ -124,9 +126,9 @@ def open_eelf_logs():
stdout_logger(msg)
cleanup_and_exit(1, tds.pid_file_name)
-
try:
- tds.eelf_debug_file_name = (tds.c_config['files.eelf_base_dir'] + "/" + tds.c_config['files.eelf_debug'])
+ tds.eelf_debug_file_name = (
+ tds.c_config['files.eelf_base_dir'] + "/" + tds.c_config['files.eelf_debug'])
tds.eelf_debug_fd = open_file(tds.eelf_debug_file_name)
except Exception as e:
@@ -135,7 +137,8 @@ def open_eelf_logs():
cleanup_and_exit(1, tds.pid_file_name)
try:
- tds.eelf_audit_file_name = (tds.c_config['files.eelf_base_dir'] + "/" + tds.c_config['files.eelf_audit'])
+ tds.eelf_audit_file_name = (
+ tds.c_config['files.eelf_base_dir'] + "/" + tds.c_config['files.eelf_audit'])
tds.eelf_audit_fd = open_file(tds.eelf_audit_file_name)
except Exception as e:
msg = "Error opening eelf audit log : " + str(e)
@@ -143,7 +146,8 @@ def open_eelf_logs():
cleanup_and_exit(1, tds.pid_file_name)
try:
- tds.eelf_metrics_file_name = (tds.c_config['files.eelf_base_dir'] + "/" + tds.c_config['files.eelf_metrics'])
+ tds.eelf_metrics_file_name = (
+ tds.c_config['files.eelf_base_dir'] + "/" + tds.c_config['files.eelf_metrics'])
tds.eelf_metrics_fd = open_file(tds.eelf_metrics_file_name)
except Exception as e:
msg = "Error opening eelf metric log : " + str(e)
@@ -163,8 +167,8 @@ def roll_file(_loc_file_name):
"""
_file_name_suffix = "%s" % (datetime.datetime.fromtimestamp(time.time()).
- fromtimestamp(time.time()).
- strftime('%Y-%m-%dT%H:%M:%S'))
+ fromtimestamp(time.time()).
+ strftime('%Y-%m-%dT%H:%M:%S'))
_loc_file_name_bak = _loc_file_name + '.' + _file_name_suffix
@@ -172,11 +176,12 @@ def roll_file(_loc_file_name):
if os.path.isfile(_loc_file_name):
try:
os.rename(_loc_file_name, _loc_file_name_bak)
- except:
+ except Exception as e:
_msg = ("ERROR: Unable to rename %s to %s"
- % (_loc_file_name,
- _loc_file_name_bak))
- ecomp_logger(tds.LOG_TYPE_ERROR, tds.SEV_CRIT, tds.CODE_GENERAL, _msg)
+ % (_loc_file_name,
+ _loc_file_name_bak))
+ ecomp_logger(tds.LOG_TYPE_ERROR, tds.SEV_CRIT,
+ tds.CODE_GENERAL, _msg)
# # # # # # # # # # # # #
@@ -189,7 +194,6 @@ def open_file(_loc_file_name):
open _loc_file_name, return file handle
"""
-
try:
# open append mode just in case so nothing is lost, but should be
# non-existent file
@@ -215,6 +219,7 @@ def close_file(_loc_fd, _loc_filename):
_loc_fd.close()
return True
except Exception as e:
- msg = "Error closing %s : %s - results indeterminate" % (_loc_filename, str(e))
+ msg = "Error closing %s : %s - results indeterminate" % (
+ _loc_filename, str(e))
ecomp_logger(tds.LOG_TYPE_ERROR, tds.SEV_FATAL, tds.CODE_GENERAL, msg)
return False
diff --git a/bin/mod/trapd_get_cbs_config.py b/bin/mod/trapd_get_cbs_config.py
index 775e0b2..e0f5ca8 100644
--- a/bin/mod/trapd_get_cbs_config.py
+++ b/bin/mod/trapd_get_cbs_config.py
@@ -20,7 +20,7 @@
#
"""
Look for CBS broker and return application config; if not present, look for
-env variable that specifies JSON equiv of CBS config (typically used for
+env variable that specifies JSON equiv of CBS config (typically used for
testing purposes)
"""
@@ -63,54 +63,56 @@ def get_cbs_config():
if tds.c_config == {}:
msg = "Unable to fetch CBS config or it is erroneously empty - trying override/simulator config"
stdout_logger(msg)
-
+
# if no CBS present, default to JSON config specified via CBS_SIM_JSON env var
except Exception as e:
msg = "ONAP controller not present, trying json config override via CBS_SIM_JSON env variable"
stdout_logger(msg)
-
+
try:
_cbs_sim_json_file = os.getenv("CBS_SIM_JSON", "None")
except Exception as e:
msg = "CBS_SIM_JSON not defined - FATAL ERROR, exiting"
stdout_logger(msg)
cleanup_and_exit(1, pid_file_name)
-
+
if _cbs_sim_json_file == "None":
msg = "CBS_SIM_JSON not defined - FATAL ERROR, exiting"
stdout_logger(msg)
cleanup_and_exit(1, pid_file_name)
else:
- msg = ("ONAP controller override specified via CBS_SIM_JSON: %s" % _cbs_sim_json_file )
+ msg = ("ONAP controller override specified via CBS_SIM_JSON: %s" %
+ _cbs_sim_json_file)
stdout_logger(msg)
try:
tds.c_config = json.load(open(_cbs_sim_json_file))
except Exception as e:
- msg = "Unable to load CBS_SIM_JSON " + _cbs_sim_json_file + " (invalid json?) - FATAL ERROR, exiting"
+ msg = "Unable to load CBS_SIM_JSON " + _cbs_sim_json_file + \
+ " (invalid json?) - FATAL ERROR, exiting"
stdout_logger(msg)
cleanup_and_exit(1, tds.pid_file_name)
# recalc timeout, set default if not present
try:
tds.timeout_seconds = tds.c_config['publisher.http_timeout_milliseconds'] / 1000.0
- except:
+ except Exception as e:
tds.timeout_seconds = 1.5
# recalc seconds_between_retries, set default if not present
try:
tds.seconds_between_retries = tds.c_config['publisher.http_milliseconds_between_retries'] / 1000.0
- except:
+ except Exception as e:
tds.seconds_between_retries = .750
# recalc min_severity_to_log, set default if not present
try:
tds.minimum_severity_to_log = tds.c_config['files.minimum_severity_to_log']
- except:
+ except Exception as e:
tds.minimum_severity_to_log = 3
try:
tds.publisher_retries = tds.c_config['publisher.http_retries']
- except:
+ except Exception as e:
tds.publisher_retries = 3
return True
diff --git a/bin/mod/trapd_http_session.py b/bin/mod/trapd_http_session.py
index 2e0b77e..b34c19d 100644
--- a/bin/mod/trapd_http_session.py
+++ b/bin/mod/trapd_http_session.py
@@ -29,9 +29,6 @@ import os
import requests
import traceback
-# snmptrapd
-import trapd_settings
-
prog_name = os.path.basename(__file__)
diff --git a/bin/mod/trapd_logging.py b/bin/mod/trapd_logging.py
index 435f03e..ae5a1a0 100644
--- a/bin/mod/trapd_logging.py
+++ b/bin/mod/trapd_logging.py
@@ -43,7 +43,7 @@ prog_name = os.path.basename(__file__)
# # # # # # # # # # # # # # # # # # #
-# fx: ecomp_logger -> log in eelf format until standard
+# fx: ecomp_logger -> log in eelf format until standard
# is released for python via LOG-161
# # # # # # # # # # ## # # # # # # #
@@ -51,16 +51,16 @@ def ecomp_logger(_log_type, _sev, _error_code, _msg):
"""
Log to ecomp-style logfiles. Logs include:
- Note: this will be updated when https://jira.onap.org/browse/LOG-161
+ Note: this will be updated when https://jira.onap.org/browse/LOG-161
is closed/available; until then, we resort to a generic format with
valuable info in "extra=" field (?)
:Parameters:
- _msg -
+ _msg -
:Exceptions:
none
:Keywords:
- eelf logging
+ eelf logging
:Log Styles:
:error.log:
@@ -95,7 +95,7 @@ def ecomp_logger(_log_type, _sev, _error_code, _msg):
else:
d = {'begtime': endAuditTime, 'begmsecs': endAuditMsec, 'endtime': endAuditTime,
'endmsecs': endAuditMsec}
-
+
self._logger.log(50, '%s|%s|%s|%s|%s|%s|%s|%s|%s|%s|%s|%s|%s|%s|%s|%s|%s|%s|%s|%s|%s|%s|%s|%s' \
% (requestID, serviceInstanceID, threadID, serverName, serviceName, partnerName,
statusCode, responseCode, responseDescription, instanceUUID, upperLogLevel,
@@ -132,43 +132,43 @@ def ecomp_logger(_log_type, _sev, _error_code, _msg):
t_out = datetime.datetime.utcnow().strftime("%Y-%m-%dT%H:%M:%S,%f")[:-3]
calling_fx = inspect.stack()[1][3]
- # FIXME: this entire module is a hack to override concept of prog logging
- # written across multiple files (???), making diagnostics IMPOSSIBLE!
+ # FIXME: this entire module is a hack to override concept of prog logging
+ # written across multiple files (???), making diagnostics IMPOSSIBLE!
# Hoping to leverage ONAP logging libraries & standards when available
# catch invalid log type
- if _log_type < 1 or _log_type > 5:
- msg = ("INVALID log type: %s " % _log_type )
- _out_rec = ("%s|%s|%s|%s|%s|%s|%s|%s|%s|%s" \
- % ((t_out, calling_fx, "snmptrapd", unused, unused, unused, tds.SEV_TYPES[_sev], _error_code, unused, (msg + _msg))))
+ if _log_type < 1 or _log_type > 5:
+ msg = ("INVALID log type: %s " % _log_type)
+ _out_rec = ("%s|%s|%s|%s|%s|%s|%s|%s|%s|%s"
+ % ((t_out, calling_fx, "snmptrapd", unused, unused, unused, tds.SEV_TYPES[_sev], _error_code, unused, (msg + _msg))))
tds.eelf_error_fd.write('%s\n' % str(_out_rec))
return False
if _sev >= tds.minimum_severity_to_log:
# log to appropriate eelf log (different files ??)
if _log_type == tds.LOG_TYPE_ERROR:
- _out_rec = ('%s|%s|%s|%s|%s|%s|%s|%s|%s|%s' \
- % ((t_out, calling_fx, "snmptrapd", unused, unused, unused, tds.SEV_TYPES[_sev], _error_code, unused, _msg)))
+ _out_rec = ('%s|%s|%s|%s|%s|%s|%s|%s|%s|%s'
+ % ((t_out, calling_fx, "snmptrapd", unused, unused, unused, tds.SEV_TYPES[_sev], _error_code, unused, _msg)))
tds.eelf_error_fd.write('%s\n' % str(_out_rec))
elif _log_type == tds.LOG_TYPE_AUDIT:
# log message in AUDIT format
- _out_rec = ('%s|%s|%s|%s|%s|%s|%s|%s|%s|%s' \
- % ((t_out, calling_fx, "snmptrapd", unused, unused, unused, tds.SEV_TYPES[_sev], _error_code, unused, _msg)))
+ _out_rec = ('%s|%s|%s|%s|%s|%s|%s|%s|%s|%s'
+ % ((t_out, calling_fx, "snmptrapd", unused, unused, unused, tds.SEV_TYPES[_sev], _error_code, unused, _msg)))
tds.eelf_audit_fd.write('%s\n' % str(_out_rec))
elif _log_type == tds.LOG_TYPE_METRICS:
# log message in METRICS format
- _out_rec = ('%s|%s|%s|%s|%s|%s|%s|%s|%s|%s' \
- % ((t_out, calling_fx, "snmptrapd", unused, unused, unused, tds.SEV_TYPES[_sev], _error_code, unused, _msg)))
+ _out_rec = ('%s|%s|%s|%s|%s|%s|%s|%s|%s|%s'
+ % ((t_out, calling_fx, "snmptrapd", unused, unused, unused, tds.SEV_TYPES[_sev], _error_code, unused, _msg)))
tds.eelf_metrics_fd.write('%s\n' % str(_out_rec))
-
+
# DEBUG *AND* others - there *MUST BE* a single time-sequenced log for diagnostics!
# FIXME: too much I/O !!!
# always write to debug; we need ONE logfile that has time-sequence full view !!!
# if (_log_type == tds.LOG_TYPE_DEBUG and _sev >= tds.current_min_sev_log_level) or (_log_type != tds.LOG_TYPE_DEBUG):
-
+
# log message in DEBUG format
- _out_rec = ("%s|%s|%s|%s|%s|%s|%s|%s|%s|%s" \
- % ((t_out, calling_fx, "snmptrapd", unused, unused, unused, tds.SEV_TYPES[_sev], _error_code, unused, _msg)))
+ _out_rec = ("%s|%s|%s|%s|%s|%s|%s|%s|%s|%s"
+ % ((t_out, calling_fx, "snmptrapd", unused, unused, unused, tds.SEV_TYPES[_sev], _error_code, unused, _msg)))
tds.eelf_debug_fd.write('%s\n' % str(_out_rec))
return True
@@ -196,4 +196,4 @@ def stdout_logger(_msg):
t_out = datetime.datetime.utcnow().strftime("%Y-%m-%dT%H:%M:%S,%f")[:-3]
# calling_fx = inspect.stack()[1][3]
- print('%s %s' % ( t_out, _msg))
+ print('%s %s' % (t_out, _msg))
diff --git a/bin/mod/trapd_settings.py b/bin/mod/trapd_settings.py
index 735b68c..be87e26 100644
--- a/bin/mod/trapd_settings.py
+++ b/bin/mod/trapd_settings.py
@@ -23,6 +23,7 @@
__docformat__ = 'restructuredtext'
+
def init():
# <CONSUL config cache>
@@ -36,7 +37,8 @@ def init():
# dns_cache_ip_to_name
# key [ip address] -> fqdn
# dns_cache_ip_expires
- # key [ip address] -> epoch time this entry expires and must be reloaded
+ # key [ip address] -> epoch time this entry expires and must
+ # be reloaded
global dns_cache_ip_to_name
dns_cache_ip_to_name = {}
global dns_cache_ip_expires
@@ -85,9 +87,9 @@ def init():
all_vb_json_str = ""
global trap_uuids_in_buffer
trap_uuids_in_buffer = ""
- # </trap and varbind dictionaries>
+ # </trap and varbind dictionaries>
- # <publish timers and counters>
+ # <publish timers and counters>
global traps_in_minute
traps_in_minute = 0
global last_epoch_second
@@ -104,12 +106,12 @@ def init():
seconds_between_retries = 2
global publisher_retries
publisher_retries = 2
- # </publish timers and counters>
+ # </publish timers and counters>
- # <publish http request session (persistent as much as possible)>
+ # <publish http request session (persistent as much as possible)>
global http_requ_session
http_requ_session = None
- # </publish http request session>
+ # </publish http request session>
# <json log of traps published>
global json_traps_filename
@@ -159,10 +161,9 @@ def init():
SEV_FATAL = 5
global CODE_GENERAL
- CODE_GENERAL="100"
+ CODE_GENERAL = "100"
global minimum_severity_to_log
- minimum_severity_to_log=3
-
+ minimum_severity_to_log = 3
# </logging types and severities>
diff --git a/bin/snmptrapd.py b/bin/snmptrapd.py
index dde4e39..3765746 100644
--- a/bin/snmptrapd.py
+++ b/bin/snmptrapd.py
@@ -124,8 +124,9 @@ def load_all_configs(_signum, _frame):
if int(_signum) != 0:
msg = ("received signal %s at frame %s; re-reading configs"
- % (_signum, _frame))
- ecomp_logger(tds.LOG_TYPE_DEBUG, tds.SEV_DETAILED, tds.CODE_GENERAL, msg)
+ % (_signum, _frame))
+ ecomp_logger(tds.LOG_TYPE_DEBUG, tds.SEV_DETAILED,
+ tds.CODE_GENERAL, msg)
# Initialize dmaap requests session object. Close existing session
# if applicable.
@@ -133,13 +134,13 @@ def load_all_configs(_signum, _frame):
tds.http_requ_session.close()
tds.http_requ_session = init_session_obj()
- if tds.http_requ_session == None:
+ if tds.http_requ_session is None:
msg = "Unable to create new http session - FATAL ERROR, exiting"
ecomp_logger(tds.LOG_TYPE_ERROR, tds.SEV_FATAL, tds.CODE_GENERAL, msg)
stdout_logger(msg)
cleanup_and_exit(1, tds.pid_file_name)
- # re-request config from config binding service
+ # re-request config from config binding service
# (either broker, or json file override)
if not get_cbs_config():
msg = "error (re)loading CBS config - FATAL ERROR, exiting"
@@ -154,7 +155,6 @@ def load_all_configs(_signum, _frame):
def log_all_arriving_traps():
-
# roll logs as needed/defined in files.roll_frequency
if tds.c_config['files.roll_frequency'] == "minute":
curr_minute = datetime.datetime.now().minute
@@ -178,23 +178,26 @@ def log_all_arriving_traps():
# going for:
# 1520971776 Tue Mar 13 16:09:36 2018; 1520971776 2018-03-13 16:09:36 DCAE-COLLECTOR-UCSNMP 15209717760049 .1.3.6.1.4.1.2636.4.1.6 gfpmt5pcs10.oss.att.com 135.91.10.139 12.123.1.240 12.123.1.240 2 varbinds: [0] .1.3.6.1.2.1.1.3.0 {10} 1212058366 140 days, 6:49:43.66 [1] .1.3.6.1.6.3.1.1.4.1.0 {6} .1.3.6.1.4.1.2636.4.1.6 [2] .1.3.6.1.4.1.2636.3.1.15.1.1.2.4.0.0 {2} 2 [3] .1.3.6.1.4.1.2636.3.1.15.1.2.2.4.0.0 {2} 4 [4] .1.3.6.1.4.1.2636.3.1.15.1.3.2.4.0.0 {2} 0 [5] .1.3.6.1.4.1.2636.3.1.15.1.4.2.4.0.0 {2} 0 [6] .1.3.6.1.4.1.2636.3.1.15.1.5.2.4.0.0 {4} PEM 3 [7] .1.3.6.1.4.1.2636.3.1.15.1.6.2.4.0.0 {2} 7 [8] .1.3.6.1.4.1.2636.3.1.15.1.7.2.4.0.0 {2} 4 [9] .1.3.6.1.6.3.18.1.3.0 {7} 12.123.1.240
- tds.arriving_traps_fd.write('%s %s; %s %s %s %s %s %s %s %s %s %s %s\n' %
- (tds.trap_dict["time received"],
- time.strftime("%a %b %d %H:%M:%S %Y", time.localtime(time.time())),
- time.strftime("%a %b %d %H:%M:%S %Y", time.localtime(tds.trap_dict["time received"])),
- tds.trap_dict["trap category"],
- tds.trap_dict["epoch_serno"],
- tds.trap_dict["notify OID"],
- tds.trap_dict["agent name"],
- tds.trap_dict["agent address"],
- tds.trap_dict["cambria.partition"],
- tds.trap_dict["protocol version"],
- tds.trap_dict["sysUptime"],
- tds.trap_dict["uuid"],
- tds.all_vb_json_str))
+ tds.arriving_traps_fd.write('%s %s; %s %s %s %s %s %s %s %s %s %s %s\n' %
+ (tds.trap_dict["time received"],
+ time.strftime(
+ "%a %b %d %H:%M:%S %Y", time.localtime(time.time())),
+ time.strftime("%a %b %d %H:%M:%S %Y", time.localtime(
+ tds.trap_dict["time received"])),
+ tds.trap_dict["trap category"],
+ tds.trap_dict["epoch_serno"],
+ tds.trap_dict["notify OID"],
+ tds.trap_dict["agent name"],
+ tds.trap_dict["agent address"],
+ tds.trap_dict["cambria.partition"],
+ tds.trap_dict["protocol version"],
+ tds.trap_dict["sysUptime"],
+ tds.trap_dict["uuid"],
+ tds.all_vb_json_str))
except Exception as e:
- msg = "Error writing to %s : %s - arriving trap %s NOT LOGGED" %(tds.arriving_traps_filename, str(e), tds.trap_dict["uuid"])
+ msg = "Error writing to %s : %s - arriving trap %s NOT LOGGED" % (
+ tds.arriving_traps_filename, str(e), tds.trap_dict["uuid"])
ecomp_logger(tds.LOG_TYPE_ERROR, tds.SEV_CRIT, tds.CODE_GENERAL, msg)
@@ -213,10 +216,13 @@ def log_published_messages(_post_data_enclosed):
try:
tds.json_traps_fd.write('%s\n' % _post_data_enclosed)
- msg = "successfully logged json for %s to %s" % (tds.trap_dict["uuid"], tds.json_traps_filename)
- ecomp_logger(tds.LOG_TYPE_DEBUG, tds.SEV_DETAILED, tds.CODE_GENERAL, msg)
+ msg = "successfully logged json for %s to %s" % (
+ tds.trap_dict["uuid"], tds.json_traps_filename)
+ ecomp_logger(tds.LOG_TYPE_DEBUG, tds.SEV_DETAILED,
+ tds.CODE_GENERAL, msg)
except Exception as e:
- msg = "Error writing to %s : %s - trap %s NOT LOGGED" %(tds.json_traps_filename, str(e), tds.trap_dict["uuid"])
+ msg = "Error writing to %s : %s - trap %s NOT LOGGED" % (
+ tds.json_traps_filename, str(e), tds.trap_dict["uuid"])
ecomp_logger(tds.LOG_TYPE_ERROR, tds.SEV_CRIT, tds.CODE_GENERAL, msg)
@@ -240,7 +246,8 @@ def post_dmaap():
if tds.http_requ_session is None:
msg = "tds.http_requ_session is None - getting new (%s)" % tds.http_requ_session
- ecomp_logger(tds.LOG_TYPE_DEBUG, tds.SEV_DETAILED, tds.CODE_GENERAL, msg)
+ ecomp_logger(tds.LOG_TYPE_DEBUG, tds.SEV_DETAILED,
+ tds.CODE_GENERAL, msg)
tds.http_requ_session = init_session_obj()
# if only 1 trap, ship as-is
@@ -258,64 +265,80 @@ def post_dmaap():
while not dmaap_pub_success and k < (int(tds.c_config['publisher.http_retries'])):
try:
- if tds.c_config['streams_publishes']['sec_fault_unsecure']['aaf_username'] == "" or tds.c_config['streams_publishes']['sec_fault_unsecure']['aaf_username'] == None:
- msg = "%d trap(s) : %s - attempt %d (unsecure)" % (tds.traps_since_last_publish, tds.trap_uuids_in_buffer, k)
- ecomp_logger(tds.LOG_TYPE_DEBUG, tds.SEV_DETAILED, tds.CODE_GENERAL, msg)
+ if tds.c_config['streams_publishes']['sec_fault_unsecure']['aaf_username'] == "" or tds.c_config['streams_publishes']['sec_fault_unsecure']['aaf_username'] is None:
+ msg = "%d trap(s) : %s - attempt %d (unsecure)" % (
+ tds.traps_since_last_publish, tds.trap_uuids_in_buffer, k)
+ ecomp_logger(tds.LOG_TYPE_DEBUG, tds.SEV_DETAILED,
+ tds.CODE_GENERAL, msg)
http_resp = tds.http_requ_session.post(tds.c_config['streams_publishes']['sec_fault_unsecure']['dmaap_info']['topic_url'], post_data_enclosed,
- headers=http_headers,
- timeout=tds.timeout_seconds)
+ headers=http_headers,
+ timeout=tds.timeout_seconds)
else:
- msg = "%d trap(s) : %s - attempt %d (secure)" % (tds.traps_since_last_publish, tds.trap_uuids_in_buffer, k)
- ecomp_logger(tds.LOG_TYPE_DEBUG, tds.SEV_DETAILED, tds.CODE_GENERAL, msg)
+ msg = "%d trap(s) : %s - attempt %d (secure)" % (
+ tds.traps_since_last_publish, tds.trap_uuids_in_buffer, k)
+ ecomp_logger(tds.LOG_TYPE_DEBUG, tds.SEV_DETAILED,
+ tds.CODE_GENERAL, msg)
http_resp = tds.http_requ_session.post(tds.c_config['streams_publishes']['sec_fault_unsecure']['dmaap_info']['topic_url'], post_data_enclosed,
- auth=(tds.c_config['streams_publishes']['sec_fault_unsecure']['aaf_username'],
- tds.c_config['streams_publishes']['sec_fault_unsecure']['aaf_password']),
- headers=http_headers,
- timeout=tds.timeout_seconds)
+ auth=(tds.c_config['streams_publishes']['sec_fault_unsecure']['aaf_username'],
+ tds.c_config['streams_publishes']['sec_fault_unsecure']['aaf_password']),
+ headers=http_headers,
+ timeout=tds.timeout_seconds)
if http_resp.status_code == requests.codes.ok:
# msg = "%d trap(s) : %s successfully published - response from %s: %d %s" % (traps_since_last_publish, trap_uuids_in_buffer, ((c_config['streams_publishes']['sec_fault_unsecure']['dmaap_info']['topic_url']).split('/')[2][:-5]) ,http_resp.status_code, http_resp.text)
- msg = "%d trap(s) successfully published: %s" % (tds.traps_since_last_publish, tds.trap_uuids_in_buffer)
- ecomp_logger(tds.LOG_TYPE_METRICS, tds.SEV_INFO, tds.CODE_GENERAL, msg)
+ msg = "%d trap(s) successfully published: %s" % (
+ tds.traps_since_last_publish, tds.trap_uuids_in_buffer)
+ ecomp_logger(tds.LOG_TYPE_METRICS, tds.SEV_INFO,
+ tds.CODE_GENERAL, msg)
log_published_messages(post_data_enclosed)
tds.last_pub_time = time.time()
dmaap_pub_success = True
break
else:
- msg = "Trap(s) %s publish attempt %d returned non-normal: %d %s" % (tds.trap_uuids_in_buffer, k, http_resp.status_code, http_resp.text)
- ecomp_logger(tds.LOG_TYPE_ERROR, tds.SEV_WARN, tds.CODE_GENERAL, msg)
+ msg = "Trap(s) %s publish attempt %d returned non-normal: %d %s" % (
+ tds.trap_uuids_in_buffer, k, http_resp.status_code, http_resp.text)
+ ecomp_logger(tds.LOG_TYPE_ERROR, tds.SEV_WARN,
+ tds.CODE_GENERAL, msg)
except OSError as e:
- msg = "OS exception while attempting to post %s attempt %s: (%s) %s %s" % (tds.trap_uuids_in_buffer, k, e.errno, e.strerror, str(e))
- ecomp_logger(tds.LOG_TYPE_ERROR, tds.SEV_WARN, tds.CODE_GENERAL, msg)
+ msg = "OS exception while attempting to post %s attempt %s: (%s) %s %s" % (
+ tds.trap_uuids_in_buffer, k, e.errno, e.strerror, str(e))
+ ecomp_logger(tds.LOG_TYPE_ERROR, tds.SEV_WARN,
+ tds.CODE_GENERAL, msg)
except requests.exceptions.RequestException as e:
- msg = "Requests exception while attempting to post %s attempt %d: (%d) %s" % (tds.trap_uuids_in_buffer, int(k), int(e.errno), str(e.strerror))
- ecomp_logger(tds.LOG_TYPE_ERROR, tds.SEV_WARN, tds.CODE_GENERAL, msg)
+ msg = "Requests exception while attempting to post %s attempt %d: (%d) %s" % (
+ tds.trap_uuids_in_buffer, int(k), int(e.errno), str(e.strerror))
+ ecomp_logger(tds.LOG_TYPE_ERROR, tds.SEV_WARN,
+ tds.CODE_GENERAL, msg)
k += 1
if k < tds.c_config['publisher.http_retries']:
- msg = "sleeping %.4f seconds and retrying" % (tds.seconds_between_retries)
- ecomp_logger(tds.LOG_TYPE_DEBUG, tds.SEV_DETAILED, tds.CODE_GENERAL, msg)
+ msg = "sleeping %.4f seconds and retrying" % (
+ tds.seconds_between_retries)
+ ecomp_logger(tds.LOG_TYPE_DEBUG, tds.SEV_DETAILED,
+ tds.CODE_GENERAL, msg)
time.sleep(tds.seconds_between_retries)
else:
break
if not dmaap_pub_success:
msg = "ALL publish attempts failed for traps %s to URL %s "\
- % (tds.trap_uuids_in_buffer, tds.c_config['streams_publishes']['sec_fault_unsecure']['dmaap_info']['topic_url'])
+ % (tds.trap_uuids_in_buffer, tds.c_config['streams_publishes']['sec_fault_unsecure']['dmaap_info']['topic_url'])
ecomp_logger(tds.LOG_TYPE_ERROR, tds.SEV_CRIT, tds.CODE_GENERAL, msg)
- # FIXME: This currently tries, then logs error and trashes buffer if all dmaap attempts fail. Better way?
+ # FIXME: This currently tries, then logs error and trashes buffer if all dmaap attempts fail. Better way?
tds.traps_since_last_publish = 0
- tds.trap_uuids_in_buffer=""
+ tds.trap_uuids_in_buffer = ""
tds.all_traps_str = ""
tds.first_trap = True
# # # # # # # # # # # # # # # # # # #
# fx: request_observer for community string rewrite
# # # # # # # # # # # # # # # # # # #
+
+
def comm_string_rewrite_observer(snmpEngine, execpoint, variables, cbCtx):
# match ALL community strings
@@ -362,7 +385,7 @@ def snmp_engine_observer_cb(snmp_engine, execpoint, variables, cbCtx):
tds.trap_dict["agent address"] = ip_addr_str
msg = 'snmp trap arrived from %s, assigned uuid: %s' % \
- (ip_addr_str, tds.trap_dict["uuid"])
+ (ip_addr_str, tds.trap_dict["uuid"])
ecomp_logger(tds.LOG_TYPE_DEBUG, tds.SEV_DETAILED, tds.CODE_GENERAL, msg)
try:
@@ -371,12 +394,13 @@ def snmp_engine_observer_cb(snmp_engine, execpoint, variables, cbCtx):
(ip_addr_str, (tds.dns_cache_ip_expires[ip_addr_str])))
else:
tds.trap_dict["agent name"] = tds.dns_cache_ip_to_name[ip_addr_str]
- except:
+ except Exception as e:
msg = "dns cache expired or missing for %s - refreshing" % ip_addr_str
- ecomp_logger(tds.LOG_TYPE_DEBUG, tds.SEV_DETAILED, tds.CODE_GENERAL, msg)
+ ecomp_logger(tds.LOG_TYPE_DEBUG, tds.SEV_DETAILED,
+ tds.CODE_GENERAL, msg)
try:
- agent_fqdn,alias,addresslist = socket.gethostbyaddr(ip_addr_str)
- except:
+ agent_fqdn, alias, addresslist = socket.gethostbyaddr(ip_addr_str)
+ except Exception as e:
agent_fqdn = ip_addr_str
tds.trap_dict["agent name"] = agent_fqdn
@@ -385,11 +409,13 @@ def snmp_engine_observer_cb(snmp_engine, execpoint, variables, cbCtx):
tds.dns_cache_ip_expires[ip_addr_str] = (
time.time() + tds.c_config['cache.dns_cache_ttl_seconds'])
msg = "cache for %s (%s) updated - set to expire at %d" % \
- (agent_fqdn, ip_addr_str, tds.dns_cache_ip_expires[ip_addr_str])
- ecomp_logger(tds.LOG_TYPE_DEBUG, tds.SEV_DETAILED, tds.CODE_GENERAL, msg)
+ (agent_fqdn, ip_addr_str, tds.dns_cache_ip_expires[ip_addr_str])
+ ecomp_logger(tds.LOG_TYPE_DEBUG, tds.SEV_DETAILED,
+ tds.CODE_GENERAL, msg)
tds.trap_dict["cambria.partition"] = str(tds.trap_dict["agent name"])
- tds.trap_dict["community"] = "" # do not include cleartext community in pub
+ # do not include cleartext community in pub
+ tds.trap_dict["community"] = ""
tds.trap_dict["community len"] = 0
# FIXME.CHECK_WITH_DOWNSTREAM_CONSUMERS: get rid of round for millisecond val
@@ -420,12 +446,14 @@ def snmp_engine_observer_cb(snmp_engine, execpoint, variables, cbCtx):
if snmp_version == 3:
tds.trap_dict["protocol version"] = "v3"
tds.trap_dict["security level"] = str(variables['securityLevel'])
- tds.trap_dict["context name"] = str(variables['contextName'].prettyPrint())
+ tds.trap_dict["context name"] = str(
+ variables['contextName'].prettyPrint())
tds.trap_dict["security name"] = str(variables['securityName'])
tds.trap_dict["security engine"] = str(
variables['contextEngineId'].prettyPrint())
tds.trap_dict['time received'] = epoch_msecond
- tds.trap_dict['trap category'] = (tds.c_config['streams_publishes']['sec_fault_unsecure']['dmaap_info']['topic_url']).split('/')[-1]
+ tds.trap_dict['trap category'] = (
+ tds.c_config['streams_publishes']['sec_fault_unsecure']['dmaap_info']['topic_url']).split('/')[-1]
# # # # # # # # # # # # # # # # # # #
@@ -469,7 +497,7 @@ def add_varbind_to_json(vb_idx, vb_oid, vb_type, vb_val):
else:
# all_vb_json_str = ''.join([all_vb_json_str, ' ,'])
# all_vb_json_str = "%s ," % all_vb_json_str
- tds.all_vb_json_str = tds.all_vb_json_str + " ,"
+ tds.all_vb_json_str = tds.all_vb_json_str + " ,"
_individual_vb_dict.clear()
_individual_vb_dict['varbind_oid'] = vb_oid.prettyPrint()
@@ -513,7 +541,7 @@ def notif_receiver_cb(snmp_engine, stateReference, contextEngineId, contextName,
# FIXME update reset location when batching publishes
vb_idx = 0
- # For reference:
+ # For reference:
#
# print('\nvarBinds ==> %s' % (varBinds))
#
@@ -567,24 +595,28 @@ def notif_receiver_cb(snmp_engine, stateReference, contextEngineId, contextName,
msg = "adding %s to buffer" % (tds.trap_dict["uuid"])
ecomp_logger(tds.LOG_TYPE_DEBUG, tds.SEV_DETAILED, tds.CODE_GENERAL, msg)
if tds.first_trap:
- tds.all_traps_str = curr_trap_json_str
- tds.trap_uuids_in_buffer = tds.trap_dict["uuid"]
- tds.first_trap = False
+ tds.all_traps_str = curr_trap_json_str
+ tds.trap_uuids_in_buffer = tds.trap_dict["uuid"]
+ tds.first_trap = False
else:
- tds.trap_uuids_in_buffer = tds.trap_uuids_in_buffer + ', ' + tds.trap_dict["uuid"]
- tds.all_traps_str = tds.all_traps_str + ', ' + curr_trap_json_str
+ tds.trap_uuids_in_buffer = tds.trap_uuids_in_buffer + \
+ ', ' + tds.trap_dict["uuid"]
+ tds.all_traps_str = tds.all_traps_str + ', ' + curr_trap_json_str
# always log arriving traps
log_all_arriving_traps()
# publish to dmaap after last varbind is processed
if tds.traps_since_last_publish >= tds.c_config['publisher.max_traps_between_publishes']:
- msg = "num traps since last publish (%d) exceeds threshold (%d) - publish traps" % (tds.traps_since_last_publish, tds.c_config['publisher.max_traps_between_publishes'])
- ecomp_logger(tds.LOG_TYPE_DEBUG, tds.SEV_DETAILED, tds.CODE_GENERAL, msg)
+ msg = "num traps since last publish (%d) exceeds threshold (%d) - publish traps" % (
+ tds.traps_since_last_publish, tds.c_config['publisher.max_traps_between_publishes'])
+ ecomp_logger(tds.LOG_TYPE_DEBUG, tds.SEV_DETAILED,
+ tds.CODE_GENERAL, msg)
post_dmaap()
elif milliseconds_since_last_publish >= tds.c_config['publisher.max_milliseconds_between_publishes']:
- msg = "num milliseconds since last publish (%.0f) exceeds threshold - publish traps"% milliseconds_since_last_publish
- ecomp_logger(tds.LOG_TYPE_DEBUG, tds.SEV_DETAILED, tds.CODE_GENERAL, msg)
+ msg = "num milliseconds since last publish (%.0f) exceeds threshold - publish traps" % milliseconds_since_last_publish
+ ecomp_logger(tds.LOG_TYPE_DEBUG, tds.SEV_DETAILED,
+ tds.CODE_GENERAL, msg)
post_dmaap()
@@ -617,8 +649,9 @@ tds.init()
tds.last_hour = datetime.datetime.now().hour
# get config binding service (CBS) values (either broker, or json file override)
-load_all_configs(0,0)
-msg = "%s : %s version %s starting" % (prog_name, tds.c_config['snmptrap.title'], tds.c_config['snmptrap.version'])
+load_all_configs(0, 0)
+msg = "%s : %s version %s starting" % (
+ prog_name, tds.c_config['snmptrap.title'], tds.c_config['snmptrap.version'])
stdout_logger(msg)
# Avoid this unless needed for testing; it prints sensitive data to log
@@ -634,18 +667,21 @@ open_eelf_logs()
# bump up logging level if overridden at command line
if verbose:
msg = "WARNING: '-v' argument present. All messages will be logged. This can slow things down, use only when needed."
- tds.minimum_severity_to_log=0
+ tds.minimum_severity_to_log = 0
stdout_logger(msg)
# name and open arriving trap log
-tds.arriving_traps_filename = tds.c_config['files.runtime_base_dir'] + "/" + tds.c_config['files.log_dir'] + "/" + (tds.c_config['files.arriving_traps_log'])
+tds.arriving_traps_filename = tds.c_config['files.runtime_base_dir'] + "/" + \
+ tds.c_config['files.log_dir'] + "/" + \
+ (tds.c_config['files.arriving_traps_log'])
tds.arriving_traps_fd = open_file(tds.arriving_traps_filename)
msg = ("arriving traps logged to: %s" % tds.arriving_traps_filename)
stdout_logger(msg)
ecomp_logger(tds.LOG_TYPE_DEBUG, tds.SEV_INFO, tds.CODE_GENERAL, msg)
# name and open json trap log
-tds.json_traps_filename = tds.c_config['files.runtime_base_dir'] + "/" + tds.c_config['files.log_dir'] + "/" + "DMAAP_" + (tds.c_config['streams_publishes']['sec_fault_unsecure']['dmaap_info']['topic_url'].split('/')[-1]) + ".json"
+tds.json_traps_filename = tds.c_config['files.runtime_base_dir'] + "/" + tds.c_config['files.log_dir'] + "/" + "DMAAP_" + (
+ tds.c_config['streams_publishes']['sec_fault_unsecure']['dmaap_info']['topic_url'].split('/')[-1]) + ".json"
tds.json_traps_fd = open_file(tds.json_traps_filename)
msg = ("published traps logged to: %s" % tds.json_traps_filename)
stdout_logger(msg)
@@ -686,7 +722,8 @@ try:
(ipv4_interface, ipv4_port))
)
except Exception as e:
- msg = "Unable to bind to %s:%s - %s" % (ipv4_interface, ipv4_port, str(e))
+ msg = "Unable to bind to %s:%s - %s" % (
+ ipv4_interface, ipv4_port, str(e))
ecomp_logger(tds.LOG_TYPE_ERROR, tds.SEV_FATAL, tds.CODE_GENERAL, msg)
stdout_logger(msg)
cleanup_and_exit(1, tds.pid_file_name)
@@ -719,7 +756,8 @@ try:
(ipv6_interface, ipv6_port))
)
except Exception as e:
- msg = "Unable to bind to %s:%s - %s" % (ipv6_interface,ipv6_port, str(e))
+ msg = "Unable to bind to %s:%s - %s" % (
+ ipv6_interface, ipv6_port, str(e))
stdout_logger(msg)
ecomp_logger(tds.LOG_TYPE_ERROR, tds.SEV_FATAL, tds.CODE_GENERAL, msg)
cleanup_and_exit(1, tds.pid_file_name)
@@ -760,7 +798,7 @@ snmp_engine.transportDispatcher.jobStarted(1) # loop forever
# Run I/O dispatcher which will receive traps
try:
snmp_engine.transportDispatcher.runDispatcher()
-except:
+except Exception as e:
snmp_engine.observer.unregisterObserver()
snmp_engine.transportDispatcher.closeDispatcher()
cleanup_and_exit(1, tds.pid_file_name)
diff --git a/etc/snmptrapd.json b/etc/snmptrapd.json
index c9caa76..053f364 100644
--- a/etc/snmptrapd.json
+++ b/etc/snmptrapd.json
@@ -3,9 +3,9 @@
"snmptrap.title": "ONAP SNMP Trap Receiver" ,
"protocols.transport": "udp",
"protocols.ipv4_interface": "0.0.0.0",
-"protocols.ipv4_port": 6164,
+"protocols.ipv4_port": 162,
"protocols.ipv6_interface": "::1",
-"protocols.ipv6_port": 6164,
+"protocols.ipv6_port": 162,
"cache.dns_cache_ttl_seconds": 60,
"publisher.http_timeout_milliseconds": 1500,
"publisher.http_retries": 3,
diff --git a/requirements.txt b/requirements.txt
index a0b06ee..56b2141 100644
--- a/requirements.txt
+++ b/requirements.txt
@@ -1,3 +1,4 @@
snmptrap
pysnmp==4.4.2
+requests=2.18.3
onap_dcae_cbs_docker_client==0.0.3