aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorBartek Grzybowski <b.grzybowski@partner.samsung.com>2020-03-24 12:11:29 +0100
committerBartek Grzybowski <b.grzybowski@partner.samsung.com>2020-03-24 12:13:14 +0100
commit3069ed97314d524296e31fab39833d4af9cf1e19 (patch)
tree815497c3c545f373577404ce19288fffbae928e2
parent7c69a61c7db9d75d5bee3529533295bba37779ab (diff)
Fix unused varbiales pylint warnings
Change-Id: I3f76a82d276c75f20453a9306b4f9964d90b5179 Issue-ID: INT-1482 Signed-off-by: Bartek Grzybowski <b.grzybowski@partner.samsung.com>
-rwxr-xr-xtutorials/vFWDT/workflow/workflow.py8
-rw-r--r--vnfs/DAaaS/microservices/PythonApps/python-kafkaConsumer-inference-app/src/producer/CustomKafkaProducer.py3
-rwxr-xr-xvnfs/DAaaS/sample-apps/sample-minio-save-model/build/save-model-to-minio.py4
3 files changed, 7 insertions, 8 deletions
diff --git a/tutorials/vFWDT/workflow/workflow.py b/tutorials/vFWDT/workflow/workflow.py
index 8703be73..0b2b6039 100755
--- a/tutorials/vFWDT/workflow/workflow.py
+++ b/tutorials/vFWDT/workflow/workflow.py
@@ -159,7 +159,7 @@ def _get_aai_rel_link_data(data, related_to, search_key=None, match_dict=None):
m_value = None
rel_dict = data.get(rel_lst)
if rel_dict: # check if data has relationship lists
- for key, rel_list in rel_dict.items():
+ for key, rel_list in rel_dict.items(): # pylint: disable=W0612
for rel in rel_list:
if rel.get("related-to") == related_to:
dval = None
@@ -398,7 +398,7 @@ def _osdf_request(rancher_ip, onap_ip, aai_data, exclude, use_oof_cache):
#print(json.dumps(template, indent=4))
with _no_ssl_verification():
- response = api.osdf.placement(body=template, params={}, headers={})
+ response = api.osdf.placement(body=template, params={}, headers={}) # pylint: disable=W0612
#if response.body.get('error_message') is not None:
# raise Exception(response.body['error_message']['explanation'])
@@ -505,7 +505,7 @@ def _extract_has_appc_identifiers(has_result, demand, onap_ip):
v_server['vserver-name'] = v_server['vserver-name'].replace("01", "02")
hostname_cache.append(v_server['vserver-name'])
- api = _init_python_aai_api(onap_ip)
+ api = _init_python_aai_api(onap_ip) # pylint: disable=W0612
vnfc_type = demand.lower()
# with _no_ssl_verification():
# response = api.aai.vnfc(v_server['vserver-name'], body=None, params={}, headers={})
@@ -1142,7 +1142,7 @@ def _generate_cdt_payloads_for_vnf(vnf_info, vnfc_type, actions):
for action_artifact in artifacts[action]:
artifact_list.append({'artifact-name': action_artifact['name'], 'artifact-type': action_artifact['type']})
if action != 'AllAction':
- req = _generate_cdt_artifact_request(req_id, action_artifact, action, vnfc_type)
+ req = _generate_cdt_artifact_request(req_id, action_artifact, action, vnfc_type) # pylint: disable=W0612
#print(json.dumps(req, indent=4))
#print(json.dumps(action_info, indent=4))
diff --git a/vnfs/DAaaS/microservices/PythonApps/python-kafkaConsumer-inference-app/src/producer/CustomKafkaProducer.py b/vnfs/DAaaS/microservices/PythonApps/python-kafkaConsumer-inference-app/src/producer/CustomKafkaProducer.py
index 43303761..6de868b7 100644
--- a/vnfs/DAaaS/microservices/PythonApps/python-kafkaConsumer-inference-app/src/producer/CustomKafkaProducer.py
+++ b/vnfs/DAaaS/microservices/PythonApps/python-kafkaConsumer-inference-app/src/producer/CustomKafkaProducer.py
@@ -23,8 +23,7 @@ class CustomKafkaProducer:
)
self.producer.flush()
- except Exception as e:
- #print("Error during producing to kafka topic. Stacktrace is %s",e)
+ except Exception:
logging.error("Error during producing to kafka topic.")
traceback.print_exc()
diff --git a/vnfs/DAaaS/sample-apps/sample-minio-save-model/build/save-model-to-minio.py b/vnfs/DAaaS/sample-apps/sample-minio-save-model/build/save-model-to-minio.py
index c5198471..91207640 100755
--- a/vnfs/DAaaS/sample-apps/sample-minio-save-model/build/save-model-to-minio.py
+++ b/vnfs/DAaaS/sample-apps/sample-minio-save-model/build/save-model-to-minio.py
@@ -49,7 +49,7 @@ if __name__ == "__main__":
found = False
try:
- client.stat_object(bucket_name, model);
+ client.stat_object(bucket_name, model)
found = True
except Exception as err:
found = False
@@ -62,5 +62,5 @@ if __name__ == "__main__":
if not found:
try:
client.fput_object(bucket_name, model, filepath, metadata=metadata)
- except expression as identifier:
+ except Exception as identifier:
print(err)