1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
|
#
# Copyright (C) 2019 - 2020 Bell Canada.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
from google.protobuf.timestamp_pb2 import Timestamp
import proto.CommandExecutor_pb2 as CommandExecutor_pb2
import json
import email.parser
CDS_IS_SUCCESSFUL_KEY = "cds_is_successful"
ERR_MSG_KEY = "err_msg"
RESULTS_KEY = "results"
RESULTS_LOG_KEY = "results_log"
RESPONSE_MAX_SIZE = 4 * 1024 * 1024 # 4Mb
def get_blueprint_id(request):
blueprint_name = request.identifiers.blueprintName
blueprint_version = request.identifiers.blueprintVersion
return blueprint_name + '/' + blueprint_version
def get_blueprint_timeout(request):
return request.timeOut
# Create a response for grpc. Fills in the timestamp as well as removes cds_is_successful element
def build_grpc_response(request_id, response):
if response[CDS_IS_SUCCESSFUL_KEY]:
status = CommandExecutor_pb2.SUCCESS
else:
status = CommandExecutor_pb2.FAILURE
response.pop(CDS_IS_SUCCESSFUL_KEY)
logs = response.pop(RESULTS_LOG_KEY)
# Payload should only contains response data returned from the executed script and/or the error message
payload = json.dumps(response)
timestamp = Timestamp()
timestamp.GetCurrentTime()
execution_output = CommandExecutor_pb2.ExecutionOutput(requestId=request_id,
response=logs,
status=status,
payload=payload,
timestamp=timestamp)
return truncate_execution_output(execution_output)
# build a ret data structure used to populate the ExecutionOutput
def build_ret_data(cds_is_successful, results_log=[], error=None):
ret_data = {
CDS_IS_SUCCESSFUL_KEY: cds_is_successful,
RESULTS_LOG_KEY: results_log
}
if error:
ret_data[ERR_MSG_KEY] = error
return ret_data
# Truncate execution logs to make sure gRPC response doesn't exceed the gRPC buffer capacity
def truncate_execution_output(execution_output):
sum_truncated_chars = 0
if execution_output.ByteSize() > RESPONSE_MAX_SIZE:
while execution_output.ByteSize() > RESPONSE_MAX_SIZE:
removed_item = execution_output.response.pop()
sum_truncated_chars += len(removed_item)
execution_output.response.append("[...] TRUNCATED CHARS : {}".format(sum_truncated_chars))
return execution_output
# Read temp file 'outputfile' into results_log and split out the returned payload into payload_result
def parse_cmd_exec_output(outputfile, logger, payload_result, results_log, extra):
payload_section = []
is_payload_section = False
outputfile.seek(0)
while True:
output = outputfile.readline()
if output == '':
break
if output.startswith('BEGIN_EXTRA_PAYLOAD'):
is_payload_section = True
output = outputfile.readline()
if output.startswith('END_EXTRA_PAYLOAD'):
is_payload_section = False
output = ''
payload = '\n'.join(payload_section)
msg = email.parser.Parser().parsestr(payload)
for part in msg.get_payload():
payload_result.update(json.loads(part.get_payload()))
if output and not is_payload_section:
logger.info(output.strip(), extra=extra)
results_log.append(output.strip())
else:
payload_section.append(output.strip())
def getExtraLogData(request=None):
extra = {'request_id' : '', 'subrequest_id' : '', 'originator_id': ''}
if request is not None:
extra = {
'request_id' : request.requestId,
'subrequest_id' : request.subRequestId,
'originator_id': request.originatorId
}
return extra
|