diff options
author | Haibin Huang <haibin.huang@intel.com> | 2019-01-28 14:00:38 +0800 |
---|---|---|
committer | Haibin Huang <haibin.huang@intel.com> | 2019-01-28 15:46:05 +0800 |
commit | f2138cc59231602b31e60a3b9f39cc59a1860892 (patch) | |
tree | 8b0eeff773d46b498fe371faa072c783621b663a /windriver/titanium_cloud | |
parent | bfa1edf6c3838a710ceedc27a7aef83ebbe4529d (diff) |
Extract some code to share
Change-Id: I6dffc1bf47c5839f077f3c7f93f021ed0b414d24
Issue-ID: MULTICLOUD-428
Signed-off-by: Haibin Huang <haibin.huang@intel.com>
Diffstat (limited to 'windriver/titanium_cloud')
9 files changed, 10 insertions, 1078 deletions
diff --git a/windriver/titanium_cloud/pub/config/log.yml b/windriver/titanium_cloud/pub/config/log.yml index 5cf448d4..ca3a5497 100644 --- a/windriver/titanium_cloud/pub/config/log.yml +++ b/windriver/titanium_cloud/pub/config/log.yml @@ -14,6 +14,10 @@ loggers: handlers: [titanium_cloud_handler] level: "DEBUG" propagate: False + starlingx_base: + handlers: [titanium_cloud_handler] + level: "DEBUG" + propagate: False handlers: titanium_cloud_handler: level: "DEBUG" diff --git a/windriver/titanium_cloud/registration/tests/test_registration2.py b/windriver/titanium_cloud/registration/tests/test_registration2.py index 9f0cd4d2..840f982f 100644 --- a/windriver/titanium_cloud/registration/tests/test_registration2.py +++ b/windriver/titanium_cloud/registration/tests/test_registration2.py @@ -18,17 +18,8 @@ import unittest # import json from django.test import Client from rest_framework import status - -# from django.core.cache import cache -# from common.msapi import extsys - from common.utils import restcall -# from newton_base.tests import mock_info -# from newton_base.tests import test_base -# from newton_base.util import VimDriverUtils - -# from newton_base.registration import registration as newton_registration -from titanium_cloud.registration.views import registration +from starlingx_base.registration import registration from newton_base.tests import test_base MOCK_VIM_INFO = { diff --git a/windriver/titanium_cloud/registration/views/__init__.py b/windriver/titanium_cloud/registration/views/__init__.py deleted file mode 100644 index ae1ce9db..00000000 --- a/windriver/titanium_cloud/registration/views/__init__.py +++ /dev/null @@ -1,13 +0,0 @@ -# Copyright (c) 2017-2018 Wind River Systems, Inc. -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. diff --git a/windriver/titanium_cloud/registration/views/registration.py b/windriver/titanium_cloud/registration/views/registration.py deleted file mode 100644 index 17992af7..00000000 --- a/windriver/titanium_cloud/registration/views/registration.py +++ /dev/null @@ -1,322 +0,0 @@ -# Copyright (c) 2017-2018 Wind River Systems, Inc. -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -import logging -import json -import uuid -import traceback - -from django.conf import settings - -from newton_base.registration import registration as newton_registration -# from common.exceptions import VimDriverNewtonException -from rest_framework import status -from rest_framework.response import Response -from common.msapi import extsys -from keystoneauth1.exceptions import HttpError -from newton_base.util import VimDriverUtils -from common.utils import restcall - -logger = logging.getLogger(__name__) - - -# DEBUG=True - - -class Registry(newton_registration.Registry): - def __init__(self): - super(Registry, self).__init__() - self.proxy_prefix = settings.MULTICLOUD_PREFIX - self.aai_base_url = settings.AAI_BASE_URL - # self._logger = logger - - def _get_ovsdpdk_capabilities(self, extra_specs, viminfo): - instruction_capability = {} - feature_uuid = uuid.uuid4() - - instruction_capability['hpa-capability-id'] = str(feature_uuid) - instruction_capability['hpa-feature'] = 'ovsDpdk' - instruction_capability['architecture'] = 'Intel64' - instruction_capability['hpa-version'] = 'v1' - - instruction_capability['hpa-feature-attributes'] = [] - instruction_capability['hpa-feature-attributes'].append( - {'hpa-attribute-key': 'dataProcessingAccelerationLibrary', - 'hpa-attribute-value': - '{{\"value\":\"{0}\"}}'.format("v17.02") - }) - return instruction_capability - - -class APIv1Registry(newton_registration.Registry): - def __init__(self): - super(APIv1Registry, self).__init__() - self.proxy_prefix = settings.MULTICLOUD_API_V1_PREFIX - self.aai_base_url = settings.AAI_BASE_URL - # self._logger = logger - - def _get_ovsdpdk_capabilities(self, extra_specs, viminfo): - instruction_capability = {} - feature_uuid = uuid.uuid4() - - instruction_capability['hpa-capability-id'] = str(feature_uuid) - instruction_capability['hpa-feature'] = 'ovsDpdk' - instruction_capability['architecture'] = 'Intel64' - instruction_capability['hpa-version'] = 'v1' - - instruction_capability['hpa-feature-attributes'] = [] - instruction_capability['hpa-feature-attributes'].append( - {'hpa-attribute-key': 'dataProcessingAccelerationLibrary', - 'hpa-attribute-value': - '{{\"value\":\"{0}\"}}'.format("v17.02") - }) - return instruction_capability - - def _update_cloud_region(self, cloud_owner, cloud_region_id, openstack_region_id, viminfo, session=None): - if cloud_owner and cloud_region_id: - self._logger.debug( - ("_update_cloud_region, %(cloud_owner)s" - "_%(cloud_region_id)s ") - % { - "cloud_owner": cloud_owner, - "cloud_region_id": cloud_region_id - }) - - # Note1: The intent is to populate the openstack region id into property: cloud-region.esr-system-info.openstackRegionId - # Note2: As temp solution: the openstack region id was put into AAI cloud-region["cloud-epa-caps"] - - resource_info = { - "cloud-owner": cloud_owner, - "cloud-region-id": cloud_region_id, - "cloud-type": viminfo["type"], - "cloud-region-version": viminfo["version"], - "identity-url": - self.proxy_prefix + "/%s_%s/identity/v2.0" % (cloud_owner, cloud_region_id) - if self.proxy_prefix[-3:] == "/v0" else - self.proxy_prefix + "/%s/%s/identity/v2.0" % (cloud_owner, cloud_region_id), - "complex-name": viminfo["complex-name"], - "cloud-extra-info": viminfo["cloud_extra_info"], - "cloud-epa-caps": openstack_region_id, - "esr-system-info-list": { - "esr-system-info": [ - { - "esr-system-info-id": str(uuid.uuid4()), - "service-url": viminfo["url"], - "user-name": viminfo["userName"], - "password": viminfo["password"], - "system-type": "VIM", - "ssl-cacert": viminfo["cacert"], - "ssl-insecure": viminfo["insecure"], - "cloud-domain": viminfo["domain"], - "default-tenant": viminfo["tenant"] - - } - ] - } - } - - # get the resource first - resource_url = ("/cloud-infrastructure/cloud-regions/" - "cloud-region/%(cloud_owner)s/%(cloud_region_id)s" - % { - "cloud_owner": cloud_owner, - "cloud_region_id": cloud_region_id - }) - - # get cloud-region - retcode, content, status_code = \ - restcall.req_to_aai(resource_url, "GET") - - # add resource-version - if retcode == 0 and content: - content = json.JSONDecoder().decode(content) - # resource_info["resource-version"] = content["resource-version"] - content.update(resource_info) - resource_info = content - - # then update the resource - retcode, content, status_code = \ - restcall.req_to_aai(resource_url, "PUT", content=resource_info) - - self._logger.debug( - ("_update_cloud_region,%(cloud_owner)s" - "_%(cloud_region_id)s , " - "return %(retcode)s, %(content)s, %(status_code)s") - % { - "cloud_owner": cloud_owner, - "cloud_region_id": cloud_region_id, - "retcode": retcode, - "content": content, - "status_code": status_code, - }) - - # wait and confirm the update has been available for next AAI calls - while True: - # get cloud-region - retcode2, content2, status_code2 = \ - restcall.req_to_aai(resource_url, "GET") - if retcode2 == 0 and content2: - content2 = json.JSONDecoder().decode(content2) - if content2.get("identity-url", None)\ - == resource_info.get("identity-url", None): - break - - return retcode - return 1 # unknown cloud owner,region_id - - def _discover_regions(self, cloud_owner="", cloud_region_id="", - session=None, viminfo=None): - try: - regions = [] - vimid = extsys.encode_vim_id(cloud_owner, cloud_region_id) - isDistributedCloud = False - openstackregions = self._get_list_resources( - "/regions", "identity", session, viminfo, vimid, - "regions") - - for region in openstackregions: - if region['id'] == 'SystemController': - isDistributedCloud = True - break - else: - continue - - for region in openstackregions: - if region['id'] == 'SystemController': - continue - elif region['id'] == 'RegionOne' and isDistributedCloud: - continue - else: - regions.append(region['id']) - - self._logger.info("Discovered Regions :%s" % regions) - return regions - - except HttpError as e: - self._logger.error("HttpError: status:%s, response:%s" - % (e.http_status, e.response.json())) - return [] - except Exception: - self._logger.error(traceback.format_exc()) - return [] - - def post(self, request, cloud_owner="", cloud_region_id=""): - self._logger.info("registration with : %s, %s" - % (cloud_owner, cloud_region_id)) - - try: - vimid = extsys.encode_vim_id(cloud_owner, cloud_region_id) - - viminfo = VimDriverUtils.get_vim_info(vimid) - cloud_extra_info_str = viminfo['cloud_extra_info'] - cloud_extra_info = None - try: - cloud_extra_info = json.loads(cloud_extra_info_str)\ - if cloud_extra_info_str else None - except Exception as ex: - logger.error("Can not convert cloud extra info %s %s" % ( - str(ex), cloud_extra_info_str)) - pass - - region_specified = cloud_extra_info.get( - "openstack-region-id", None) if cloud_extra_info else None - multi_region_discovery = cloud_extra_info.get( - "multi-region-discovery", None) if cloud_extra_info else None - - # set the default tenant since there is no tenant info in the VIM yet - sess = VimDriverUtils.get_session( - viminfo, tenant_name=viminfo['tenant']) - - # discover the regions, expect it always returns a list (even empty list) - region_ids = self._discover_regions(cloud_owner, cloud_region_id, sess, viminfo) - - if len(region_ids) == 0: - self._logger.warn("failed to get region id") - - # compare the regions with region_specified and then cloud_region_id - if region_specified in region_ids: - pass - elif cloud_region_id in region_ids: - region_specified = cloud_region_id - pass - else: - # assume the first region be the primary region - # since we have no other way to determine it. - region_specified = region_ids.pop(0) - - # update cloud region and discover/register resource - if multi_region_discovery: - # no input for specified cloud region, - # so discover all cloud region - for regionid in region_ids: - # do not update the specified region here - if region_specified == regionid: - continue - - # create cloud region with composed AAI cloud_region_id - # except for the one onboarded externally (e.g. ESR) - gen_cloud_region_id = cloud_region_id + "_" + regionid - self._logger.info("create a cloud region: %s,%s,%s" - % (cloud_owner, gen_cloud_region_id, regionid)) - - self._update_cloud_region( - cloud_owner, gen_cloud_region_id, regionid, viminfo) - new_vimid = extsys.encode_vim_id( - cloud_owner, gen_cloud_region_id) - super(APIv1Registry, self).post(request, new_vimid) - - # update the specified region - self._update_cloud_region(cloud_owner, cloud_region_id, - region_specified, viminfo) - return super(APIv1Registry, self).post(request, vimid) - - except HttpError as e: - self._logger.error("HttpError: status:%s, response:%s" - % (e.http_status, e.response.json())) - return Response(data=e.response.json(), status=e.http_status) - except Exception as e: - self._logger.error(traceback.format_exc()) - return Response( - data={'error': str(e)}, - status=status.HTTP_500_INTERNAL_SERVER_ERROR) - - def delete(self, request, cloud_owner="", cloud_region_id=""): - self._logger.debug("unregister cloud region: %s, %s" - % (cloud_owner, cloud_region_id)) - - vimid = extsys.encode_vim_id(cloud_owner, cloud_region_id) - return super(APIv1Registry, self).delete(request, vimid) - - -# APIv0 handler upgrading: leverage APIv1 handler -class APIv0Registry(APIv1Registry): - def __init__(self): - super(APIv0Registry, self).__init__() - self.proxy_prefix = settings.MULTICLOUD_PREFIX - self.aai_base_url = settings.AAI_BASE_URL - # self._logger = logger - - def post(self, request, vimid=""): - self._logger.info("registration with : %s" % vimid) - - cloud_owner, cloud_region_id = extsys.decode_vim_id(vimid) - return super(APIv0Registry, self).post( - request, cloud_owner, cloud_region_id) - - def delete(self, request, vimid=""): - self._logger.debug("unregister cloud region: %s" % vimid) - - cloud_owner, cloud_region_id = extsys.decode_vim_id(vimid) - return super(APIv0Registry, self).delete( - request, cloud_owner, cloud_region_id) diff --git a/windriver/titanium_cloud/resource/tests/tests_infra_workload.py b/windriver/titanium_cloud/resource/tests/tests_infra_workload.py index c6991837..cd8fe558 100644 --- a/windriver/titanium_cloud/resource/tests/tests_infra_workload.py +++ b/windriver/titanium_cloud/resource/tests/tests_infra_workload.py @@ -15,13 +15,11 @@ import mock import unittest -# import json from rest_framework import status -# from common.utils import restcall from common.msapi.helper import Helper as helper -from titanium_cloud.resource.views.infra_workload import InfraWorkload -from titanium_cloud.resource.views.infra_workload import APIv1InfraWorkload +from newton_base.resource.infra_workload import InfraWorkload +from newton_base.resource.infra_workload import APIv1InfraWorkload MOCK_TOKEN_RESPONSE = { "access": diff --git a/windriver/titanium_cloud/resource/views/__init__.py b/windriver/titanium_cloud/resource/views/__init__.py deleted file mode 100644 index ae1ce9db..00000000 --- a/windriver/titanium_cloud/resource/views/__init__.py +++ /dev/null @@ -1,13 +0,0 @@ -# Copyright (c) 2017-2018 Wind River Systems, Inc. -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. diff --git a/windriver/titanium_cloud/resource/views/capacity.py b/windriver/titanium_cloud/resource/views/capacity.py deleted file mode 100644 index 669cb659..00000000 --- a/windriver/titanium_cloud/resource/views/capacity.py +++ /dev/null @@ -1,159 +0,0 @@ -# Copyright (c) 2017-2018 Wind River Systems, Inc. -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -import logging -# import json -import traceback - -# from rest_framework import status - -# from django.conf import settings -from common.exceptions import VimDriverNewtonException -from newton_base.util import VimDriverUtils - -from keystoneauth1.exceptions import HttpError -from rest_framework import status -from rest_framework.response import Response -from rest_framework.views import APIView -from common.msapi import extsys - -logger = logging.getLogger(__name__) - - -class CapacityCheck(APIView): - def __init__(self): - self._logger = logger - - def post(self, request, vimid=""): - self._logger.info("vimid, data> %s, %s" % (vimid, request.data)) - self._logger.debug("META> %s" % request.META) - - hasEnoughResource = False - try: - resource_demand = request.data - - tenant_name = None - vim = VimDriverUtils.get_vim_info(vimid) - sess = VimDriverUtils.get_session(vim, tenant_name) - - # get token: - cloud_owner, regionid = extsys.decode_vim_id(vimid) - interface = 'public' - service = {'service_type': 'compute', - 'interface': interface, - 'region_name': vim['openstack_region_id'] - if vim.get('openstack_region_id') - else vim['cloud_region_id'] - } - - # get limit for this tenant - req_resouce = "/limits" - self._logger.info("check limits> URI:%s" % req_resouce) - resp = sess.get(req_resouce, endpoint_filter=service) - self._logger.info("check limits> status:%s" % resp.status_code) - content = resp.json() - compute_limits = content['limits']['absolute'] - self._logger.debug("check limits> resp data:%s" % content) - - # get total resource of this cloud region - try: - req_resouce = "/os-hypervisors/statistics" - self._logger.info("check os-hypervisors statistics> URI:%s" % req_resouce) - resp = sess.get(req_resouce, endpoint_filter=service) - self._logger.info("check os-hypervisors statistics> status:%s" % resp.status_code) - content = resp.json() - hypervisor_statistics = content['hypervisor_statistics'] - self._logger.debug("check os-hypervisors statistics> resp data:%s" % content) - except HttpError as e: - if e.http_status == status.HTTP_403_FORBIDDEN: - # Due to non administrator account cannot get hypervisor data, - # so construct enough resource data - conVCPUS = int(resource_demand['vCPU']) - conFreeRamMB = int(resource_demand['Memory']) - conFreeDiskGB = int(resource_demand['Storage']) - self._logger.info("Non administator forbidden to access hypervisor statistics data") - hypervisor_statistics = {'vcpus_used': 0, - 'vcpus': conVCPUS, - 'free_ram_mb': conFreeRamMB, - 'free_disk_gb': conFreeDiskGB} - else: - # non forbiden exeption will be redirected - raise e - - # get storage limit for this tenant - service['service_type'] = 'volumev2' - req_resouce = "/limits" - self._logger.info("check volumev2 limits> URI:%s" % req_resouce) - resp = sess.get(req_resouce, endpoint_filter=service) - self._logger.info("check volumev2> status:%s" % resp.status_code) - content = resp.json() - storage_limits = content['limits']['absolute'] - self._logger.debug("check volumev2> resp data:%s" % content) - - # compute actual available resource for this tenant - remainVCPU = compute_limits['maxTotalCores'] - compute_limits['totalCoresUsed'] - remainHypervisorVCPU = hypervisor_statistics['vcpus'] - hypervisor_statistics['vcpus_used'] - - if (remainVCPU > remainHypervisorVCPU): - remainVCPU = remainHypervisorVCPU - - remainMEM = compute_limits['maxTotalRAMSize'] - compute_limits['totalRAMUsed'] - remainHypervisorMEM = hypervisor_statistics['free_ram_mb'] - if remainMEM > remainHypervisorMEM: - remainMEM = remainHypervisorMEM - - remainStorage = storage_limits['maxTotalVolumeGigabytes'] - storage_limits['totalGigabytesUsed'] - remainHypervisorStorage = hypervisor_statistics['free_disk_gb'] - if (remainStorage > remainHypervisorStorage): - remainStorage = remainHypervisorStorage - - # compare resource demanded with available - if (int(resource_demand['vCPU']) > remainVCPU): - hasEnoughResource = False - elif (int(resource_demand['Memory']) > remainMEM): - hasEnoughResource = False - elif (int(resource_demand['Storage']) > remainStorage): - hasEnoughResource = False - else: - hasEnoughResource = True - - self._logger.info("RESP with data> result:%s" % hasEnoughResource) - return Response(data={'result': hasEnoughResource}, status=status.HTTP_200_OK) - except VimDriverNewtonException as e: - self._logger.error("Plugin exception> status:%s,error:%s" - % (e.status_code, e.content)) - return Response(data={'result': hasEnoughResource, - 'error': e.content}, status=e.status_code) - except HttpError as e: - self._logger.error("HttpError: status:%s, response:%s" % (e.http_status, e.response.json())) - resp = e.response.json() - resp.update({'result': hasEnoughResource}) - return Response(data=e.response.json(), status=e.http_status) - except Exception as e: - self._logger.error(traceback.format_exc()) - return Response(data={'result': hasEnoughResource, 'error': str(e)}, - status=status.HTTP_500_INTERNAL_SERVER_ERROR) - - -class APIv1CapacityCheck(CapacityCheck): - def __init__(self): - super(APIv1CapacityCheck, self).__init__() - # self._logger = logger - - def post(self, request, cloud_owner="", cloud_region_id=""): - self._logger.info("vimid, data> %s,%s, %s" % (cloud_owner, cloud_region_id, request.data)) - self._logger.debug("META> %s" % request.META) - - vimid = extsys.encode_vim_id(cloud_owner, cloud_region_id) - return super(APIv1CapacityCheck, self).post(request, vimid) diff --git a/windriver/titanium_cloud/resource/views/infra_workload.py b/windriver/titanium_cloud/resource/views/infra_workload.py deleted file mode 100644 index 74c187fe..00000000 --- a/windriver/titanium_cloud/resource/views/infra_workload.py +++ /dev/null @@ -1,554 +0,0 @@ -# Copyright (c) 2017-2018 Wind River Systems, Inc. -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -import logging -import json -import traceback - -from django.conf import settings -from common.exceptions import VimDriverNewtonException -# from newton_base.util import VimDriverUtils - -from keystoneauth1.exceptions import HttpError -from rest_framework import status -from rest_framework.response import Response -from rest_framework.views import APIView -from common.msapi import extsys -from common.msapi.helper import Helper as helper - -from common.utils import restcall - -logger = logging.getLogger(__name__) - - -class InfraWorkload(APIView): - def __init__(self): - self._logger = logger - - def post(self, request, vimid=""): - self._logger.info("vimid: %s" % (vimid)) - self._logger.info("data: %s" % (request.data)) - self._logger.debug("META: %s" % request.META) - - try: - - data = request.data - oof_directive = data.get("oof_directives", {}) - template_type = data.get("template_type", None) - template_data = data.get("template_data", {}) - resp_template = None - if template_type and "heat" == template_type.lower(): - # update heat parameters from oof_directive - parameters = template_data.get("parameters", {}) - - for directive in oof_directive.get("directives", []): - if directive["type"] == "vnfc": - for directive2 in directive.get("directives", []): - if directive2["type"] in ["flavor_directives", - "sriovNICNetwork_directives"]: - for attr in directive2.get("attributes", []): - flavor_label = attr.get("attribute_name", None) - flavor_value = attr.get("attribute_value", None) - if flavor_label in parameters: - parameters[flavor_label] = flavor_value - else: - self._logger.warn( - "There is no parameter exist: %s" % - flavor_label) - - # update parameters - template_data["parameters"] = parameters - - # reset to make sure "files" are empty - template_data["files"] = {} - - # authenticate - cloud_owner, regionid = extsys.decode_vim_id(vimid) - # should go via multicloud proxy so that - # the selflink is updated by multicloud - retcode, v2_token_resp_json, os_status = \ - helper.MultiCloudIdentityHelper( - settings.MULTICLOUD_API_V1_PREFIX, - cloud_owner, regionid, "/v2.0/tokens") - - if retcode > 0 or not v2_token_resp_json: - logger.error("authenticate fails:%s,%s, %s" % - (cloud_owner, regionid, v2_token_resp_json)) - return - # tenant_id = v2_token_resp_json["access"]["token"]["tenant"]["id"] - - service_type = "orchestration" - resource_uri = "/stacks" - self._logger.info("retrieve stack resources, URI:%s" % resource_uri) - retcode, content, os_status = \ - helper.MultiCloudServiceHelper(cloud_owner, regionid, - v2_token_resp_json, - service_type, resource_uri, - template_data, "POST") - - stack1 = content.get('stack', None) \ - if retcode == 0 and content else None - - resp_template = { - "template_type": template_type, - "workload_id": stack1["id"] if stack1 else "", - "template_response": content - } - self._logger.info("RESP with data> result:%s" % resp_template) - - return Response(data=resp_template, status=os_status) - - else: - msg = "The template type %s is not supported" % (template_type) - self._logger.warn(msg) - return Response(data={"error": msg}, - status=status.HTTP_500_INTERNAL_SERVER_ERROR) - - except VimDriverNewtonException as e: - self._logger.error("Plugin exception> status:%s,error:%s" - % (e.status_code, e.content)) - return Response(data={'error': e.content}, status=e.status_code) - except HttpError as e: - self._logger.error("HttpError: status:%s, response:%s" % - (e.http_status, e.response.json())) - return Response(data=e.response.json(), status=e.http_status) - except Exception as e: - self._logger.error(traceback.format_exc()) - return Response(data={'error': str(e)}, - status=status.HTTP_500_INTERNAL_SERVER_ERROR) - - def get(self, request, vimid="", requri=""): - self._logger.info("vimid,requri: %s, %s" % (vimid, requri)) - self._logger.debug("META: %s" % request.META) - - try: - # assume the workload_type is heat - stack_id = requri - cloud_owner, regionid = extsys.decode_vim_id(vimid) - # should go via multicloud proxy so that the selflink is updated by multicloud - retcode, v2_token_resp_json, os_status = \ - helper.MultiCloudIdentityHelper( - settings.MULTICLOUD_API_V1_PREFIX, - cloud_owner, regionid, "/v2.0/tokens") - - if retcode > 0 or not v2_token_resp_json: - logger.error("authenticate fails:%s, %s, %s" % - (cloud_owner, regionid, v2_token_resp_json)) - return - # tenant_id = v2_token_resp_json["access"]["token"]["tenant"]["id"] - # tenant_name = v2_token_resp_json["access"]["token"]["tenant"]["name"] - - # get stack status - service_type = "orchestration" - resource_uri = "/stacks?id=%s" % stack_id if stack_id else "/stacks" - self._logger.info("retrieve stack resources, URI:%s" % resource_uri) - retcode, content, os_status = \ - helper.MultiCloudServiceHelper(cloud_owner, regionid, - v2_token_resp_json, - service_type, resource_uri, - None, "GET") - - stacks = content.get('stacks', []) if retcode == 0 and content else [] - stack_status = stacks[0]["stack_status"] if len(stacks) > 0 else "" - - # stub response - resp_template = { - "template_type": "HEAT", - "workload_id": stack_id, - "workload_status": stack_status - } - - if retcode > 0: - # return error messsages - resp_template['workload_response'] = content - - if ('CREATE_COMPLETE' == stack_status): - self.heatbridge_update(request, vimid, stack_id) - - self._logger.info("RESP with data> result:%s" % resp_template) - return Response(data=resp_template, status=os_status) - except VimDriverNewtonException as e: - self._logger.error("Plugin exception> status:%s,error:%s" - % (e.status_code, e.content)) - return Response(data={'error': e.content}, status=e.status_code) - except HttpError as e: - self._logger.error("HttpError: status:%s, response:%s" % - (e.http_status, e.response.json())) - return Response(data=e.response.json(), status=e.http_status) - except Exception as e: - self._logger.error(traceback.format_exc()) - return Response(data={'error': str(e)}, - status=status.HTTP_500_INTERNAL_SERVER_ERROR) - - def delete(self, request, vimid="", requri=""): - self._logger.info("vimid,requri: %s, %s" % (vimid, requri)) - self._logger.debug("META: %s" % request.META) - - try: - if requri == "": - raise VimDriverNewtonException( - message="workload_id is not specified", - content="workload_id must be specified to delete the workload", - status_code=400) - - # assume the workload_type is heat - stack_id = requri - cloud_owner, regionid = extsys.decode_vim_id(vimid) - # should go via multicloud proxy so that - # the selflink is updated by multicloud - retcode, v2_token_resp_json, os_status = \ - helper.MultiCloudIdentityHelper( - settings.MULTICLOUD_API_V1_PREFIX, - cloud_owner, regionid, "/v2.0/tokens") - - if retcode > 0 or not v2_token_resp_json: - logger.error("authenticate fails:%s, %s, %s" % - (cloud_owner, regionid, v2_token_resp_json)) - return - # tenant_id = v2_token_resp_json["access"]["token"]["tenant"]["id"] - # tenant_name = v2_token_resp_json["access"]["token"]["tenant"]["name"] - - # get stack status - service_type = "orchestration" - resource_uri = "/stacks?id=%s" % stack_id if stack_id else "/stacks" - self._logger.info("retrieve stack resources, URI:%s" % resource_uri) - retcode, content, os_status = \ - helper.MultiCloudServiceHelper(cloud_owner, regionid, - v2_token_resp_json, - service_type, resource_uri, - None, "GET") - - stacks = content.get('stacks', []) \ - if retcode == 0 and content else [] - # assume there is at most 1 stack returned - # since it was filtered by id - stack1 = stacks[0] if stacks else None - stack_status = "" - - if stack1 and 'CREATE_COMPLETE' == stack1['stack_status']: - # delete the stack - resource_uri = "/stacks/%s/%s" % \ - (stack1['stack_name'], stack1['id']) - self._logger.info("delete stack, URI:%s" % resource_uri) - retcode, content, os_status = \ - helper.MultiCloudServiceHelper(cloud_owner, regionid, - v2_token_resp_json, - service_type, resource_uri, - None, "DELETE") - if retcode == 0: - stack_status = "DELETE_IN_PROCESS" - # and update AAI inventory by heatbridge-delete - self.heatbridge_delete(request, vimid, stack1['id']) - - # stub response - resp_template = { - "template_type": "HEAT", - "workload_id": stack_id, - "workload_status": stack_status - } - - if retcode > 0: - resp_template["workload_response"] = content - - self._logger.info("RESP with data> result:%s" % resp_template) - return Response(status=os_status) - except VimDriverNewtonException as e: - self._logger.error("Plugin exception> status:%s,error:%s" - % (e.status_code, e.content)) - return Response(data={'error': e.content}, status=e.status_code) - except HttpError as e: - self._logger.error("HttpError: status:%s, response:%s" % - (e.http_status, e.response.json())) - return Response(data=e.response.json(), status=e.http_status) - except Exception as e: - self._logger.error(traceback.format_exc()) - return Response(data={'error': str(e)}, - status=status.HTTP_500_INTERNAL_SERVER_ERROR) - - def heatbridge_update(self, request, vimid, stack_id): - ''' - update heat resource to AAI for the specified cloud region and tenant - The resources includes: vserver, vserver/l-interface, - :param request: - :param vimid: - :param stack_id: - :return: - ''' - - cloud_owner, regionid = extsys.decode_vim_id(vimid) - # should go via multicloud proxy so that the selflink is updated by multicloud - retcode, v2_token_resp_json, os_status = \ - helper.MultiCloudIdentityHelper(settings.MULTICLOUD_API_V1_PREFIX, - cloud_owner, regionid, "/v2.0/tokens") - if retcode > 0: - logger.error("authenticate fails:%s, %s, %s" % - (cloud_owner, regionid, v2_token_resp_json)) - - return None - tenant_id = v2_token_resp_json["access"]["token"]["tenant"]["id"] - # tenant_name = v2_token_resp_json["access"]["token"]["tenant"]["name"] - - # common prefix - aai_cloud_region = \ - "/cloud-infrastructure/cloud-regions/cloud-region/%s/%s/tenants/tenant/%s" \ - % (cloud_owner, regionid, tenant_id) - - # get stack resource - service_type = "orchestration" - resource_uri = "/stacks/%s/resources" % (stack_id) - self._logger.info("retrieve stack resources, URI:%s" % resource_uri) - retcode, content, os_status = \ - helper.MultiCloudServiceHelper(cloud_owner, regionid, - v2_token_resp_json, - service_type, resource_uri, - None, "GET") - - resources = content.get('resources', []) if retcode == 0 and content else [] - - # find and update resources - transactions = [] - for resource in resources: - if resource.get('resource_status', None) != "CREATE_COMPLETE": - continue - if resource.get('resource_type', None) == 'OS::Nova::Server': - # retrieve vserver details - service_type = "compute" - resource_uri = "/servers/%s" % (resource['physical_resource_id']) - self._logger.info("retrieve vserver detail, URI:%s" % resource_uri) - retcode, content, os_status = \ - helper.MultiCloudServiceHelper(cloud_owner, regionid, - v2_token_resp_json, - service_type, resource_uri, - None, "GET") - - self._logger.debug(" resp data:%s" % content) - vserver_detail = content.get('server', None) if retcode == 0 and content else None - if vserver_detail: - # compose inventory entry for vserver - vserver_link = "" - for link in vserver_detail['links']: - if link['rel'] == 'self': - vserver_link = link['href'] - break - pass - - # note: relationship-list to flavor/image is not be update yet - # note: volumes is not updated yet - # note: relationship-list to vnf will be handled somewhere else - aai_resource = { - 'body': { - 'vserver-name': vserver_detail['name'], - 'vserver-name2': vserver_detail['name'], - "vserver-id": vserver_detail['id'], - "vserver-selflink": vserver_link, - "prov-status": vserver_detail['status'] - }, - "uri": aai_cloud_region + "/vservers/vserver/%s" % (vserver_detail['id']) - } - - try: - # then update the resource - retcode, content, status_code = \ - restcall.req_to_aai(aai_resource['uri'], - "PUT", content=aai_resource['body']) - - if retcode == 0 and content: - content = json.JSONDecoder().decode(content) - self._logger.debug("AAI update %s response: %s" % - (aai_resource['uri'], content)) - except Exception: - self._logger.error(traceback.format_exc()) - pass - - aai_resource_transactions = {"put": [aai_resource]} - transactions.append(aai_resource_transactions) - # self._logger.debug("aai_resource :%s" % aai_resource_transactions) - pass - - for resource in resources: - if resource.get('resource_status', None) != "CREATE_COMPLETE": - continue - if resource.get('resource_type', None) == 'OS::Neutron::Port': - # retrieve vport details - service_type = "network" - resource_uri = "/v2.0/ports/%s" % (resource['physical_resource_id']) - self._logger.info("retrieve vport detail, URI:%s" % resource_uri) - retcode, content, os_status = \ - helper.MultiCloudServiceHelper(cloud_owner, regionid, - v2_token_resp_json, - service_type, resource_uri, - None, "GET") - - self._logger.debug(" resp data:%s" % content) - - vport_detail = content.get('port', None) if retcode == 0 and content else None - if vport_detail: - # compose inventory entry for vport - # note: l3-interface-ipv4-address-list, - # l3-interface-ipv6-address-list are not updated yet - # note: network-name is not update yet since the detail - # coming with network-id - aai_resource = { - "body": { - "interface-name": vport_detail['name'], - "interface-id": vport_detail['id'], - "macaddr": vport_detail['mac_address'] - }, - 'uri': - aai_cloud_region + "/vservers/vserver/%s/l-interfaces/l-interface/%s" - % (vport_detail['device_id'], vport_detail['name']) - } - try: - # then update the resource - retcode, content, status_code = \ - restcall.req_to_aai(aai_resource['uri'], "PUT", - content=aai_resource['body']) - - if retcode == 0 and content: - content = json.JSONDecoder().decode(content) - self._logger.debug("AAI update %s response: %s" % - (aai_resource['uri'], content)) - except Exception: - self._logger.error(traceback.format_exc()) - pass - - aai_resource_transactions = {"put": [aai_resource]} - transactions.append(aai_resource_transactions) - # self._logger.debug("aai_resource :%s" % aai_resource_transactions) - - pass - - aai_transactions = {"transactions": transactions} - self._logger.debug("aai_transactions :%s" % aai_transactions) - - return aai_transactions - - def heatbridge_delete(self, request, vimid, stack_id): - ''' - remove heat resource from AAI for the specified cloud region and tenant - The resources includes: vserver, vserver/l-interface, - :param request: - :param vimid: - :param stack_id: - :param tenant_id: - :return: - ''' - - # enumerate the resources - cloud_owner, regionid = extsys.decode_vim_id(vimid) - # should go via multicloud proxy so that the selflink is updated by multicloud - retcode, v2_token_resp_json, os_status = \ - helper.MultiCloudIdentityHelper(settings.MULTICLOUD_API_V1_PREFIX, - cloud_owner, regionid, "/v2.0/tokens") - if retcode > 0: - logger.error("authenticate fails:%s, %s, %s" % - (cloud_owner, regionid, v2_token_resp_json)) - return None - - tenant_id = v2_token_resp_json["access"]["token"]["tenant"]["id"] - # tenant_name = v2_token_resp_json["access"]["token"]["tenant"]["name"] - - # common prefix - aai_cloud_region = \ - "/cloud-infrastructure/cloud-regions/cloud-region/%s/%s/tenants/tenant/%s" \ - % (cloud_owner, regionid, tenant_id) - - # get stack resource - service_type = "orchestration" - resource_uri = "/stacks/%s/resources" % (stack_id) - self._logger.info("retrieve stack resources, URI:%s" % resource_uri) - retcode, content, os_status = \ - helper.MultiCloudServiceHelper(cloud_owner, regionid, - v2_token_resp_json, - service_type, resource_uri, - None, "GET") - resources = content.get('resources', []) \ - if retcode == 0 and content else [] - - vserver_list = [resource['physical_resource_id'] for resource in resources - if resource.get('resource_type', None) == 'OS::Nova::Server'] - - try: - # get list of vservers - vserver_list_url = aai_cloud_region + "/vservers?depth=all" - retcode, content, status_code = \ - restcall.req_to_aai(vserver_list_url, "GET") - if retcode > 0 or not content: - self._logger.debug("AAI get %s response: %s" % (vserver_list_url, content)) - return None - content = json.JSONDecoder().decode(content) - vservers = content['vserver'] - for vserver in vservers: - if vserver['vserver-id'] not in vserver_list: - continue - - try: - # iterate vport, except will be raised if no l-interface exist - for vport in vserver['l-interfaces']['l-interface']: - # delete vport - vport_delete_url = \ - aai_cloud_region + \ - "/vservers/vserver/%s/l-interfaces/l-interface/%s?resource-version=%s" \ - % (vserver['vserver-id'], vport['interface-name'], - vport['resource-version']) - - restcall.req_to_aai(vport_delete_url, "DELETE") - except Exception: - pass - - try: - # delete vserver - vserver_delete_url = \ - aai_cloud_region + \ - "/vservers/vserver/%s?resource-version=%s" \ - % (vserver['vserver-id'], vserver['resource-version']) - - restcall.req_to_aai(vserver_delete_url, "DELETE") - except Exception: - continue - - except Exception: - self._logger.error(traceback.format_exc()) - return None - pass - - -class APIv1InfraWorkload(InfraWorkload): - def __init__(self): - super(APIv1InfraWorkload, self).__init__() - # self._logger = logger - - def post(self, request, cloud_owner="", cloud_region_id=""): - # self._logger.info("cloud owner, cloud region id, data: %s,%s, %s" % - # (cloud_owner, cloud_region_id, request.data)) - # self._logger.debug("META: %s" % request.META) - - vimid = extsys.encode_vim_id(cloud_owner, cloud_region_id) - return super(APIv1InfraWorkload, self).post(request, vimid) - - def get(self, request, cloud_owner="", cloud_region_id="", requri=""): - # self._logger.info("cloud owner, cloud region id, data: %s,%s, %s" % - # (cloud_owner, cloud_region_id, request.data)) - # self._logger.debug("META: %s" % request.META) - - vimid = extsys.encode_vim_id(cloud_owner, cloud_region_id) - return super(APIv1InfraWorkload, self).get(request, vimid, requri) - - def delete(self, request, cloud_owner="", cloud_region_id="", requri=""): - # self._logger.info("cloud owner, cloud region id, data: %s,%s, %s" % - # (cloud_owner, cloud_region_id, request.data)) - # self._logger.debug("META: %s" % request.META) - - vimid = extsys.encode_vim_id(cloud_owner, cloud_region_id) - return super(APIv1InfraWorkload, self).delete(request, vimid, requri) diff --git a/windriver/titanium_cloud/urls.py b/windriver/titanium_cloud/urls.py index e953140e..f4adc10b 100644 --- a/windriver/titanium_cloud/urls.py +++ b/windriver/titanium_cloud/urls.py @@ -14,11 +14,11 @@ from django.conf.urls import include, url -from titanium_cloud.registration.views import registration +from starlingx_base.registration import registration from newton_base.openoapi import tenants -from titanium_cloud.resource.views import capacity +from newton_base.resource import capacity from titanium_cloud.vesagent import vesagent_ctrl -from titanium_cloud.resource.views import infra_workload +from newton_base.resource import infra_workload urlpatterns = [ url(r'^', include('titanium_cloud.swagger.urls')), |