diff options
author | Michael Lando <ml636r@att.com> | 2018-02-27 16:10:32 +0200 |
---|---|---|
committer | Michael Lando <ml636r@att.com> | 2018-02-27 16:12:11 +0200 |
commit | 121a7a375d1cea2764dfc34d82f4e2f252099a33 (patch) | |
tree | 8b56296d6a856007969b5a18a3a41a8fe5cc04bf /src/main | |
parent | 5e7130be2350a29d8dbc65f52cdca596a188114f (diff) |
Revert "vLAN Tagging - Support Policies"
This reverts commit 5e7130be2350a29d8dbc65f52cdca596a188114f.
Issue-ID: SDC-1056
Change-Id: I8526a8965e2a69c9d9189e6628cbc36e92282228
Signed-off-by: Michael Lando <ml636r@att.com>
Diffstat (limited to 'src/main')
3 files changed, 242 insertions, 53 deletions
diff --git a/src/main/java/org/openecomp/sdc/toscaparser/api/ImportsLoader.java b/src/main/java/org/openecomp/sdc/toscaparser/api/ImportsLoader.java index 5e94378..b2a0da7 100644 --- a/src/main/java/org/openecomp/sdc/toscaparser/api/ImportsLoader.java +++ b/src/main/java/org/openecomp/sdc/toscaparser/api/ImportsLoader.java @@ -28,6 +28,7 @@ public class ImportsLoader { private ArrayList<String> typeDefinitionList; private LinkedHashMap<String,Object> customDefs; + private LinkedHashMap<String,Object> allCustomDefs; private ArrayList<LinkedHashMap<String,Object>> nestedToscaTpls; private LinkedHashMap<String,Object> repositories; @@ -39,6 +40,7 @@ public class ImportsLoader { this.importslist = _importslist; customDefs = new LinkedHashMap<String,Object>(); + allCustomDefs = new LinkedHashMap<String,Object>(); nestedToscaTpls = new ArrayList<LinkedHashMap<String,Object>>(); if((_path == null || _path.isEmpty()) && tpl == null) { //msg = _('Input tosca template is not provided.') @@ -65,7 +67,7 @@ public class ImportsLoader { } public LinkedHashMap<String,Object> getCustomDefs() { - return customDefs; + return allCustomDefs; } public ArrayList<LinkedHashMap<String,Object>> getNestedToscaTpls() { @@ -131,33 +133,50 @@ public class ImportsLoader { } } - @SuppressWarnings("unchecked") + /** + * This method is used to get consolidated custom definitions by passing custom Types from + * each import. The resultant collection is then passed back which contains all import + * definitions + * + * @param customType the custom type + * @param namespacePrefix the namespace prefix + */ + @SuppressWarnings("unchecked") private void _updateCustomDefs(LinkedHashMap<String,Object> customType, String namespacePrefix) { - LinkedHashMap<String,Object> outerCustomTypes;// = new LinkedHashMap<String,Object>(); - for(String typeDef: typeDefinitionList) { - if(typeDef.equals("imports")) { - // imports are ArrayList... - customDefs.put("imports",(ArrayList<Object>)customType.get(typeDef)); - } - else { - outerCustomTypes = (LinkedHashMap<String,Object>)customType.get(typeDef); - if(outerCustomTypes != null) { - if(namespacePrefix != null && !namespacePrefix.isEmpty()) { - LinkedHashMap<String,Object> prefixCustomTypes = new LinkedHashMap<String,Object>(); - for(Map.Entry<String,Object> me: outerCustomTypes.entrySet()) { - String typeDefKey = me.getKey(); - String nameSpacePrefixToKey = namespacePrefix + "." + typeDefKey; - prefixCustomTypes.put(nameSpacePrefixToKey, outerCustomTypes.get(typeDefKey)); - } - customDefs.putAll(prefixCustomTypes); - } - else { - customDefs.putAll(outerCustomTypes); - } - } - } - } - } + LinkedHashMap<String,Object> outerCustomTypes; + for(String typeDef: typeDefinitionList) { + if(typeDef.equals("imports")) { + customDefs.put("imports", customType.get(typeDef)); + if (allCustomDefs.isEmpty() || allCustomDefs.get("imports") == null){ + allCustomDefs.put("imports",customType.get(typeDef)); + } + else if (customType.get(typeDef) != null){ + Set<Object> allCustomImports = new HashSet<>((ArrayList<Object>)allCustomDefs.get("imports")); + allCustomImports.addAll((ArrayList<Object>) customType.get(typeDef)); + allCustomDefs.put("imports", new ArrayList<>(allCustomImports)); + } + } + else { + outerCustomTypes = (LinkedHashMap<String,Object>)customType.get(typeDef); + if(outerCustomTypes != null) { + if(namespacePrefix != null && !namespacePrefix.isEmpty()) { + LinkedHashMap<String,Object> prefixCustomTypes = new LinkedHashMap<String,Object>(); + for(Map.Entry<String,Object> me: outerCustomTypes.entrySet()) { + String typeDefKey = me.getKey(); + String nameSpacePrefixToKey = namespacePrefix + "." + typeDefKey; + prefixCustomTypes.put(nameSpacePrefixToKey, outerCustomTypes.get(typeDefKey)); + } + customDefs.putAll(prefixCustomTypes); + allCustomDefs.putAll(prefixCustomTypes); + } + else { + customDefs.putAll(outerCustomTypes); + allCustomDefs.putAll(outerCustomTypes); + } + } + } + } + } private void _updateNestedToscaTpls(String fullFileName,LinkedHashMap<String,Object> customTpl) { if(fullFileName != null && customTpl != null) { diff --git a/src/main/java/org/openecomp/sdc/toscaparser/api/Policy.java b/src/main/java/org/openecomp/sdc/toscaparser/api/Policy.java index 1f536f8..26805bd 100644 --- a/src/main/java/org/openecomp/sdc/toscaparser/api/Policy.java +++ b/src/main/java/org/openecomp/sdc/toscaparser/api/Policy.java @@ -119,12 +119,6 @@ public class Policy extends EntityTemplate { ", properties=" + properties + '}'; } - - public int compareTo(Policy other){ - if(this.equals(other)) - return 0; - return this.getName().compareTo(other.getName()) == 0 ? this.getType().compareTo(other.getType()) : this.getName().compareTo(other.getName()); - } } /*python diff --git a/src/main/java/org/openecomp/sdc/toscaparser/api/ToscaTemplate.java b/src/main/java/org/openecomp/sdc/toscaparser/api/ToscaTemplate.java index 7553414..e96ca56 100644 --- a/src/main/java/org/openecomp/sdc/toscaparser/api/ToscaTemplate.java +++ b/src/main/java/org/openecomp/sdc/toscaparser/api/ToscaTemplate.java @@ -9,6 +9,9 @@ import java.io.IOException; import java.io.InputStream; import java.util.*; import java.util.concurrent.ConcurrentHashMap; +import java.nio.file.Files; +import java.util.function.Predicate; +import java.nio.file.Paths; import org.openecomp.sdc.toscaparser.api.common.ValidationIssueCollector; import org.openecomp.sdc.toscaparser.api.common.JToscaException; @@ -70,6 +73,7 @@ public class ToscaTemplate extends Object { private boolean isFile; private String path; private String inputPath; + private String rootPath; private LinkedHashMap<String,Object> parsedParams; private boolean resolveGetInput; private LinkedHashMap<String,Object> tpl; @@ -91,6 +95,7 @@ public class ToscaTemplate extends Object { private String csarTempDir; private int nestingLoopCounter; private LinkedHashMap<String, LinkedHashMap<String, Object>> metaProperties; + private Set<String> processedImports; public ToscaTemplate(String _path, LinkedHashMap<String,Object> _parsedParams, @@ -193,6 +198,9 @@ public class ToscaTemplate extends Object { if(tpl != null) { parsedParams = _parsedParams; _validateField(); + this.rootPath = path; + this.processedImports = new HashSet<String>(); + this.imports = _tplImports(); this.version = _tplVersion(); this.metaData = _tplMetaData(); this.relationshipTypes = _tplRelationshipTypes(); @@ -305,30 +313,200 @@ public class ToscaTemplate extends Object { private ArrayList<Policy> _policies() { return topologyTemplate.getPolicies(); } - - private LinkedHashMap<String,Object> _getAllCustomDefs(ArrayList<Object> alImports) { - + + /** + * This method is used to get consolidated custom definitions from all imports + * It is logically divided in two parts to handle imports; map and list formats. + * Before processing the imports; it sorts them to make sure the current directory imports are + * being processed first and then others. Once sorted; it processes each import one by one in + * recursive manner. + * To avoid cyclic dependency among imports; this method uses a set to keep track of all + * imports which are already processed and filters the imports which occurs more than once. + * + * @param alImports all imports which needs to be processed + * @return the linked hash map containing all import definitions + */ + private LinkedHashMap<String,Object> _getAllCustomDefs(Object alImports) { + String types[] = { - IMPORTS, NODE_TYPES, CAPABILITY_TYPES, RELATIONSHIP_TYPES, - DATA_TYPES, INTERFACE_TYPES, POLICY_TYPES, GROUP_TYPES + IMPORTS, NODE_TYPES, CAPABILITY_TYPES, RELATIONSHIP_TYPES, + DATA_TYPES, INTERFACE_TYPES, POLICY_TYPES, GROUP_TYPES }; - LinkedHashMap<String,Object> customDefsFinal = new LinkedHashMap<String,Object>(); - LinkedHashMap<String,Object> customDefs = _getCustomTypes(types,alImports); - if(customDefs != null) { - customDefsFinal.putAll(customDefs); - if(customDefs.get(IMPORTS) != null) { - @SuppressWarnings("unchecked") - LinkedHashMap<String,Object> importDefs = _getAllCustomDefs((ArrayList<Object>)customDefs.get(IMPORTS)); - customDefsFinal.putAll(importDefs); + LinkedHashMap<String,Object> customDefsFinal = new LinkedHashMap<>(); + + List<Map<String, Object>> imports = (List<Map<String, Object>>) alImports; + if (imports != null && !imports.isEmpty()) { + if (imports.get(0) instanceof LinkedHashMap) { + imports = sortImports(imports); + + for (Map<String, Object> map : imports) { + List<Map<String, Object>> singleImportList = new ArrayList(); + singleImportList.add(map); + + Map<String, String> importNameDetails = getValidFileNameForImportReference(singleImportList); + singleImportList = filterImportsForRecursion(singleImportList, importNameDetails); + + if(!singleImportList.get(0).isEmpty()){ + LinkedHashMap<String, Object> customDefs = _getCustomTypes(types, new ArrayList<>(singleImportList)); + processedImports.add(importNameDetails.get("importFileName")); + + if (customDefs != null) { + customDefsFinal.putAll(customDefs); + + if (customDefs.get(IMPORTS) != null) { + resetPathForRecursiveImports(importNameDetails.get("importRelativeName")); + LinkedHashMap<String, Object> importDefs = _getAllCustomDefs(customDefs.get(IMPORTS)); + customDefsFinal.putAll(importDefs); + } + } + } + } + } else { + LinkedHashMap<String, Object> customDefs = _getCustomTypes(types, new ArrayList<>(imports)); + if (customDefs != null) { + customDefsFinal.putAll(customDefs); + + if (customDefs.get(IMPORTS) != null) { + LinkedHashMap<String, Object> importDefs = _getAllCustomDefs(customDefs.get(IMPORTS)); + customDefsFinal.putAll(importDefs); + } + } } } - - // As imports are not custom_types, remove from the dict - customDefsFinal.remove(IMPORTS); + + // As imports are not custom_types, remove from the dict + customDefsFinal.remove(IMPORTS); return customDefsFinal; } + /** + * This method is used to sort the imports in order so that same directory + * imports will be processed first + * + * @param customImports the custom imports + * @return the sorted list of imports + */ + private List<Map<String, Object>> sortImports(List<Map<String, Object>> customImports){ + List<Map<String, Object>> finalList1 = new ArrayList<>(); + List<Map<String, Object>> finalList2 = new ArrayList<>(); + Iterator<Map<String, Object>> itr = customImports.iterator(); + while(itr.hasNext()) { + Map innerMap = itr.next(); + if (innerMap.toString().contains("../")) { + finalList2.add(innerMap); + itr.remove(); + } + else if (innerMap.toString().contains("/")) { + finalList1.add(innerMap); + itr.remove(); + } + } + + customImports.addAll(finalList1); + customImports.addAll(finalList2); + return customImports; + } + + /** + * This method is used to reset PATH variable after processing of current import file is done + * This is required because of relative path nature of imports present in files. + * + * @param currImportRelativeName the current import relative name + */ + private void resetPathForRecursiveImports(String currImportRelativeName){ + path = getPath(path, currImportRelativeName); + } + + /** + * This is a recursive method which starts from current import and then recursively finds a + * valid path relative to current import file name. + * By doing this it handles all nested hierarchy of imports defined in CSARs + * + * @param path the path + * @param importFileName the import file name + * @return the string containing updated path value + */ + private String getPath(String path, String importFileName){ + String tempFullPath = (Paths.get(path).toAbsolutePath().getParent() + .toString() + File.separator + importFileName.replace("../", "")).replace('\\', '/'); + String tempPartialPath = (Paths.get(path).toAbsolutePath().getParent().toString()).replace('\\', '/'); + if(Files.exists(Paths.get(tempFullPath))) + return tempFullPath; + else + return getPath(tempPartialPath, importFileName); + } + + /** + * This method is used to get full path name for the file which needs to be processed. It helps + * in situation where files are present in different directory and are references as relative + * paths. + * + * @param customImports the custom imports + * @return the map containing import file full and relative paths + */ + private Map<String, String> getValidFileNameForImportReference(List<Map<String, Object>> + customImports){ + String importFileName; + Map<String, String> retMap = new HashMap<>(); + for (Map<String, Object> map1 : customImports) { + for (Map.Entry<String, Object> entry : map1.entrySet()) { + Map innerMostMap = (Map) entry.getValue(); + Iterator<Map.Entry<String, String>> it = innerMostMap.entrySet().iterator(); + while (it.hasNext()) { + Map.Entry<String, String> val = it.next(); + if(val.getValue().contains("/")){ + importFileName = (Paths.get(rootPath).toAbsolutePath().getParent().toString() + File + .separator + val.getValue().replace("../", "")).replace('\\', '/'); + } + else { + importFileName = (Paths.get(path).toAbsolutePath().getParent().toString() + File + .separator + val.getValue().replace("../", "")).replace('\\', '/'); + } + retMap.put("importFileName", importFileName); + retMap.put("importRelativeName", val.getValue()); + } + } + } + return retMap; + } + + /** + * This method is used to filter the imports which already gets processed in previous step. + * It handles the use case of cyclic dependency in imports which may cause Stack Overflow + * exception + * + * @param customImports the custom imports + * @param importNameDetails the import name details + * @return the list containing filtered imports + */ + private List<Map<String, Object>> filterImportsForRecursion(List<Map<String, Object>> + customImports, Map<String, + String> importNameDetails){ + for (Map<String, Object> map1 : customImports) { + for (Map.Entry<String, Object> entry : map1.entrySet()) { + Map innerMostMap = (Map) entry.getValue(); + Iterator<Map.Entry<String, String>> it = innerMostMap.entrySet().iterator(); + while (it.hasNext()) { + it.next(); + if (processedImports.contains(importNameDetails.get("importFileName"))) { + it.remove(); + } + } + } + } + + // Remove Empty elements + Iterator<Map<String, Object>> itr = customImports.iterator(); + while(itr.hasNext()) { + Map innerMap = itr.next(); + Predicate<Map> predicate = p-> p.values().isEmpty(); + innerMap.values().removeIf(predicate); + } + + return customImports; + } + @SuppressWarnings("unchecked") private LinkedHashMap<String,Object> _getCustomTypes(Object typeDefinitions,ArrayList<Object> alImports) { @@ -396,6 +574,8 @@ public class ToscaTemplate extends Object { log.error("ToscaTemplate - _handleNestedToscaTemplatesWithTopology - Nested Topologies Loop: too many levels, aborting"); return; } + // Reset Processed Imports for nested templates + this.processedImports = new HashSet<>(); for(Map.Entry<String,Object> me: nestedToscaTplsWithTopology.entrySet()) { String fname = me.getKey(); LinkedHashMap<String,Object> toscaTpl = @@ -656,10 +836,6 @@ public class ToscaTemplate extends Object { return nestedToscaTemplatesWithTopology; } - public ConcurrentHashMap<String, Object> getNestedTopologyTemplates() { - return nestedToscaTplsWithTopology; - } - @Override public String toString() { return "ToscaTemplate{" + |