aboutsummaryrefslogtreecommitdiffstats
path: root/asdctool/src/main/java/org
diff options
context:
space:
mode:
authorChris André <chris.andre@yoppworks.com>2020-04-15 20:05:32 -0400
committerOfir Sonsino <ofir.sonsino@intl.att.com>2020-04-20 07:17:49 +0000
commita0c6e45f2085ba1f26573f5823d37b55268a1191 (patch)
treefc3ed0e0bdaf3012db800741b02afda01c04f6df /asdctool/src/main/java/org
parent659556d07ad5b928cb0032f040dc86fcb11944c8 (diff)
Added a try-with-resource to properly close a FileInputStream
Issue-ID: SDC-2898 Signed-off-by: Chris Andre <chris.andre@yoppworks.com> Change-Id: I0a0163937ee2e760f3074d858dfc68a7b262ecd9
Diffstat (limited to 'asdctool/src/main/java/org')
-rw-r--r--asdctool/src/main/java/org/openecomp/sdc/asdctool/main/SdcSchemaFileImport.java455
1 files changed, 234 insertions, 221 deletions
diff --git a/asdctool/src/main/java/org/openecomp/sdc/asdctool/main/SdcSchemaFileImport.java b/asdctool/src/main/java/org/openecomp/sdc/asdctool/main/SdcSchemaFileImport.java
index 676c3b45c0..57939df910 100644
--- a/asdctool/src/main/java/org/openecomp/sdc/asdctool/main/SdcSchemaFileImport.java
+++ b/asdctool/src/main/java/org/openecomp/sdc/asdctool/main/SdcSchemaFileImport.java
@@ -7,9 +7,9 @@
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
- *
+ *
* http://www.apache.org/licenses/LICENSE-2.0
- *
+ *
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
@@ -55,228 +55,241 @@ import java.util.zip.ZipOutputStream;
public class SdcSchemaFileImport {
-
- private static final String SEPARATOR = FileSystems.getDefault().getSeparator();
-
- private static final String TOSCA_VERSION = "tosca_simple_yaml_1_1";
-
- private static String importToscaPath;
-
- private static final byte[] buffer = new byte[1024];
-
- private static final String YAML_EXTENSION = ".yml";
-
- private static final String DEPLOYMENT_TYPE_ONAP = "onap";
-
- private static String LICENSE_TXT;
-
- private static ZipOutputStream zos;
-
- public static void main(String[] args) throws Exception {
-
- //Generation flow start - generating SDC from normatives
- System.out.println("Starting SdcSchemaFileImport procedure...");
- final String FILE_NAME = "SDC.zip";
-
- if (args == null || !(args.length ==4 || args.length == 5 )) {
- usageAndExit();
- }
-
- importToscaPath = args[0];
- String sdcReleaseNum = args[1];
- String conformanceLevel = args[2];
- String appConfigDir = args[3];
- String deploymentType=null;
- if(args.length==5){
- deploymentType=args[4];
- }
-
-
- ByteArrayOutputStream baos = new ByteArrayOutputStream();
-
- zos = new ZipOutputStream(baos);
-
- //Initialize the license text
- try {
- LICENSE_TXT = new String(Files.readAllBytes(Paths.get(appConfigDir + SEPARATOR+"license.txt")));
- }
- catch(Exception e) {
- System.err.println("Couldn't read license.txt in location :" + appConfigDir+", error: "+e);
- System.exit(1);
- }
-
- //Loop over schema file list and create each yaml file from /import/tosca folder
- SchemaZipFileEnum[] schemaFileList = SchemaZipFileEnum.values();
- for (SchemaZipFileEnum schemaZipFileEnum : schemaFileList) {
- try {
- //get the source yaml file
- String pathname = importToscaPath + SEPARATOR + schemaZipFileEnum.getSourceFolderName() + SEPARATOR + schemaZipFileEnum.getSourceFileName() + YAML_EXTENSION;
- System.out.println("Processing file "+pathname+"....");
- InputStream input = new FileInputStream(new File(pathname));
- //Convert the content of file to yaml
- Yaml yamlFileSource = new Yaml();
- Object content = yamlFileSource.load(input);
-
- createAndSaveSchemaFileYaml(schemaZipFileEnum, content);
- }
- catch(Exception e) {
- System.err.println("Error in file creation : " + schemaZipFileEnum.getFileName() + ", " + e.getMessage());
- System.exit(1);
- }
- }
-
- createAndSaveNodeSchemaFile(deploymentType);
-
- try {
- //close the ZipOutputStream
- zos.close();
- System.out.println("File SDC.zip creation successful");
-
- } catch(Exception ex) {
- System.err.println("Failed to pack SDC.zip file, error: "+ex);
- System.exit(1);
- }
-
- //Generation flow end - generating SDC from narratives
-
- AnnotationConfigApplicationContext context = initContext(appConfigDir);
- SdcSchemaFilesCassandraDao schemaFilesCassandraDao = (SdcSchemaFilesCassandraDao) context.getBean("sdc-schema-files-cassandra-dao");
-
- byte[] fileBytes = baos.toByteArray();
-
- Date date = new Date();
- String md5Hex = DigestUtils.md5Hex(fileBytes);
-
- SdcSchemaFilesData schemeFileData = new SdcSchemaFilesData(sdcReleaseNum, date, conformanceLevel, FILE_NAME, fileBytes, md5Hex);
- CassandraOperationStatus saveSchemaFile = schemaFilesCassandraDao.saveSchemaFile(schemeFileData);
-
- if(!saveSchemaFile.equals(CassandraOperationStatus.OK)) {
- System.err.println("SdcSchemaFileImport failed cassandra error" + saveSchemaFile);
- System.exit(1);
- }
-
- System.out.println("SdcSchemaFileImport successfully completed");
-
- System.exit(0);
- }
-
- public static void createAndSaveSchemaFileYaml(SchemaZipFileEnum schemaZipFileEnum, Object content) {
- createAndSaveSchemaFileYaml(schemaZipFileEnum.getFileName(), schemaZipFileEnum.getImportFileList(), schemaZipFileEnum.getCollectionTitle(), content);
- }
-
- public static void createAndSaveSchemaFileYaml(String fileName, String[] importFileList, String collectionTitle, Object content) {
-
- //Initialize the snake yaml dumper option
- DumperOptions options = new DumperOptions();
- options.setDefaultFlowStyle(DumperOptions.FlowStyle.BLOCK);
-
- //Create the new yaml
- Yaml yaml = new Yaml(options);
- yaml.setName(fileName);
-
- //Initialize the yaml contents
- final Map<String, Object> data = new LinkedHashMap<>();
- data.put(ToscaTagNamesEnum.TOSCA_VERSION.getElementName(), TOSCA_VERSION);
-
- if (importFileList.length > 0) {
- data.put("imports", importFileList);
- }
-
- data.put(collectionTitle, content);
-
- //Save the new yaml to file
- try {
-
- FileWriter writer;
- File file = File.createTempFile(fileName, YAML_EXTENSION);
- writer = new FileWriter(file);
-
- //Add the license as comment in top of file
- writer.write(LICENSE_TXT);
-
- yaml.dump(data, writer);
-
- writer.close();
-
- // begin writing a new ZIP entry, positions the stream to the start of the entry data
- ZipEntry entry = new ZipEntry(yaml.getName() + YAML_EXTENSION);
- zos.putNextEntry(entry);
- FileInputStream stream = new FileInputStream(file.getAbsolutePath());
- int len;
- while ((len = stream.read(buffer)) > 0) {
- zos.write(buffer, 0, len);
- }
- //close the InputStream
+
+ private static final String SEPARATOR = FileSystems.getDefault().getSeparator();
+
+ private static final String TOSCA_VERSION = "tosca_simple_yaml_1_1";
+
+ private static String importToscaPath;
+
+ private static final byte[] buffer = new byte[1024];
+
+ private static final String YAML_EXTENSION = ".yml";
+
+ private static final String DEPLOYMENT_TYPE_ONAP = "onap";
+
+ private static String LICENSE_TXT;
+
+ private static ZipOutputStream zos;
+
+ public static void main(String[] args) throws Exception {
+
+ //Generation flow start - generating SDC from normatives
+ System.out.println("Starting SdcSchemaFileImport procedure...");
+ final String FILE_NAME = "SDC.zip";
+
+ if (args == null || !(args.length == 4 || args.length == 5)) {
+ usageAndExit();
+ }
+
+ importToscaPath = args[0];
+ String sdcReleaseNum = args[1];
+ String conformanceLevel = args[2];
+ String appConfigDir = args[3];
+ String deploymentType = null;
+ if (args.length == 5) {
+ deploymentType = args[4];
+ }
+
+ ByteArrayOutputStream baos = new ByteArrayOutputStream();
+
+ zos = new ZipOutputStream(baos);
+
+ //Initialize the license text
+ try {
+ LICENSE_TXT = new String(Files.readAllBytes(Paths.get(appConfigDir + SEPARATOR + "license.txt")));
+ } catch (Exception e) {
+ System.err.println("Couldn't read license.txt in location :" + appConfigDir + ", error: " + e);
+ System.exit(1);
+ }
+
+ //Loop over schema file list and create each yaml file from /import/tosca folder
+ SchemaZipFileEnum[] schemaFileList = SchemaZipFileEnum.values();
+ for (SchemaZipFileEnum schemaZipFileEnum : schemaFileList) {
+ String folderName = schemaZipFileEnum.getSourceFolderName();
+ String fileName = schemaZipFileEnum.getFileName();
+
+ if ((folderName != null) && (fileName != null)) {
+ File folder = new File(importToscaPath, folderName);
+ File path = new File(folder, fileName + YAML_EXTENSION);
+
+ try (InputStream input = new FileInputStream(path)) {
+ // Convert the content of file to yaml
+ Yaml yamlFileSource = new Yaml();
+ Object content = yamlFileSource.load(input);
+
+ createAndSaveSchemaFileYaml(schemaZipFileEnum, content);
+ } catch (Exception e) {
+ System.err.println(
+ "Error in file creation : "
+ + schemaZipFileEnum.getFileName()
+ + ", "
+ + e.getMessage());
+ System.exit(1);
+ }
+ }
+ }
+
+ createAndSaveNodeSchemaFile(deploymentType);
+
+ try {
+ //close the ZipOutputStream
+ zos.close();
+ System.out.println("File SDC.zip creation successful");
+
+ } catch (Exception ex) {
+ System.err.println("Failed to pack SDC.zip file, error: " + ex);
+ System.exit(1);
+ }
+
+ //Generation flow end - generating SDC from narratives
+
+ AnnotationConfigApplicationContext context = initContext(appConfigDir);
+ SdcSchemaFilesCassandraDao schemaFilesCassandraDao = (SdcSchemaFilesCassandraDao) context
+ .getBean("sdc-schema-files-cassandra-dao");
+
+ byte[] fileBytes = baos.toByteArray();
+
+ Date date = new Date();
+ String md5Hex = DigestUtils.md5Hex(fileBytes);
+
+ SdcSchemaFilesData schemeFileData = new SdcSchemaFilesData(sdcReleaseNum, date, conformanceLevel, FILE_NAME,
+ fileBytes, md5Hex);
+ CassandraOperationStatus saveSchemaFile = schemaFilesCassandraDao.saveSchemaFile(schemeFileData);
+
+ if (!saveSchemaFile.equals(CassandraOperationStatus.OK)) {
+ System.err.println("SdcSchemaFileImport failed cassandra error" + saveSchemaFile);
+ System.exit(1);
+ }
+
+ System.out.println("SdcSchemaFileImport successfully completed");
+
+ System.exit(0);
+ }
+
+ public static void createAndSaveSchemaFileYaml(SchemaZipFileEnum schemaZipFileEnum, Object content) {
+ createAndSaveSchemaFileYaml(schemaZipFileEnum.getFileName(), schemaZipFileEnum.getImportFileList(),
+ schemaZipFileEnum.getCollectionTitle(), content);
+ }
+
+ public static void createAndSaveSchemaFileYaml(String fileName, String[] importFileList, String collectionTitle,
+ Object content) {
+
+ //Initialize the snake yaml dumper option
+ DumperOptions options = new DumperOptions();
+ options.setDefaultFlowStyle(DumperOptions.FlowStyle.BLOCK);
+
+ //Create the new yaml
+ Yaml yaml = new Yaml(options);
+ yaml.setName(fileName);
+
+ //Initialize the yaml contents
+ final Map<String, Object> data = new LinkedHashMap<>();
+ data.put(ToscaTagNamesEnum.TOSCA_VERSION.getElementName(), TOSCA_VERSION);
+
+ if (importFileList.length > 0) {
+ data.put("imports", importFileList);
+ }
+
+ data.put(collectionTitle, content);
+
+ //Save the new yaml to file
+ try {
+
+ FileWriter writer;
+ File file = File.createTempFile(fileName, YAML_EXTENSION);
+ writer = new FileWriter(file);
+
+ //Add the license as comment in top of file
+ writer.write(LICENSE_TXT);
+
+ yaml.dump(data, writer);
+
+ writer.close();
+
+ // begin writing a new ZIP entry, positions the stream to the start of the entry data
+ ZipEntry entry = new ZipEntry(yaml.getName() + YAML_EXTENSION);
+ zos.putNextEntry(entry);
+ FileInputStream stream = new FileInputStream(file.getAbsolutePath());
+ int len;
+ while ((len = stream.read(buffer)) > 0) {
+ zos.write(buffer, 0, len);
+ }
+ //close the InputStream
file.delete();
stream.close();
- zos.closeEntry();
-
-
- } catch (IOException e) {
- System.out.println("Error in file creation : " + fileName + ", " + e.getMessage());
- System.exit(1);
- }
- }
-
- /**
- *the method is responsible for creating and storing the sdc normatives in the DB
- * @param deploymentType if the deployments type is onap the onap narratives will be add to the zip
- * @throws IOException thrown in case of issues in reding files.
- */
- public static void createAndSaveNodeSchemaFile(String deploymentType) throws IOException {
-
- //Initialize the snake yaml dumper option
- DumperOptions options = new DumperOptions();
- options.setDefaultFlowStyle(DumperOptions.FlowStyle.BLOCK);
-
- Map<String, Object> nodeTypeList = new LinkedHashMap<>();
-
- String[] importFileList = new String[]{"data.yml", "artifacts.yml", "capabilities.yml", "interfaces.yml", "relationships.yml"};
- String collectionTitle = "node_types";
-
- //Create node.yaml - collect all types from normative-types and heat-types directories
- String[] nodeTypesMainFolders = new String[]{"normative-types", "heat-types"};
-
- if(DEPLOYMENT_TYPE_ONAP.equals(deploymentType)){
+ zos.closeEntry();
+
+
+ } catch (IOException e) {
+ System.out.println("Error in file creation : " + fileName + ", " + e.getMessage());
+ System.exit(1);
+ }
+ }
+
+ /**
+ * the method is responsible for creating and storing the sdc normatives in the DB
+ *
+ * @param deploymentType if the deployments type is onap the onap narratives will be add to the zip
+ * @throws IOException thrown in case of issues in reding files.
+ */
+ public static void createAndSaveNodeSchemaFile(String deploymentType) throws IOException {
+
+ //Initialize the snake yaml dumper option
+ DumperOptions options = new DumperOptions();
+ options.setDefaultFlowStyle(DumperOptions.FlowStyle.BLOCK);
+
+ Map<String, Object> nodeTypeList = new LinkedHashMap<>();
+
+ String[] importFileList = new String[]{"data.yml", "artifacts.yml", "capabilities.yml", "interfaces.yml",
+ "relationships.yml"};
+ String collectionTitle = "node_types";
+
+ //Create node.yaml - collect all types from normative-types and heat-types directories
+ String[] nodeTypesMainFolders = new String[]{"normative-types", "heat-types"};
+
+ if (DEPLOYMENT_TYPE_ONAP.equals(deploymentType)) {
String[] onapNodeTypesMainFolders = new String[]{"nfv-types"};
- nodeTypesMainFolders=ArrayUtils.addAll(nodeTypesMainFolders,onapNodeTypesMainFolders);
- }
-
- for (String nodeTypesMainFolder : nodeTypesMainFolders) {
- try (Stream<Path> paths = Files.walk(Paths.get(importToscaPath + SEPARATOR + nodeTypesMainFolder))) {
- paths.filter(path -> path.getFileName().toString().toLowerCase().endsWith(YAML_EXTENSION))
- .forEach(yamlFile -> {
- try {
- String path = yamlFile.toAbsolutePath().toString();
- System.out.println("Processing node type file " + path + "...");
- FileInputStream inputStream = new FileInputStream(path);
- Yaml yaml = new Yaml();
- Map<String, Object> load = yaml.loadAs(inputStream, Map.class);
- Map<String, Object> nodeType = (Map<String, Object>) load.get(collectionTitle);
- nodeTypeList.putAll(nodeType);
-
- } catch (Exception e) {
- System.err.println("Error in opening file " + yamlFile.toAbsolutePath().toString());
- System.exit(1);
- }
- });
+ nodeTypesMainFolders = ArrayUtils.addAll(nodeTypesMainFolders, onapNodeTypesMainFolders);
}
+
+ for (String nodeTypesMainFolder : nodeTypesMainFolders) {
+ try (Stream<Path> paths = Files.walk(Paths.get(importToscaPath + SEPARATOR + nodeTypesMainFolder))) {
+ paths.filter(path -> path.getFileName().toString().toLowerCase().endsWith(YAML_EXTENSION))
+ .forEach(yamlFile -> {
+ try {
+ String path = yamlFile.toAbsolutePath().toString();
+ System.out.println("Processing node type file " + path + "...");
+ FileInputStream inputStream = new FileInputStream(path);
+ Yaml yaml = new Yaml();
+ Map<String, Object> load = yaml.loadAs(inputStream, Map.class);
+ Map<String, Object> nodeType = (Map<String, Object>) load.get(collectionTitle);
+ nodeTypeList.putAll(nodeType);
+
+ } catch (Exception e) {
+ System.err.println("Error in opening file " + yamlFile.toAbsolutePath().toString());
+ System.exit(1);
+ }
+ });
+ }
+ }
+ createAndSaveSchemaFileYaml("nodes", importFileList, collectionTitle, nodeTypeList);
+ }
+
+ private static void usageAndExit() {
+ SdcSchemaFileImportUsage();
+ System.exit(1);
+ }
+
+ private static void SdcSchemaFileImportUsage() {
+ System.err.println(
+ "Usage: <file dir/filename> <SDC release number> <Schema conformance level> <configuration dir> <deployment type optional>");
+ }
+
+ private static AnnotationConfigApplicationContext initContext(String appConfigDir) {
+ ConfigurationSource configurationSource = new FSConfigurationSource(ExternalConfiguration.getChangeListener(),
+ appConfigDir);
+ new ConfigurationManager(configurationSource);
+ return new AnnotationConfigApplicationContext(SdcSchemaFileImportConfiguration.class);
}
- createAndSaveSchemaFileYaml("nodes", importFileList, collectionTitle, nodeTypeList);
- }
-
- private static void usageAndExit() {
- SdcSchemaFileImportUsage();
- System.exit(1);
- }
-
- private static void SdcSchemaFileImportUsage() {
- System.err.println("Usage: <file dir/filename> <SDC release number> <Schema conformance level> <configuration dir> <deployment type optional>");
- }
-
- private static AnnotationConfigApplicationContext initContext(String appConfigDir) {
- ConfigurationSource configurationSource = new FSConfigurationSource(ExternalConfiguration.getChangeListener(), appConfigDir);
- new ConfigurationManager(configurationSource);
- return new AnnotationConfigApplicationContext(SdcSchemaFileImportConfiguration.class);
- }
}