diff options
author | Kajur, Harish (vk250x) <vk250x@att.com> | 2018-09-13 10:45:23 -0400 |
---|---|---|
committer | Kajur, Harish (vk250x) <vk250x@att.com> | 2018-09-18 14:44:58 -0400 |
commit | d13dd0ed905109546950b4fdd1e3f9e0471ec8e7 (patch) | |
tree | a8985e647c86ba2dc7e4e37e154cc4f52abb304f /src/main | |
parent | 1e3c9bde6ceb3920aa1695d59e6f677510ca649c (diff) |
Add the dataSnapshot unit test and datagrooming
to get to the 50% coverage
Issue-ID: AAI-1633
Change-Id: Ic25baa59be8945cc3b07b346d44e1d35168c3b9c
Signed-off-by: Kajur, Harish (vk250x) <vk250x@att.com>
Diffstat (limited to 'src/main')
-rw-r--r-- | src/main/java/org/onap/aai/datagrooming/DataGrooming.java | 265 |
1 files changed, 211 insertions, 54 deletions
diff --git a/src/main/java/org/onap/aai/datagrooming/DataGrooming.java b/src/main/java/org/onap/aai/datagrooming/DataGrooming.java index 6149dd9..199e704 100644 --- a/src/main/java/org/onap/aai/datagrooming/DataGrooming.java +++ b/src/main/java/org/onap/aai/datagrooming/DataGrooming.java @@ -48,7 +48,6 @@ import org.apache.tinkerpop.gremlin.structure.Vertex; import org.apache.tinkerpop.gremlin.structure.VertexProperty; import org.onap.aai.GraphAdminApp; import org.onap.aai.dbmap.AAIGraph; -import org.onap.aai.dbmap.AAIGraphConfig; import org.onap.aai.exceptions.AAIException; import org.onap.aai.introspection.Introspector; import org.onap.aai.introspection.Loader; @@ -67,7 +66,9 @@ import org.onap.aai.logging.LoggingContext.StatusCode; import com.att.eelf.configuration.Configuration; import com.att.eelf.configuration.EELFLogger; import com.att.eelf.configuration.EELFManager; -import org.janusgraph.core.JanusGraphFactory; +import com.beust.jcommander.JCommander; +import com.beust.jcommander.Parameter; + import org.janusgraph.core.JanusGraph; import org.springframework.context.annotation.AnnotationConfigApplicationContext; @@ -82,6 +83,16 @@ public class DataGrooming { private LoaderFactory loaderFactory; private SchemaVersions schemaVersions; + + private CommandLineArgs cArgs; + + HashMap<String, Vertex> orphanNodeHash ; + HashMap<String, Vertex> missingAaiNtNodeHash ; + HashMap<String, Edge> oneArmedEdgeHash ; + HashMap<String, Vertex> ghostNodeHash ; + ArrayList<String> dupeGroups; + Set<String> deleteCandidateList; + private int deleteCount = 0; public DataGrooming(LoaderFactory loaderFactory, SchemaVersions schemaVersions){ this.loaderFactory = loaderFactory; @@ -91,18 +102,6 @@ public class DataGrooming { public void execute(String[] args){ String ver = "version"; // Placeholder - Boolean doAutoFix = false; - Boolean edgesOnlyFlag = false; - Boolean dontFixOrphansFlag = false; - Boolean skipHostCheck = false; - Boolean singleCommits = false; - Boolean dupeCheckOff = false; - Boolean dupeFixOn = false; - Boolean ghost2CheckOff = false; - Boolean ghost2FixOn = false; - Boolean neverUseCache = false; - Boolean skipEdgeCheckFlag = false; - Boolean skipIndexUpdateFix = false; // A value of 0 means that we will not have a time-window -- we will look // at all nodes of the passed-in nodeType. @@ -131,7 +130,26 @@ public class DataGrooming { FormatDate fd = new FormatDate("yyyyMMddHHmm", "GMT"); String dteStr = fd.getDateTime(); - if (args.length > 0) { + cArgs = new CommandLineArgs(); + JCommander jCommander = new JCommander(cArgs, args); + jCommander.setProgramName(DataGrooming.class.getSimpleName()); + + //Print Defaults + LOGGER.info("EdgesOnlyFlag is" + cArgs.edgesOnlyFlag); + LOGGER.info("DoAutoFix is" + cArgs.doAutoFix); + LOGGER.info("skipHostCheck is" + cArgs.skipHostCheck); + LOGGER.info("dontFixOrphansFlag is" + cArgs.dontFixOrphansFlag); + LOGGER.info("singleCommits is" + cArgs.singleCommits); + LOGGER.info("dupeCheckOff is" + cArgs.dupeCheckOff); + LOGGER.info("dupeFixOn is" + cArgs.dupeFixOn); + LOGGER.info("ghost2CheckOff is" + cArgs.ghost2CheckOff); + LOGGER.info("ghost2FixOn is" + cArgs.ghost2FixOn); + LOGGER.info("neverUseCache is" + cArgs.neverUseCache); + LOGGER.info("skipEdgeChecks is" + cArgs.skipEdgeCheckFlag); + LOGGER.info("skipIndexUpdateFix is" + cArgs.skipIndexUpdateFix); + LOGGER.info("maxFix is" + cArgs.maxRecordsToFix); + + /*if (args.length > 0) { // They passed some arguments in that will affect processing for (int i = 0; i < args.length; i++) { String thisArg = args[i]; @@ -241,9 +259,10 @@ public class DataGrooming { AAISystemExitUtil.systemExitCloseAAIGraph(0); } } - } + } */ String windowTag = "FULL"; + //TODO??? if( timeWindowMinutes > 0 ){ windowTag = "PARTIAL"; } @@ -267,13 +286,14 @@ public class DataGrooming { + prevFileName + "] for cleanup. "); Boolean finalShutdownFlag = true; Boolean cacheDbOkFlag = false; - doTheGrooming(prevFileName, edgesOnlyFlag, dontFixOrphansFlag, - maxRecordsToFix, groomOutFileName, ver, singleCommits, - dupeCheckOff, dupeFixOn, ghost2CheckOff, ghost2FixOn, - finalShutdownFlag, cacheDbOkFlag, - skipEdgeCheckFlag, timeWindowMinutes, - singleNodeType, skipIndexUpdateFix ); - } else if (doAutoFix) { + doTheGrooming(prevFileName, cArgs.edgesOnlyFlag, cArgs.dontFixOrphansFlag, + cArgs.maxRecordsToFix, groomOutFileName, ver, cArgs.singleCommits, + cArgs.dupeCheckOff, cArgs.dupeFixOn, cArgs.ghost2CheckOff, cArgs.ghost2FixOn, + cArgs.finalShutdownFlag, cArgs.cacheDbOkFlag, + cArgs.skipEdgeCheckFlag, cArgs.timeWindowMinutes, + cArgs.singleNodeType, cArgs.skipIndexUpdateFix ); + + } else if (cArgs.doAutoFix) { // They want us to run the processing twice -- first to look for // delete candidates, then after // napping for a while, run it again and delete any candidates @@ -284,21 +304,21 @@ public class DataGrooming { LOGGER.info(" First, Call doTheGrooming() to look at what's out there. "); Boolean finalShutdownFlag = false; Boolean cacheDbOkFlag = true; - int fixCandCount = doTheGrooming("", edgesOnlyFlag, - dontFixOrphansFlag, maxRecordsToFix, groomOutFileName, - ver, singleCommits, dupeCheckOff, dupeFixOn, ghost2CheckOff, ghost2FixOn, - finalShutdownFlag, cacheDbOkFlag, - skipEdgeCheckFlag, timeWindowMinutes, - singleNodeType, skipIndexUpdateFix ); + int fixCandCount = doTheGrooming("", cArgs.edgesOnlyFlag, + cArgs.dontFixOrphansFlag, cArgs.maxRecordsToFix, groomOutFileName, + ver, cArgs.singleCommits, cArgs.dupeCheckOff, cArgs.dupeFixOn, cArgs.ghost2CheckOff, cArgs.ghost2FixOn, + cArgs.finalShutdownFlag, cArgs.cacheDbOkFlag, + cArgs.skipEdgeCheckFlag, cArgs.timeWindowMinutes, + cArgs.singleNodeType, cArgs.skipIndexUpdateFix ); if (fixCandCount == 0) { LOGGER.info(" No fix-Candidates were found by the first pass, so no second/fix-pass is needed. "); } else { // We'll sleep a little and then run a fix-pass based on the // first-run's output file. try { - LOGGER.info("About to sleep for " + sleepMinutes + LOGGER.info("About to sleep for " + cArgs.sleepMinutes + " minutes."); - int sleepMsec = sleepMinutes * 60 * 1000; + int sleepMsec = cArgs.sleepMinutes * 60 * 1000; Thread.sleep(sleepMsec); } catch (InterruptedException ie) { LOGGER.info("\n >>> Sleep Thread has been Interrupted <<< "); @@ -312,13 +332,13 @@ public class DataGrooming { + groomOutFileName + "]"); finalShutdownFlag = true; cacheDbOkFlag = false; - doTheGrooming(groomOutFileName, edgesOnlyFlag, - dontFixOrphansFlag, maxRecordsToFix, - secondGroomOutFileName, ver, singleCommits, - dupeCheckOff, dupeFixOn, ghost2CheckOff, ghost2FixOn, - finalShutdownFlag, cacheDbOkFlag, - skipEdgeCheckFlag, timeWindowMinutes, - singleNodeType, skipIndexUpdateFix ); + doTheGrooming(groomOutFileName, cArgs.edgesOnlyFlag, + cArgs.dontFixOrphansFlag, cArgs.maxRecordsToFix, + secondGroomOutFileName, ver, cArgs.singleCommits, + cArgs.dupeCheckOff, cArgs.dupeFixOn, cArgs.ghost2CheckOff, cArgs.ghost2FixOn, + cArgs.finalShutdownFlag, cArgs.cacheDbOkFlag, + cArgs.skipEdgeCheckFlag, cArgs.timeWindowMinutes, + cArgs.singleNodeType, cArgs.skipIndexUpdateFix ); } } else { // Do the grooming - plain vanilla (no fix-it-file, no @@ -326,16 +346,16 @@ public class DataGrooming { Boolean finalShutdownFlag = true; LOGGER.info(" Call doTheGrooming() "); Boolean cacheDbOkFlag = true; - if( neverUseCache ){ + if( cArgs.neverUseCache ){ // They have forbidden us from using a cached db connection. - cacheDbOkFlag = false; + cArgs.cacheDbOkFlag = false; } - doTheGrooming("", edgesOnlyFlag, dontFixOrphansFlag, - maxRecordsToFix, groomOutFileName, ver, singleCommits, - dupeCheckOff, dupeFixOn, ghost2CheckOff, ghost2FixOn, - finalShutdownFlag, cacheDbOkFlag, - skipEdgeCheckFlag, timeWindowMinutes, - singleNodeType, skipIndexUpdateFix ); + doTheGrooming("", cArgs.edgesOnlyFlag, cArgs.dontFixOrphansFlag, + cArgs.maxRecordsToFix, groomOutFileName, ver, cArgs.singleCommits, + cArgs.dupeCheckOff, cArgs.dupeFixOn, cArgs.ghost2CheckOff, cArgs.ghost2FixOn, + cArgs.finalShutdownFlag, cArgs.cacheDbOkFlag, + cArgs.skipEdgeCheckFlag, cArgs.timeWindowMinutes, + cArgs.singleNodeType, cArgs.skipIndexUpdateFix ); } } catch (Exception ex) { LoggingContext.statusCode(StatusCode.ERROR); @@ -416,10 +436,10 @@ public class DataGrooming { BufferedWriter bw = null; JanusGraph graph = null; JanusGraph graph2 = null; - int deleteCount = 0; + deleteCount = 0; int dummyUpdCount = 0; boolean executeFinalCommit = false; - Set<String> deleteCandidateList = new LinkedHashSet<>(); + deleteCandidateList = new LinkedHashSet<>(); Set<String> processedVertices = new LinkedHashSet<>(); Set<String> postCommitRemoveList = new LinkedHashSet<>(); @@ -505,12 +525,12 @@ public class DataGrooming { ArrayList<String> errArr = new ArrayList<>(); int totalNodeCount = 0; HashMap<String, String> misMatchedHash = new HashMap<String, String>(); - HashMap<String, Vertex> orphanNodeHash = new HashMap<String, Vertex>(); - HashMap<String, Vertex> missingAaiNtNodeHash = new HashMap<String, Vertex>(); - HashMap<String, Edge> oneArmedEdgeHash = new HashMap<String, Edge>(); + orphanNodeHash = new HashMap<String, Vertex>(); + missingAaiNtNodeHash = new HashMap<String, Vertex>(); + oneArmedEdgeHash = new HashMap<String, Edge>(); HashMap<String, String> emptyVertexHash = new HashMap<String, String>(); - HashMap<String, Vertex> ghostNodeHash = new HashMap<String, Vertex>(); - ArrayList<String> dupeGroups = new ArrayList<>(); + ghostNodeHash = new HashMap<String, Vertex>(); + dupeGroups = new ArrayList<>(); Loader loader = loaderFactory.createLoaderForVersion(ModelType.MOXY, schemaVersions.getDefaultVersion()); @@ -1627,7 +1647,7 @@ public class DataGrooming { }// end of doTheGrooming() - private void updateIndexedProps(Vertex thisVtx, String thisVidStr, String nType, + public void updateIndexedProps(Vertex thisVtx, String thisVidStr, String nType, HashMap <String,String>propTypeHash, ArrayList <String> indexedProps) { // This is a "missing-aai-node-type" scenario. // Other indexes may also be messed up, so we will update all of them on @@ -2849,5 +2869,142 @@ public class DataGrooming { return returnVid; }// End of findJustOneUsingIndex() + +class CommandLineArgs { + + + @Parameter(names = "--help", help = true) + public boolean help; + + @Parameter(names = "-edgesOnly", description = "Check grooming on edges only", arity = 1) + public Boolean edgesOnlyFlag = false; + + @Parameter(names = "-autoFix", description = "doautofix", arity = 1) + public Boolean doAutoFix = false; + + @Parameter(names = "-skipHostCheck", description = "skipHostCheck", arity = 1) + public Boolean skipHostCheck = false; + + @Parameter(names = "-dontFixOrphans", description = "dontFixOrphans", arity = 1) + public Boolean dontFixOrphansFlag = false; + + @Parameter(names = "-singleCommits", description = "singleCommits", arity = 1) + public Boolean singleCommits = false; + + @Parameter(names = "-dupeCheckOff", description = "dupeCheckOff", arity = 1) + public Boolean dupeCheckOff = false; + + @Parameter(names = "-dupeFixOn", description = "dupeFixOn", arity = 1) + public Boolean dupeFixOn = false; + + @Parameter(names = "-ghost2CheckOff", description = "ghost2CheckOff", arity = 1) + public Boolean ghost2CheckOff = false; + + @Parameter(names = "-ghost2FixOn", description = "ghost2FixOn", arity = 1) + public Boolean ghost2FixOn = false; + + @Parameter(names = "-neverUseCache", description = "neverUseCache", arity = 1) + public Boolean neverUseCache = false; + + @Parameter(names = "-skipEdgeChecks", description = "skipEdgeChecks", arity = 1) + public Boolean skipEdgeCheckFlag = false; + + @Parameter(names = "-skipIndexUpdateFix", description = "skipIndexUpdateFix", arity = 1) + public Boolean skipIndexUpdateFix = false; + + @Parameter(names = "-maxFix", description = "maxFix") + public int maxRecordsToFix = AAIConstants.AAI_GROOMING_DEFAULT_MAX_FIX; + + @Parameter(names = "-sleepMinutes", description = "sleepMinutes") + public int sleepMinutes = AAIConstants.AAI_GROOMING_DEFAULT_SLEEP_MINUTES; + + // A value of 0 means that we will not have a time-window -- we will look + // at all nodes of the passed-in nodeType. + @Parameter(names = "-timeWindowMinutes", description = "timeWindowMinutes") + public int timeWindowMinutes = 0; + + @Parameter(names = "-f", description = "file") + public String prevFileName = ""; + + @Parameter(names = "-singleNodeType", description = "sleepMinutes") + public String singleNodeType = ""; + + Boolean finalShutdownFlag = true; + Boolean cacheDbOkFlag = true; + } + + public HashMap<String, Vertex> getGhostNodeHash() { + return ghostNodeHash; + } + + public void setGhostNodeHash(HashMap<String, Vertex> ghostNodeHash) { + this.ghostNodeHash = ghostNodeHash; + } + + public int getGhostNodeCount(){ + return getGhostNodeHash().size(); + } + + public HashMap<String, Vertex> getOrphanNodeHash() { + return orphanNodeHash; + } + + public void setOrphanNodeHash(HashMap<String, Vertex> orphanNodeHash) { + this.orphanNodeHash = orphanNodeHash; + } + + public int getOrphanNodeCount(){ + return getOrphanNodeHash().size(); + } + + public HashMap<String, Vertex> getMissingAaiNtNodeHash() { + return missingAaiNtNodeHash; + } + + public void setMissingAaiNtNodeHash(HashMap<String, Vertex> missingAaiNtNodeHash) { + this.missingAaiNtNodeHash = missingAaiNtNodeHash; + } + + public int getMissingAaiNtNodeCount(){ + return getMissingAaiNtNodeHash().size(); + } + + public HashMap<String, Edge> getOneArmedEdgeHash() { + return oneArmedEdgeHash; + } + + public void setOneArmedEdgeHash(HashMap<String, Edge> oneArmedEdgeHash) { + this.oneArmedEdgeHash = oneArmedEdgeHash; + } + + public int getOneArmedEdgeHashCount(){ + return getOneArmedEdgeHash().size(); + } + + public Set<String> getDeleteCandidateList() { + return deleteCandidateList; + } + + public void setDeleteCandidateList(Set<String> deleteCandidateList) { + this.deleteCandidateList = deleteCandidateList; + } + + public int getDeleteCount() { + return deleteCount; + } + + public void setDeleteCount(int deleteCount) { + this.deleteCount = deleteCount; + } + + public ArrayList<String> getDupeGroups() { + return dupeGroups; + } + + public void setDupeGroups(ArrayList<String> dupeGroups) { + this.dupeGroups = dupeGroups; + } + + }
\ No newline at end of file |