diff options
author | Benjamin Jones <bjones@galois.com> | 2012-12-03 15:59:51 -0800 |
---|---|---|
committer | Benjamin Jones <bjones@galois.com> | 2012-12-03 15:59:51 -0800 |
commit | 5e8dd37e31da9c1b0c50d3df1768c6792ed3faf9 (patch) | |
tree | 184e493012f3503119ebdc0ed2f0bb7c89fa20e6 /headless | |
parent | ad9120922134705397c3a14036ae91e3745a2ef8 (diff) |
skip webcrawler initialization and execution when crawlType == none
Diffstat (limited to 'headless')
-rw-r--r-- | headless/src/main/java/com/galois/fiveui/BatchRunner.java | 45 |
1 files changed, 25 insertions, 20 deletions
diff --git a/headless/src/main/java/com/galois/fiveui/BatchRunner.java b/headless/src/main/java/com/galois/fiveui/BatchRunner.java index c3a0948..7721a66 100644 --- a/headless/src/main/java/com/galois/fiveui/BatchRunner.java +++ b/headless/src/main/java/com/galois/fiveui/BatchRunner.java @@ -106,26 +106,31 @@ public class BatchRunner { File tmpPath = Files.createTempDir(); logger.debug("tmp directory for crawl data: " + tmpPath.toString()); - // Crawl starting at the seed page - logger.debug("starting web crawl controller ..."); - BasicCrawlerController con = - new BasicCrawlerController(seedUrl, - params.match, - params.depth, params.maxFetch, - params.politeness, - 1, // TODO only one thread is currently supported - tmpPath.getAbsolutePath()); - try { - urls = con.go(); - } catch (Exception e) { - String errStr = "failed to complete webcrawl of" + seedUrl + "\n"; - errStr += e.toString(); - builder.add(Result.exception(_driver, errStr)); - logger.error(errStr); - continue; - } finally { - IO.deleteFolder(tmpPath); // does its own logging - } + if (params.isNone()) { + urls = ImmutableList.of(seedUrl); + logger.debug("skipping webcrawl"); + } else { + // Crawl starting at the seed page + logger.debug("starting webcrawl controller ..."); + BasicCrawlerController con = + new BasicCrawlerController(seedUrl, + params.match, + params.depth, params.maxFetch, + params.politeness, + 1, // TODO only one thread is currently supported + tmpPath.getAbsolutePath()); + try { + urls = con.go(); + } catch (Exception e) { + String errStr = "failed to complete webcrawl of" + seedUrl + "\n"; + errStr += e.toString(); + builder.add(Result.exception(_driver, errStr)); + logger.error(errStr); + continue; + } finally { + IO.deleteFolder(tmpPath); // does its own logging + } + } // run ruleset on each discovered URL for (String url: urls) { |