aboutsummaryrefslogtreecommitdiff
path: root/headless
diff options
context:
space:
mode:
authorGravatar Benjamin Jones <bjones@galois.com>2012-12-03 15:59:51 -0800
committerGravatar Benjamin Jones <bjones@galois.com>2012-12-03 15:59:51 -0800
commit5e8dd37e31da9c1b0c50d3df1768c6792ed3faf9 (patch)
tree184e493012f3503119ebdc0ed2f0bb7c89fa20e6 /headless
parentad9120922134705397c3a14036ae91e3745a2ef8 (diff)
skip webcrawler initialization and execution when crawlType == none
Diffstat (limited to 'headless')
-rw-r--r--headless/src/main/java/com/galois/fiveui/BatchRunner.java45
1 files changed, 25 insertions, 20 deletions
diff --git a/headless/src/main/java/com/galois/fiveui/BatchRunner.java b/headless/src/main/java/com/galois/fiveui/BatchRunner.java
index c3a0948..7721a66 100644
--- a/headless/src/main/java/com/galois/fiveui/BatchRunner.java
+++ b/headless/src/main/java/com/galois/fiveui/BatchRunner.java
@@ -106,26 +106,31 @@ public class BatchRunner {
File tmpPath = Files.createTempDir();
logger.debug("tmp directory for crawl data: " + tmpPath.toString());
- // Crawl starting at the seed page
- logger.debug("starting web crawl controller ...");
- BasicCrawlerController con =
- new BasicCrawlerController(seedUrl,
- params.match,
- params.depth, params.maxFetch,
- params.politeness,
- 1, // TODO only one thread is currently supported
- tmpPath.getAbsolutePath());
- try {
- urls = con.go();
- } catch (Exception e) {
- String errStr = "failed to complete webcrawl of" + seedUrl + "\n";
- errStr += e.toString();
- builder.add(Result.exception(_driver, errStr));
- logger.error(errStr);
- continue;
- } finally {
- IO.deleteFolder(tmpPath); // does its own logging
- }
+ if (params.isNone()) {
+ urls = ImmutableList.of(seedUrl);
+ logger.debug("skipping webcrawl");
+ } else {
+ // Crawl starting at the seed page
+ logger.debug("starting webcrawl controller ...");
+ BasicCrawlerController con =
+ new BasicCrawlerController(seedUrl,
+ params.match,
+ params.depth, params.maxFetch,
+ params.politeness,
+ 1, // TODO only one thread is currently supported
+ tmpPath.getAbsolutePath());
+ try {
+ urls = con.go();
+ } catch (Exception e) {
+ String errStr = "failed to complete webcrawl of" + seedUrl + "\n";
+ errStr += e.toString();
+ builder.add(Result.exception(_driver, errStr));
+ logger.error(errStr);
+ continue;
+ } finally {
+ IO.deleteFolder(tmpPath); // does its own logging
+ }
+ }
// run ruleset on each discovered URL
for (String url: urls) {