命令如下:
Administrator@f523540 ~ $ cd /cygdrive/d/nutch/apache-nutch-1.4-bin/runtime/local/ Administrator@f523540 /cygdrive/d/nutch/apache-nutch-1.4-bin/runtime/local $ ./bin/nutch crawl urls -dir crawl -topN 5 -depth 3 cygpath: can't convert empty path solrUrl is not set, indexing will be skipped... crawl started in: crawl rootUrlDir = urls threads = 10 depth = 3 solrUrl=null topN = 5 Injector: starting at 2012-06-17 13:47:45 Injector: crawlDb: crawl/crawldb Injector: urlDir: urls Injector: Converting injected urls to crawl db entries. Exception in thread "main" java.io.IOException: Job failed! at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:1252) at org.apache.nutch.crawl.Injector.inject(Injector.java:217) at org.apache.nutch.crawl.Crawl.run(Crawl.java:127) at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) at org.apache.nutch.crawl.Crawl.main(Crawl.java:55)
环境:cygwin windows xp java 1.6 nutch 1.4。不知道哪位有没有遇到过此问题,期待您的回答!