⭐ 欢迎来到虫虫下载站! | 📦 资源下载 📁 资源专辑 ℹ️ 关于我们
⭐ 虫虫下载站

📄 crawl.java

📁 nutch0.8源码
💻 JAVA
字号:
/** * Copyright 2005 The Apache Software Foundation * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * *     http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */package org.apache.nutch.crawl;import java.util.*;import java.text.*;// Commons Logging importsimport org.apache.commons.logging.Log;import org.apache.commons.logging.LogFactory;import org.apache.nutch.fetcher.Fetcher;import org.apache.hadoop.fs.*;import org.apache.hadoop.conf.*;import org.apache.hadoop.mapred.*;import org.apache.nutch.parse.ParseSegment;import org.apache.nutch.indexer.DeleteDuplicates;import org.apache.nutch.indexer.IndexMerger;import org.apache.nutch.indexer.Indexer;import org.apache.nutch.util.NutchConfiguration;import org.apache.nutch.util.NutchJob;public class Crawl {  public static final Log LOG = LogFactory.getLog(Crawl.class);  private static String getDate() {    return new SimpleDateFormat("yyyyMMddHHmmss").format      (new Date(System.currentTimeMillis()));  }  /* Perform complete crawling and indexing given a set of root urls. */  public static void main(String args[]) throws Exception {    if (args.length < 1) {      System.out.println        ("Usage: Crawl <urlDir> [-dir d] [-threads n] [-depth i] [-topN N]");      return;    }    Configuration conf = NutchConfiguration.create();    conf.addDefaultResource("crawl-tool.xml");    JobConf job = new NutchJob(conf);    Path rootUrlDir = null;    Path dir = new Path("crawl-" + getDate());    int threads = job.getInt("fetcher.threads.fetch", 10);    int depth = 5;    int topN = Integer.MAX_VALUE;    for (int i = 0; i < args.length; i++) {      if ("-dir".equals(args[i])) {        dir = new Path(args[i+1]);        i++;      } else if ("-threads".equals(args[i])) {        threads = Integer.parseInt(args[i+1]);        i++;      } else if ("-depth".equals(args[i])) {        depth = Integer.parseInt(args[i+1]);        i++;      } else if ("-topN".equals(args[i])) {        topN = Integer.parseInt(args[i+1]);        i++;      } else if (args[i] != null) {        rootUrlDir = new Path(args[i]);      }    }    FileSystem fs = FileSystem.get(job);    if (fs.exists(dir)) {      throw new RuntimeException(dir + " already exists.");    }    if (LOG.isInfoEnabled()) {      LOG.info("crawl started in: " + dir);      LOG.info("rootUrlDir = " + rootUrlDir);      LOG.info("threads = " + threads);      LOG.info("depth = " + depth);      if (topN != Integer.MAX_VALUE)        LOG.info("topN = " + topN);    }        Path crawlDb = new Path(dir + "/crawldb");    Path linkDb = new Path(dir + "/linkdb");    Path segments = new Path(dir + "/segments");    Path indexes = new Path(dir + "/indexes");    Path index = new Path(dir + "/index");    Path tmpDir = job.getLocalPath("crawl"+Path.SEPARATOR+getDate());          // initialize crawlDb    new Injector(job).inject(crawlDb, rootUrlDir);          for (int i = 0; i < depth; i++) {             // generate new segment      Path segment =        new Generator(job).generate(crawlDb, segments, -1,                                     topN, System.currentTimeMillis());      new Fetcher(job).fetch(segment, threads);  // fetch it      if (!Fetcher.isParsing(job)) {        new ParseSegment(job).parse(segment);    // parse it, if needed      }      new CrawlDb(job).update(crawlDb, segment); // update crawldb    }          new LinkDb(job).invert(linkDb, segments); // invert links    // index, dedup & merge    new Indexer(job).index(indexes, crawlDb, linkDb, fs.listPaths(segments));    new DeleteDuplicates(job).dedup(new Path[] { indexes });    new IndexMerger(fs, fs.listPaths(indexes), index, tmpDir, job).merge();    if (LOG.isInfoEnabled()) { LOG.info("crawl finished: " + dir); }  }}

⌨️ 快捷键说明

复制代码 Ctrl + C
搜索代码 Ctrl + F
全屏模式 F11
切换主题 Ctrl + Shift + D
显示快捷键 ?
增大字号 Ctrl + =
减小字号 Ctrl + -