⭐ 欢迎来到虫虫下载站! | 📦 资源下载 📁 资源专辑 ℹ️ 关于我们
⭐ 虫虫下载站

📄 parseoutputformat.java

📁 nutch0.8源码
💻 JAVA
字号:
/** * Copyright 2005 The Apache Software Foundation * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * *     http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */package org.apache.nutch.parse;// Commons Logging importsimport org.apache.commons.logging.Log;import org.apache.commons.logging.LogFactory;import org.apache.hadoop.io.*;import org.apache.nutch.crawl.CrawlDatum;import org.apache.nutch.fetcher.Fetcher;import org.apache.hadoop.fs.*;import org.apache.hadoop.mapred.*;import org.apache.nutch.scoring.ScoringFilterException;import org.apache.nutch.scoring.ScoringFilters;import org.apache.nutch.util.StringUtil;import org.apache.nutch.net.*;import java.io.*;import java.net.MalformedURLException;import java.net.URL;import org.apache.hadoop.util.Progressable;/* Parse content in a segment. */public class ParseOutputFormat implements OutputFormat {  private static final Log LOG = LogFactory.getLog(ParseOutputFormat.class);  private UrlNormalizer urlNormalizer;  private URLFilters filters;  private ScoringFilters scfilters;  public void checkOutputSpecs(FileSystem fs, JobConf job) throws IOException {    if (fs.exists(new Path(job.getOutputPath(), CrawlDatum.PARSE_DIR_NAME)))      throw new IOException("Segment already parsed!");  }  public RecordWriter getRecordWriter(FileSystem fs, JobConf job,                                      String name, Progressable progress) throws IOException {    this.urlNormalizer = new UrlNormalizerFactory(job).getNormalizer();    this.filters = new URLFilters(job);    this.scfilters = new ScoringFilters(job);    final float interval = job.getFloat("db.default.fetch.interval", 30f);    final boolean ignoreExternalLinks = job.getBoolean("db.ignore.external.links", false);        Path text =      new Path(new Path(job.getOutputPath(), ParseText.DIR_NAME), name);    Path data =      new Path(new Path(job.getOutputPath(), ParseData.DIR_NAME), name);    Path crawl =      new Path(new Path(job.getOutputPath(), CrawlDatum.PARSE_DIR_NAME), name);        final MapFile.Writer textOut =      new MapFile.Writer(fs, text.toString(), UTF8.class, ParseText.class);        final MapFile.Writer dataOut =      new MapFile.Writer(fs, data.toString(), UTF8.class,ParseData.class,true);        final SequenceFile.Writer crawlOut =      new SequenceFile.Writer(fs, crawl, UTF8.class, CrawlDatum.class);        return new RecordWriter() {        public void write(WritableComparable key, Writable value)          throws IOException {                    Parse parse = (Parse)value;          String fromUrl = key.toString();          String fromHost = null;           String toHost = null;                    textOut.append(key, new ParseText(parse.getText()));                    ParseData parseData = parse.getData();          // recover the signature prepared by Fetcher or ParseSegment          String sig = parseData.getContentMeta().get(Fetcher.SIGNATURE_KEY);          if (sig != null) {            byte[] signature = StringUtil.fromHexString(sig);            if (signature != null) {              // append a CrawlDatum with a signature              CrawlDatum d = new CrawlDatum(CrawlDatum.STATUS_SIGNATURE, 0.0f);              d.setSignature(signature);              crawlOut.append(key, d);            }          }          // collect outlinks for subsequent db update          Outlink[] links = parseData.getOutlinks();          if (ignoreExternalLinks) {            try {              fromHost = new URL(fromUrl).getHost().toLowerCase();            } catch (MalformedURLException e) {              fromHost = null;            }          } else {            fromHost = null;          }          String[] toUrls = new String[links.length];          int validCount = 0;          for (int i = 0; i < links.length; i++) {            String toUrl = links[i].getToUrl();            try {              toUrl = urlNormalizer.normalize(toUrl); // normalize the url              toUrl = filters.filter(toUrl);   // filter the url            } catch (Exception e) {              toUrl = null;            }            // ignore links to self (or anchors within the page)            if (fromUrl.equals(toUrl)) toUrl = null;            if (toUrl != null) validCount++;            toUrls[i] = toUrl;          }          CrawlDatum adjust = null;          // compute score contributions and adjustment to the original score          for (int i = 0; i < toUrls.length; i++) {            if (toUrls[i] == null) continue;            if (ignoreExternalLinks) {              try {                toHost = new URL(toUrls[i]).getHost().toLowerCase();              } catch (MalformedURLException e) {                toHost = null;              }              if (toHost == null || !toHost.equals(fromHost)) { // external links                continue; // skip it              }            }            CrawlDatum target = new CrawlDatum(CrawlDatum.STATUS_LINKED, interval);            UTF8 targetUrl = new UTF8(toUrls[i]);            adjust = null;            try {              adjust = scfilters.distributeScoreToOutlink((UTF8)key, targetUrl,                      parseData, target, null, links.length, validCount);            } catch (ScoringFilterException e) {              if (LOG.isWarnEnabled()) {                LOG.warn("Cannot distribute score from " + key + " to " +                         targetUrl + " - skipped (" + e.getMessage());              }              continue;            }            crawlOut.append(targetUrl, target);            if (adjust != null) crawlOut.append(key, adjust);          }          dataOut.append(key, parseData);        }                public void close(Reporter reporter) throws IOException {          textOut.close();          dataOut.close();          crawlOut.close();        }              };      }}

⌨️ 快捷键说明

复制代码 Ctrl + C
搜索代码 Ctrl + F
全屏模式 F11
切换主题 Ctrl + Shift + D
显示快捷键 ?
增大字号 Ctrl + =
减小字号 Ctrl + -