⭐ 欢迎来到虫虫下载站! | 📦 资源下载 📁 资源专辑 ℹ️ 关于我们
⭐ 虫虫下载站

📄 dutchanalyzer.java

📁 lucene2.2.0版本
💻 JAVA
字号:
package org.apache.lucene.analysis.nl;/** * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements.  See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License.  You may obtain a copy of the License at * *     http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */import org.apache.lucene.analysis.Analyzer;import org.apache.lucene.analysis.StopFilter;import org.apache.lucene.analysis.TokenStream;import org.apache.lucene.analysis.standard.StandardFilter;import org.apache.lucene.analysis.standard.StandardTokenizer;import java.io.File;import java.io.IOException;import java.io.Reader;import java.util.HashMap;import java.util.HashSet;import java.util.Set;import java.util.Map;/** * Analyzer for Dutch language. Supports an external list of stopwords (words that * will not be indexed at all), an external list of exclusions (word that will * not be stemmed, but indexed) and an external list of word-stem pairs that overrule * the algorithm (dictionary stemming). * A default set of stopwords is used unless an alternative list is specified, the * exclusion list is empty by default. *  * @author Edwin de Jonge */public class DutchAnalyzer extends Analyzer {  /**   * List of typical Dutch stopwords.   */  public final static String[] DUTCH_STOP_WORDS =      {        "de", "en", "van", "ik", "te", "dat", "die", "in", "een",        "hij", "het", "niet", "zijn", "is", "was", "op", "aan", "met", "als", "voor", "had",        "er", "maar", "om", "hem", "dan", "zou", "of", "wat", "mijn", "men", "dit", "zo",        "door", "over", "ze", "zich", "bij", "ook", "tot", "je", "mij", "uit", "der", "daar",        "haar", "naar", "heb", "hoe", "heeft", "hebben", "deze", "u", "want", "nog", "zal",        "me", "zij", "nu", "ge", "geen", "omdat", "iets", "worden", "toch", "al", "waren",        "veel", "meer", "doen", "toen", "moet", "ben", "zonder", "kan", "hun", "dus",        "alles", "onder", "ja", "eens", "hier", "wie", "werd", "altijd", "doch", "wordt",        "wezen", "kunnen", "ons", "zelf", "tegen", "na", "reeds", "wil", "kon", "niets",        "uw", "iemand", "geweest", "andere"      };  /**   * Contains the stopwords used with the StopFilter.   */  private Set stoptable = new HashSet();  /**   * Contains words that should be indexed but not stemmed.   */  private Set excltable = new HashSet();  private Map stemdict = new HashMap();  /**   * Builds an analyzer with the default stop words ({@link #DUTCH_STOP_WORDS})    * and a few default entries for the stem exclusion table.   *    */  public DutchAnalyzer() {    stoptable = StopFilter.makeStopSet(DUTCH_STOP_WORDS);    stemdict.put("fiets", "fiets"); //otherwise fiet    stemdict.put("bromfiets", "bromfiets"); //otherwise bromfiet    stemdict.put("ei", "eier");    stemdict.put("kind", "kinder");  }  /**   * Builds an analyzer with the given stop words.   *   * @param stopwords   */  public DutchAnalyzer(String[] stopwords) {    stoptable = StopFilter.makeStopSet(stopwords);  }  /**   * Builds an analyzer with the given stop words.   *   * @param stopwords   */  public DutchAnalyzer(HashSet stopwords) {    stoptable = stopwords;  }  /**   * Builds an analyzer with the given stop words.   *   * @param stopwords   */  public DutchAnalyzer(File stopwords) {    try {      stoptable = org.apache.lucene.analysis.WordlistLoader.getWordSet(stopwords);    } catch (IOException e) {      // TODO: throw IOException      throw new RuntimeException(e);    }  }  /**   * Builds an exclusionlist from an array of Strings.   *   * @param exclusionlist   */  public void setStemExclusionTable(String[] exclusionlist) {    excltable = StopFilter.makeStopSet(exclusionlist);  }  /**   * Builds an exclusionlist from a Hashtable.   */  public void setStemExclusionTable(HashSet exclusionlist) {    excltable = exclusionlist;  }  /**   * Builds an exclusionlist from the words contained in the given file.   */  public void setStemExclusionTable(File exclusionlist) {    try {      excltable = org.apache.lucene.analysis.WordlistLoader.getWordSet(exclusionlist);    } catch (IOException e) {      // TODO: throw IOException      throw new RuntimeException(e);    }  }  /**   * Reads a stemdictionary file , that overrules the stemming algorithm   * This is a textfile that contains per line   * <tt>word<b>\t</b>stem</tt>, i.e: two tab seperated words   */  public void setStemDictionary(File stemdictFile) {    try {      stemdict = org.apache.lucene.analysis.WordlistLoader.getStemDict(stemdictFile);    } catch (IOException e) {      // TODO: throw IOException      throw new RuntimeException(e);    }  }  /**   * Creates a TokenStream which tokenizes all the text in the provided TextReader.   *   * @return A TokenStream build from a StandardTokenizer filtered with StandardFilter,   * StopFilter, DutchStemFilter   */  public TokenStream tokenStream(String fieldName, Reader reader) {    TokenStream result = new StandardTokenizer(reader);    result = new StandardFilter(result);    result = new StopFilter(result, stoptable);    result = new DutchStemFilter(result, excltable, stemdict);    return result;  }}

⌨️ 快捷键说明

复制代码 Ctrl + C
搜索代码 Ctrl + F
全屏模式 F11
切换主题 Ctrl + Shift + D
显示快捷键 ?
增大字号 Ctrl + =
减小字号 Ctrl + -