📄 robotdecoratormapper.java
字号:
/* * Title: RobotDecoratorMapper * Description: * * This software is published under the terms of the OpenSymphony Software * License version 1.1, of which a copy has been included with this * distribution in the LICENSE.txt file. */package com.opensymphony.module.sitemesh.mapper;import com.opensymphony.module.sitemesh.Config;import com.opensymphony.module.sitemesh.Decorator;import com.opensymphony.module.sitemesh.DecoratorMapper;import com.opensymphony.module.sitemesh.Page;import com.opensymphony.module.sitemesh.RequestConstants;import javax.servlet.http.HttpServletRequest;import javax.servlet.http.HttpSession;import java.util.Properties;/** * The RobotDecoratorMapper will use the specified decorator when the requester * is identified as a robot (also known as spider, crawler, ferret) of a search engine. * * <p>The name of this decorator should be supplied in the <code>decorator</code> * property.</p> * * @author <a href="mailto:pathos@pandora.be">Mathias Bogaert</a> * @version $Revision: 1.3 $ * * @see com.opensymphony.module.sitemesh.DecoratorMapper */public class RobotDecoratorMapper extends AbstractDecoratorMapper { private String decoratorName = null; /** All known robot hosts (list can be found <a href="http://www.spiderhunter.com">here</a>). */ private static final String[] botHosts = {"alltheweb.com", "alta-vista.net", "altavista.com", "atext.com", "euroseek.net", "excite.com", "fast-search.net", "google.com", "googlebot.com", "infoseek.co.jp", "infoseek.com", "inktomi.com", "inktomisearch.com", "linuxtoday.com.au", "lycos.com", "lycos.com", "northernlight.com", "pa-x.dec.com"}; /** * All known robot user-agent headers (list can be found * <a href="http://www.robotstxt.org/wc/active.html">here</a>). * * <p>NOTE: To avoid bad detection:</p> * * <ul> * <li>Robots with ID of 2 letters only were removed</li> * <li>Robot called "webs" were removed</li> * <li>directhit was changed in direct_hit (its real id)</li> * </ul> */ private static final String[] botAgents = { "acme.spider", "ahoythehomepagefinder", "alkaline", "appie", "arachnophilia", "architext", "aretha", "ariadne", "aspider", "atn.txt", "atomz", "auresys", "backrub", "bigbrother", "bjaaland", "blackwidow", "blindekuh", "bloodhound", "brightnet", "bspider", "cactvschemistryspider", "calif", "cassandra", "cgireader", "checkbot", "churl", "cmc", "collective", "combine", "conceptbot", "core", "cshkust", "cusco", "cyberspyder", "deweb", "dienstspider", "diibot", "direct_hit", "dnabot", "download_express", "dragonbot", "dwcp", "ebiness", "eit", "emacs", "emcspider", "esther", "evliyacelebi", "fdse", "felix", "ferret", "fetchrover", "fido", "finnish", "fireball", "fish", "fouineur", "francoroute", "freecrawl", "funnelweb", "gazz", "gcreep", "getbot", "geturl", "golem", "googlebot", "grapnel", "griffon", "gromit", "gulliver", "hambot", "harvest", "havindex", "hometown", "wired-digital", "htdig", "htmlgobble", "hyperdecontextualizer", "ibm", "iconoclast", "ilse", "imagelock", "incywincy", "informant", "infoseek", "infoseeksidewinder", "infospider", "inspectorwww", "intelliagent", "iron33", "israelisearch", "javabee", "jcrawler", "jeeves", "jobot", "joebot", "jubii", "jumpstation", "katipo", "kdd", "kilroy", "ko_yappo_robot", "labelgrabber.txt", "larbin", "legs", "linkscan", "linkwalker", "lockon", "logo_gif", "lycos", "macworm", "magpie", "mediafox", "merzscope", "meshexplorer", "mindcrawler", "moget", "momspider", "monster", "motor", "muscatferret", "mwdsearch", "myweb", "netcarta", "netmechanic", "netscoop", "newscan-online", "nhse", "nomad", "northstar", "nzexplorer", "occam", "octopus", "orb_search", "packrat", "pageboy", "parasite", "patric", "perignator", "perlcrawler", "phantom", "piltdownman", "pioneer", "pitkow", "pjspider", "pka", "plumtreewebaccessor", "poppi", "portalb", "puu", "python", "raven", "rbse", "resumerobot", "rhcs", "roadrunner", "robbie", "robi", "roverbot", "safetynetrobot", "scooter", "search_au", "searchprocess", "senrigan", "sgscout", "shaggy", "shaihulud", "sift", "simbot", "site-valet", "sitegrabber", "sitetech", "slurp", "smartspider", "snooper", "solbot", "spanner", "speedy", "spider_monkey", "spiderbot", "spiderman", "spry", "ssearcher", "suke", "sven", "tach_bw", "tarantula", "tarspider", "tcl", "techbot", "templeton", "titin", "titan", "tkwww", "tlspider", "ucsd", "udmsearch", "urlck", "valkyrie", "victoria", "visionsearch", "voyager", "vwbot", "w3index", "w3m2", "wanderer", "webbandit", "webcatcher", "webcopy", "webfetcher", "webfoot", "weblayers", "weblinker", "webmirror", "webmoose", "webquest", "webreader", "webreaper", "websnarf", "webspider", "webvac", "webwalk", "webwalker", "webwatch", "wget", "whowhere", "wmir", "wolp", "wombat", "worm", "wwwc", "wz101", "xget", "nederland.zoek" }; public void init(Config config, Properties properties, DecoratorMapper parent) throws InstantiationException { super.init(config, properties, parent); decoratorName = properties.getProperty("decorator"); } public Decorator getDecorator(HttpServletRequest request, Page page) { Decorator result = null; if (decoratorName != null && isBot(request)) { result = getNamedDecorator(request, decoratorName); } return result == null ? super.getDecorator(request, page) : result; } /** Check if the current request came from a robot (also known as spider, crawler, ferret) */ private static boolean isBot(HttpServletRequest request) { if (request == null) return false; // force creation of a session HttpSession session = request.getSession(true); if (Boolean.FALSE.equals(session.getAttribute(RequestConstants.ROBOT))) { return false; } else if (Boolean.TRUE.equals(session.getAttribute(RequestConstants.ROBOT))) { // a key was found in the session indicating it is a robot return true; } else { if ("robots.txt".indexOf(request.getRequestURI()) != -1) { // there is a specific request for the robots.txt file, so we assume // it must be a robot (only robots request robots.txt) // set a key in the session, so the next time we don't have to manually // detect the robot again session.setAttribute(RequestConstants.ROBOT, Boolean.TRUE); return true; } else { String userAgent = request.getHeader("User-Agent"); if (userAgent != null && userAgent.trim().length() > 2) { // first check for common user-agent headers, so that we can speed // this thing up, hopefully clever spiders will not send a fake header if (userAgent.indexOf("MSIE") != -1 || userAgent.indexOf("Gecko") != -1 // MSIE and Mozilla || userAgent.indexOf("Opera") != -1 || userAgent.indexOf("iCab") != -1 // Opera and iCab (mac browser) || userAgent.indexOf("Konqueror") != -1 || userAgent.indexOf("KMeleon") != -1 // Konqueror and KMeleon || userAgent.indexOf("4.7") != -1 || userAgent.indexOf("Lynx") != -1) { // NS 4.78 and Lynx // indicate this session is not a robot session.setAttribute(RequestConstants.ROBOT, Boolean.FALSE); return false; } for (int i = 0; i < botAgents.length; i++) { if (userAgent.indexOf(botAgents[i]) != -1) { // set a key in the session, so the next time we don't have to manually // detect the robot again session.setAttribute(RequestConstants.ROBOT, Boolean.TRUE); return true; } } } // detect the robot from the host or user-agent String remoteHost = request.getRemoteHost(); // requires one DNS lookup // if the DNS server didn't return a hostname, getRemoteHost returns the // IP address, which is ignored here (the last char is checked, because some // remote hosts begin with the IP) if (remoteHost != null && remoteHost.length() > 0 && remoteHost.charAt(remoteHost.length() - 1) > 64) { for (int i = 0; i < botHosts.length; i++) { if (remoteHost.indexOf(botHosts[i]) != -1) { // set a key in the session, so the next time we don't have to manually // detect the robot again session.setAttribute(RequestConstants.ROBOT, Boolean.TRUE); return true; } } } // remote host and user agent are not in the predefined list, // so it must be an unknown robot or not a robot // indicate this session is not a robot session.setAttribute(RequestConstants.ROBOT, Boolean.FALSE); return false; } } }}
⌨️ 快捷键说明
复制代码
Ctrl + C
搜索代码
Ctrl + F
全屏模式
F11
切换主题
Ctrl + Shift + D
显示快捷键
?
增大字号
Ctrl + =
减小字号
Ctrl + -