stdlexical.scala

来自「JAVA 语言的函数式编程扩展」· SCALA 代码 · 共 91 行

SCALA
91
字号
/*                     __                                               *\**     ________ ___   / /  ___     Scala API                            ****    / __/ __// _ | / /  / _ |    (c) 2006-2007, LAMP/EPFL             ****  __\ \/ /__/ __ |/ /__/ __ |    http://scala-lang.org/               **** /____/\___/_/ |_/____/_/ | |                                         ****                          |/                                          **\*                                                                      */// $Id: StdLexical.scala 14415 2008-03-19 00:53:09Z mihaylov $package scala.util.parsing.combinator.lexicalimport scala.util.parsing.syntax._import scala.util.parsing.input.CharArrayReader.EofChimport collection.mutable.HashSet/** <p> *    This component provides a standard lexical parser for a simple, Scala-like language.  *    It parses keywords and identifiers, numeric literals (integers), strings, and delimiters.  *  </p> *  <p> *    To distinguish between identifiers and keywords, it uses a set of reserved identifiers:  *    every string contained in `reserved' is returned as a keyword token. *    (Note that "=>" is hard-coded as a keyword.)  *    Additionally, the kinds of delimiters can be specified by the `delimiters' set. *  </p> *  <p> *    Usually this component is used to break character-based input into bigger tokens, *    which are then passed to a token-parser {@see TokenParsers}. *  </p> * * @author Martin Odersky, Iulian Dragos, Adriaan Moors  */class StdLexical extends Lexical with StdTokens {  // see `token' in `Scanners'  def token: Parser[Token] =     ( letter ~ rep( letter | digit )                    ^^ { case first ~ rest => processIdent(first :: rest mkString "") }    | digit ~ rep( digit )                              ^^ { case first ~ rest => NumericLit(first :: rest mkString "") }    | '\'' ~ rep( chrExcept('\'', '\n', EofCh) ) ~ '\'' ^^ { case '\'' ~ chars ~ '\'' => StringLit(chars mkString "") }    | '\"' ~ rep( chrExcept('\"', '\n', EofCh) ) ~ '\"' ^^ { case '\"' ~ chars ~ '\"' => StringLit(chars mkString "") }    | EofCh                                             ^^^ EOF    | '\'' ~> failure("unclosed string literal")            | '\"' ~> failure("unclosed string literal")            | delim                                                 | failure("illegal character")    )  // see `whitespace in `Scanners'  def whitespace: Parser[Any] = rep(      whitespaceChar    | '/' ~ '*' ~ comment    | '/' ~ '/' ~ rep( chrExcept(EofCh, '\n') )    | '/' ~ '*' ~ failure("unclosed comment")    )  protected def comment: Parser[Any] = (      '*' ~ '/'  ^^ { case _ => ' '  }    | chrExcept(EofCh) ~ comment    )  /** The set of reserved identifiers: these will be returned as `Keyword's */  val reserved = new HashSet[String]  /** The set of delimiters (ordering does not matter) */  val delimiters = new HashSet[String]  protected def processIdent(name: String) =     if (reserved contains name) Keyword(name) else Identifier(name)  private var _delim: Parser[Token] = null  protected def delim: Parser[Token] = {    if (_delim eq null) { // construct parser for delimiters by |'ing together the parsers for the individual delimiters,     // starting with the longest one (hence the sort + reverse) -- otherwise a delimiter D will never be matched if     // there is another delimiter that is a prefix of D         def parseDelim(s: String): Parser[Token] = accept(s.toList) ^^ { x => Keyword(s) }            val d = new Array[String](delimiters.size)      delimiters.copyToArray(d,0)      scala.util.Sorting.quickSort(d)       _delim = d.toList.reverse.map(parseDelim).reduceRight[Parser[Token]](_ | _) // no offence :-)          }        _delim  }  private def lift[T](f: String => T)(xs: List[Char]): T = f(xs.mkString("", "", ""))  private def lift2[T](f: String => T)(p: ~[Char, List[Char]]): T = lift(f)(p._1 :: p._2)}

⌨️ 快捷键说明

复制代码Ctrl + C
搜索代码Ctrl + F
全屏模式F11
增大字号Ctrl + =
减小字号Ctrl + -
显示快捷键?