IPnom Home • Manuals • FreeBSD

 FreeBSD Man Pages

Man Sections:Commands (1)System Calls (2)Library Functions (3)Device Drivers (4)File Formats (5)Miscellaneous (7)System Utilities (8)
Keyword Live Search (10 results max):
 Type in part of a command in the search box.
 
Index:
  [(1)
  addftinfo(1)
  addr2line(1)
  afmtodit(1)
  alias(1)
  alloc(1)
  apply(1)
  apropos(1)
  ar(1)
  as(1)
  asa(1)
  at(1)
  atq(1)
  atrm(1)
  awk(1)
  b64decode(1)
  b64encode(1)
  basename(1)
  batch(1)
  bc(1)
  bdes(1)
  bg(1)
  biff(1)
  bind(1)
  bindkey(1)
  brandelf(1)
  break(1)
  breaksw(1)
  bsdtar(1)
  bsnmpd(1)
  bthost(1)
  btsockstat(1)
  builtin(1)
  builtins(1)
  bunzip2(1)
  byacc(1)
  bzcat(1)
  bzegrep(1)
  bzfgrep(1)
  bzgrep(1)
  bzip2(1)
  c++(1)
  c89(1)
  c99(1)
  cal(1)
  calendar(1)
  cap_mkdb(1)
  case(1)
  cat(1)
  catman(1)
  cc(1)
  cd(1)
  cdcontrol(1)
  chdir(1)
  checknr(1)
  chflags(1)
  chfn(1)
  chgrp(1)
  chio(1)
  chkey(1)
  chmod(1)
  chpass(1)
  chsh(1)
  ci(1)
  ckdist(1)
  cksum(1)
  clear(1)
  cmp(1)
  co(1)
  col(1)
  colcrt(1)
  colldef(1)
  colrm(1)
  column(1)
  comm(1)
  command(1)
  compile_et(1)
  complete(1)
  compress(1)
  continue(1)
  cp(1)
  cpio(1)
  cpp(1)
  crontab(1)
  crunchgen(1)
  crunchide(1)
  crypt(1)
  csh(1)
  csplit(1)
  ctags(1)
  ctm(1)
  ctm_dequeue(1)
  ctm_rmail(1)
  ctm_smail(1)
  cu(1)
  cursor(1)
  cut(1)
  cvs(1)
  date(1)
  dc(1)
  dd(1)
  default(1)
  df(1)
  dialog(1)
  diff(1)
  diff3(1)
  dig(1)
  dirname(1)
  dirs(1)
  do(1)
  domainname(1)
  done(1)
  dtmfdecode(1)
  du(1)
  echo(1)
  echotc(1)
  ed(1)
  edit(1)
  ee(1)
  egrep(1)
  elfdump(1)
  elif(1)
  else(1)
  end(1)
  endif(1)
  endsw(1)
  enigma(1)
  env(1)
  eqn(1)
  esac(1)
  eval(1)
  ex(1)
  exec(1)
  exit(1)
  expand(1)
  export(1)
  expr(1)
  f77(1)
  false(1)
  fc(1)
  fdformat(1)
  fdread(1)
  fdwrite(1)
  fetch(1)
  fg(1)
  fgrep(1)
  fi(1)
  file(1)
  file2c(1)
  filetest(1)
  find(1)
  finger(1)
  flex++(1)
  flex(1)
  fmt(1)
  fold(1)
  fontedit(1)
  for(1)
  foreach(1)
  from(1)
  fstat(1)
  fsync(1)
  ftp(1)
  g++(1)
  g711conv(1)
  gate-ftp(1)
  gcc(1)
  gcore(1)
  gcov(1)
  gdb(1)
  gencat(1)
  gensnmptree(1)
  getNAME(1)
  getconf(1)
  getfacl(1)
  getopt(1)
  getopts(1)
  glob(1)
  goto(1)
  gperf(1)
  gprof(1)
  grep(1)
  grn(1)
  grodvi(1)
  groff(1)
  grog(1)
  grolbp(1)
  grolj4(1)
  grops(1)
  grotty(1)
  groups(1)
  gtar(1)
  gunzip(1)
  gzcat(1)
  gzexe(1)
  gzip(1)
  hash(1)
  hashstat(1)
  hd(1)
  head(1)
  hesinfo(1)
  hexdump(1)
  history(1)
  host(1)
  hostname(1)
  hpftodit(1)
  hup(1)
  id(1)
  ident(1)
  idprio(1)
  if(1)
  indent(1)
  indxbib(1)
  info(1)
  install-info(1)
  install(1)
  intro(1)
  introduction(1)
  ipcrm(1)
  ipcs(1)
  ipftest(1)
  ipnat(1)
  ipresend(1)
  ipsend(1)
  iptest(1)
  jobid(1)
  jobs(1)
  join(1)
  jot(1)
  kbdcontrol(1)
  kbdmap(1)
  kcon(1)
  kdestroy(1)
  kdump(1)
  kenv(1)
  keylogin(1)
  keylogout(1)
  kgdb(1)
  kill(1)
  killall(1)
  kinit(1)
  klist(1)
  kpasswd(1)
  krb5-config(1)
  ktrace(1)
  lam(1)
  last(1)
  lastcomm(1)
  ld-elf.so.1(1)
  ld(1)
  ld(1)
  ldd(1)
  leave(1)
  less(1)
  lesskey(1)
  lex++(1)
  lex(1)
  limit(1)
  limits(1)
  link(1)
  lint(1)
  lkbib(1)
  ln(1)
  loadfont(1)
  locale(1)
  locate(1)
  lock(1)
  lockf(1)
  log(1)
  logger(1)
  login(1)
  logins(1)
  logname(1)
  logout(1)
  look(1)
  lookbib(1)
  lorder(1)
  lp(1)
  lpq(1)
  lpr(1)
  lprm(1)
  lptest(1)
  ls-F(1)
  ls(1)
  lsvfs(1)
  m4(1)
  mail(1)
  mailq(1)
  mailx(1)
  make(1)
  makeinfo(1)
  makewhatis(1)
  man(1)
  manpath(1)
  md5(1)
  merge(1)
  mesg(1)
  minigzip(1)
  mkdep(1)
  mkdir(1)
  mkfifo(1)
  mklocale(1)
  mkstr(1)
  mktemp(1)
  mmroff(1)
  more(1)
  mptable(1)
  msgs(1)
  mt(1)
  mv(1)
  nawk(1)
  nc(1)
  ncal(1)
  ncplist(1)
  ncplogin(1)
  ncplogout(1)
  neqn(1)
  netstat(1)
  newaliases(1)
  newgrp(1)
  nex(1)
  nfsstat(1)
  nice(1)
  nl(1)
  nm(1)
  nohup(1)
  notify(1)
  nroff(1)
  nslookup(1)
  nvi(1)
  nview(1)
  objcopy(1)
  objdump(1)
  objformat(1)
  od(1)
  omshell(1)
  onintr(1)
  opieinfo(1)
  opiekey(1)
  opiepasswd(1)
  otp-md4(1)
  otp-md5(1)
  otp-sha(1)
  pagesize(1)
  passwd(1)
  paste(1)
  patch(1)
  pathchk(1)
  pawd(1)
  pax(1)
  pfbtops(1)
  pftp(1)
  pgrep(1)
  pic(1)
  pkg_add(1)
  pkg_check(1)
  pkg_create(1)
  pkg_delete(1)
  pkg_info(1)
  pkg_sign(1)
  pkg_version(1)
  pkill(1)
  popd(1)
  pr(1)
  printenv(1)
  printf(1)
  ps(1)
  psroff(1)
  pushd(1)
  pwd(1)
  quota(1)
  ranlib(1)
  rcp(1)
  rcs(1)
  rcsclean(1)
  rcsdiff(1)
  rcsfreeze(1)
  rcsintro(1)
  rcsmerge(1)
  read(1)
  readelf(1)
  readlink(1)
  readonly(1)
  realpath(1)
  red(1)
  ree(1)
  refer(1)
  rehash(1)
  repeat(1)
  reset(1)
  rev(1)
  rfcomm_sppd(1)
  rlog(1)
  rlogin(1)
  rm(1)
  rmd160(1)
  rmdir(1)
  rpcgen(1)
  rs(1)
  rsh(1)
  rtld(1)
  rtprio(1)
  rup(1)
  ruptime(1)
  rusers(1)
  rwall(1)
  rwho(1)
  sched(1)
  scon(1)
  scp(1)
  script(1)
  sdiff(1)
  sed(1)
  send-pr(1)
  sendbug(1)
  set(1)
  setenv(1)
  setfacl(1)
  settc(1)
  setty(1)
  setvar(1)
  sftp(1)
  sh(1)
  sha1(1)
  shar(1)
  shift(1)
  size(1)
  sleep(1)
  slogin(1)
  smbutil(1)
  sockstat(1)
  soelim(1)
  sort(1)
  source(1)
  split(1)
  sscop(1)
  ssh-add(1)
  ssh-agent(1)
  ssh-keygen(1)
  ssh-keyscan(1)
  ssh(1)
  startslip(1)
  stat(1)
  stop(1)
  strings(1)
  strip(1)
  stty(1)
  su(1)
  sum(1)
  suspend(1)
  switch(1)
  systat(1)
  tabs(1)
  tail(1)
  talk(1)
  tar(1)
  tbl(1)
  tcopy(1)
  tcpdump(1)
  tcpslice(1)
  tcsh(1)
  tee(1)
  telltc(1)
  telnet(1)
  test(1)
  texindex(1)
  tfmtodit(1)
  tftp(1)
  then(1)
  time(1)
  tip(1)
  top(1)
  touch(1)
  tput(1)
  tr(1)
  trace(1)
  trap(1)
  troff(1)
  true(1)
  truncate(1)
  truss(1)
  tset(1)
  tsort(1)
  tty(1)
  type(1)
  ul(1)
  ulimit(1)
  umask(1)
  unalias(1)
  uname(1)
  uncomplete(1)
  uncompress(1)
  unexpand(1)
  unhash(1)
  unifdef(1)
  unifdefall(1)
  uniq(1)
  units(1)
  unlimit(1)
  unlink(1)
  unset(1)
  unsetenv(1)
  until(1)
  unvis(1)
  uptime(1)
  usbhidaction(1)
  usbhidctl(1)
  users(1)
  uudecode(1)
  uuencode(1)
  uuidgen(1)
  vacation(1)
  vgrind(1)
  vi(1)
  vidcontrol(1)
  vidfont(1)
  view(1)
  vis(1)
  vt220keys(1)
  vttest(1)
  w(1)
  wait(1)
  wall(1)
  wc(1)
  what(1)
  whatis(1)
  where(1)
  whereis(1)
  which(1)
  while(1)
  who(1)
  whoami(1)
  whois(1)
  window(1)
  write(1)
  xargs(1)
  xstr(1)
  yacc(1)
  yes(1)
  ypcat(1)
  ypchfn(1)
  ypchpass(1)
  ypchsh(1)
  ypmatch(1)
  yppasswd(1)
  ypwhich(1)
  yyfix(1)
  zcat(1)
  zcmp(1)
  zdiff(1)
  zegrep(1)
  zfgrep(1)
  zforce(1)
  zgrep(1)
  zmore(1)
  znew(1)

flex(1)

NAME

       flex - fast lexical analyzer generator


SYNOPSIS

       flex [-bcdfhilnpstvwBFILTV78+? -C[aefFmr] -ooutput -Pprefix -Sskeleton]
       [--help --version] [filename ...]


OVERVIEW

       This manual describes flex, a tool for generating programs that perform
       pattern-matching on text.  The manual includes both tutorial and refer-
       ence sections:

	   Description
	       a brief overview of the tool

	   Some Simple Examples

	   Format Of The Input File

	   Patterns
	       the extended regular expressions used by flex

	   How The Input Is Matched
	       the rules for determining what has been matched

	   Actions
	       how to specify what to do when a pattern is matched

	   The Generated Scanner
	       details regarding the scanner that flex produces;
	       how to control the input source

	   Start Conditions
	       introducing context into your scanners, and
	       managing "mini-scanners"

	   Multiple Input Buffers
	       how to manipulate multiple input sources; how to
	       scan from strings instead of files

	   End-of-file Rules
	       special rules for matching the end of the input

	   Miscellaneous Macros
	       a summary of macros available to the actions

	   Values Available To The User
	       a summary of values available to the actions

	   Interfacing With Yacc
	       connecting flex scanners together with yacc parsers

	   Options
	       flex command-line options, and the "%option"
	       directive

	   Incompatibilities With Lex And POSIX
	       how flex differs from AT&T lex and the POSIX lex
	       standard

	   Diagnostics
	       those error messages produced by flex (or scanners
	       it generates) whose meanings might not be apparent

	   Files
	       files used by flex

	   Deficiencies / Bugs
	       known problems with flex

	   See Also
	       other documentation, related tools

	   Author
	       includes contact information


DESCRIPTION

       flex is a tool for generating scanners: programs which recognize  lexi-
       cal  patterns  in text.	flex reads the given input files, or its stan-
       dard input if no file names are given, for a description of  a  scanner
       to  generate.   The  description  is  in  the  form of pairs of regular
       expressions and C code, called rules. flex  generates  as  output  a  C
       source  file,  lex.yy.c, which defines a routine yylex().  This file is
       compiled and linked with the -ll  library  to  produce  an  executable.
       When  the  executable  is run, it analyzes its input for occurrences of
       the regular expressions.  Whenever it finds one, it executes the corre-
       sponding C code.


SOME SIMPLE EXAMPLES

       First some simple examples to get the flavor of how one uses flex.  The
       following flex input specifies a scanner which whenever	it  encounters
       the string "username" will replace it with the user's login name:

	   %%
	   username    printf( "%s", getlogin() );

       By  default,  any  text	not matched by a flex scanner is copied to the
       output, so the net effect of this scanner is to copy its input file  to
       its output with each occurrence of "username" expanded.	In this input,
       there is just one rule.	"username" is the pattern and the "printf"  is
       the action.  The "%%" marks the beginning of the rules.

       Here's another simple example:

	   %{
		   int num_lines = 0, num_chars = 0;
	   %}

	   %%
	   \n	   ++num_lines; ++num_chars;
	   .	   ++num_chars;

	   %%
       its input (it produces no output other than the	final  report  on  the
       counts).    The	first  line  declares  two  globals,  "num_lines"  and
       "num_chars", which are accessible both inside yylex() and in the main()
       routine declared after the second "%%".	There are two rules, one which
       matches a newline ("\n") and increments both the  line  count  and  the
       character  count, and one which matches any character other than a new-
       line (indicated by the "." regular expression).

       A somewhat more complicated example:

	   /* scanner for a toy Pascal-like language */

	   %{
	   /* need this for the call to atof() below */
	   #include <math.h>
	   %}

	   DIGIT    [0-9]
	   ID	    [a-z][a-z0-9]*

	   %%

	   {DIGIT}+    {
		       printf( "An integer: %s (%d)\n", yytext,
			       atoi( yytext ) );
		       }

	   {DIGIT}+"."{DIGIT}*	      {
		       printf( "A float: %s (%g)\n", yytext,
			       atof( yytext ) );
		       }

	   if|then|begin|end|procedure|function        {
		       printf( "A keyword: %s\n", yytext );
		       }

	   {ID}        printf( "An identifier: %s\n", yytext );

	   "+"|"-"|"*"|"/"   printf( "An operator: %s\n", yytext );

	   "{"[^}\n]*"}"     /* eat up one-line comments */

	   [ \t\n]+	     /* eat up whitespace */

	   .	       printf( "Unrecognized character: %s\n", yytext );

	   %%

	   main( argc, argv )
	   int argc;
	   char **argv;
	       {
	       ++argv, --argc;	/* skip over program name */
	       if ( argc > 0 )
		       yyin = fopen( argv[0], "r" );
	       else
		       yyin = stdin;

       tions.


FORMAT OF THE INPUT FILE

       The  flex  input  file  consists of three sections, separated by a line
       with just %% in it:

	   definitions
	   %%
	   rules
	   %%
	   user code

       The definitions section contains declarations of  simple  name  defini-
       tions  to simplify the scanner specification, and declarations of start
       conditions, which are explained in a later section.

       Name definitions have the form:

	   name definition

       The "name" is a word beginning with a letter  or  an  underscore  ('_')
       followed by zero or more letters, digits, '_', or '-' (dash).  The def-
       inition is taken to begin at the first non-white-space  character  fol-
       lowing  the name and continuing to the end of the line.	The definition
       can subsequently be referred to using "{name}", which  will  expand  to
       "(definition)".	For example,

	   DIGIT    [0-9]
	   ID	    [a-z][a-z0-9]*

       defines	"DIGIT"  to  be  a  regular  expression which matches a single
       digit, and "ID" to be a regular expression which matches a letter  fol-
       lowed by zero-or-more letters-or-digits.  A subsequent reference to

	   {DIGIT}+"."{DIGIT}*

       is identical to

	   ([0-9])+"."([0-9])*

       and  matches  one-or-more digits followed by a '.' followed by zero-or-
       more digits.

       The rules section of the flex input contains a series of rules  of  the
       form:

	   pattern   action

       where  the  pattern must be unindented and the action must begin on the
       same line.

       See below for a further description of patterns and actions.

       Finally, the user code section is simply copied to  lex.yy.c  verbatim.
       It is used for companion routines which call or are called by the scan-
       ner.  The presence of this section is optional; if it is  missing,  the
       second %% in the input file may be skipped, too.

       text in the rule section is still copied to the output, but its meaning
       is not well-defined and it may well  cause  compile-time  errors  (this
       feature	is present for POSIX compliance; see below for other such fea-
       tures).

       In the definitions section (but not in the  rules  section),  an  unin-
       dented comment (i.e., a line beginning with "/*") is also copied verba-
       tim to the output up to the next "*/".


PATTERNS

       The patterns in the input are written using an extended set of  regular
       expressions.  These are:

	   x	      match the character 'x'
	   .	      any character (byte) except newline
	   [xyz]      a "character class"; in this case, the pattern
			matches either an 'x', a 'y', or a 'z'
	   [abj-oZ]   a "character class" with a range in it; matches
			an 'a', a 'b', any letter from 'j' through 'o',
			or a 'Z'
	   [^A-Z]     a "negated character class", i.e., any character
			but those in the class.  In this case, any
			character EXCEPT an uppercase letter.
	   [^A-Z\n]   any character EXCEPT an uppercase letter or
			a newline
	   r*	      zero or more r's, where r is any regular expression
	   r+	      one or more r's
	   r?	      zero or one r's (that is, "an optional r")
	   r{2,5}     anywhere from two to five r's
	   r{2,}      two or more r's
	   r{4}       exactly 4 r's
	   {name}     the expansion of the "name" definition
		      (see above)
	   "[xyz]\"foo"
		      the literal string: [xyz]"foo
	   \X	      if X is an 'a', 'b', 'f', 'n', 'r', 't', or 'v',
			then the ANSI-C interpretation of \x.
			Otherwise, a literal 'X' (used to escape
			operators such as '*')
	   \0	      a NUL character (ASCII code 0)
	   \123       the character with octal value 123
	   \x2a       the character with hexadecimal value 2a
	   (r)	      match an r; parentheses are used to override
			precedence (see below)

	   rs	      the regular expression r followed by the
			regular expression s; called "concatenation"

	   r|s	      either an r or an s

	   r/s	      an r but only if it is followed by an s.	The
			text matched by s is included when determining
			whether this rule is the "longest match",
			but is then returned to the input before
			the action is executed.  So the action only
			newline has been scanned).
	   r$	      an r, but only at the end of a line (i.e., just
			before a newline).  Equivalent to "r/\n".

		      Note that flex's notion of "newline" is exactly
		      whatever the C compiler used to compile flex
		      interprets '\n' as; in particular, on some DOS
		      systems you must either filter out \r's in the
		      input yourself, or explicitly use r/\r\n for "r$".

	   <s>r       an r, but only in start condition s (see
			below for discussion of start conditions)
	   <s1,s2,s3>r
		      same, but in any of start conditions s1,
			s2, or s3
	   <*>r       an r in any start condition, even an exclusive one.

	   <<EOF>>    an end-of-file
	   <s1,s2><<EOF>>
		      an end-of-file when in start condition s1 or s2

       Note that inside of a character class, all regular expression operators
       lose their special meaning except escape ('\') and the character  class
       operators, '-', ']', and, at the beginning of the class, '^'.

       The  regular  expressions  listed above are grouped according to prece-
       dence, from highest precedence at the top  to  lowest  at  the  bottom.
       Those grouped together have equal precedence.  For example,

	   foo|bar*

       is the same as

	   (foo)|(ba(r*))

       since  the  '*'	operator has higher precedence than concatenation, and
       concatenation higher than alternation ('|').   This  pattern  therefore
       matches either the string "foo" or the string "ba" followed by zero-or-
       more r's.  To match "foo" or zero-or-more "bar"'s, use:

	   foo|(bar)*

       and to match zero-or-more "foo"'s-or-"bar"'s:

	   (foo|bar)*

       In addition to characters and ranges of characters,  character  classes
       can  also  contain  character class expressions.  These are expressions
       enclosed inside [: and :]  delimiters  (which  themselves  must	appear
       between	the  '['  and  ']'  of the character class; other elements may
       occur inside the character class, too).	The valid expressions are:

	   [:alnum:] [:alpha:] [:blank:]
	   [:cntrl:] [:digit:] [:graph:]
	   [:lower:] [:print:] [:punct:]
       For example, the following character classes are all equivalent:

	   [[:alnum:]]
	   [[:alpha:][:digit:]]
	   [[:alpha:]0-9]
	   [a-zA-Z0-9]

       If your scanner is case-insensitive (the -i flag), then	[:upper:]  and
       [:lower:] are equivalent to [:alpha:].

       Some notes on patterns:

       -      A  negated  character  class  such as the example "[^A-Z]" above
	      will match a  newline  unless  "\n"  (or	an  equivalent	escape
	      sequence)  is  one  of  the characters explicitly present in the
	      negated character class (e.g., "[^A-Z\n]").  This is unlike  how
	      many  other  regular  expression	tools  treat negated character
	      classes, but unfortunately  the  inconsistency  is  historically
	      entrenched.   Matching  newlines means that a pattern like [^"]*
	      can match the entire input unless there's another quote  in  the
	      input.

       -      A  rule  can  have at most one instance of trailing context (the
	      '/' operator or the '$' operator).  The  start  condition,  '^',
	      and "<<EOF>>" patterns can only occur at the beginning of a pat-
	      tern, and, as well as with '/' and '$', cannot be grouped inside
	      parentheses.   A	'^' which does not occur at the beginning of a
	      rule or a '$' which does not occur at the end of	a  rule  loses
	      its special properties and is treated as a normal character.

	      The following are illegal:

		  foo/bar$
		  <sc1>foo<sc2>bar

	      Note that the first of these, can be written "foo/bar\n".

	      The  following will result in '$' or '^' being treated as a nor-
	      mal character:

		  foo|(bar$)
		  foo|^bar

	      If what's wanted is a "foo" or a bar-followed-by-a-newline,  the
	      following  could	be  used  (the special '|' action is explained
	      below):

		  foo	   |
		  bar$	   /* action goes here */

	      A similar trick will work for matching a foo  or	a  bar-at-the-
	      beginning-of-a-line.


HOW THE INPUT IS MATCHED

       When  the  generated  scanner is run, it analyzes its input looking for
       strings which match any of its patterns.  If it	finds  more  than  one
       match,  it  takes  the one matching the most text (for trailing context
       rules, this includes the length of the trailing part,  even  though  it
       description  of	actions  follows),  and  then  the  remaining input is
       scanned for another match.

       If no match is found, then the default rule is executed: the next char-
       acter  in  the  input  is considered matched and copied to the standard
       output.	Thus, the simplest legal flex input is:

	   %%

       which generates a scanner that simply copies its input  (one  character
       at a time) to its output.

       Note  that  yytext  can	be  defined in two different ways: either as a
       character pointer or as a character array.  You can control which defi-
       nition flex uses by including one of the special directives %pointer or
       %array in the first (definitions) section  of  your  flex  input.   The
       default is %pointer, unless you use the -l lex compatibility option, in
       which case yytext will be an array.  The advantage of using %pointer is
       substantially faster scanning and no buffer overflow when matching very
       large tokens (unless you run out of dynamic memory).  The  disadvantage
       is  that  you are restricted in how your actions can modify yytext (see
       the next section), and calls  to  the  unput()  function  destroys  the
       present	contents  of  yytext,  which  can  be  a  considerable porting
       headache when moving between different lex versions.

       The advantage of %array is that you can	then  modify  yytext  to  your
       heart's	content,  and  calls  to  unput()  do  not destroy yytext (see
       below).	Furthermore, existing lex  programs  sometimes	access	yytext
       externally using declarations of the form:
	   extern char yytext[];
       This  definition  is erroneous when used with %pointer, but correct for
       %array.

       %array defines yytext to  be  an  array	of  YYLMAX  characters,  which
       defaults  to  a	fairly large value.  You can change the size by simply
       #define'ing YYLMAX to a different value in the first  section  of  your
       flex input.  As mentioned above, with %pointer yytext grows dynamically
       to accommodate large tokens.  While this means  your  %pointer  scanner
       can  accommodate  very  large tokens (such as matching entire blocks of
       comments), bear in mind that each time the scanner must	resize	yytext
       it  also  must  rescan the entire token from the beginning, so matching
       such tokens can prove slow.  yytext presently does not dynamically grow
       if  a  call  to	unput()  results  in  too much text being pushed back;
       instead, a run-time error results.

       Also note that you cannot use %array with C++ scanner classes (the  c++
       option; see below).


ACTIONS

       Each  pattern  in  a  rule has a corresponding action, which can be any
       arbitrary C statement.  The  pattern  ends  at  the  first  non-escaped
       whitespace  character; the remainder of the line is its action.	If the
       action is empty, then when the pattern is matched the  input  token  is
       simply discarded.  For example, here is the specification for a program
       which deletes all occurrences of "zap me" from its input:

	   %%
	   "zap me"
	   [ \t]+	 putchar( ' ' );
	   [ \t]+$	 /* ignore this token */

       If the action contains a '{', then the action spans till the  balancing
       '}'  is	found,	and  the  action may cross multiple lines.  flex knows
       about C strings and comments and won't be fooled by braces found within
       them,  but  also  allows actions to begin with %{ and will consider the
       action to be all the text up to the next  %}  (regardless  of  ordinary
       braces inside the action).

       An  action consisting solely of a vertical bar ('|') means "same as the
       action for the next rule."  See below for an illustration.

       Actions can include arbitrary C code, including	return	statements  to
       return  a  value to whatever routine called yylex().  Each time yylex()
       is called it continues processing tokens from where it  last  left  off
       until it either reaches the end of the file or executes a return.

       Actions	are  free  to  modify yytext except for lengthening it (adding
       characters to its end--these will overwrite  later  characters  in  the
       input  stream).	 This  however	does  not apply when using %array (see
       above); in that case, yytext may be freely modified in any way.

       Actions are free to modify yyleng except they should not do so  if  the
       action also includes use of yymore() (see below).

       There  are  a number of special directives which can be included within
       an action:

       -      ECHO copies yytext to the scanner's output.

       -      BEGIN followed by the name of a start condition places the scan-
	      ner in the corresponding start condition (see below).

       -      REJECT  directs  the  scanner to proceed on to the "second best"
	      rule which matched the input (or a prefix of  the  input).   The
	      rule is chosen as described above in "How the Input is Matched",
	      and yytext and yyleng set up appropriately.  It  may  either  be
	      one which matched as much text as the originally chosen rule but
	      came later in the flex input file, or  one  which  matched  less
	      text.   For  example, the following will both count the words in
	      the input and call the  routine  special()  whenever  "frob"  is
	      seen:

			  int word_count = 0;
		  %%

		  frob	      special(); REJECT;
		  [^ \t\n]+   ++word_count;

	      Without  the  REJECT,  any  "frob"'s  in	the input would not be
	      counted as words, since the scanner normally executes  only  one
	      action per token.  Multiple REJECT's are allowed, each one find-
	      ing the next best choice to  the	currently  active  rule.   For
	      example,	when  the following scanner scans the token "abcd", it
	      will write "abcdabcaba" to the output:

	      the  special  '|'  action.)   REJECT is a particularly expensive
	      feature in terms of scanner performance; if it is used in any of
	      the  scanner's  actions  it  will slow down all of the scanner's
	      matching.  Furthermore, REJECT cannot be used with  the  -Cf  or
	      -CF options (see below).

	      Note  also  that	unlike	the other special actions, REJECT is a
	      branch; code immediately following it in the action will not  be
	      executed.

       -      yymore() tells the scanner that the next time it matches a rule,
	      the corresponding token should  be  appended  onto  the  current
	      value  of  yytext  rather than replacing it.  For example, given
	      the input "mega-kludge" the  following  will  write  "mega-mega-
	      kludge" to the output:

		  %%
		  mega-    ECHO; yymore();
		  kludge   ECHO;

	      First  "mega-"  is  matched  and	echoed	to  the  output.  Then
	      "kludge" is matched, but the previous "mega-" is	still  hanging
	      around  at  the beginning of yytext so the ECHO for the "kludge"
	      rule will actually write "mega-kludge".

       Two notes regarding use of yymore().  First, yymore()  depends  on  the
       value  of yyleng correctly reflecting the size of the current token, so
       you must not modify yyleng if you  are  using  yymore().   Second,  the
       presence  of  yymore()  in the scanner's action entails a minor perfor-
       mance penalty in the scanner's matching speed.

       -      yyless(n) returns all but the first n characters of the  current
	      token  back  to  the  input stream, where they will be rescanned
	      when the scanner looks for the next match.   yytext  and	yyleng
	      are  adjusted appropriately (e.g., yyleng will now be equal to n
	      ).  For example, on the input "foobar" the following will  write
	      out "foobarbar":

		  %%
		  foobar    ECHO; yyless(3);
		  [a-z]+    ECHO;

	      An  argument  of 0 to yyless will cause the entire current input
	      string to be scanned again.  Unless you've changed how the scan-
	      ner  will subsequently process its input (using BEGIN, for exam-
	      ple), this will result in an endless loop.

       Note that yyless is a macro and can only be  used  in  the  flex  input
       file, not from other source files.

       -      unput(c)	puts  the  character c back onto the input stream.  It
	      will be the next character scanned.  The following  action  will
	      take  the current token and cause it to be rescanned enclosed in
	      parentheses.

		  {
		  int i;
		  /* Copy yytext because unput() trashes yytext */
	      Note that since each unput() puts the given  character  back  at
	      the  beginning of the input stream, pushing back strings must be
	      done back-to-front.

       An important potential problem when using unput() is that  if  you  are
       using  %pointer	(the default), a call to unput() destroys the contents
       of yytext, starting with its  rightmost	character  and	devouring  one
       character  to the left with each call.  If you need the value of yytext
       preserved after a call to unput() (as in the above example),  you  must
       either  first  copy  it	elsewhere,  or build your scanner using %array
       instead (see How The Input Is Matched).

       Finally, note that you cannot put back EOF to attempt to mark the input
       stream with an end-of-file.

       -      input()  reads  the  next  character from the input stream.  For
	      example, the following is one way to eat up C comments:

		  %%
		  "/*"	      {
			      register int c;

			      for ( ; ; )
				  {
				  while ( (c = input()) != '*' &&
					  c != EOF )
				      ;    /* eat up text of comment */

				  if ( c == '*' )
				      {
				      while ( (c = input()) == '*' )
					  ;
				      if ( c == '/' )
					  break;    /* found the end */
				      }

				  if ( c == EOF )
				      {
				      error( "EOF in comment" );
				      break;
				      }
				  }
			      }

	      (Note that if the scanner is compiled using C++, then input() is
	      instead referred to as yyinput(), in order to avoid a name clash
	      with the C++ stream by the name of input.)

       -      YY_FLUSH_BUFFER flushes the scanner's internal  buffer  so  that
	      the  next  time  the  scanner attempts to match a token, it will
	      first refill the buffer using YY_INPUT (see The Generated  Scan-
	      ner,  below).  This action is a special case of the more general
	      yy_flush_buffer() function, described below in the section  Mul-
	      tiple Input Buffers.

       -      yyterminate()  can  be  used in lieu of a return statement in an
	      action.  It terminates the scanner and returns a 0 to the  scan-
	      ner's  caller, indicating "all done".  By default, yyterminate()

	   int yylex()
	       {
	       ... various definitions and the actions in here ...
	       }

       (If your environment supports function prototypes, then it will be "int
       yylex(  void  )".)   This  definition  may  be  changed by defining the
       "YY_DECL" macro.  For example, you could use:

	   #define YY_DECL float lexscan( a, b ) float a, b;

       to give the scanning routine the name lexscan, returning a  float,  and
       taking two floats as arguments.	Note that if you give arguments to the
       scanning routine using a K&R-style/non-prototyped function declaration,
       you must terminate the definition with a semi-colon (;).

       Whenever  yylex() is called, it scans tokens from the global input file
       yyin (which defaults to stdin).	It continues until it  either  reaches
       an  end-of-file	(at  which point it returns the value 0) or one of its
       actions executes a return statement.

       If the scanner reaches an end-of-file, subsequent calls	are  undefined
       unless  either yyin is pointed at a new input file (in which case scan-
       ning continues from that file), or yyrestart() is called.   yyrestart()
       takes  one  argument, a FILE * pointer (which can be nil, if you've set
       up YY_INPUT to scan from a source other	than  yyin),  and  initializes
       yyin  for  scanning from that file.  Essentially there is no difference
       between just assigning yyin to a new input file or using yyrestart() to
       do so; the latter is available for compatibility with previous versions
       of flex, and because it can be used to switch input files in the middle
       of  scanning.   It  can	also  be  used to throw away the current input
       buffer, by calling it with an argument of yyin; but better  is  to  use
       YY_FLUSH_BUFFER	(see above).  Note that yyrestart() does not reset the
       start condition to INITIAL (see Start Conditions, below).

       If yylex() stops scanning due to executing a return statement in one of
       the  actions,  the  scanner may then be called again and it will resume
       scanning where it left off.

       By default (and for purposes of efficiency), the  scanner  uses	block-
       reads  rather  than  simple  getc() calls to read characters from yyin.
       The nature of how it gets its input can be controlled by  defining  the
       YY_INPUT      macro.	  YY_INPUT's	  calling      sequence     is
       "YY_INPUT(buf,result,max_size)".  Its action is to place up to max_size
       characters  in  the character array buf and return in the integer vari-
       able result either the  number  of  characters  read  or  the  constant
       YY_NULL	(0  on	Unix  systems)	to indicate EOF.  The default YY_INPUT
       reads from the global file-pointer "yyin".

       A sample definition of YY_INPUT (in  the  definitions  section  of  the
       input file):

	   %{
	   #define YY_INPUT(buf,result,max_size) \
	       { \
	       int c = getchar(); \
	       result = (c == EOF) ? YY_NULL : (buf[0] = c, 1); \

       then it is assumed that the function has gone ahead and set up yyin  to
       point  to  another  input  file, and scanning continues.  If it returns
       true (non-zero), then  the  scanner  terminates,  returning  0  to  its
       caller.	 Note  that  in  either  case,	the  start  condition  remains
       unchanged; it does not revert to INITIAL.

       If you do not supply your own version of yywrap(), then you must either
       use  %option  noyywrap  (in  which  case  the scanner behaves as though
       yywrap() returned 1), or you must link with -ll to obtain  the  default
       version of the routine, which always returns 1.

       Three routines are available for scanning from in-memory buffers rather
       than files: yy_scan_string(),  yy_scan_bytes(),	and  yy_scan_buffer().
       See the discussion of them below in the section Multiple Input Buffers.

       The scanner writes its ECHO output to the yyout global  (default,  std-
       out), which may be redefined by the user simply by assigning it to some
       other FILE pointer.


START CONDITIONS

       flex provides a mechanism for conditionally activating rules.  Any rule
       whose  pattern  is  prefixed  with  "<sc>" will only be active when the
       scanner is in the start condition named "sc".  For example,

	   <STRING>[^"]*	{ /* eat up the string body ... */
		       ...
		       }

       will be active only when the scanner is in the  "STRING"  start	condi-
       tion, and

	   <INITIAL,STRING,QUOTE>\.	   { /* handle an escape ... */
		       ...
		       }

       will  be  active  only when the current start condition is either "INI-
       TIAL", "STRING", or "QUOTE".

       Start conditions are declared in the definitions (first) section of the
       input using unindented lines beginning with either %s or %x followed by
       a list of names.  The former declares inclusive start  conditions,  the
       latter  exclusive  start  conditions.   A  start condition is activated
       using the BEGIN action.	Until the next BEGIN action is executed, rules
       with  the  given  start	condition  will be active and rules with other
       start conditions will be inactive.  If the start  condition  is	inclu-
       sive,  then  rules with no start conditions at all will also be active.
       If it is exclusive, then only rules qualified with the start  condition
       will  be active.  A set of rules contingent on the same exclusive start
       condition describe a scanner which is independent of any of  the  other
       rules  in  the flex input.  Because of this, exclusive start conditions
       make it easy to specify "mini-scanners"	which  scan  portions  of  the
       input  that are syntactically different from the rest (e.g., comments).

       If the distinction between inclusive and exclusive start conditions  is
       still  a little vague, here's a simple example illustrating the connec-
       tion between the two.  The set of rules:

	   %s example
	   %x example
	   %%

	   <example>foo   do_something();

	   <INITIAL,example>bar    something_else();

       Without the <INITIAL,example> qualifier, the bar pattern in the	second
       example	wouldn't be active (i.e., couldn't match) when in start condi-
       tion example.  If we just used <example> to qualify bar,  though,  then
       it  would  only	be  active in example and not in INITIAL, while in the
       first example it's active in both, because in  the  first  example  the
       example start condition is an inclusive (%s) start condition.

       Also  note that the special start-condition specifier <*> matches every
       start condition.  Thus, the above example could also have been written;

	   %x example
	   %%

	   <example>foo   do_something();

	   <*>bar    something_else();

       The  default  rule  (to ECHO any unmatched character) remains active in
       start conditions.  It is equivalent to:

	   <*>.|\n     ECHO;

       BEGIN(0) returns to the original state where only  the  rules  with  no
       start conditions are active.  This state can also be referred to as the
       start-condition "INITIAL", so BEGIN(INITIAL) is equivalent to BEGIN(0).
       (The  parentheses  around the start condition name are not required but
       are considered good style.)

       BEGIN actions can also be given as indented code at  the  beginning  of
       the  rules  section.  For example, the following will cause the scanner
       to enter the "SPECIAL" start condition whenever yylex() is  called  and
       the global variable enter_special is true:

		   int enter_special;

	   %x SPECIAL
	   %%
		   if ( enter_special )
		       BEGIN(SPECIAL);

	   <SPECIAL>blahblahblah
	   ...more rules follow...

       To  illustrate  the  uses  of start conditions, here is a scanner which
       provides two different interpretations of a string like "123.456".   By
       default	it  will  treat  it  as three tokens, the integer "123", a dot
       ('.'), and the integer "456".  But if the string is preceded earlier in
       the  line  by  the  string "expect-floats" it will treat it as a single

	   expect-floats	BEGIN(expect);

	   <expect>[0-9]+"."[0-9]+	{
		       printf( "found a float, = %f\n",
			       atof( yytext ) );
		       }
	   <expect>\n		{
		       /* that's the end of the line, so
			* we need another "expect-number"
			* before we'll recognize any more
			* numbers
			*/
		       BEGIN(INITIAL);
		       }

	   [0-9]+      {
		       printf( "found an integer, = %d\n",
			       atoi( yytext ) );
		       }

	   "."	       printf( "found a dot\n" );

       Here is a scanner which recognizes  (and  discards)  C  comments  while
       maintaining a count of the current input line.

	   %x comment
	   %%
		   int line_num = 1;

	   "/*" 	BEGIN(comment);

	   <comment>[^*\n]*	   /* eat anything that's not a '*' */
	   <comment>"*"+[^*/\n]*   /* eat up '*'s not followed by '/'s */
	   <comment>\n		   ++line_num;
	   <comment>"*"+"/"	   BEGIN(INITIAL);

       This scanner goes to a bit of trouble to match as much text as possible
       with each rule.	In general, when  attempting  to  write  a  high-speed
       scanner	try to match as much possible in each rule, as it's a big win.

       Note that start-conditions names are really integer values and  can  be
       stored  as  such.   Thus,  the above could be extended in the following
       fashion:

	   %x comment foo
	   %%
		   int line_num = 1;
		   int comment_caller;

	   "/*" 	{
			comment_caller = INITIAL;
			BEGIN(comment);
			}

	   ...

	   <foo>"/*"	{
			comment_caller = foo;
       Furthermore, you can access the current start condition using the inte-
       ger-valued  YY_START macro.  For example, the above assignments to com-
       ment_caller could instead be written

	   comment_caller = YY_START;

       Flex provides YYSTATE as an alias for YY_START (since  that  is	what's
       used by AT&T lex).

       Note  that  start conditions do not have their own name-space; %s's and
       %x's declare names in the same fashion as #define's.

       Finally, here's an example of how to match C-style quoted strings using
       exclusive  start  conditions,  including expanded escape sequences (but
       not including checking for a string that's too long):

	   %x str

	   %%
		   char string_buf[MAX_STR_CONST];
		   char *string_buf_ptr;

	   \"	   string_buf_ptr = string_buf; BEGIN(str);

	   <str>\"	  { /* saw closing quote - all done */
		   BEGIN(INITIAL);
		   *string_buf_ptr = '\0';
		   /* return string constant token type and
		    * value to parser
		    */
		   }

	   <str>\n	  {
		   /* error - unterminated string constant */
		   /* generate error message */
		   }

	   <str>\\[0-7]{1,3} {
		   /* octal escape sequence */
		   int result;

		   (void) sscanf( yytext + 1, "%o", &result );

		   if ( result > 0xff )
			   /* error, constant is out-of-bounds */

		   *string_buf_ptr++ = result;
		   }

	   <str>\\[0-9]+ {
		   /* generate error - bad escape sequence; something
		    * like '\48' or '\0777777'
		    */
		   }

	   <str>\\n  *string_buf_ptr++ = '\n';
	   <str>\\t  *string_buf_ptr++ = '\t';

		   while ( *yptr )
			   *string_buf_ptr++ = *yptr++;
		   }

       Often, such as in some of the examples above, you  wind	up  writing  a
       whole bunch of rules all preceded by the same start condition(s).  Flex
       makes this a little easier and cleaner by introducing a notion of start
       condition scope.  A start condition scope is begun with:

	   <SCs>{

       where  SCs is a list of one or more start conditions.  Inside the start
       condition scope, every rule automatically has the prefix <SCs>  applied
       to it, until a '}' which matches the initial '{'.  So, for example,

	   <ESC>{
	       "\\n"   return '\n';
	       "\\r"   return '\r';
	       "\\f"   return '\f';
	       "\\0"   return '\0';
	   }

       is equivalent to:

	   <ESC>"\\n"  return '\n';
	   <ESC>"\\r"  return '\r';
	   <ESC>"\\f"  return '\f';
	   <ESC>"\\0"  return '\0';

       Start condition scopes may be nested.

       Three  routines	are  available for manipulating stacks of start condi-
       tions:

       void yy_push_state(int new_state)
	      pushes the current start condition onto the  top	of  the  start
	      condition stack and switches to new_state as though you had used
	      BEGIN new_state (recall that  start  condition  names  are  also
	      integers).

       void yy_pop_state()
	      pops the top of the stack and switches to it via BEGIN.

       int yy_top_state()
	      returns  the  top of the stack without altering the stack's con-
	      tents.

       The start condition stack grows dynamically and so has no built-in size
       limitation.  If memory is exhausted, program execution aborts.

       To  use	start  condition  stacks,  your scanner must include a %option
       stack directive (see Options below).


MULTIPLE INPUT BUFFERS

       Some scanners (such as those which  support  "include"  files)  require
       reading from several input streams.  As flex scanners do a large amount
       is created by using:

	   YY_BUFFER_STATE yy_create_buffer( FILE *file, int size )

       which takes a FILE pointer and a size and creates a  buffer  associated
       with  the  given file and large enough to hold size characters (when in
       doubt, use YY_BUF_SIZE for the size).   It  returns  a  YY_BUFFER_STATE
       handle,	which  may  then be passed to other routines (see below).  The
       YY_BUFFER_STATE type is a pointer to an opaque  struct  yy_buffer_state
       structure,  so  you  may safely initialize YY_BUFFER_STATE variables to
       ((YY_BUFFER_STATE) 0) if you wish, and also refer to the opaque	struc-
       ture  in order to correctly declare input buffers in source files other
       than that of your scanner.  Note that the FILE pointer in the  call  to
       yy_create_buffer is only used as the value of yyin seen by YY_INPUT; if
       you redefine YY_INPUT so it no longer uses yyin, then  you  can	safely
       pass  a	nil FILE pointer to yy_create_buffer.  You select a particular
       buffer to scan from using:

	   void yy_switch_to_buffer( YY_BUFFER_STATE new_buffer )

       switches the scanner's input buffer so subsequent tokens will come from
       new_buffer.  Note that yy_switch_to_buffer() may be used by yywrap() to
       set things up for continued scanning, instead of opening a new file and
       pointing yyin at it.  Note also that switching input sources via either
       yy_switch_to_buffer() or yywrap() does not change the start  condition.

	   void yy_delete_buffer( YY_BUFFER_STATE buffer )

       is  used to reclaim the storage associated with a buffer.  ( buffer can
       be nil, in which case the routine does nothing.)  You  can  also  clear
       the current contents of a buffer using:

	   void yy_flush_buffer( YY_BUFFER_STATE buffer )

       This  function  discards  the  buffer's	contents, so the next time the
       scanner attempts to match a token from the buffer, it will  first  fill
       the buffer anew using YY_INPUT.

       yy_new_buffer()	is  an alias for yy_create_buffer(), provided for com-
       patibility with the C++ use of new and delete for creating and destroy-
       ing dynamic objects.

       Finally,  the  YY_CURRENT_BUFFER macro returns a YY_BUFFER_STATE handle
       to the current buffer.

       Here is an example of using these features for writing a scanner  which
       expands include files (the <<EOF>> feature is discussed below):

	   /* the "incl" state is used for picking up the name
	    * of an include file
	    */
	   %x incl

	   %{
	   #define MAX_INCLUDE_DEPTH 10
	   YY_BUFFER_STATE include_stack[MAX_INCLUDE_DEPTH];
	   int include_stack_ptr = 0;
	   %}
	   <incl>[^ \t\n]+   { /* got the include file name */
		   if ( include_stack_ptr >= MAX_INCLUDE_DEPTH )
		       {
		       fprintf( stderr, "Includes nested too deeply" );
		       exit( 1 );
		       }

		   include_stack[include_stack_ptr++] =
		       YY_CURRENT_BUFFER;

		   yyin = fopen( yytext, "r" );

		   if ( ! yyin )
		       error( ... );

		   yy_switch_to_buffer(
		       yy_create_buffer( yyin, YY_BUF_SIZE ) );

		   BEGIN(INITIAL);
		   }

	   <<EOF>> {
		   if ( --include_stack_ptr < 0 )
		       {
		       yyterminate();
		       }

		   else
		       {
		       yy_delete_buffer( YY_CURRENT_BUFFER );
		       yy_switch_to_buffer(
			    include_stack[include_stack_ptr] );
		       }
		   }

       Three  routines are available for setting up input buffers for scanning
       in-memory strings instead of files.  All of them  create  a  new  input
       buffer	for   scanning	 the   string,	 and  return  a  corresponding
       YY_BUFFER_STATE handle (which you should delete with yy_delete_buffer()
       when  done  with  it).	They  also  switch  to	the  new  buffer using
       yy_switch_to_buffer(), so the next call to yylex() will start  scanning
       the string.

       yy_scan_string(const char *str)
	      scans a NUL-terminated string.

       yy_scan_bytes(const char *bytes, int len)
	      scans  len bytes (including possibly NUL's) starting at location
	      bytes.

       Note that both of these functions create and scan a copy of the	string
       or  bytes.  (This may be desirable, since yylex() modifies the contents
       of the buffer it is scanning.)  You can avoid the copy by using:

       yy_scan_buffer(char *base, yy_size_t size)
	      which scans in place the buffer starting at base, consisting  of
	      size   bytes,   the   last   two	 bytes	 of   which   must  be
	      YY_END_OF_BUFFER_CHAR (ASCII NUL).  These last two bytes are not
	      integer expression reflecting the size of the buffer.


END-OF-FILE RULES

       The special rule "<<EOF>>" indicates actions which are to be taken when
       an end-of-file is encountered  and  yywrap()  returns  non-zero	(i.e.,
       indicates  no  further  files  to  process).  The action must finish by
       doing one of four things:

       -      assigning yyin to a new input  file  (in	previous  versions  of
	      flex,  after  doing  the	assignment you had to call the special
	      action YY_NEW_FILE; this is no longer necessary);

       -      executing a return statement;

       -      executing the special yyterminate() action;

       -      or, switching to a new  buffer  using  yy_switch_to_buffer()  as
	      shown in the example above.

       <<EOF>>	rules  may  not  be used with other patterns; they may only be
       qualified with a list of start conditions.  If an  unqualified  <<EOF>>
       rule  is given, it applies to all start conditions which do not already
       have <<EOF>> actions.  To specify an <<EOF>> rule for only the  initial
       start condition, use

	   <INITIAL><<EOF>>

       These  rules are useful for catching things like unclosed comments.  An
       example:

	   %x quote
	   %%

	   ...other rules for dealing with quotes...

	   <quote><<EOF>>   {
		    error( "unterminated quote" );
		    yyterminate();
		    }
	   <<EOF>>  {
		    if ( *++filelist )
			yyin = fopen( *filelist, "r" );
		    else
		       yyterminate();
		    }


MISCELLANEOUS MACROS

       The macro YY_USER_ACTION can be defined to provide an action  which  is
       always  executed  prior	to the matched rule's action.  For example, it
       could be #define'd to call a routine to convert yytext  to  lower-case.
       When YY_USER_ACTION is invoked, the variable yy_act gives the number of
       the matched rule (rules are numbered starting  with  1).   Suppose  you
       want to profile how often each of your rules is matched.  The following
       would do the trick:

	   #define YY_USER_ACTION ++ctr[yy_act]

       The  macro  YY_USER_INIT  may  be defined to provide an action which is
       always executed before the first scan (and before the scanner's	inter-
       nal initializations are done).  For example, it could be used to call a
       routine to read in a data table or open a logging file.

       The macro yy_set_interactive(is_interactive) can  be  used  to  control
       whether	the  current buffer is considered interactive.	An interactive
       buffer is processed more slowly, but must be used  when	the  scanner's
       input  source is indeed interactive to avoid problems due to waiting to
       fill buffers (see the discussion of the -I  flag  below).   A  non-zero
       value  in  the macro invocation marks the buffer as interactive, a zero
       value as non-interactive.   Note  that  use  of	this  macro  overrides
       %option	interactive  ,	%option  always-interactive  or %option never-
       interactive (see Options below).  yy_set_interactive() must be  invoked
       prior to beginning to scan the buffer that is (or is not) to be consid-
       ered interactive.

       The macro yy_set_bol(at_bol) can be used to control whether the current
       buffer's scanning context for the next token match is done as though at
       the beginning of  a  line.   A  non-zero  macro	argument  makes  rules
       anchored  with  '^' active, while a zero argument makes '^' rules inac-
       tive.

       The macro YY_AT_BOL() returns true if the next token scanned  from  the
       current buffer will have '^' rules active, false otherwise.

       In  the	generated  scanner,  the actions are all gathered in one large
       switch statement and separated using YY_BREAK, which may be  redefined.
       By default, it is simply a "break", to separate each rule's action from
       the following rule's.  Redefining YY_BREAK  allows,  for  example,  C++
       users  to #define YY_BREAK to do nothing (while being very careful that
       every rule ends with a "break" or a "return"!) to avoid suffering  from
       unreachable  statement warnings where because a rule's action ends with
       "return", the YY_BREAK is inaccessible.


VALUES AVAILABLE TO THE USER

       This section summarizes the various values available to the user in the
       rule actions.

       -      char  *yytext  holds  the  text of the current token.  It may be
	      modified but not lengthened (you cannot append characters to the
	      end).

	      If  the special directive %array appears in the first section of
	      the scanner description, then yytext is  instead	declared  char
	      yytext[YYLMAX],  where YYLMAX is a macro definition that you can
	      redefine in the first section if	you  don't  like  the  default
	      value  (generally 8KB).  Using %array results in somewhat slower
	      scanners, but the value of yytext becomes  immune  to  calls  to
	      input()  and  unput(),  which potentially destroy its value when
	      yytext is a  character  pointer.	 The  opposite	of  %array  is
	      %pointer, which is the default.

	      You  cannot  use %array when generating C++ scanner classes (the
	      -+ flag).

       -      int yyleng holds the length of the current token.
	      continue scanning.

       -      void  yyrestart( FILE *new_file ) may be called to point yyin at
	      the new input file.  The switch-over to the new file is  immedi-
	      ate (any previously buffered-up input is lost).  Note that call-
	      ing yyrestart() with yyin as an argument thus  throws  away  the
	      current input buffer and continues scanning the same input file.

       -      FILE *yyout is the file to which ECHO actions are done.  It  can
	      be reassigned by the user.

       -      YY_CURRENT_BUFFER  returns  a YY_BUFFER_STATE handle to the cur-
	      rent buffer.

       -      YY_START returns an integer value corresponding to  the  current
	      start condition.	You can subsequently use this value with BEGIN
	      to return to that start condition.


INTERFACING WITH YACC

       One of the main uses of flex is as a companion to the yacc  parser-gen-
       erator.	 yacc  parsers	expect to call a routine named yylex() to find
       the next input token.  The routine is supposed to return  the  type  of
       the  next  token  as well as putting any associated value in the global
       yylval.	To use flex with yacc, one specifies the -d option to yacc  to
       instruct  it to generate the file y.tab.h containing definitions of all
       the %tokens appearing in the yacc input.  This file is then included in
       the  flex  scanner.  For example, if one of the tokens is "TOK_NUMBER",
       part of the scanner might look like:

	   %{
	   #include "y.tab.h"
	   %}

	   %%

	   [0-9]+	 yylval = atoi( yytext ); return TOK_NUMBER;


OPTIONS

       flex has the following options:

       -b     Generate backing-up information to lex.backup.  This is  a  list
	      of scanner states which require backing up and the input charac-
	      ters on which they do so.  By adding rules one can remove  back-
	      ing-up  states.  If all backing-up states are eliminated and -Cf
	      or -CF is used, the generated scanner will run faster  (see  the
	      -p  flag).   Only users who wish to squeeze every last cycle out
	      of their scanners need worry about this option.  (See  the  sec-
	      tion on Performance Considerations below.)

       -c     is  a  do-nothing,  deprecated option included for POSIX compli-
	      ance.

       -d     makes the generated scanner run in debug mode.  Whenever a  pat-
	      tern  is	recognized  and  the  global yy_flex_debug is non-zero
	      (which is the default), the scanner will write to stderr a  line
	      of the form:


       -f     specifies fast scanner.  No table compression is done and  stdio
	      is  bypassed.   The  result  is  large but fast.	This option is
	      equivalent to -Cfr (see below).

       -h     generates a "help" summary of flex's options to stdout and  then
	      exits.  -?  and --help are synonyms for -h.

       -i     instructs flex to generate a case-insensitive scanner.  The case
	      of letters given in the flex input patterns will be ignored, and
	      tokens  in  the  input  will be matched regardless of case.  The
	      matched text given in yytext will have the preserved case (i.e.,
	      it will not be folded).

       -l     turns on maximum compatibility with the original AT&T lex imple-
	      mentation.  Note that this does  not  mean  full	compatibility.
	      Use  of  this option costs a considerable amount of performance,
	      and it cannot be used with the -+, -f, -F, -Cf, or -CF  options.
	      For  details on the compatibilities it provides, see the section
	      "Incompatibilities With Lex And POSIX" below.  This option  also
	      results  in  the	name YY_FLEX_LEX_COMPAT being #define'd in the
	      generated scanner.

       -n     is another do-nothing, deprecated option included only for POSIX
	      compliance.

       -p     generates  a  performance report to stderr.  The report consists
	      of comments regarding features of the flex input file which will
	      cause  a	serious  loss of performance in the resulting scanner.
	      If you give the flag twice, you will also get comments regarding
	      features that lead to minor performance losses.

	      Note  that  the  use  of	REJECT, %option yylineno, and variable
	      trailing context (see the Deficiencies  /  Bugs  section	below)
	      entails  a substantial performance penalty; use of yymore(), the
	      ^ operator, and the -I flag entail minor performance  penalties.

       -s     causes  the default rule (that unmatched scanner input is echoed
	      to stdout) to be suppressed.  If the  scanner  encounters  input
	      that  does  not match any of its rules, it aborts with an error.
	      This option is useful for finding holes in a scanner's rule set.

       -t     instructs  flex  to  write  the scanner it generates to standard
	      output instead of lex.yy.c.

       -v     specifies that flex should write to stderr a summary of  statis-
	      tics regarding the scanner it generates.	Most of the statistics
	      are meaningless to the casual flex  user,  but  the  first  line
	      identifies the version of flex (same as reported by -V), and the
	      next line the flags used when generating the scanner,  including
	      those that are on by default.

       -w     suppresses warning messages.

       -B     instructs  flex  to  generate  a	batch scanner, the opposite of
	      interactive scanners generated by -I (see below).   In  general,
	      you  use -B when you are certain that your scanner will never be
	      used interactively, and you want to squeeze a little  more  per-
	      terns will be considerably smaller (and for others, larger).  In
	      general, if the pattern  set  contains  both  "keywords"	and  a
	      catch-all, "identifier" rule, such as in the set:

		  "case"    return TOK_CASE;
		  "switch"  return TOK_SWITCH;
		  ...
		  "default" return TOK_DEFAULT;
		  [a-z]+    return TOK_ID;

	      then  you're better off using the full table representation.  If
	      only the "identifier" rule is present and you then  use  a  hash
	      table  or  some  such  to detect the keywords, you're better off
	      using -F.

	      This option is equivalent to -CFr (see  below).	It  cannot  be
	      used with -+.

       -I     instructs  flex to generate an interactive scanner.  An interac-
	      tive scanner is one that only looks ahead to decide  what  token
	      has  been  matched  if  it  absolutely  must.  It turns out that
	      always looking one extra character ahead, even  if  the  scanner
	      has  already seen enough text to disambiguate the current token,
	      is a bit faster than only looking  ahead	when  necessary.   But
	      scanners	that  always look ahead give dreadful interactive per-
	      formance; for example, when a user types a newline,  it  is  not
	      recognized  as  a  newline token until they enter another token,
	      which often means typing in another whole line.

	      Flex scanners default to interactive unless you use the  -Cf  or
	      -CF  table-compression  options  (see below).  That's because if
	      you're looking for high-performance you should be using  one  of
	      these options, so if you didn't, flex assumes you'd rather trade
	      off a bit of  run-time  performance  for	intuitive  interactive
	      behavior.   Note also that you cannot use -I in conjunction with
	      -Cf or -CF.  Thus, this option is not really needed; it is on by
	      default for all those cases in which it is allowed.

	      Note  that if isatty() returns false for the scanner input, flex
	      will revert to batch mode, even if -I was specified.   To  force
	      interactive  mode no matter what, use %option always-interactive
	      (see Options below).

	      You can force a scanner to not be interactive by using  -B  (see
	      above).

       -L     instructs  flex  not to generate #line directives.  Without this
	      option, flex peppers the generated scanner with #line directives
	      so  error messages in the actions will be correctly located with
	      respect to either the original flex input file  (if  the	errors
	      are  due	to code in the input file), or lex.yy.c (if the errors
	      are flex's fault -- you should report these sorts of  errors  to
	      the email address given below).

       -T     makes  flex  run	in trace mode.	It will generate a lot of mes-
	      sages to stderr concerning the form of the input and the	resul-
	      tant  non-deterministic and deterministic finite automata.  This
	      option is mostly for use in maintaining flex.
	      vantage is that such scanners often hang or crash if their input
	      contains an 8-bit character.

	      Note, however, that unless you generate your scanner  using  the
	      -Cf or -CF table compression options, use of -7 will save only a
	      small amount of table space, and make your scanner  considerably
	      less  portable.  Flex's default behavior is to generate an 8-bit
	      scanner unless you use the  -Cf  or  -CF,  in  which  case  flex
	      defaults	to  generating	7-bit  scanners  unless  your site was
	      always configured to generate 8-bit scanners (as will  often  be
	      the  case with non-USA sites).  You can tell whether flex gener-
	      ated a 7-bit or an 8-bit scanner by inspecting the flag  summary
	      in the -v output as described above.

	      Note  that  if  you  use	-Cfe  or -CFe (those table compression
	      options, but also using equivalence  classes  as	discussed  see
	      below),  flex  still  defaults  to  generating an 8-bit scanner,
	      since usually with these compression options full  8-bit	tables
	      are not much more expensive than 7-bit tables.

       -8     instructs flex to generate an 8-bit scanner, i.e., one which can
	      recognize 8-bit characters.  This flag is only needed for  scan-
	      ners  generated  using -Cf or -CF, as otherwise flex defaults to
	      generating an 8-bit scanner anyway.

	      See the discussion of -7 above for flex's default  behavior  and
	      the tradeoffs between 7-bit and 8-bit scanners.

       -+     specifies  that  you  want flex to generate a C++ scanner class.
	      See the section on Generating C++ Scanners below for details.

       -C[aefFmr]
	      controls the degree of table compression	and,  more  generally,
	      trade-offs between small scanners and fast scanners.

	      -Ca  ("align")  instructs flex to trade off larger tables in the
	      generated scanner for faster performance because the elements of
	      the tables are better aligned for memory access and computation.
	      On some RISC architectures, fetching and manipulating  longwords
	      is  more	efficient than with smaller-sized units such as short-
	      words.  This option can double the size of the  tables  used  by
	      your scanner.

	      -Ce directs flex to construct equivalence classes, i.e., sets of
	      characters which have identical lexical properties (for example,
	      if  the  only  appearance  of digits in the flex input is in the
	      character class "[0-9]" then the digits '0', '1', ..., '9'  will
	      all  be put in the same equivalence class).  Equivalence classes
	      usually give dramatic reductions in the final table/object  file
	      sizes  (typically  a factor of 2-5) and are pretty cheap perfor-
	      mance-wise (one array look-up per character scanned).

	      -Cf specifies that the full scanner tables should be generated -
	      flex should not compress the tables by taking advantages of sim-
	      ilar transition functions for different states.

	      -CF specifies that the  alternate  fast  scanner	representation
	      (described above under the -F flag) should be used.  This option

	      -Cr causes the generated scanner to bypass use of  the  standard
	      I/O  library  (stdio)  for input.  Instead of calling fread() or
	      getc(), the scanner will use the read() system  call,  resulting
	      in a performance gain which varies from system to system, but in
	      general is probably negligible unless you are also using -Cf  or
	      -CF.   Using -Cr can cause strange behavior if, for example, you
	      read from yyin using stdio prior to calling the scanner (because
	      the  scanner will miss whatever text your previous reads left in
	      the stdio input buffer).

	      -Cr has no effect if you	define	YY_INPUT  (see	The  Generated
	      Scanner above).

	      A lone -C specifies that the scanner tables should be compressed
	      but neither equivalence  classes	nor  meta-equivalence  classes
	      should be used.

	      The  options  -Cf  or  -CF  and -Cm do not make sense together -
	      there is no opportunity for meta-equivalence classes if the  ta-
	      ble  is  not  being  compressed.	 Otherwise  the options may be
	      freely mixed, and are cumulative.

	      The default setting is -Cem, which specifies  that  flex	should
	      generate equivalence classes and meta-equivalence classes.  This
	      setting provides the highest degree of table  compression.   You
	      can  trade  off  faster-executing scanners at the cost of larger
	      tables with the following generally being true:

		  slowest & smallest
			-Cem
			-Cm
			-Ce
			-C
			-C{f,F}e
			-C{f,F}
			-C{f,F}a
		  fastest & largest

	      Note that scanners with the smallest tables are  usually	gener-
	      ated  and  compiled the quickest, so during development you will
	      usually want to use the default, maximal compression.

	      -Cfe is often a good compromise between speed and size for  pro-
	      duction scanners.

       -ooutput
	      directs  flex to write the scanner to the file output instead of
	      lex.yy.c.  If you combine -o with the -t option, then the  scan-
	      ner  is  written	to stdout but its #line directives (see the -L
	      option above) refer to the file output.

       -Pprefix
	      changes the default yy prefix used by flex for all globally-vis-
	      ible  variable  and  function  names  to instead be prefix.  For
	      example, -Pfoo changes the name of yytext to footext.   It  also
	      changes  the  name  of  the default output file from lex.yy.c to
	      lex.foo.c.  Here are all of the names affected:

		  yyin
		  yyleng
		  yylex
		  yylineno
		  yyout
		  yyrestart
		  yytext
		  yywrap

	      (If  you	are  using  a  C++  scanner,  then  only  yywrap   and
	      yyFlexLexer  are affected.)  Within your scanner itself, you can
	      still refer to the global variables and functions  using	either
	      version  of  their  name; but externally, they have the modified
	      name.

	      This option lets you easily link together multiple flex programs
	      into  the same executable.  Note, though, that using this option
	      also renames yywrap(), so you now must either provide  your  own
	      (appropriately-named)  version  of the routine for your scanner,
	      or use %option noyywrap, as linking with -ll no longer  provides
	      one for you by default.

       -Sskeleton_file
	      overrides  the  default skeleton file from which flex constructs
	      its scanners.  You'll never need	this  option  unless  you  are
	      doing flex maintenance or development.

       flex also provides a mechanism for controlling options within the scan-
       ner specification itself, rather than from the flex command-line.  This
       is  done  by  including	%option directives in the first section of the
       scanner specification.  You can specify multiple options with a	single
       %option directive, and multiple directives in the first section of your
       flex input file.

       Most options are given simply as names, optionally preceded by the word
       "no"  (with no intervening whitespace) to negate their meaning.	A num-
       ber are equivalent to flex flags or their negation:

	   7bit 	   -7 option
	   8bit 	   -8 option
	   align	   -Ca option
	   backup	   -b option
	   batch	   -B option
	   c++		   -+ option

	   caseful or
	   case-sensitive  opposite of -i (default)

	   case-insensitive or
	   caseless	   -i option

	   debug	   -d option
	   default	   opposite of -s option
	   ecs		   -Ce option
	   fast 	   -F option
	   full 	   -f option
	   interactive	   -I option
	   lex-compat	   -l option
	   array	   equivalent to "%array"
	   pointer	   equivalent to "%pointer" (default)

       Some %option's provide features otherwise not available:

       always-interactive
	      instructs flex to generate a scanner which always considers  its
	      input "interactive".  Normally, on each new input file the scan-
	      ner calls isatty() in an attempt to determine whether the  scan-
	      ner's  input  source  is	interactive  and thus should be read a
	      character at a time.  When this option is used, however, then no
	      such call is made.

       main   directs  flex  to provide a default main() program for the scan-
	      ner, which simply calls yylex().	This option  implies  noyywrap
	      (see below).

       never-interactive
	      instructs  flex  to generate a scanner which never considers its
	      input "interactive" (again, no call made to isatty()).  This  is
	      the opposite of always-interactive.

       stack  enables  the use of start condition stacks (see Start Conditions
	      above).

       stdinit
	      if set (i.e., %option stdinit) initializes  yyin	and  yyout  to
	      stdin  and stdout, instead of the default of nil.  Some existing
	      lex programs depend on this behavior, even though it is not com-
	      pliant  with  ANSI C, which does not require stdin and stdout to
	      be compile-time constant.

       yylineno
	      directs flex to generate a scanner that maintains the number  of
	      the  current  line  read	from  its input in the global variable
	      yylineno.  This option is implied by %option lex-compat.

       yywrap if unset (i.e., %option noyywrap), makes the  scanner  not  call
	      yywrap()	upon  an end-of-file, but simply assume that there are
	      no more files to scan (until the user points yyin at a new  file
	      and calls yylex() again).

       flex scans your rule actions to determine whether you use the REJECT or
       yymore() features.  The reject and  yymore  options  are  available  to
       override its decision as to whether you use the options, either by set-
       ting them (e.g., %option reject) to  indicate  the  feature  is	indeed
       used,  or  unsetting  them  to  indicate it actually is not used (e.g.,
       %option noyymore).

       Three options take string-delimited values, offset with '=':

	   %option outfile="ABC"

       is equivalent to -oABC, and

	   %option prefix="XYZ"

       is equivalent to -PXYZ.	Finally,
       Error()) if called.  See Generating C++ Scanners, below, for additional
       information.

       A number of options are available for lint purists who want to suppress
       the  appearance of unneeded routines in the generated scanner.  Each of
       the following, if unset (e.g., %option nounput ), results in the corre-
       sponding routine not appearing in the generated scanner:

	   input, unput
	   yy_push_state, yy_pop_state, yy_top_state
	   yy_scan_buffer, yy_scan_bytes, yy_scan_string

       (though	yy_push_state() and friends won't appear anyway unless you use
       %option stack).


PERFORMANCE CONSIDERATIONS

       The main design goal of flex is that it generate high-performance scan-
       ners.  It has been optimized for dealing well with large sets of rules.
       Aside from the effects on scanner speed of  the	table  compression  -C
       options	outlined  above,  there  are a number of options/actions which
       degrade performance.  These are, from most expensive to least:

	   REJECT
	   %option yylineno
	   arbitrary trailing context

	   pattern sets that require backing up
	   %array
	   %option interactive
	   %option always-interactive

	   '^' beginning-of-line operator
	   yymore()

       with the first three all being quite expensive and the last  two  being
       quite  cheap.   Note also that unput() is implemented as a routine call
       that potentially does quite a bit of work, while yyless() is  a	quite-
       cheap  macro; so if just putting back some excess text you scanned, use
       yyless().

       REJECT should be avoided at all costs when  performance	is  important.
       It is a particularly expensive option.

       Getting	rid of backing up is messy and often may be an enormous amount
       of work for a complicated scanner.  In principal, one begins  by  using
       the -b flag to generate a lex.backup file.  For example, on the input

	   %%
	   foo	      return TOK_KEYWORD;
	   foobar     return TOK_KEYWORD;

       the file looks like:

	   State #6 is non-accepting -
	    associated rule line numbers:
		  2	  3
	    out-transitions: [ o ]
	    jam-transitions: EOF [ \001-n  p-\177 ]
	    associated rule line numbers:
		  3
	    out-transitions: [ r ]
	    jam-transitions: EOF [ \001-q  s-\177 ]

	   Compressed tables always back up.

       The  first  few	lines tell us that there's a scanner state in which it
       can make a transition on an 'o' but not on  any	other  character,  and
       that  in that state the currently scanned text does not match any rule.
       The state occurs when trying to match the rules found at lines 2 and  3
       in  the	input  file.   If  the scanner is in that state and then reads
       something other than an 'o', it will have to back up  to  find  a  rule
       which  is  matched.  With a bit of headscratching one can see that this
       must be the state it's in when it has seen "fo".  When  this  has  hap-
       pened,  if  anything  other  than another 'o' is seen, the scanner will
       have to back up to simply match the 'f' (by the default rule).

       The comment regarding State #8 indicates there's a problem when	"foob"
       has  been  scanned.   Indeed,  on  any character other than an 'a', the
       scanner will have to back up to accept "foo".  Similarly,  the  comment
       for State #9 concerns when "fooba" has been scanned and an 'r' does not
       follow.

       The final comment reminds us that there's no point  going  to  all  the
       trouble of removing backing up from the rules unless we're using -Cf or
       -CF, since there's no performance gain doing so with  compressed  scan-
       ners.

       The way to remove the backing up is to add "error" rules:

	   %%
	   foo	       return TOK_KEYWORD;
	   foobar      return TOK_KEYWORD;

	   fooba       |
	   foob        |
	   fo	       {
		       /* false alarm, not really a keyword */
		       return TOK_ID;
		       }

       Eliminating  backing up among a list of keywords can also be done using
       a "catch-all" rule:

	   %%
	   foo	       return TOK_KEYWORD;
	   foobar      return TOK_KEYWORD;

	   [a-z]+      return TOK_ID;

       This is usually the best solution when appropriate.

       Backing up messages tend to cascade.  With a complicated set  of  rules
       it's  not  uncommon  to	get hundreds of messages.  If one can decipher
       them, though, it often only takes a dozen or so rules to eliminate  the
       backing	up  (though it's easy to make a mistake and have an error rule
       not have a fixed length) entails almost the same  performance  loss  as
       REJECT (i.e., substantial).  So when possible a rule like:

	   %%
	   mouse|rat/(cat|dog)	 run();

       is better written:

	   %%
	   mouse/cat|dog	 run();
	   rat/cat|dog		 run();

       or as

	   %%
	   mouse|rat/cat	 run();
	   mouse|rat/dog	 run();

       Note that here the special '|' action does not provide any savings, and
       can even make things worse (see Deficiencies / Bugs below).

       Another area where the user can increase a scanner's  performance  (and
       one  that's  easier  to implement) arises from the fact that the longer
       the tokens matched, the faster the scanner will run.  This  is  because
       with long tokens the processing of most input characters takes place in
       the (short) inner scanning loop, and does not often have to go  through
       the  additional	work  of  setting  up  the scanning environment (e.g.,
       yytext) for the action.	Recall the scanner for C comments:

	   %x comment
	   %%
		   int line_num = 1;

	   "/*" 	BEGIN(comment);

	   <comment>[^*\n]*
	   <comment>"*"+[^*/\n]*
	   <comment>\n		   ++line_num;
	   <comment>"*"+"/"	   BEGIN(INITIAL);

       This could be sped up by writing it as:

	   %x comment
	   %%
		   int line_num = 1;

	   "/*" 	BEGIN(comment);

	   <comment>[^*\n]*
	   <comment>[^*\n]*\n	   ++line_num;
	   <comment>"*"+[^*/\n]*
	   <comment>"*"+[^*/\n]*\n ++line_num;
	   <comment>"*"+"/"	   BEGIN(INITIAL);

       Now instead of each newline requiring the processing of another action,
       recognizing  the newlines is "distributed" over the other rules to keep
       the matched text as long as possible.  Note that adding rules does  not
       slow  down the scanner!	The speed of the scanner is independent of the

	   %%
	   asm	    |
	   auto     |
	   break    |
	   ... etc ...
	   volatile |
	   while    /* it's a keyword */

	   .|\n     /* it's not a keyword */

       To eliminate the back-tracking, introduce a catch-all rule:

	   %%
	   asm	    |
	   auto     |
	   break    |
	   ... etc ...
	   volatile |
	   while    /* it's a keyword */

	   [a-z]+   |
	   .|\n     /* it's not a keyword */

       Now, if it's guaranteed that there's exactly one word per line, then we
       can reduce the total number of matches by a  half  by  merging  in  the
       recognition of newlines with that of the other tokens:

	   %%
	   asm\n    |
	   auto\n   |
	   break\n  |
	   ... etc ...
	   volatile\n |
	   while\n  /* it's a keyword */

	   [a-z]+\n |
	   .|\n     /* it's not a keyword */

       One has to be careful here, as we have now reintroduced backing up into
       the scanner.  In particular, while we know that there will never be any
       characters  in  the  input  stream other than letters or newlines, flex
       can't figure this out, and it will plan for possibly needing to back up
       when  it has scanned a token like "auto" and then the next character is
       something other than a newline or a letter.  Previously it  would  then
       just  match the "auto" rule and be done, but now it has no "auto" rule,
       only an "auto\n" rule.  To eliminate the possibility of backing up,  we
       could  either duplicate all rules but without final newlines, or, since
       we never expect to encounter such an input and therefore don't how it's
       classified,  we	can  introduce one more catch-all rule, this one which
       doesn't include a newline:

	   %%
	   asm\n    |
	   auto\n   |
	   break\n  |
	   ... etc ...
	   volatile\n |

       A  final  note:	flex  is slow when matching NUL's, particularly when a
       token contains multiple NUL's.  It's best to write  rules  which  match
       short  amounts  of  text  if  it's anticipated that the text will often
       include NUL's.

       Another final note regarding performance: as  mentioned	above  in  the
       section How the Input is Matched, dynamically resizing yytext to accom-
       modate huge tokens is a slow process because it presently requires that
       the  (huge) token be rescanned from the beginning.  Thus if performance
       is vital, you should attempt to match "large" quantities  of  text  but
       not  "huge" quantities, where the cutoff between the two is at about 8K
       characters/token.


GENERATING C++ SCANNERS

       flex provides two different ways to generate scanners for use with C++.
       The  first way is to simply compile a scanner generated by flex using a
       C++ compiler instead of a C compiler.  You  should  not	encounter  any
       compilations  errors  (please  report any you find to the email address
       given in the Author section below).  You can then use C++ code in  your
       rule actions instead of C code.	Note that the default input source for
       your scanner remains yyin, and default echoing is still done to	yyout.
       Both of these remain FILE * variables and not C++ streams.

       You  can  also  use  flex to generate a C++ scanner class, using the -+
       option (or, equivalently, %option c++), which is  automatically	speci-
       fied  if the name of the flex executable ends in a '+', such as flex++.
       When using this option, flex defaults to generating the scanner to  the
       file lex.yy.cc instead of lex.yy.c.  The generated scanner includes the
       header file  FlexLexer.h,  which  defines  the  interface  to  two  C++
       classes.

       The  first  class,  FlexLexer, provides an abstract base class defining
       the general scanner class interface.  It provides the following	member
       functions:

       const char* YYText()
	      returns the text of the most recently matched token, the equiva-
	      lent of yytext.

       int YYLeng()
	      returns the length of  the  most	recently  matched  token,  the
	      equivalent of yyleng.

       int lineno() const
	      returns the current input line number (see %option yylineno), or
	      1 if %option yylineno was not used.

       void set_debug( int flag )
	      sets the debugging flag for the scanner, equivalent to assigning
	      to yy_flex_debug (see the Options section above).  Note that you
	      must build the scanner using %option debug to include  debugging
	      information in it.

       int debug() const
	      returns the current setting of the debugging flag.

       Also provided are member functions equivalent to yy_switch_to_buffer(),
       yyFlexLexer( istream* arg_yyin = 0, ostream* arg_yyout = 0 )
	      constructs a yyFlexLexer object  using  the  given  streams  for
	      input  and output.  If not specified, the streams default to cin
	      and cout, respectively.

       virtual int yylex()
	      performs the same role is yylex() does for ordinary  flex  scan-
	      ners:  it  scans	the  input  stream,  consuming tokens, until a
	      rule's action returns a value.  If you derive a subclass S  from
	      yyFlexLexer  and	want  to access the member functions and vari-
	      ables of	S  inside  yylex(),  then  you	need  to  use  %option
	      yyclass="S"  to inform flex that you will be using that subclass
	      instead of yyFlexLexer.  In this case,  rather  than  generating
	      yyFlexLexer::yylex(), flex generates S::yylex() (and also gener-
	      ates a dummy yyFlexLexer::yylex() that calls yyFlexLexer::Lexer-
	      Error() if called).

       virtual void switch_streams(istream* new_in = 0,
	      ostream*	new_out = 0) reassigns yyin to new_in (if non-nil) and
	      yyout to new_out (ditto), deleting the previous input buffer  if
	      yyin is reassigned.

       int yylex( istream* new_in, ostream* new_out = 0 )
	      first  switches  the  input  streams via switch_streams( new_in,
	      new_out ) and then returns the value of yylex().

       In addition, yyFlexLexer defines the following protected virtual  func-
       tions which you can redefine in derived classes to tailor the scanner:

       virtual int LexerInput( char* buf, int max_size )
	      reads  up to max_size characters into buf and returns the number
	      of characters read.  To indicate end-of-input, return 0  charac-
	      ters.   Note  that  "interactive"  scanners  (see  the -B and -I
	      flags) define the macro YY_INTERACTIVE.  If  you	redefine  Lex-
	      erInput()  and  need  to	take  different  actions  depending on
	      whether or not the scanner  might  be  scanning  an  interactive
	      input  source,  you  can	test for the presence of this name via
	      #ifdef.

       virtual void LexerOutput( const char* buf, int size )
	      writes out size characters from the  buffer  buf,  which,  while
	      NUL-terminated,  may  also contain "internal" NUL's if the scan-
	      ner's rules can match text with NUL's in them.

       virtual void LexerError( const char* msg )
	      reports a fatal error message.   The  default  version  of  this
	      function writes the message to the stream cerr and exits.

       Note  that  a  yyFlexLexer  object  contains its entire scanning state.
       Thus you can use such objects to create reentrant  scanners.   You  can
       instantiate  multiple  instances of the same yyFlexLexer class, and you
       can also combine multiple C++ scanner classes together in the same pro-
       gram using the -P option discussed above.

       Finally,  note  that the %array feature is not available to C++ scanner
       classes; you must use %pointer (the default).

       Here is an example of a simple C++ scanner:

	   ws	   [ \t]+

	   alpha   [A-Za-z]
	   dig	   [0-9]
	   name    ({alpha}|{dig}|\$)({alpha}|{dig}|[_.\-/$])*
	   num1    [-+]?{dig}+\.?([eE][-+]?{dig}+)?
	   num2    [-+]?{dig}*\.{dig}+([eE][-+]?{dig}+)?
	   number  {num1}|{num2}

	   %%

	   {ws}    /* skip blanks and tabs */

	   "/*"    {
		   int c;

		   while((c = yyinput()) != 0)
		       {
		       if(c == '\n')
			   ++mylineno;

		       else if(c == '*')
			   {
			   if((c = yyinput()) == '/')
			       break;
			   else
			       unput(c);
			   }
		       }
		   }

	   {number}  cout << "number " << YYText() << '\n';

	   \n	     mylineno++;

	   {name}    cout << "name " << YYText() << '\n';

	   {string}  cout << "string " << YYText() << '\n';

	   %%

	   int main( int /* argc */, char** /* argv */ )
	       {
	       FlexLexer* lexer = new yyFlexLexer;
	       while(lexer->yylex() != 0)
		   ;
	       return 0;
	       }
       If you want to create multiple (different) lexer classes, you  use  the
       -P  flag  (or  the  prefix=  option) to rename each yyFlexLexer to some
       other xxFlexLexer.  You then can include <FlexLexer.h>  in  your  other
       sources once per lexer class, first renaming yyFlexLexer as follows:

	   #undef yyFlexLexer
	   #define yyFlexLexer xxFlexLexer
	   #include <FlexLexer.h>

       may change considerably between major releases.


INCOMPATIBILITIES WITH LEX AND POSIX

       flex is a rewrite of the AT&T Unix lex tool (the two implementations do
       not share any code, though), with some  extensions  and	incompatibili-
       ties,  both of which are of concern to those who wish to write scanners
       acceptable to either implementation.  Flex is fully compliant with  the
       POSIX lex specification, except that when using %pointer (the default),
       a call to unput() destroys the contents of yytext, which is counter  to
       the POSIX specification.

       In  this  section  we discuss all of the known areas of incompatibility
       between flex, AT&T lex, and the POSIX specification.

       flex's -l option turns on maximum compatibility with the original  AT&T
       lex  implementation, at the cost of a major loss in the generated scan-
       ner's performance.  We note below which incompatibilities can be  over-
       come using the -l option.

       flex is fully compatible with lex with the following exceptions:

       -      The  undocumented  lex scanner internal variable yylineno is not
	      supported unless -l or %option yylineno is used.

	      yylineno should be maintained on a per-buffer basis, rather than
	      a per-scanner (single global variable) basis.

	      yylineno is not part of the POSIX specification.

       -      The  input() routine is not redefinable, though it may be called
	      to read characters following whatever  has  been	matched  by  a
	      rule.   If input() encounters an end-of-file the normal yywrap()
	      processing is done.   A  ``real''  end-of-file  is  returned  by
	      input() as EOF.

	      Input is instead controlled by defining the YY_INPUT macro.

	      The  flex  restriction  that  input()  cannot be redefined is in
	      accordance with the POSIX specification, which simply  does  not
	      specify any way of controlling the scanner's input other than by
	      making an initial assignment to yyin.

       -      The unput() routine is not redefinable.  This restriction is  in
	      accordance with POSIX.

       -      flex scanners are not as reentrant as lex scanners.  In particu-
	      lar, if you have an interactive scanner and an interrupt handler
	      which  long-jumps  out of the scanner, and the scanner is subse-
	      quently called again, you may get the following message:

		  fatal flex scanner internal error--end of buffer missed

	      To reenter the scanner, first use

		  yyrestart( yyin );

	      Note that this call will throw away any buffered input;  usually
	      this isn't a problem with an interactive scanner.
	      output() is not part of the POSIX specification.

       -      lex does not support exclusive  start  conditions  (%x),	though
	      they are in the POSIX specification.

       -      When  definitions  are expanded, flex encloses them in parenthe-
	      ses.  With lex, the following:

		  NAME	  [A-Z][A-Z0-9]*
		  %%
		  foo{NAME}?	  printf( "Found it\n" );
		  %%

	      will not match the  string  "foo"  because  when	the  macro  is
	      expanded the rule is equivalent to "foo[A-Z][A-Z0-9]*?"  and the
	      precedence is such that the '?' is associated with  "[A-Z0-9]*".
	      With  flex,  the rule will be expanded to "foo([A-Z][A-Z0-9]*)?"
	      and so the string "foo" will match.

	      Note that if the definition begins with ^ or ends with $ then it
	      is  not  expanded  with parentheses, to allow these operators to
	      appear in definitions without  losing  their  special  meanings.
	      But  the	<s>, /, and <<EOF>> operators cannot be used in a flex
	      definition.

	      Using -l results in the lex behavior of  no  parentheses	around
	      the definition.

	      The  POSIX  specification  is that the definition be enclosed in
	      parentheses.

       -      Some implementations of lex allow a rule's action to begin on  a
	      separate line, if the rule's pattern has trailing whitespace:

		  %%
		  foo|bar<space here>
		    { foobar_action(); }

	      flex does not support this feature.

       -      The  lex %r (generate a Ratfor scanner) option is not supported.
	      It is not part of the POSIX specification.

       -      After a call to unput(), yytext  is  undefined  until  the  next
	      token  is  matched,  unless  the scanner was built using %array.
	      This is not the case with lex or the POSIX  specification.   The
	      -l option does away with this incompatibility.

       -      The  precedence of the {} (numeric range) operator is different.
	      lex interprets "abc{1,3}" as "match one, two,  or  three	occur-
	      rences of 'abc'", whereas flex interprets it as "match 'ab' fol-
	      lowed by one, two, or three occurrences of 'c'".	The latter  is
	      in agreement with the POSIX specification.

       -      The  precedence  of the ^ operator is different.	lex interprets
	      "^foo|bar" as "match either 'foo' at the beginning of a line, or
	      'bar'  anywhere",  whereas  flex	interprets it as "match either
	      'foo' or 'bar' if they come at the beginning of  a  line".   The
	      version  of flex generated the scanner (for example, for the 2.5
	      release, these defines would be 2 and 5 respectively).

       The following flex features are not included in lex or the POSIX speci-
       fication:

	   C++ scanners
	   %option
	   start condition scopes
	   start condition stacks
	   interactive/non-interactive scanners
	   yy_scan_string() and friends
	   yyterminate()
	   yy_set_interactive()
	   yy_set_bol()
	   YY_AT_BOL()
	   <<EOF>>
	   <*>
	   YY_DECL
	   YY_START
	   YY_USER_ACTION
	   YY_USER_INIT
	   #line directives
	   %{}'s around actions
	   multiple actions on a line

       plus almost all of the flex flags.  The last feature in the list refers
       to the fact that with flex you can put multiple	actions  on  the  same
       line, separated with semi-colons, while with lex, the following

	   foo	  handle_foo(); ++num_foos_seen;

       is (rather surprisingly) truncated to

	   foo	  handle_foo();

       flex  does  not	truncate the action.  Actions that are not enclosed in
       braces are simply terminated at the end of the line.


DIAGNOSTICS

       warning, rule cannot be matched indicates that the given rule cannot be
       matched	because it follows other rules that will always match the same
       text as it.  For example, in the  following  "foo"  cannot  be  matched
       because it comes after an identifier "catch-all" rule:

	   [a-z]+    got_identifier();
	   foo	     got_foo();

       Using REJECT in a scanner suppresses this warning.

       warning,  -s option given but default rule can be matched means that it
       is possible (perhaps only in a particular  start  condition)  that  the
       default	rule  (match  any  single character) is the only one that will
       match a particular input.  Since -s was given, presumably this  is  not
       intended.

       reject_used_but_not_detected  undefined or yymore_used_but_not_detected
       undefined - These errors can occur at compile time.  They indicate that
       input  string which wasn't matched by any of its rules.	This error can
       also occur due to internal problems.

       token too large, exceeds YYLMAX - your scanner uses %array and  one  of
       its rules matched a string longer than the YYLMAX constant (8K bytes by
       default).  You can increase the value by #define'ing YYLMAX in the def-
       initions section of your flex input.

       scanner requires -8 flag to use the character 'x' - Your scanner speci-
       fication includes recognizing the 8-bit character 'x' and you  did  not
       specify	the  -8  flag, and your scanner defaulted to 7-bit because you
       used the -Cf or -CF table compression options.  See the	discussion  of
       the -7 flag for details.

       flex scanner push-back overflow - you used unput() to push back so much
       text that the scanner's buffer could not hold both the pushed-back text
       and  the  current  token in yytext.  Ideally the scanner should dynami-
       cally resize the buffer in this case, but at present it does not.

       input buffer overflow, can't enlarge buffer because scanner uses REJECT
       -  the  scanner	was  working  on matching an extremely large token and
       needed to expand the input buffer.  This  doesn't  work	with  scanners
       that use REJECT.

       fatal  flex  scanner  internal  error--end  of buffer missed - This can
       occur in a scanner which is reentered after a long-jump has jumped  out
       (or  over) the scanner's activation frame.  Before reentering the scan-
       ner, use:

	   yyrestart( yyin );

       or, as noted above, switch to using the C++ scanner class.

       too many start conditions in <> construct! - you listed more start con-
       ditions	in a <> construct than exist (so you must have listed at least
       one of them twice).


FILES

       -ll    library with which scanners must be linked.

       lex.yy.c
	      generated scanner (called lexyy.c on some systems).

       lex.yy.cc
	      generated C++ scanner class, when using -+.

       <FlexLexer.h>
	      header file defining the C++ scanner base class, FlexLexer,  and
	      its derived class, yyFlexLexer.

       flex.skl
	      skeleton	scanner.   This  file is only used when building flex,
	      not when flex executes.

       lex.backup
	      backing-up information for -b flag (called lex.bck on some  sys-
	      tems).

       For some trailing context rules, parts which are actually  fixed-length
       are  not recognized as such, leading to the above mentioned performance
       loss.  In particular, parts using '|' or {n}  (such  as	"foo{3}")  are
       always considered variable-length.

       Combining  trailing  context  with the special '|' action can result in
       fixed trailing context being turned into the  more  expensive  variable
       trailing context.  For example, in the following:

	   %%
	   abc	    |
	   xyz/def

       Use  of unput() invalidates yytext and yyleng, unless the %array direc-
       tive or the -l option has been used.

       Pattern-matching of NUL's is substantially slower than  matching  other
       characters.

       Dynamic	resizing of the input buffer is slow, as it entails rescanning
       all the text matched so far by the current (generally huge) token.

       Due to both buffering of input  and  read-ahead,  you  cannot  intermix
       calls to <stdio.h> routines, such as, for example, getchar(), with flex
       rules and expect it to work.  Call input() instead.

       The total table entries listed by the -v flag excludes  the  number  of
       table entries needed to determine what rule has been matched.  The num-
       ber of entries is equal to the number of DFA states if the scanner does
       not  use  REJECT,  and somewhat greater than the number of states if it
       does.

       REJECT cannot be used with the -f or -F options.

       The flex internal algorithms need documentation.


SEE ALSO

       lex(1), yacc(1), sed(1), awk(1).

       John Levine, Tony Mason, and Doug Brown, Lex & Yacc, O'Reilly and Asso-
       ciates.	Be sure to get the 2nd edition.

       M. E. Lesk and E. Schmidt, LEX - Lexical Analyzer Generator

       Alfred Aho, Ravi Sethi and Jeffrey Ullman, Compilers: Principles, Tech-
       niques and Tools, Addison-Wesley (1986).  Describes the	pattern-match-
       ing techniques used by flex (deterministic finite automata).


AUTHOR

       Vern  Paxson, with the help of many ideas and much inspiration from Van
       Jacobson.  Original version by Jef Poskanzer.  The fast table represen-
       tation  is  a  partial implementation of a design done by Van Jacobson.
       The implementation was done by Kevin Gong and Vern Paxson.

       Thanks to the many flex beta-testers,  feedbackers,  and  contributors,
       especially Francois Pinard, Casey Leedom, Robert Abramovitz, Stan Ader-
       mann, Terry Allen, David Barker-Plummer, John Basrai, Neal Becker, Nel-
       Jarkko  Hietaniemi, Scott Hofmann, Jeff Honig, Dana Hudes, Eric Hughes,
       John Interrante, Ceriel Jacobs, Michal  Jaegermann,  Sakari  Jalovaara,
       Jeffrey R. Jones, Henry Juengst, Klaus Kaempf, Jonathan I. Kamens, Ter-
       rence O Kane, Amir  Katz,  ken@ken.hilco.com,  Kevin  B.  Kenny,  Steve
       Kirsch,	Winfried  Koenig, Marq Kole, Ronald Lamprecht, Greg Lee, Rohan
       Lenard, Craig Leres, John Levine, Steve Liddle,	David  Loffredo,  Mike
       Long,  Mohamed  el  Lozy,  Brian  Madsen,  Malte,  Joe  Marshall, Bengt
       Martensson, Chris Metcalf, Luke Mewburn,  Jim  Meyering,  R.  Alexander
       Milowski,  Erik	Naggum,  G.T.  Nicol,  Landon Noll, James Nordby, Marc
       Nozell, Richard Ohnemus, Karsten Pahnke, Sven Panne, Roland Pesch, Wal-
       ter  Pelissero, Gaumond Pierre, Esmond Pitt, Jef Poskanzer, Joe Rahmeh,
       Jarmo Raiha, Frederic Raimbault, Pat  Rankin,  Rick  Richardson,  Kevin
       Rodgers, Kai Uwe Rommel, Jim Roskind, Alberto Santini, Andreas Scherer,
       Darrell Schiebel, Raf Schietekat, Doug Schmidt,	Philippe  Schnoebelen,
       Andreas	Schwab, Larry Schwimmer, Alex Siegel, Eckehard Stolz, Jan-Erik
       Strvmquist, Mike Stump, Paul Stuart, Dave Tallman,  Ian	Lance  Taylor,
       Chris Thewalt, Richard M. Timoney, Jodi Tsai, Paul Tuinenga, Gary Weik,
       Frank Whaley, Gerhard Wilhelms, Kent Williams,  Ken  Yap,  Ron  Zellar,
       Nathan Zelle, David Zuhn, and those whose names have slipped my margin-
       al mail-archiving skills but whose contributions  are  appreciated  all
       the same.

       Thanks to Keith Bostic, Jon Forrest, Noah Friedman, John Gilmore, Craig
       Leres, John Levine, Bob Mulcahy, G.T.   Nicol,  Francois  Pinard,  Rich
       Salz,   and   Richard  Stallman	for  help  with  various  distribution
       headaches.

       Thanks to Esmond Pitt and Earle Horton for 8-bit character support;  to
       Benson  Margulies  and Fred Burke for C++ support; to Kent Williams and
       Tom Epperly for C++ class support; to Ove Ewerlid for support of NUL's;
       and to Eric Hughes for support of multiple buffers.

       This  work  was	primarily  done  when I was with the Real Time Systems
       Group at the Lawrence Berkeley Laboratory in Berkeley, CA.  Many thanks
       to all there for the support I received.

       Send comments to vern@ee.lbl.gov.

Version 2.5			  April 1995			       FLEX(1)

SPONSORED LINKS




Man(1) output converted with man2html , sed , awk