forked from jasder/antlr
decided to clean up a bit, backtracking from updating current to have inner/outer context. will try with full ctx again.
[git-p4: depot-paths = "//depot/code/antlr4/main/": change = 8908]
This commit is contained in:
parent
76ca9ba0d5
commit
c6fa2fce6e
|
@ -211,10 +211,14 @@ public class RuleContext {
|
|||
}
|
||||
|
||||
public String toString(BaseRecognizer recog) {
|
||||
return toString(recog, null);
|
||||
}
|
||||
|
||||
public String toString(BaseRecognizer recog, RuleContext stop) {
|
||||
StringBuffer buf = new StringBuffer();
|
||||
RuleContext p = this;
|
||||
buf.append("[");
|
||||
while ( p != null ) {
|
||||
while ( p != null && p != stop ) {
|
||||
if ( recog!=null ) {
|
||||
ATN atn = recog.getATN();
|
||||
ATNState s = atn.states.get(p.s);
|
||||
|
|
|
@ -45,10 +45,29 @@ public class ATNConfig {
|
|||
public int alt;
|
||||
|
||||
/** The stack of invoking states leading to the rule/states associated
|
||||
* wit this config.
|
||||
* with this config. We track only those contexts pushed during
|
||||
* execution of the ATN simulator.
|
||||
*/
|
||||
public RuleContext context;
|
||||
|
||||
/** The stack acquired from parser invoking the ATN interpreter.
|
||||
* Initially it's the rule stack upon entry into interp.adaptivePredict().
|
||||
*
|
||||
* We only pop from outerContext if we hit rule stop state
|
||||
* of rule that initiates the adaptivePredict(). outerContext
|
||||
* tracks as we go back up the entry context. If we do a call,
|
||||
* we again push onto the regular context, rather than pushing onto
|
||||
* outerContext.
|
||||
*
|
||||
* At an accept state, we ignore outerContext unless we had to use it
|
||||
* during prediction. If outerContext != originalContext
|
||||
* (stored in the ATN interpreter), then this config needed context
|
||||
* to find another symbol.
|
||||
*
|
||||
* Lexer matches with ATN so there is no external context; this is null.
|
||||
*/
|
||||
//public RuleContext outerContext;
|
||||
|
||||
/**
|
||||
Indicates that we have reached this ATN configuration after
|
||||
traversing a predicate transition. This is important because we
|
||||
|
@ -60,7 +79,7 @@ public class ATNConfig {
|
|||
|
||||
/**
|
||||
Indicates that we have reached this ATN configuration after
|
||||
traversing a non-force action transition. We do not execute
|
||||
traversing a non-forced action transition. We do not execute
|
||||
predicates after such actions because the predicates could be
|
||||
functions of the side effects. Force actions must be either side
|
||||
effect free or automatically undone as the parse continues.
|
||||
|
@ -80,6 +99,7 @@ public class ATNConfig {
|
|||
this.state = c.state;
|
||||
this.alt = c.alt;
|
||||
this.context = c.context;
|
||||
// this.outerContext = c.outerContext;
|
||||
this.traversedPredicate = c.traversedPredicate;
|
||||
this.traversedAction = c.traversedAction;
|
||||
}
|
||||
|
|
|
@ -50,6 +50,17 @@ public class ParserATNSimulator extends ATNSimulator {
|
|||
public DFA[] decisionToDFA;
|
||||
protected boolean userWantsCtxSensitive = false;
|
||||
|
||||
/** This is the original context upon entry to the ATN simulator.
|
||||
* ATNConfig objects carry on tracking the new context derived from
|
||||
* the decision point. This field is used instead of passing the value
|
||||
* around to the various functions, which would be confusing. Its
|
||||
* value is reset upon prediction call to adaptivePredict() or the
|
||||
* predictATN/DFA methods.
|
||||
*
|
||||
* The full stack at any moment is [config.outerContext + config.context].
|
||||
*/
|
||||
protected RuleContext originalContext;
|
||||
|
||||
protected Set<ATNConfig> closureBusy = new HashSet<ATNConfig>();
|
||||
|
||||
public ParserATNSimulator(ATN atn) {
|
||||
|
@ -96,7 +107,7 @@ public class ParserATNSimulator extends ATNSimulator {
|
|||
RuleContext ctx = RuleContext.EMPTY;
|
||||
if ( useContext ) ctx = originalContext;
|
||||
OrderedHashSet<ATNConfig> s0_closure =
|
||||
computeStartState(dfa.atnStartState, ctx, originalContext);
|
||||
computeStartState(dfa.atnStartState, ctx);
|
||||
dfa.s0 = addDFAState(dfa, s0_closure);
|
||||
if ( prevAccept!=null ) {
|
||||
dfa.s0.isAcceptState = true;
|
||||
|
@ -120,7 +131,7 @@ public class ParserATNSimulator extends ATNSimulator {
|
|||
public int matchATN(TokenStream input, ATNState startState) {
|
||||
DFA dfa = new DFA(startState);
|
||||
RuleContext ctx = new ParserRuleContext();
|
||||
OrderedHashSet<ATNConfig> s0_closure = computeStartState(startState, ctx, RuleContext.EMPTY);
|
||||
OrderedHashSet<ATNConfig> s0_closure = computeStartState(startState, ctx);
|
||||
return execATN(input, dfa, input.index(), s0_closure, ctx, false);
|
||||
}
|
||||
|
||||
|
@ -244,7 +255,7 @@ public class ParserATNSimulator extends ATNSimulator {
|
|||
Transition trans = c.state.transition(ti);
|
||||
ATNState target = getReachableTarget(trans, t);
|
||||
if ( target!=null ) {
|
||||
closure(new ATNConfig(c, target), reach, originalContext);
|
||||
closure(new ATNConfig(c, target), reach);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -368,9 +379,7 @@ public class ParserATNSimulator extends ATNSimulator {
|
|||
return predictedAlt;
|
||||
}
|
||||
|
||||
public OrderedHashSet<ATNConfig> computeStartState(ATNState p, RuleContext ctx,
|
||||
RuleContext originalContext)
|
||||
{
|
||||
public OrderedHashSet<ATNConfig> computeStartState(ATNState p, RuleContext ctx) {
|
||||
RuleContext initialContext = ctx; // always at least the implicit call to start rule
|
||||
OrderedHashSet<ATNConfig> configs = new OrderedHashSet<ATNConfig>();
|
||||
prevAccept = null; // might reach end rule; track
|
||||
|
@ -379,7 +388,7 @@ public class ParserATNSimulator extends ATNSimulator {
|
|||
for (int i=0; i<p.getNumberOfTransitions(); i++) {
|
||||
ATNState target = p.transition(i).target;
|
||||
ATNConfig c = new ATNConfig(target, i+1, initialContext);
|
||||
closure(c, configs, originalContext);
|
||||
closure(c, configs);
|
||||
}
|
||||
|
||||
return configs;
|
||||
|
@ -407,15 +416,13 @@ public class ParserATNSimulator extends ATNSimulator {
|
|||
return null;
|
||||
}
|
||||
|
||||
protected void closure(ATNConfig config, OrderedHashSet<ATNConfig> configs,
|
||||
RuleContext originalContext) {
|
||||
protected void closure(ATNConfig config, OrderedHashSet<ATNConfig> configs) {
|
||||
closureBusy.clear();
|
||||
closure(config, configs, originalContext, closureBusy);
|
||||
closure(config, configs, closureBusy);
|
||||
}
|
||||
|
||||
protected void closure(ATNConfig config,
|
||||
OrderedHashSet<ATNConfig> configs,
|
||||
RuleContext originalContext,
|
||||
Set<ATNConfig> closureBusy)
|
||||
{
|
||||
if ( debug ) System.out.println("closure("+config+")");
|
||||
|
@ -431,7 +438,7 @@ public class ParserATNSimulator extends ATNSimulator {
|
|||
RuleTransition rt = (RuleTransition)invokingState.transition(0);
|
||||
ATNState retState = rt.followState;
|
||||
ATNConfig c = new ATNConfig(retState, config.alt, newContext);
|
||||
closure(c, configs, originalContext, closureBusy);
|
||||
closure(c, configs, closureBusy);
|
||||
return;
|
||||
}
|
||||
// else if we have no context info, just chase follow links
|
||||
|
@ -444,14 +451,12 @@ public class ParserATNSimulator extends ATNSimulator {
|
|||
for (int i=0; i<p.getNumberOfTransitions(); i++) {
|
||||
Transition t = p.transition(i);
|
||||
boolean evalPreds = !config.traversedAction;
|
||||
ATNConfig c = getEpsilonTarget(config, t, originalContext, evalPreds);
|
||||
if ( c!=null ) closure(c, configs, originalContext, closureBusy);
|
||||
ATNConfig c = getEpsilonTarget(config, t, evalPreds);
|
||||
if ( c!=null ) closure(c, configs, closureBusy);
|
||||
}
|
||||
}
|
||||
|
||||
public ATNConfig getEpsilonTarget(ATNConfig config, Transition t,
|
||||
RuleContext originalContext, boolean evalPreds)
|
||||
{
|
||||
public ATNConfig getEpsilonTarget(ATNConfig config, Transition t, boolean evalPreds) {
|
||||
ATNConfig c = null;
|
||||
if ( t instanceof RuleTransition ) {
|
||||
ATNState p = config.state;
|
||||
|
|
|
@ -0,0 +1,751 @@
|
|||
/*
|
||||
[The "BSD license"]
|
||||
Copyright (c) 2011 Terence Parr
|
||||
All rights reserved.
|
||||
|
||||
Redistribution and use in source and binary forms, with or without
|
||||
modification, are permitted provided that the following conditions
|
||||
are met:
|
||||
|
||||
1. Redistributions of source code must retain the above copyright
|
||||
notice, this list of conditions and the following disclaimer.
|
||||
2. Redistributions in binary form must reproduce the above copyright
|
||||
notice, this list of conditions and the following disclaimer in the
|
||||
documentation and/or other materials provided with the distribution.
|
||||
3. The name of the author may not be used to endorse or promote products
|
||||
derived from this software without specific prior written permission.
|
||||
|
||||
THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
|
||||
IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
|
||||
OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
|
||||
IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
|
||||
INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
|
||||
NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
||||
DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
||||
THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
||||
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
|
||||
THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||
*/
|
||||
|
||||
/**********************************
|
||||
* This version I was trying to keep an outer context as well as the
|
||||
* context used during prediction, but it started getting complicated.
|
||||
* For example, s0.configs start state configurations were computed based
|
||||
* upon no an initial context. Later, if I need context I can't start
|
||||
* from these computations. It has to be started over completely.
|
||||
*
|
||||
**********************************/
|
||||
package org.antlr.v4.runtime.atn;
|
||||
|
||||
import org.antlr.v4.runtime.*;
|
||||
import org.antlr.v4.runtime.dfa.*;
|
||||
import org.antlr.v4.runtime.misc.OrderedHashSet;
|
||||
import org.stringtemplate.v4.misc.MultiMap;
|
||||
|
||||
import java.util.*;
|
||||
|
||||
public class ParserATNSimulatorVariationInnerOuterContexts extends ATNSimulator {
|
||||
public static boolean debug = true;
|
||||
public static boolean dfa_debug = true;
|
||||
|
||||
public static int ATN_failover = 0;
|
||||
public static int predict_calls = 0;
|
||||
|
||||
protected BaseRecognizer parser;
|
||||
|
||||
public Map<RuleContext, DFA[]> ctxToDFAs;
|
||||
public Map<RuleContext, DFA>[] decisionToDFAPerCtx; // TODO: USE THIS ONE
|
||||
public DFA[] decisionToDFA;
|
||||
protected boolean userWantsCtxSensitive = false;
|
||||
|
||||
/** This is the original context upon entry to the ATN simulator.
|
||||
* ATNConfig objects carry on tracking the new context derived from
|
||||
* the decision point. This field is used instead of passing the value
|
||||
* around to the various functions, which would be confusing. Its
|
||||
* value is reset upon prediction call to adaptivePredict() or the
|
||||
* predictATN/DFA methods.
|
||||
*
|
||||
* The full stack at any moment is [config.outerContext + config.context].
|
||||
*/
|
||||
protected RuleContext originalContext;
|
||||
|
||||
protected Set<ATNConfig> closureBusy = new HashSet<ATNConfig>();
|
||||
|
||||
public ParserATNSimulatorVariationInnerOuterContexts(ATN atn) {
|
||||
super(atn);
|
||||
ctxToDFAs = new HashMap<RuleContext, DFA[]>();
|
||||
decisionToDFA = new DFA[atn.getNumberOfDecisions()];
|
||||
}
|
||||
|
||||
public ParserATNSimulatorVariationInnerOuterContexts(BaseRecognizer parser, ATN atn) {
|
||||
super(atn);
|
||||
this.parser = parser;
|
||||
ctxToDFAs = new HashMap<RuleContext, DFA[]>();
|
||||
decisionToDFA = new DFA[atn.getNumberOfDecisions()+1];
|
||||
// DOTGenerator dot = new DOTGenerator(null);
|
||||
// System.out.println(dot.getDOT(atn.rules.get(0), parser.getRuleNames()));
|
||||
// System.out.println(dot.getDOT(atn.rules.get(1), parser.getRuleNames()));
|
||||
}
|
||||
|
||||
public int adaptivePredict(TokenStream input, int decision, RuleContext originalContext) {
|
||||
predict_calls++;
|
||||
DFA dfa = decisionToDFA[decision];
|
||||
if ( dfa==null || dfa.s0==null ) {
|
||||
ATNState startState = atn.decisionToState.get(decision);
|
||||
decisionToDFA[decision] = dfa = new DFA(startState);
|
||||
dfa.decision = decision;
|
||||
return predictATN(dfa, input, decision, originalContext, false);
|
||||
}
|
||||
else {
|
||||
//dump(dfa);
|
||||
// start with the DFA
|
||||
int m = input.mark();
|
||||
int alt = execDFA(input, dfa, dfa.s0, originalContext);
|
||||
input.seek(m);
|
||||
return alt;
|
||||
}
|
||||
}
|
||||
|
||||
public int predictATN(DFA dfa, TokenStream input,
|
||||
int decision,
|
||||
RuleContext originalContext,
|
||||
boolean useContext)
|
||||
{
|
||||
if ( originalContext==null ) originalContext = RuleContext.EMPTY;
|
||||
this.originalContext = originalContext;
|
||||
RuleContext ctx = RuleContext.EMPTY;
|
||||
if ( useContext ) ctx = originalContext;
|
||||
OrderedHashSet<ATNConfig> s0_closure =
|
||||
computeStartState(dfa.atnStartState, ctx, originalContext);
|
||||
dfa.s0 = addDFAState(dfa, s0_closure);
|
||||
if ( prevAccept!=null ) {
|
||||
dfa.s0.isAcceptState = true;
|
||||
dfa.s0.prediction = prevAccept.alt;
|
||||
}
|
||||
|
||||
int alt = 0;
|
||||
int m = input.mark();
|
||||
try {
|
||||
alt = execATN(input, dfa, m, s0_closure, originalContext, useContext);
|
||||
}
|
||||
catch (NoViableAltException nvae) { dumpDeadEndConfigs(nvae); throw nvae; }
|
||||
finally {
|
||||
input.seek(m);
|
||||
}
|
||||
if ( debug ) System.out.println("DFA after predictATN: "+dfa.toString());
|
||||
return alt;
|
||||
}
|
||||
|
||||
// doesn't create DFA when matching
|
||||
public int matchATN(TokenStream input, ATNState startState) {
|
||||
DFA dfa = new DFA(startState);
|
||||
RuleContext ctx = new ParserRuleContext();
|
||||
OrderedHashSet<ATNConfig> s0_closure = computeStartState(startState, ctx, RuleContext.EMPTY);
|
||||
return execATN(input, dfa, input.index(), s0_closure, ctx, false);
|
||||
}
|
||||
|
||||
public int execDFA(TokenStream input, DFA dfa, DFAState s0, RuleContext originalContext) {
|
||||
if ( dfa_debug ) System.out.println("DFA decision "+dfa.decision+" exec LA(1)=="+input.LT(1));
|
||||
// dump(dfa);
|
||||
if ( originalContext==null ) originalContext = RuleContext.EMPTY;
|
||||
this.originalContext = originalContext;
|
||||
DFAState prevAcceptState = null;
|
||||
DFAState s = s0;
|
||||
int t = input.LA(1);
|
||||
int start = input.index();
|
||||
loop:
|
||||
while ( true ) {
|
||||
if ( dfa_debug ) System.out.println("DFA state "+s.stateNumber+" LA(1)=="+t);
|
||||
// TODO: ctxSensitive
|
||||
if ( s.isCtxSensitive ) {
|
||||
Integer predI = s.ctxToPrediction.get(originalContext);
|
||||
if ( dfa_debug ) System.out.println("ctx sensitive state "+originalContext+"->"+predI+
|
||||
" in "+s);
|
||||
if ( predI!=null ) return predI;
|
||||
// System.out.println("start all over with ATN; can't use DFA");
|
||||
// start all over with ATN; can't use DFA
|
||||
input.seek(start);
|
||||
DFA throwAwayDFA = new DFA(dfa.atnStartState);
|
||||
int alt = predictATN(throwAwayDFA, input, dfa.decision, originalContext, false);
|
||||
// int alt = execATN(input, throwAwayDFA, start, s0.configs, originalContext, false);
|
||||
// we might get to a different input sequence by retrying with this new context
|
||||
// so don't set the prediction for this accept state.
|
||||
// wait: we are talking about the same input here. it can match more
|
||||
// or less but not different input.Oh, that updates a throwaway DFA.
|
||||
// It will not update our DFA
|
||||
s.ctxToPrediction.put(originalContext, alt);
|
||||
if ( dfa_debug ) System.out.println("after retry, ctx sensitive state mapping "+originalContext+"->"+alt);
|
||||
return alt;
|
||||
}
|
||||
if ( s.isAcceptState ) {
|
||||
if ( dfa_debug ) System.out.println("accept; predict "+s.prediction +" in state "+s.stateNumber);
|
||||
prevAcceptState = s;
|
||||
if ( s.ctxToPrediction!=null ) {
|
||||
Integer predI = s.ctxToPrediction.get(originalContext);
|
||||
System.out.println("ctx based prediction: "+predI);
|
||||
}
|
||||
|
||||
// keep going unless we're at EOF or state only has one alt number
|
||||
// mentioned in configs; check if something else could match
|
||||
// TODO: for prediction, we must stop here I believe; there's no way to get
|
||||
// past the stop state since we will never look for the longest match.
|
||||
// during prediction, we always look for enough input to get a unique prediction.
|
||||
// nothing beyond that will help
|
||||
if ( s.complete || t==CharStream.EOF ) break;
|
||||
}
|
||||
// if no edge, pop over to ATN interpreter, update DFA and return
|
||||
if ( s.edges == null || t >= s.edges.length || s.edges[t+1] == null ) {
|
||||
if ( dfa_debug ) System.out.println("no edge for "+t);
|
||||
int alt = -1;
|
||||
if ( dfa_debug ) {
|
||||
System.out.println("ATN exec upon "+
|
||||
input.toString(start,input.index())+
|
||||
" at DFA state "+s.stateNumber);
|
||||
}
|
||||
try {
|
||||
alt = execATN(input, dfa, start, s.configs, originalContext, false);
|
||||
// this adds edge even if next state is accept for
|
||||
// same alt; e.g., s0-A->:s1=>2-B->:s2=>2
|
||||
// TODO: This next stuff kills edge, but extra states remain. :(
|
||||
if ( s.isAcceptState && alt!=-1 ) {
|
||||
DFAState d = s.edges[input.LA(1)+1];
|
||||
if ( d.isAcceptState && d.prediction==s.prediction ) {
|
||||
// we can carve it out.
|
||||
s.edges[input.LA(1)+1] = ERROR; // IGNORE really not error
|
||||
}
|
||||
}
|
||||
}
|
||||
catch (NoViableAltException nvae) {
|
||||
alt = -1;
|
||||
}
|
||||
if ( dfa_debug ) {
|
||||
System.out.println("back from DFA update, alt="+alt+", dfa=\n"+dfa);
|
||||
//dump(dfa);
|
||||
}
|
||||
if ( alt==-1 ) {
|
||||
addDFAEdge(s, t, ERROR);
|
||||
break loop; // dead end; no where to go, fall back on prev if any
|
||||
}
|
||||
// action already executed
|
||||
if ( dfa_debug ) System.out.println("DFA decision "+dfa.decision+
|
||||
" predicts "+alt);
|
||||
return alt; // we've updated DFA, exec'd action, and have our deepest answer
|
||||
}
|
||||
DFAState target = s.edges[t+1];
|
||||
if ( target == ERROR ) break;
|
||||
s = target;
|
||||
input.consume();
|
||||
t = input.LA(1);
|
||||
}
|
||||
if ( prevAcceptState==null ) {
|
||||
System.out.println("!!! no viable alt in dfa");
|
||||
return -1;
|
||||
}
|
||||
if ( dfa_debug ) System.out.println("DFA decision "+dfa.decision+
|
||||
" predicts "+prevAcceptState.prediction);
|
||||
return prevAcceptState.prediction;
|
||||
}
|
||||
|
||||
public int execATN(TokenStream input,
|
||||
DFA dfa,
|
||||
int startIndex,
|
||||
OrderedHashSet<ATNConfig> s0,
|
||||
RuleContext originalContext,
|
||||
boolean useContext)
|
||||
{
|
||||
if ( debug ) System.out.println("ATN decision "+dfa.decision+" exec LA(1)=="+input.LT(1));
|
||||
ATN_failover++;
|
||||
OrderedHashSet<ATNConfig> closure = new OrderedHashSet<ATNConfig>();
|
||||
|
||||
closure.addAll(s0);
|
||||
|
||||
if ( debug ) System.out.println("start state closure="+closure);
|
||||
|
||||
int t = input.LA(1);
|
||||
if ( t==Token.EOF && prevAccept!=null ) {
|
||||
// computeStartState must have reached end of rule
|
||||
return prevAccept.alt;
|
||||
}
|
||||
|
||||
prevAccept = null;
|
||||
prevAcceptIndex = -1;
|
||||
OrderedHashSet<ATNConfig> reach = new OrderedHashSet<ATNConfig>();
|
||||
|
||||
do { // while more work
|
||||
if ( debug ) System.out.println("in reach starting closure: " + closure);
|
||||
int ncl = closure.size();
|
||||
for (int ci=0; ci<ncl; ci++) { // TODO: foreach
|
||||
ATNConfig c = closure.get(ci);
|
||||
if ( debug ) System.out.println("testing "+getTokenName(t)+" at "+c.toString());
|
||||
int n = c.state.getNumberOfTransitions();
|
||||
for (int ti=0; ti<n; ti++) { // for each transition
|
||||
Transition trans = c.state.transition(ti);
|
||||
ATNState target = getReachableTarget(trans, t);
|
||||
if ( target!=null ) {
|
||||
closure(new ATNConfig(c, target), reach);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// resolve ambig in DFAState for reach
|
||||
Set<Integer> ambigAlts = getAmbiguousAlts(reach);
|
||||
if ( ambigAlts!=null ) {
|
||||
if ( debug ) {
|
||||
ATNState loc = atn.states.get(originalContext.s);
|
||||
String rname = "n/a";
|
||||
if ( parser !=null ) rname = parser.getRuleNames()[loc.ruleIndex];
|
||||
System.out.println("AMBIG in "+rname+" for alt "+ambigAlts+" upon "+
|
||||
input.toString(startIndex, input.index()));
|
||||
}
|
||||
dfa.conflict = true; // at least one DFA state is ambiguous
|
||||
if ( !userWantsCtxSensitive ) reportConflict(startIndex, input.index(), ambigAlts, reach);
|
||||
|
||||
// ATNState loc = atn.states.get(originalContext.s);
|
||||
// String rname = recog.getRuleNames()[loc.ruleIndex];
|
||||
// System.out.println("AMBIG orig="+originalContext.toString((BaseRecognizer)recog)+" for alt "+ambigAlts+" upon "+
|
||||
// input.toString(startIndex, input.index()));
|
||||
if ( !userWantsCtxSensitive || useContext ) {
|
||||
resolveToMinAlt(reach, ambigAlts);
|
||||
}
|
||||
else {
|
||||
return retryWithContext(input, dfa, startIndex, originalContext,
|
||||
closure, t, reach, ambigAlts);
|
||||
}
|
||||
}
|
||||
|
||||
// if reach predicts single alt, can stop
|
||||
|
||||
int uniqueAlt = getUniqueAlt(reach);
|
||||
if ( uniqueAlt!=ATN.INVALID_ALT_NUMBER ) {
|
||||
if ( debug ) System.out.println("PREDICT alt "+uniqueAlt+
|
||||
" decision "+dfa.decision+
|
||||
" at index "+input.index());
|
||||
addDFAEdge(dfa, closure, t, reach);
|
||||
makeAcceptState(dfa, reach, uniqueAlt);
|
||||
return uniqueAlt;
|
||||
}
|
||||
|
||||
if ( reach.size()==0 ) {
|
||||
break;
|
||||
}
|
||||
|
||||
// If we matched t anywhere, need to consume and add closer-t->reach DFA edge
|
||||
// else error if no previous accept
|
||||
input.consume();
|
||||
addDFAEdge(dfa, closure, t, reach);
|
||||
t = input.LA(1);
|
||||
|
||||
// swap to avoid reallocating space
|
||||
OrderedHashSet<ATNConfig> tmp = reach;
|
||||
reach = closure;
|
||||
closure = tmp;
|
||||
reach.clear(); // THIS MIGHT BE SLOW! kills each element; realloc might be faster
|
||||
} while ( true );
|
||||
|
||||
if ( prevAccept==null ) {
|
||||
System.out.println("no viable token at input "+input.LT(1)+", index "+input.index());
|
||||
NoViableAltException nvae = new NoViableAltException(parser, input, closure, originalContext);
|
||||
nvae.startIndex = startIndex;
|
||||
throw nvae;
|
||||
}
|
||||
|
||||
if ( debug ) System.out.println("PREDICT " + prevAccept + " index " + prevAccept.alt);
|
||||
return prevAccept.alt;
|
||||
}
|
||||
|
||||
protected int resolveToMinAlt(OrderedHashSet<ATNConfig> reach, Set<Integer> ambigAlts) {
|
||||
int min = getMinAlt(ambigAlts);
|
||||
// if predicting, create DFA accept state for resolved alt
|
||||
ambigAlts.remove(min);
|
||||
// kill dead alts so we don't chase them ever
|
||||
killAlts(ambigAlts, reach);
|
||||
if ( debug ) System.out.println("RESOLVED TO "+reach);
|
||||
return min;
|
||||
}
|
||||
|
||||
public int retryWithContext(TokenStream input,
|
||||
DFA dfa,
|
||||
int startIndex,
|
||||
RuleContext originalContext,
|
||||
OrderedHashSet<ATNConfig> closure,
|
||||
int t,
|
||||
OrderedHashSet<ATNConfig> reach,
|
||||
Set<Integer> ambigAlts)
|
||||
{
|
||||
// ASSUMES PREDICT ONLY
|
||||
// retry using context, if any; if none, kill all but min as before
|
||||
if ( debug ) System.out.println("RETRY with ctx="+ originalContext);
|
||||
int min = getMinAlt(ambigAlts);
|
||||
if ( originalContext==RuleContext.EMPTY ) {
|
||||
// no point in retrying with ctx since it's same.
|
||||
// this implies that we have a true ambiguity
|
||||
reportAmbiguity(startIndex, input.index(), ambigAlts, reach);
|
||||
return min;
|
||||
}
|
||||
// otherwise we have to retry with context, filling in tmp DFA.
|
||||
// if it comes back with conflict, we have a true ambiguity
|
||||
input.seek(startIndex); // rewind
|
||||
DFA ctx_dfa = new DFA(dfa.atnStartState);
|
||||
int ctx_alt = predictATN(ctx_dfa, input, dfa.decision, originalContext, true);
|
||||
if ( debug ) System.out.println("retry predicts "+ctx_alt+" vs "+getMinAlt(ambigAlts)+
|
||||
" with conflict="+ctx_dfa.conflict+
|
||||
" dfa="+ctx_dfa);
|
||||
if ( ctx_dfa.conflict ) reportAmbiguity(startIndex, input.index(), ambigAlts, reach);
|
||||
else reportContextSensitivity(startIndex, input.index(), ambigAlts, reach);
|
||||
// it's not context-sensitive; true ambig. fall thru to strip dead alts
|
||||
|
||||
int predictedAlt = ctx_alt;
|
||||
DFAState reachTarget = addDFAEdge(dfa, closure, t, reach);
|
||||
reachTarget.isCtxSensitive = true;
|
||||
if ( reachTarget.ctxToPrediction==null ) {
|
||||
reachTarget.ctxToPrediction = new LinkedHashMap<RuleContext, Integer>();
|
||||
}
|
||||
reachTarget.ctxToPrediction.put(originalContext, predictedAlt);
|
||||
// System.out.println("RESOLVE to "+predictedAlt);
|
||||
//System.out.println(reachTarget.ctxToPrediction.size()+" size of ctx map");
|
||||
return predictedAlt;
|
||||
}
|
||||
|
||||
public OrderedHashSet<ATNConfig> computeStartState(ATNState p, RuleContext ctx,
|
||||
RuleContext originalContext)
|
||||
{
|
||||
RuleContext initialContext = ctx; // always at least the implicit call to start rule
|
||||
OrderedHashSet<ATNConfig> configs = new OrderedHashSet<ATNConfig>();
|
||||
prevAccept = null; // might reach end rule; track
|
||||
prevAcceptIndex = -1;
|
||||
|
||||
for (int i=0; i<p.getNumberOfTransitions(); i++) {
|
||||
ATNState target = p.transition(i).target;
|
||||
ATNConfig c = new ATNConfig(target, i+1, initialContext);
|
||||
// c.outerContext = originalContext; COMMENTED OUT TO COMPILE
|
||||
closure(c, configs);
|
||||
}
|
||||
|
||||
return configs;
|
||||
}
|
||||
|
||||
public ATNState getReachableTarget(Transition trans, int ttype) {
|
||||
if ( trans instanceof AtomTransition ) {
|
||||
AtomTransition at = (AtomTransition)trans;
|
||||
// boolean not = trans instanceof NotAtomTransition;
|
||||
if ( at.label == ttype ) {
|
||||
return at.target;
|
||||
}
|
||||
}
|
||||
else if ( trans instanceof SetTransition ) {
|
||||
SetTransition st = (SetTransition)trans;
|
||||
boolean not = trans instanceof NotSetTransition;
|
||||
if ( !not && st.set.member(ttype) || not && !st.set.member(ttype) ) {
|
||||
return st.target;
|
||||
}
|
||||
}
|
||||
// TODO else if ( trans instanceof WildcardTransition && t!=Token.EOF ) {
|
||||
// ATNConfig targetConfig = new ATNConfig(c, trans.target);
|
||||
// closure(input, targetConfig, reach);
|
||||
// }
|
||||
return null;
|
||||
}
|
||||
|
||||
protected void closure(ATNConfig config, OrderedHashSet<ATNConfig> configs) {
|
||||
closureBusy.clear();
|
||||
closure(config, configs, closureBusy);
|
||||
}
|
||||
|
||||
protected void closure(ATNConfig config,
|
||||
OrderedHashSet<ATNConfig> configs,
|
||||
Set<ATNConfig> closureBusy)
|
||||
{
|
||||
if ( debug ) System.out.println("closure("+config+")");
|
||||
|
||||
if ( closureBusy.contains(config) ) return; // avoid infinite recursion
|
||||
closureBusy.add(config);
|
||||
|
||||
if ( config.state instanceof RuleStopState ) {
|
||||
// if ( debug ) System.out.println("RuleStopState "+config+", originalContext: "+config.outerContext);
|
||||
// We hit rule end. If we have context info, use it
|
||||
if ( config.context!=null && !config.context.isEmpty() ) {
|
||||
RuleContext newContext = config.context.parent; // "pop" invoking state
|
||||
ATNState invokingState = atn.states.get(config.context.invokingState);
|
||||
RuleTransition rt = (RuleTransition)invokingState.transition(0);
|
||||
ATNState retState = rt.followState;
|
||||
//ATNConfig c = new ATNConfig(retState, config.alt, newContext);
|
||||
ATNConfig c = new ATNConfig(config, retState, newContext);
|
||||
closure(c, configs, closureBusy);
|
||||
return;
|
||||
}
|
||||
// else use original context info
|
||||
/* COMMENTED OUT TO COMPILE
|
||||
if ( !config.outerContext.isEmpty() ) {
|
||||
System.out.println("context stack empty, using originalContext");
|
||||
RuleContext newContext = config.outerContext.parent; // "pop" invoking state
|
||||
ATNState invokingState = atn.states.get(config.outerContext.invokingState);
|
||||
RuleTransition rt = (RuleTransition)invokingState.transition(0);
|
||||
ATNState retState = rt.followState;
|
||||
ATNConfig c = new ATNConfig(config, retState, newContext);
|
||||
// pop one from originalContext too for this thread of computation
|
||||
c.outerContext = newContext;
|
||||
closure(c, configs, closureBusy);
|
||||
return;
|
||||
}
|
||||
*/
|
||||
// else if we have no context info, just chase follow links
|
||||
// TODO: should be EOF now that we're using originalContext right?
|
||||
}
|
||||
|
||||
ATNState p = config.state;
|
||||
// optimization
|
||||
if ( !p.onlyHasEpsilonTransitions() ) configs.add(config);
|
||||
|
||||
for (int i=0; i<p.getNumberOfTransitions(); i++) {
|
||||
Transition t = p.transition(i);
|
||||
boolean evalPreds = !config.traversedAction;
|
||||
ATNConfig c = getEpsilonTarget(config, t, evalPreds);
|
||||
if ( c!=null ) closure(c, configs, closureBusy);
|
||||
}
|
||||
}
|
||||
|
||||
public ATNConfig getEpsilonTarget(ATNConfig config, Transition t, boolean evalPreds) {
|
||||
ATNConfig c = null;
|
||||
if ( t instanceof RuleTransition ) {
|
||||
ATNState p = config.state;
|
||||
RuleContext newContext;
|
||||
if ( parser != null ) {
|
||||
System.out.println("rule trans to rule "+parser.getRuleNames()[t.target.ruleIndex]);
|
||||
newContext = parser.newContext(config.context, t.target.stateNumber, t.target.ruleIndex, -999);
|
||||
newContext.invokingState = p.stateNumber;
|
||||
System.out.println("new ctx type is "+newContext.getClass().getSimpleName());
|
||||
}
|
||||
else {
|
||||
newContext = new RuleContext(config.context, p.stateNumber, t.target.stateNumber);
|
||||
}
|
||||
c = new ATNConfig(config, t.target, newContext);
|
||||
}
|
||||
else if ( t instanceof PredicateTransition ) {
|
||||
PredicateTransition pt = (PredicateTransition)t;
|
||||
if ( debug ) System.out.println("PRED (eval="+evalPreds+") "+pt.ruleIndex+":"+pt.predIndex);
|
||||
if ( parser != null ) System.out.println("rule surrounding pred is "+
|
||||
parser.getRuleNames()[pt.ruleIndex]);
|
||||
// preds are epsilon if we're not doing preds.
|
||||
// if we are doing preds, pred must eval to true
|
||||
/* COMMENTED OUT TO COMPILE
|
||||
if ( !evalPreds ||
|
||||
(evalPreds && parser.sempred(config.outerContext, pt.ruleIndex, pt.predIndex)) ) {
|
||||
c = new ATNConfig(config, t.target);
|
||||
c.traversedPredicate = true;
|
||||
}
|
||||
*/
|
||||
}
|
||||
else if ( t instanceof ActionTransition ) {
|
||||
c = new ATNConfig(config, t.target);
|
||||
ActionTransition at = (ActionTransition)t;
|
||||
if ( debug ) System.out.println("ACTION edge "+at.ruleIndex+":"+at.actionIndex);
|
||||
if ( at.actionIndex>=0 ) {
|
||||
if ( debug ) System.out.println("DO ACTION "+at.ruleIndex+":"+at.actionIndex);
|
||||
// COMMENTED OUT TO COMPILE parser.action(config.outerContext, at.ruleIndex, at.actionIndex);
|
||||
}
|
||||
else {
|
||||
// non-forced action traversed to get to t.target
|
||||
if ( debug && !config.traversedAction ) {
|
||||
System.out.println("NONFORCED; pruning future pred eval derived from s"+
|
||||
config.state.stateNumber);
|
||||
}
|
||||
c.traversedAction = true;
|
||||
}
|
||||
}
|
||||
else if ( t.isEpsilon() ) {
|
||||
c = new ATNConfig(config, t.target);
|
||||
}
|
||||
return c;
|
||||
}
|
||||
|
||||
public void reportConflict(int startIndex, int stopIndex, Set<Integer> alts, OrderedHashSet<ATNConfig> configs) {
|
||||
if ( parser!=null ) parser.reportConflict(startIndex, stopIndex, alts, configs);
|
||||
}
|
||||
|
||||
public void reportContextSensitivity(int startIndex, int stopIndex, Set<Integer> alts, OrderedHashSet<ATNConfig> configs) {
|
||||
if ( parser!=null ) parser.reportContextSensitivity(startIndex, stopIndex, alts, configs);
|
||||
}
|
||||
|
||||
/** If context sensitive parsing, we know it's ambiguity not conflict */
|
||||
public void reportAmbiguity(int startIndex, int stopIndex, Set<Integer> alts, OrderedHashSet<ATNConfig> configs) {
|
||||
if ( parser!=null ) parser.reportAmbiguity(startIndex, stopIndex, alts, configs);
|
||||
}
|
||||
|
||||
public static int getUniqueAlt(Collection<ATNConfig> configs) {
|
||||
int alt = ATN.INVALID_ALT_NUMBER;
|
||||
for (ATNConfig c : configs) {
|
||||
if ( alt == ATN.INVALID_ALT_NUMBER ) {
|
||||
alt = c.alt; // found first alt
|
||||
}
|
||||
else if ( c.alt!=alt ) {
|
||||
return ATN.INVALID_ALT_NUMBER;
|
||||
}
|
||||
}
|
||||
return alt;
|
||||
}
|
||||
|
||||
public Set<Integer> getAmbiguousAlts(OrderedHashSet<ATNConfig> configs) {
|
||||
// System.err.println("check ambiguous "+configs);
|
||||
Set<Integer> ambigAlts = null;
|
||||
int numConfigs = configs.size();
|
||||
// First get a list of configurations for each state.
|
||||
// Most of the time, each state will have one associated configuration.
|
||||
MultiMap<Integer, ATNConfig> stateToConfigListMap =
|
||||
new MultiMap<Integer, ATNConfig>();
|
||||
for (ATNConfig c : configs) {
|
||||
stateToConfigListMap.map(c.state.stateNumber, c);
|
||||
}
|
||||
// potential conflicts are states with > 1 configuration and diff alts
|
||||
for (List<ATNConfig> configsPerAlt : stateToConfigListMap.values()) {
|
||||
ATNConfig goal = configsPerAlt.get(0);
|
||||
int size = configsPerAlt.size();
|
||||
for (int i=1; i< size; i++) {
|
||||
ATNConfig c = configsPerAlt.get(i);
|
||||
if ( c.alt!=goal.alt ) {
|
||||
//System.out.println("chk stack "+goal+", "+c);
|
||||
boolean sameCtx =
|
||||
(goal.context==null&&c.context==null) ||
|
||||
goal.context.equals(c.context) ||
|
||||
c.context.conflictsWith(goal.context);
|
||||
if ( sameCtx ) {
|
||||
if ( debug ) {
|
||||
System.out.println("we reach state "+c.state.stateNumber+
|
||||
" in rule "+
|
||||
(parser !=null ? parser.getRuleNames()[c.state.ruleIndex]:"n/a")+
|
||||
" alts "+goal.alt+","+c.alt+" from ctx "+goal.context.toString((BaseRecognizer) parser)
|
||||
+" and "+
|
||||
c.context.toString((BaseRecognizer) parser));
|
||||
}
|
||||
if ( ambigAlts==null ) ambigAlts = new HashSet<Integer>();
|
||||
ambigAlts.add(goal.alt);
|
||||
ambigAlts.add(c.alt);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
if ( ambigAlts!=null ) {
|
||||
//System.err.println("ambig upon "+input.toString(startIndex, input.index()));
|
||||
}
|
||||
return ambigAlts;
|
||||
}
|
||||
|
||||
public static int getMinAlt(Set<Integer> ambigAlts) {
|
||||
int min = Integer.MAX_VALUE;
|
||||
for (int alt : ambigAlts) {
|
||||
if ( alt < min ) min = alt;
|
||||
}
|
||||
return min;
|
||||
}
|
||||
|
||||
public static void killAlts(Set<Integer> alts, OrderedHashSet<ATNConfig> configs) {
|
||||
int i = 0;
|
||||
while ( i<configs.size() ) {
|
||||
ATNConfig c = configs.get(i);
|
||||
if ( alts.contains(c.alt) ) {
|
||||
configs.remove(i);
|
||||
}
|
||||
else i++;
|
||||
}
|
||||
}
|
||||
|
||||
protected DFAState addDFAEdge(DFA dfa,
|
||||
OrderedHashSet<ATNConfig> p,
|
||||
int t,
|
||||
OrderedHashSet<ATNConfig> q)
|
||||
{
|
||||
System.out.println("MOVE "+p+" -> "+q+" upon "+getTokenName(t));
|
||||
DFAState from = addDFAState(dfa, p);
|
||||
DFAState to = addDFAState(dfa, q);
|
||||
addDFAEdge(from, t, to);
|
||||
return to;
|
||||
}
|
||||
|
||||
protected void addDFAEdge(DFAState p, int t, DFAState q) {
|
||||
if ( p==null ) return;
|
||||
if ( p.edges==null ) {
|
||||
p.edges = new DFAState[atn.maxTokenType+1+1]; // TODO: make adaptive
|
||||
}
|
||||
p.edges[t+1] = q; // connect
|
||||
}
|
||||
|
||||
/** See comment on LexerInterpreter.addDFAState. */
|
||||
protected DFAState addDFAState(DFA dfa, OrderedHashSet<ATNConfig> configs) {
|
||||
DFAState proposed = new DFAState(configs);
|
||||
DFAState existing = dfa.states.get(proposed);
|
||||
if ( existing!=null ) return existing;
|
||||
|
||||
DFAState newState = proposed;
|
||||
|
||||
boolean traversedPredicate = false;
|
||||
for (ATNConfig c : configs) {
|
||||
if ( c.traversedPredicate ) {traversedPredicate = true; break;}
|
||||
}
|
||||
|
||||
if ( traversedPredicate ) return null; // cannot cache
|
||||
|
||||
newState.stateNumber = dfa.states.size();
|
||||
newState.configs = new OrderedHashSet<ATNConfig>();
|
||||
newState.configs.addAll(configs);
|
||||
dfa.states.put(newState, newState);
|
||||
return newState;
|
||||
}
|
||||
|
||||
public void makeAcceptState(DFA dfa, OrderedHashSet<ATNConfig> reach, int uniqueAlt) {
|
||||
DFAState accept = dfa.states.get(new DFAState(reach));
|
||||
if ( accept==null ) return;
|
||||
|
||||
boolean usesOuterContext;
|
||||
RuleContext outerContext = null;
|
||||
for (ATNConfig c : accept.configs) {
|
||||
/* COMMENTED OUT TO COMPILE
|
||||
if ( c.outerContext!=originalContext ) {
|
||||
System.out.println("### we used context "+
|
||||
originalContext.toString(parser, c.outerContext)+" from "+
|
||||
originalContext.toString(parser));
|
||||
// TODO:Will they ever be different context sizes for the same except state?
|
||||
// seems like they can only one context depth otherwise we'd
|
||||
// be a different except states for different input
|
||||
usesOuterContext = true;
|
||||
outerContext = c.outerContext;
|
||||
break;
|
||||
}
|
||||
*/
|
||||
}
|
||||
if ( outerContext!=null ) {
|
||||
// accept.setContextSensitivePrediction(originalContext, uniqueAlt);
|
||||
}
|
||||
else {
|
||||
accept.isAcceptState = true;
|
||||
accept.prediction = uniqueAlt;
|
||||
accept.complete = true;
|
||||
}
|
||||
}
|
||||
|
||||
public String getTokenName(int t) {
|
||||
if ( t==-1 ) return "EOF";
|
||||
if ( parser!=null && parser.getTokenNames()!=null ) return parser.getTokenNames()[t]+"<"+t+">";
|
||||
return String.valueOf(t);
|
||||
}
|
||||
|
||||
public void setContextSensitive(boolean ctxSensitive) {
|
||||
this.userWantsCtxSensitive = ctxSensitive;
|
||||
}
|
||||
|
||||
public void dumpDeadEndConfigs(NoViableAltException nvae) {
|
||||
System.err.println("dead end configs: ");
|
||||
for (ATNConfig c : nvae.deadEndConfigs) {
|
||||
Transition t = c.state.transition(0);
|
||||
String trans = "";
|
||||
if ( t instanceof AtomTransition) {
|
||||
AtomTransition at = (AtomTransition)t;
|
||||
trans = "Atom "+getTokenName(at.label);
|
||||
}
|
||||
else if ( t instanceof SetTransition ) {
|
||||
SetTransition st = (SetTransition)t;
|
||||
boolean not = st instanceof NotSetTransition;
|
||||
trans = (not?"~":"")+"Set "+st.set.toString();
|
||||
}
|
||||
System.err.println(c.toString(parser, true)+":"+trans);
|
||||
}
|
||||
}
|
||||
}
|
|
@ -100,6 +100,16 @@ public class DFAState {
|
|||
return alts;
|
||||
}
|
||||
|
||||
/*
|
||||
public void setContextSensitivePrediction(RuleContext ctx, int predictedAlt) {
|
||||
isCtxSensitive = true;
|
||||
if ( ctxToPrediction==null ) {
|
||||
ctxToPrediction = new LinkedHashMap<RuleContext, Integer>();
|
||||
}
|
||||
ctxToPrediction.put(ctx, predictedAlt);
|
||||
}
|
||||
*/
|
||||
|
||||
/** A decent hash for a DFA state is the sum of the ATN state/alt pairs. */
|
||||
public int hashCode() {
|
||||
int h = 0;
|
||||
|
|
|
@ -2,7 +2,6 @@ import org.antlr.v4.Tool;
|
|||
import org.antlr.v4.automata.ParserATNFactory;
|
||||
import org.antlr.v4.runtime.*;
|
||||
import org.antlr.v4.runtime.atn.ATN;
|
||||
import org.antlr.v4.runtime.tree.Tree;
|
||||
import org.antlr.v4.semantics.SemanticPipeline;
|
||||
import org.antlr.v4.tool.*;
|
||||
|
||||
|
@ -14,7 +13,7 @@ public class TestT {
|
|||
CommonTokenStream tokens = new CommonTokenStream(t);
|
||||
TParser p = new TParser(tokens);
|
||||
ParserRuleContext ret = p.s();
|
||||
System.out.println(((Tree)ret.tree).toStringTree());
|
||||
// System.out.println(((Tree)ret.tree).toStringTree());
|
||||
}
|
||||
|
||||
public static void dump() throws Exception {
|
||||
|
|
Loading…
Reference in New Issue