forked from jasder/antlr
decided to clean up a bit, backtracking from updating current to have inner/outer context. will try with full ctx again.
[git-p4: depot-paths = "//depot/code/antlr4/main/": change = 8908]
This commit is contained in:
parent
76ca9ba0d5
commit
c6fa2fce6e
|
@ -211,10 +211,14 @@ public class RuleContext {
|
||||||
}
|
}
|
||||||
|
|
||||||
public String toString(BaseRecognizer recog) {
|
public String toString(BaseRecognizer recog) {
|
||||||
|
return toString(recog, null);
|
||||||
|
}
|
||||||
|
|
||||||
|
public String toString(BaseRecognizer recog, RuleContext stop) {
|
||||||
StringBuffer buf = new StringBuffer();
|
StringBuffer buf = new StringBuffer();
|
||||||
RuleContext p = this;
|
RuleContext p = this;
|
||||||
buf.append("[");
|
buf.append("[");
|
||||||
while ( p != null ) {
|
while ( p != null && p != stop ) {
|
||||||
if ( recog!=null ) {
|
if ( recog!=null ) {
|
||||||
ATN atn = recog.getATN();
|
ATN atn = recog.getATN();
|
||||||
ATNState s = atn.states.get(p.s);
|
ATNState s = atn.states.get(p.s);
|
||||||
|
|
|
@ -45,10 +45,29 @@ public class ATNConfig {
|
||||||
public int alt;
|
public int alt;
|
||||||
|
|
||||||
/** The stack of invoking states leading to the rule/states associated
|
/** The stack of invoking states leading to the rule/states associated
|
||||||
* wit this config.
|
* with this config. We track only those contexts pushed during
|
||||||
|
* execution of the ATN simulator.
|
||||||
*/
|
*/
|
||||||
public RuleContext context;
|
public RuleContext context;
|
||||||
|
|
||||||
|
/** The stack acquired from parser invoking the ATN interpreter.
|
||||||
|
* Initially it's the rule stack upon entry into interp.adaptivePredict().
|
||||||
|
*
|
||||||
|
* We only pop from outerContext if we hit rule stop state
|
||||||
|
* of rule that initiates the adaptivePredict(). outerContext
|
||||||
|
* tracks as we go back up the entry context. If we do a call,
|
||||||
|
* we again push onto the regular context, rather than pushing onto
|
||||||
|
* outerContext.
|
||||||
|
*
|
||||||
|
* At an accept state, we ignore outerContext unless we had to use it
|
||||||
|
* during prediction. If outerContext != originalContext
|
||||||
|
* (stored in the ATN interpreter), then this config needed context
|
||||||
|
* to find another symbol.
|
||||||
|
*
|
||||||
|
* Lexer matches with ATN so there is no external context; this is null.
|
||||||
|
*/
|
||||||
|
//public RuleContext outerContext;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
Indicates that we have reached this ATN configuration after
|
Indicates that we have reached this ATN configuration after
|
||||||
traversing a predicate transition. This is important because we
|
traversing a predicate transition. This is important because we
|
||||||
|
@ -60,7 +79,7 @@ public class ATNConfig {
|
||||||
|
|
||||||
/**
|
/**
|
||||||
Indicates that we have reached this ATN configuration after
|
Indicates that we have reached this ATN configuration after
|
||||||
traversing a non-force action transition. We do not execute
|
traversing a non-forced action transition. We do not execute
|
||||||
predicates after such actions because the predicates could be
|
predicates after such actions because the predicates could be
|
||||||
functions of the side effects. Force actions must be either side
|
functions of the side effects. Force actions must be either side
|
||||||
effect free or automatically undone as the parse continues.
|
effect free or automatically undone as the parse continues.
|
||||||
|
@ -80,6 +99,7 @@ public class ATNConfig {
|
||||||
this.state = c.state;
|
this.state = c.state;
|
||||||
this.alt = c.alt;
|
this.alt = c.alt;
|
||||||
this.context = c.context;
|
this.context = c.context;
|
||||||
|
// this.outerContext = c.outerContext;
|
||||||
this.traversedPredicate = c.traversedPredicate;
|
this.traversedPredicate = c.traversedPredicate;
|
||||||
this.traversedAction = c.traversedAction;
|
this.traversedAction = c.traversedAction;
|
||||||
}
|
}
|
||||||
|
|
|
@ -50,6 +50,17 @@ public class ParserATNSimulator extends ATNSimulator {
|
||||||
public DFA[] decisionToDFA;
|
public DFA[] decisionToDFA;
|
||||||
protected boolean userWantsCtxSensitive = false;
|
protected boolean userWantsCtxSensitive = false;
|
||||||
|
|
||||||
|
/** This is the original context upon entry to the ATN simulator.
|
||||||
|
* ATNConfig objects carry on tracking the new context derived from
|
||||||
|
* the decision point. This field is used instead of passing the value
|
||||||
|
* around to the various functions, which would be confusing. Its
|
||||||
|
* value is reset upon prediction call to adaptivePredict() or the
|
||||||
|
* predictATN/DFA methods.
|
||||||
|
*
|
||||||
|
* The full stack at any moment is [config.outerContext + config.context].
|
||||||
|
*/
|
||||||
|
protected RuleContext originalContext;
|
||||||
|
|
||||||
protected Set<ATNConfig> closureBusy = new HashSet<ATNConfig>();
|
protected Set<ATNConfig> closureBusy = new HashSet<ATNConfig>();
|
||||||
|
|
||||||
public ParserATNSimulator(ATN atn) {
|
public ParserATNSimulator(ATN atn) {
|
||||||
|
@ -96,7 +107,7 @@ public class ParserATNSimulator extends ATNSimulator {
|
||||||
RuleContext ctx = RuleContext.EMPTY;
|
RuleContext ctx = RuleContext.EMPTY;
|
||||||
if ( useContext ) ctx = originalContext;
|
if ( useContext ) ctx = originalContext;
|
||||||
OrderedHashSet<ATNConfig> s0_closure =
|
OrderedHashSet<ATNConfig> s0_closure =
|
||||||
computeStartState(dfa.atnStartState, ctx, originalContext);
|
computeStartState(dfa.atnStartState, ctx);
|
||||||
dfa.s0 = addDFAState(dfa, s0_closure);
|
dfa.s0 = addDFAState(dfa, s0_closure);
|
||||||
if ( prevAccept!=null ) {
|
if ( prevAccept!=null ) {
|
||||||
dfa.s0.isAcceptState = true;
|
dfa.s0.isAcceptState = true;
|
||||||
|
@ -120,7 +131,7 @@ public class ParserATNSimulator extends ATNSimulator {
|
||||||
public int matchATN(TokenStream input, ATNState startState) {
|
public int matchATN(TokenStream input, ATNState startState) {
|
||||||
DFA dfa = new DFA(startState);
|
DFA dfa = new DFA(startState);
|
||||||
RuleContext ctx = new ParserRuleContext();
|
RuleContext ctx = new ParserRuleContext();
|
||||||
OrderedHashSet<ATNConfig> s0_closure = computeStartState(startState, ctx, RuleContext.EMPTY);
|
OrderedHashSet<ATNConfig> s0_closure = computeStartState(startState, ctx);
|
||||||
return execATN(input, dfa, input.index(), s0_closure, ctx, false);
|
return execATN(input, dfa, input.index(), s0_closure, ctx, false);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -244,7 +255,7 @@ public class ParserATNSimulator extends ATNSimulator {
|
||||||
Transition trans = c.state.transition(ti);
|
Transition trans = c.state.transition(ti);
|
||||||
ATNState target = getReachableTarget(trans, t);
|
ATNState target = getReachableTarget(trans, t);
|
||||||
if ( target!=null ) {
|
if ( target!=null ) {
|
||||||
closure(new ATNConfig(c, target), reach, originalContext);
|
closure(new ATNConfig(c, target), reach);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -368,9 +379,7 @@ public class ParserATNSimulator extends ATNSimulator {
|
||||||
return predictedAlt;
|
return predictedAlt;
|
||||||
}
|
}
|
||||||
|
|
||||||
public OrderedHashSet<ATNConfig> computeStartState(ATNState p, RuleContext ctx,
|
public OrderedHashSet<ATNConfig> computeStartState(ATNState p, RuleContext ctx) {
|
||||||
RuleContext originalContext)
|
|
||||||
{
|
|
||||||
RuleContext initialContext = ctx; // always at least the implicit call to start rule
|
RuleContext initialContext = ctx; // always at least the implicit call to start rule
|
||||||
OrderedHashSet<ATNConfig> configs = new OrderedHashSet<ATNConfig>();
|
OrderedHashSet<ATNConfig> configs = new OrderedHashSet<ATNConfig>();
|
||||||
prevAccept = null; // might reach end rule; track
|
prevAccept = null; // might reach end rule; track
|
||||||
|
@ -379,7 +388,7 @@ public class ParserATNSimulator extends ATNSimulator {
|
||||||
for (int i=0; i<p.getNumberOfTransitions(); i++) {
|
for (int i=0; i<p.getNumberOfTransitions(); i++) {
|
||||||
ATNState target = p.transition(i).target;
|
ATNState target = p.transition(i).target;
|
||||||
ATNConfig c = new ATNConfig(target, i+1, initialContext);
|
ATNConfig c = new ATNConfig(target, i+1, initialContext);
|
||||||
closure(c, configs, originalContext);
|
closure(c, configs);
|
||||||
}
|
}
|
||||||
|
|
||||||
return configs;
|
return configs;
|
||||||
|
@ -407,15 +416,13 @@ public class ParserATNSimulator extends ATNSimulator {
|
||||||
return null;
|
return null;
|
||||||
}
|
}
|
||||||
|
|
||||||
protected void closure(ATNConfig config, OrderedHashSet<ATNConfig> configs,
|
protected void closure(ATNConfig config, OrderedHashSet<ATNConfig> configs) {
|
||||||
RuleContext originalContext) {
|
|
||||||
closureBusy.clear();
|
closureBusy.clear();
|
||||||
closure(config, configs, originalContext, closureBusy);
|
closure(config, configs, closureBusy);
|
||||||
}
|
}
|
||||||
|
|
||||||
protected void closure(ATNConfig config,
|
protected void closure(ATNConfig config,
|
||||||
OrderedHashSet<ATNConfig> configs,
|
OrderedHashSet<ATNConfig> configs,
|
||||||
RuleContext originalContext,
|
|
||||||
Set<ATNConfig> closureBusy)
|
Set<ATNConfig> closureBusy)
|
||||||
{
|
{
|
||||||
if ( debug ) System.out.println("closure("+config+")");
|
if ( debug ) System.out.println("closure("+config+")");
|
||||||
|
@ -431,7 +438,7 @@ public class ParserATNSimulator extends ATNSimulator {
|
||||||
RuleTransition rt = (RuleTransition)invokingState.transition(0);
|
RuleTransition rt = (RuleTransition)invokingState.transition(0);
|
||||||
ATNState retState = rt.followState;
|
ATNState retState = rt.followState;
|
||||||
ATNConfig c = new ATNConfig(retState, config.alt, newContext);
|
ATNConfig c = new ATNConfig(retState, config.alt, newContext);
|
||||||
closure(c, configs, originalContext, closureBusy);
|
closure(c, configs, closureBusy);
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
// else if we have no context info, just chase follow links
|
// else if we have no context info, just chase follow links
|
||||||
|
@ -444,14 +451,12 @@ public class ParserATNSimulator extends ATNSimulator {
|
||||||
for (int i=0; i<p.getNumberOfTransitions(); i++) {
|
for (int i=0; i<p.getNumberOfTransitions(); i++) {
|
||||||
Transition t = p.transition(i);
|
Transition t = p.transition(i);
|
||||||
boolean evalPreds = !config.traversedAction;
|
boolean evalPreds = !config.traversedAction;
|
||||||
ATNConfig c = getEpsilonTarget(config, t, originalContext, evalPreds);
|
ATNConfig c = getEpsilonTarget(config, t, evalPreds);
|
||||||
if ( c!=null ) closure(c, configs, originalContext, closureBusy);
|
if ( c!=null ) closure(c, configs, closureBusy);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
public ATNConfig getEpsilonTarget(ATNConfig config, Transition t,
|
public ATNConfig getEpsilonTarget(ATNConfig config, Transition t, boolean evalPreds) {
|
||||||
RuleContext originalContext, boolean evalPreds)
|
|
||||||
{
|
|
||||||
ATNConfig c = null;
|
ATNConfig c = null;
|
||||||
if ( t instanceof RuleTransition ) {
|
if ( t instanceof RuleTransition ) {
|
||||||
ATNState p = config.state;
|
ATNState p = config.state;
|
||||||
|
|
|
@ -0,0 +1,751 @@
|
||||||
|
/*
|
||||||
|
[The "BSD license"]
|
||||||
|
Copyright (c) 2011 Terence Parr
|
||||||
|
All rights reserved.
|
||||||
|
|
||||||
|
Redistribution and use in source and binary forms, with or without
|
||||||
|
modification, are permitted provided that the following conditions
|
||||||
|
are met:
|
||||||
|
|
||||||
|
1. Redistributions of source code must retain the above copyright
|
||||||
|
notice, this list of conditions and the following disclaimer.
|
||||||
|
2. Redistributions in binary form must reproduce the above copyright
|
||||||
|
notice, this list of conditions and the following disclaimer in the
|
||||||
|
documentation and/or other materials provided with the distribution.
|
||||||
|
3. The name of the author may not be used to endorse or promote products
|
||||||
|
derived from this software without specific prior written permission.
|
||||||
|
|
||||||
|
THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
|
||||||
|
IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
|
||||||
|
OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
|
||||||
|
IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
|
||||||
|
INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
|
||||||
|
NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
||||||
|
DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
||||||
|
THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
||||||
|
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
|
||||||
|
THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||||
|
*/
|
||||||
|
|
||||||
|
/**********************************
|
||||||
|
* This version I was trying to keep an outer context as well as the
|
||||||
|
* context used during prediction, but it started getting complicated.
|
||||||
|
* For example, s0.configs start state configurations were computed based
|
||||||
|
* upon no an initial context. Later, if I need context I can't start
|
||||||
|
* from these computations. It has to be started over completely.
|
||||||
|
*
|
||||||
|
**********************************/
|
||||||
|
package org.antlr.v4.runtime.atn;
|
||||||
|
|
||||||
|
import org.antlr.v4.runtime.*;
|
||||||
|
import org.antlr.v4.runtime.dfa.*;
|
||||||
|
import org.antlr.v4.runtime.misc.OrderedHashSet;
|
||||||
|
import org.stringtemplate.v4.misc.MultiMap;
|
||||||
|
|
||||||
|
import java.util.*;
|
||||||
|
|
||||||
|
public class ParserATNSimulatorVariationInnerOuterContexts extends ATNSimulator {
|
||||||
|
public static boolean debug = true;
|
||||||
|
public static boolean dfa_debug = true;
|
||||||
|
|
||||||
|
public static int ATN_failover = 0;
|
||||||
|
public static int predict_calls = 0;
|
||||||
|
|
||||||
|
protected BaseRecognizer parser;
|
||||||
|
|
||||||
|
public Map<RuleContext, DFA[]> ctxToDFAs;
|
||||||
|
public Map<RuleContext, DFA>[] decisionToDFAPerCtx; // TODO: USE THIS ONE
|
||||||
|
public DFA[] decisionToDFA;
|
||||||
|
protected boolean userWantsCtxSensitive = false;
|
||||||
|
|
||||||
|
/** This is the original context upon entry to the ATN simulator.
|
||||||
|
* ATNConfig objects carry on tracking the new context derived from
|
||||||
|
* the decision point. This field is used instead of passing the value
|
||||||
|
* around to the various functions, which would be confusing. Its
|
||||||
|
* value is reset upon prediction call to adaptivePredict() or the
|
||||||
|
* predictATN/DFA methods.
|
||||||
|
*
|
||||||
|
* The full stack at any moment is [config.outerContext + config.context].
|
||||||
|
*/
|
||||||
|
protected RuleContext originalContext;
|
||||||
|
|
||||||
|
protected Set<ATNConfig> closureBusy = new HashSet<ATNConfig>();
|
||||||
|
|
||||||
|
public ParserATNSimulatorVariationInnerOuterContexts(ATN atn) {
|
||||||
|
super(atn);
|
||||||
|
ctxToDFAs = new HashMap<RuleContext, DFA[]>();
|
||||||
|
decisionToDFA = new DFA[atn.getNumberOfDecisions()];
|
||||||
|
}
|
||||||
|
|
||||||
|
public ParserATNSimulatorVariationInnerOuterContexts(BaseRecognizer parser, ATN atn) {
|
||||||
|
super(atn);
|
||||||
|
this.parser = parser;
|
||||||
|
ctxToDFAs = new HashMap<RuleContext, DFA[]>();
|
||||||
|
decisionToDFA = new DFA[atn.getNumberOfDecisions()+1];
|
||||||
|
// DOTGenerator dot = new DOTGenerator(null);
|
||||||
|
// System.out.println(dot.getDOT(atn.rules.get(0), parser.getRuleNames()));
|
||||||
|
// System.out.println(dot.getDOT(atn.rules.get(1), parser.getRuleNames()));
|
||||||
|
}
|
||||||
|
|
||||||
|
public int adaptivePredict(TokenStream input, int decision, RuleContext originalContext) {
|
||||||
|
predict_calls++;
|
||||||
|
DFA dfa = decisionToDFA[decision];
|
||||||
|
if ( dfa==null || dfa.s0==null ) {
|
||||||
|
ATNState startState = atn.decisionToState.get(decision);
|
||||||
|
decisionToDFA[decision] = dfa = new DFA(startState);
|
||||||
|
dfa.decision = decision;
|
||||||
|
return predictATN(dfa, input, decision, originalContext, false);
|
||||||
|
}
|
||||||
|
else {
|
||||||
|
//dump(dfa);
|
||||||
|
// start with the DFA
|
||||||
|
int m = input.mark();
|
||||||
|
int alt = execDFA(input, dfa, dfa.s0, originalContext);
|
||||||
|
input.seek(m);
|
||||||
|
return alt;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
public int predictATN(DFA dfa, TokenStream input,
|
||||||
|
int decision,
|
||||||
|
RuleContext originalContext,
|
||||||
|
boolean useContext)
|
||||||
|
{
|
||||||
|
if ( originalContext==null ) originalContext = RuleContext.EMPTY;
|
||||||
|
this.originalContext = originalContext;
|
||||||
|
RuleContext ctx = RuleContext.EMPTY;
|
||||||
|
if ( useContext ) ctx = originalContext;
|
||||||
|
OrderedHashSet<ATNConfig> s0_closure =
|
||||||
|
computeStartState(dfa.atnStartState, ctx, originalContext);
|
||||||
|
dfa.s0 = addDFAState(dfa, s0_closure);
|
||||||
|
if ( prevAccept!=null ) {
|
||||||
|
dfa.s0.isAcceptState = true;
|
||||||
|
dfa.s0.prediction = prevAccept.alt;
|
||||||
|
}
|
||||||
|
|
||||||
|
int alt = 0;
|
||||||
|
int m = input.mark();
|
||||||
|
try {
|
||||||
|
alt = execATN(input, dfa, m, s0_closure, originalContext, useContext);
|
||||||
|
}
|
||||||
|
catch (NoViableAltException nvae) { dumpDeadEndConfigs(nvae); throw nvae; }
|
||||||
|
finally {
|
||||||
|
input.seek(m);
|
||||||
|
}
|
||||||
|
if ( debug ) System.out.println("DFA after predictATN: "+dfa.toString());
|
||||||
|
return alt;
|
||||||
|
}
|
||||||
|
|
||||||
|
// doesn't create DFA when matching
|
||||||
|
public int matchATN(TokenStream input, ATNState startState) {
|
||||||
|
DFA dfa = new DFA(startState);
|
||||||
|
RuleContext ctx = new ParserRuleContext();
|
||||||
|
OrderedHashSet<ATNConfig> s0_closure = computeStartState(startState, ctx, RuleContext.EMPTY);
|
||||||
|
return execATN(input, dfa, input.index(), s0_closure, ctx, false);
|
||||||
|
}
|
||||||
|
|
||||||
|
public int execDFA(TokenStream input, DFA dfa, DFAState s0, RuleContext originalContext) {
|
||||||
|
if ( dfa_debug ) System.out.println("DFA decision "+dfa.decision+" exec LA(1)=="+input.LT(1));
|
||||||
|
// dump(dfa);
|
||||||
|
if ( originalContext==null ) originalContext = RuleContext.EMPTY;
|
||||||
|
this.originalContext = originalContext;
|
||||||
|
DFAState prevAcceptState = null;
|
||||||
|
DFAState s = s0;
|
||||||
|
int t = input.LA(1);
|
||||||
|
int start = input.index();
|
||||||
|
loop:
|
||||||
|
while ( true ) {
|
||||||
|
if ( dfa_debug ) System.out.println("DFA state "+s.stateNumber+" LA(1)=="+t);
|
||||||
|
// TODO: ctxSensitive
|
||||||
|
if ( s.isCtxSensitive ) {
|
||||||
|
Integer predI = s.ctxToPrediction.get(originalContext);
|
||||||
|
if ( dfa_debug ) System.out.println("ctx sensitive state "+originalContext+"->"+predI+
|
||||||
|
" in "+s);
|
||||||
|
if ( predI!=null ) return predI;
|
||||||
|
// System.out.println("start all over with ATN; can't use DFA");
|
||||||
|
// start all over with ATN; can't use DFA
|
||||||
|
input.seek(start);
|
||||||
|
DFA throwAwayDFA = new DFA(dfa.atnStartState);
|
||||||
|
int alt = predictATN(throwAwayDFA, input, dfa.decision, originalContext, false);
|
||||||
|
// int alt = execATN(input, throwAwayDFA, start, s0.configs, originalContext, false);
|
||||||
|
// we might get to a different input sequence by retrying with this new context
|
||||||
|
// so don't set the prediction for this accept state.
|
||||||
|
// wait: we are talking about the same input here. it can match more
|
||||||
|
// or less but not different input.Oh, that updates a throwaway DFA.
|
||||||
|
// It will not update our DFA
|
||||||
|
s.ctxToPrediction.put(originalContext, alt);
|
||||||
|
if ( dfa_debug ) System.out.println("after retry, ctx sensitive state mapping "+originalContext+"->"+alt);
|
||||||
|
return alt;
|
||||||
|
}
|
||||||
|
if ( s.isAcceptState ) {
|
||||||
|
if ( dfa_debug ) System.out.println("accept; predict "+s.prediction +" in state "+s.stateNumber);
|
||||||
|
prevAcceptState = s;
|
||||||
|
if ( s.ctxToPrediction!=null ) {
|
||||||
|
Integer predI = s.ctxToPrediction.get(originalContext);
|
||||||
|
System.out.println("ctx based prediction: "+predI);
|
||||||
|
}
|
||||||
|
|
||||||
|
// keep going unless we're at EOF or state only has one alt number
|
||||||
|
// mentioned in configs; check if something else could match
|
||||||
|
// TODO: for prediction, we must stop here I believe; there's no way to get
|
||||||
|
// past the stop state since we will never look for the longest match.
|
||||||
|
// during prediction, we always look for enough input to get a unique prediction.
|
||||||
|
// nothing beyond that will help
|
||||||
|
if ( s.complete || t==CharStream.EOF ) break;
|
||||||
|
}
|
||||||
|
// if no edge, pop over to ATN interpreter, update DFA and return
|
||||||
|
if ( s.edges == null || t >= s.edges.length || s.edges[t+1] == null ) {
|
||||||
|
if ( dfa_debug ) System.out.println("no edge for "+t);
|
||||||
|
int alt = -1;
|
||||||
|
if ( dfa_debug ) {
|
||||||
|
System.out.println("ATN exec upon "+
|
||||||
|
input.toString(start,input.index())+
|
||||||
|
" at DFA state "+s.stateNumber);
|
||||||
|
}
|
||||||
|
try {
|
||||||
|
alt = execATN(input, dfa, start, s.configs, originalContext, false);
|
||||||
|
// this adds edge even if next state is accept for
|
||||||
|
// same alt; e.g., s0-A->:s1=>2-B->:s2=>2
|
||||||
|
// TODO: This next stuff kills edge, but extra states remain. :(
|
||||||
|
if ( s.isAcceptState && alt!=-1 ) {
|
||||||
|
DFAState d = s.edges[input.LA(1)+1];
|
||||||
|
if ( d.isAcceptState && d.prediction==s.prediction ) {
|
||||||
|
// we can carve it out.
|
||||||
|
s.edges[input.LA(1)+1] = ERROR; // IGNORE really not error
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
catch (NoViableAltException nvae) {
|
||||||
|
alt = -1;
|
||||||
|
}
|
||||||
|
if ( dfa_debug ) {
|
||||||
|
System.out.println("back from DFA update, alt="+alt+", dfa=\n"+dfa);
|
||||||
|
//dump(dfa);
|
||||||
|
}
|
||||||
|
if ( alt==-1 ) {
|
||||||
|
addDFAEdge(s, t, ERROR);
|
||||||
|
break loop; // dead end; no where to go, fall back on prev if any
|
||||||
|
}
|
||||||
|
// action already executed
|
||||||
|
if ( dfa_debug ) System.out.println("DFA decision "+dfa.decision+
|
||||||
|
" predicts "+alt);
|
||||||
|
return alt; // we've updated DFA, exec'd action, and have our deepest answer
|
||||||
|
}
|
||||||
|
DFAState target = s.edges[t+1];
|
||||||
|
if ( target == ERROR ) break;
|
||||||
|
s = target;
|
||||||
|
input.consume();
|
||||||
|
t = input.LA(1);
|
||||||
|
}
|
||||||
|
if ( prevAcceptState==null ) {
|
||||||
|
System.out.println("!!! no viable alt in dfa");
|
||||||
|
return -1;
|
||||||
|
}
|
||||||
|
if ( dfa_debug ) System.out.println("DFA decision "+dfa.decision+
|
||||||
|
" predicts "+prevAcceptState.prediction);
|
||||||
|
return prevAcceptState.prediction;
|
||||||
|
}
|
||||||
|
|
||||||
|
public int execATN(TokenStream input,
|
||||||
|
DFA dfa,
|
||||||
|
int startIndex,
|
||||||
|
OrderedHashSet<ATNConfig> s0,
|
||||||
|
RuleContext originalContext,
|
||||||
|
boolean useContext)
|
||||||
|
{
|
||||||
|
if ( debug ) System.out.println("ATN decision "+dfa.decision+" exec LA(1)=="+input.LT(1));
|
||||||
|
ATN_failover++;
|
||||||
|
OrderedHashSet<ATNConfig> closure = new OrderedHashSet<ATNConfig>();
|
||||||
|
|
||||||
|
closure.addAll(s0);
|
||||||
|
|
||||||
|
if ( debug ) System.out.println("start state closure="+closure);
|
||||||
|
|
||||||
|
int t = input.LA(1);
|
||||||
|
if ( t==Token.EOF && prevAccept!=null ) {
|
||||||
|
// computeStartState must have reached end of rule
|
||||||
|
return prevAccept.alt;
|
||||||
|
}
|
||||||
|
|
||||||
|
prevAccept = null;
|
||||||
|
prevAcceptIndex = -1;
|
||||||
|
OrderedHashSet<ATNConfig> reach = new OrderedHashSet<ATNConfig>();
|
||||||
|
|
||||||
|
do { // while more work
|
||||||
|
if ( debug ) System.out.println("in reach starting closure: " + closure);
|
||||||
|
int ncl = closure.size();
|
||||||
|
for (int ci=0; ci<ncl; ci++) { // TODO: foreach
|
||||||
|
ATNConfig c = closure.get(ci);
|
||||||
|
if ( debug ) System.out.println("testing "+getTokenName(t)+" at "+c.toString());
|
||||||
|
int n = c.state.getNumberOfTransitions();
|
||||||
|
for (int ti=0; ti<n; ti++) { // for each transition
|
||||||
|
Transition trans = c.state.transition(ti);
|
||||||
|
ATNState target = getReachableTarget(trans, t);
|
||||||
|
if ( target!=null ) {
|
||||||
|
closure(new ATNConfig(c, target), reach);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// resolve ambig in DFAState for reach
|
||||||
|
Set<Integer> ambigAlts = getAmbiguousAlts(reach);
|
||||||
|
if ( ambigAlts!=null ) {
|
||||||
|
if ( debug ) {
|
||||||
|
ATNState loc = atn.states.get(originalContext.s);
|
||||||
|
String rname = "n/a";
|
||||||
|
if ( parser !=null ) rname = parser.getRuleNames()[loc.ruleIndex];
|
||||||
|
System.out.println("AMBIG in "+rname+" for alt "+ambigAlts+" upon "+
|
||||||
|
input.toString(startIndex, input.index()));
|
||||||
|
}
|
||||||
|
dfa.conflict = true; // at least one DFA state is ambiguous
|
||||||
|
if ( !userWantsCtxSensitive ) reportConflict(startIndex, input.index(), ambigAlts, reach);
|
||||||
|
|
||||||
|
// ATNState loc = atn.states.get(originalContext.s);
|
||||||
|
// String rname = recog.getRuleNames()[loc.ruleIndex];
|
||||||
|
// System.out.println("AMBIG orig="+originalContext.toString((BaseRecognizer)recog)+" for alt "+ambigAlts+" upon "+
|
||||||
|
// input.toString(startIndex, input.index()));
|
||||||
|
if ( !userWantsCtxSensitive || useContext ) {
|
||||||
|
resolveToMinAlt(reach, ambigAlts);
|
||||||
|
}
|
||||||
|
else {
|
||||||
|
return retryWithContext(input, dfa, startIndex, originalContext,
|
||||||
|
closure, t, reach, ambigAlts);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// if reach predicts single alt, can stop
|
||||||
|
|
||||||
|
int uniqueAlt = getUniqueAlt(reach);
|
||||||
|
if ( uniqueAlt!=ATN.INVALID_ALT_NUMBER ) {
|
||||||
|
if ( debug ) System.out.println("PREDICT alt "+uniqueAlt+
|
||||||
|
" decision "+dfa.decision+
|
||||||
|
" at index "+input.index());
|
||||||
|
addDFAEdge(dfa, closure, t, reach);
|
||||||
|
makeAcceptState(dfa, reach, uniqueAlt);
|
||||||
|
return uniqueAlt;
|
||||||
|
}
|
||||||
|
|
||||||
|
if ( reach.size()==0 ) {
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
|
||||||
|
// If we matched t anywhere, need to consume and add closer-t->reach DFA edge
|
||||||
|
// else error if no previous accept
|
||||||
|
input.consume();
|
||||||
|
addDFAEdge(dfa, closure, t, reach);
|
||||||
|
t = input.LA(1);
|
||||||
|
|
||||||
|
// swap to avoid reallocating space
|
||||||
|
OrderedHashSet<ATNConfig> tmp = reach;
|
||||||
|
reach = closure;
|
||||||
|
closure = tmp;
|
||||||
|
reach.clear(); // THIS MIGHT BE SLOW! kills each element; realloc might be faster
|
||||||
|
} while ( true );
|
||||||
|
|
||||||
|
if ( prevAccept==null ) {
|
||||||
|
System.out.println("no viable token at input "+input.LT(1)+", index "+input.index());
|
||||||
|
NoViableAltException nvae = new NoViableAltException(parser, input, closure, originalContext);
|
||||||
|
nvae.startIndex = startIndex;
|
||||||
|
throw nvae;
|
||||||
|
}
|
||||||
|
|
||||||
|
if ( debug ) System.out.println("PREDICT " + prevAccept + " index " + prevAccept.alt);
|
||||||
|
return prevAccept.alt;
|
||||||
|
}
|
||||||
|
|
||||||
|
protected int resolveToMinAlt(OrderedHashSet<ATNConfig> reach, Set<Integer> ambigAlts) {
|
||||||
|
int min = getMinAlt(ambigAlts);
|
||||||
|
// if predicting, create DFA accept state for resolved alt
|
||||||
|
ambigAlts.remove(min);
|
||||||
|
// kill dead alts so we don't chase them ever
|
||||||
|
killAlts(ambigAlts, reach);
|
||||||
|
if ( debug ) System.out.println("RESOLVED TO "+reach);
|
||||||
|
return min;
|
||||||
|
}
|
||||||
|
|
||||||
|
public int retryWithContext(TokenStream input,
|
||||||
|
DFA dfa,
|
||||||
|
int startIndex,
|
||||||
|
RuleContext originalContext,
|
||||||
|
OrderedHashSet<ATNConfig> closure,
|
||||||
|
int t,
|
||||||
|
OrderedHashSet<ATNConfig> reach,
|
||||||
|
Set<Integer> ambigAlts)
|
||||||
|
{
|
||||||
|
// ASSUMES PREDICT ONLY
|
||||||
|
// retry using context, if any; if none, kill all but min as before
|
||||||
|
if ( debug ) System.out.println("RETRY with ctx="+ originalContext);
|
||||||
|
int min = getMinAlt(ambigAlts);
|
||||||
|
if ( originalContext==RuleContext.EMPTY ) {
|
||||||
|
// no point in retrying with ctx since it's same.
|
||||||
|
// this implies that we have a true ambiguity
|
||||||
|
reportAmbiguity(startIndex, input.index(), ambigAlts, reach);
|
||||||
|
return min;
|
||||||
|
}
|
||||||
|
// otherwise we have to retry with context, filling in tmp DFA.
|
||||||
|
// if it comes back with conflict, we have a true ambiguity
|
||||||
|
input.seek(startIndex); // rewind
|
||||||
|
DFA ctx_dfa = new DFA(dfa.atnStartState);
|
||||||
|
int ctx_alt = predictATN(ctx_dfa, input, dfa.decision, originalContext, true);
|
||||||
|
if ( debug ) System.out.println("retry predicts "+ctx_alt+" vs "+getMinAlt(ambigAlts)+
|
||||||
|
" with conflict="+ctx_dfa.conflict+
|
||||||
|
" dfa="+ctx_dfa);
|
||||||
|
if ( ctx_dfa.conflict ) reportAmbiguity(startIndex, input.index(), ambigAlts, reach);
|
||||||
|
else reportContextSensitivity(startIndex, input.index(), ambigAlts, reach);
|
||||||
|
// it's not context-sensitive; true ambig. fall thru to strip dead alts
|
||||||
|
|
||||||
|
int predictedAlt = ctx_alt;
|
||||||
|
DFAState reachTarget = addDFAEdge(dfa, closure, t, reach);
|
||||||
|
reachTarget.isCtxSensitive = true;
|
||||||
|
if ( reachTarget.ctxToPrediction==null ) {
|
||||||
|
reachTarget.ctxToPrediction = new LinkedHashMap<RuleContext, Integer>();
|
||||||
|
}
|
||||||
|
reachTarget.ctxToPrediction.put(originalContext, predictedAlt);
|
||||||
|
// System.out.println("RESOLVE to "+predictedAlt);
|
||||||
|
//System.out.println(reachTarget.ctxToPrediction.size()+" size of ctx map");
|
||||||
|
return predictedAlt;
|
||||||
|
}
|
||||||
|
|
||||||
|
public OrderedHashSet<ATNConfig> computeStartState(ATNState p, RuleContext ctx,
|
||||||
|
RuleContext originalContext)
|
||||||
|
{
|
||||||
|
RuleContext initialContext = ctx; // always at least the implicit call to start rule
|
||||||
|
OrderedHashSet<ATNConfig> configs = new OrderedHashSet<ATNConfig>();
|
||||||
|
prevAccept = null; // might reach end rule; track
|
||||||
|
prevAcceptIndex = -1;
|
||||||
|
|
||||||
|
for (int i=0; i<p.getNumberOfTransitions(); i++) {
|
||||||
|
ATNState target = p.transition(i).target;
|
||||||
|
ATNConfig c = new ATNConfig(target, i+1, initialContext);
|
||||||
|
// c.outerContext = originalContext; COMMENTED OUT TO COMPILE
|
||||||
|
closure(c, configs);
|
||||||
|
}
|
||||||
|
|
||||||
|
return configs;
|
||||||
|
}
|
||||||
|
|
||||||
|
public ATNState getReachableTarget(Transition trans, int ttype) {
|
||||||
|
if ( trans instanceof AtomTransition ) {
|
||||||
|
AtomTransition at = (AtomTransition)trans;
|
||||||
|
// boolean not = trans instanceof NotAtomTransition;
|
||||||
|
if ( at.label == ttype ) {
|
||||||
|
return at.target;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
else if ( trans instanceof SetTransition ) {
|
||||||
|
SetTransition st = (SetTransition)trans;
|
||||||
|
boolean not = trans instanceof NotSetTransition;
|
||||||
|
if ( !not && st.set.member(ttype) || not && !st.set.member(ttype) ) {
|
||||||
|
return st.target;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// TODO else if ( trans instanceof WildcardTransition && t!=Token.EOF ) {
|
||||||
|
// ATNConfig targetConfig = new ATNConfig(c, trans.target);
|
||||||
|
// closure(input, targetConfig, reach);
|
||||||
|
// }
|
||||||
|
return null;
|
||||||
|
}
|
||||||
|
|
||||||
|
protected void closure(ATNConfig config, OrderedHashSet<ATNConfig> configs) {
|
||||||
|
closureBusy.clear();
|
||||||
|
closure(config, configs, closureBusy);
|
||||||
|
}
|
||||||
|
|
||||||
|
protected void closure(ATNConfig config,
|
||||||
|
OrderedHashSet<ATNConfig> configs,
|
||||||
|
Set<ATNConfig> closureBusy)
|
||||||
|
{
|
||||||
|
if ( debug ) System.out.println("closure("+config+")");
|
||||||
|
|
||||||
|
if ( closureBusy.contains(config) ) return; // avoid infinite recursion
|
||||||
|
closureBusy.add(config);
|
||||||
|
|
||||||
|
if ( config.state instanceof RuleStopState ) {
|
||||||
|
// if ( debug ) System.out.println("RuleStopState "+config+", originalContext: "+config.outerContext);
|
||||||
|
// We hit rule end. If we have context info, use it
|
||||||
|
if ( config.context!=null && !config.context.isEmpty() ) {
|
||||||
|
RuleContext newContext = config.context.parent; // "pop" invoking state
|
||||||
|
ATNState invokingState = atn.states.get(config.context.invokingState);
|
||||||
|
RuleTransition rt = (RuleTransition)invokingState.transition(0);
|
||||||
|
ATNState retState = rt.followState;
|
||||||
|
//ATNConfig c = new ATNConfig(retState, config.alt, newContext);
|
||||||
|
ATNConfig c = new ATNConfig(config, retState, newContext);
|
||||||
|
closure(c, configs, closureBusy);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
// else use original context info
|
||||||
|
/* COMMENTED OUT TO COMPILE
|
||||||
|
if ( !config.outerContext.isEmpty() ) {
|
||||||
|
System.out.println("context stack empty, using originalContext");
|
||||||
|
RuleContext newContext = config.outerContext.parent; // "pop" invoking state
|
||||||
|
ATNState invokingState = atn.states.get(config.outerContext.invokingState);
|
||||||
|
RuleTransition rt = (RuleTransition)invokingState.transition(0);
|
||||||
|
ATNState retState = rt.followState;
|
||||||
|
ATNConfig c = new ATNConfig(config, retState, newContext);
|
||||||
|
// pop one from originalContext too for this thread of computation
|
||||||
|
c.outerContext = newContext;
|
||||||
|
closure(c, configs, closureBusy);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
*/
|
||||||
|
// else if we have no context info, just chase follow links
|
||||||
|
// TODO: should be EOF now that we're using originalContext right?
|
||||||
|
}
|
||||||
|
|
||||||
|
ATNState p = config.state;
|
||||||
|
// optimization
|
||||||
|
if ( !p.onlyHasEpsilonTransitions() ) configs.add(config);
|
||||||
|
|
||||||
|
for (int i=0; i<p.getNumberOfTransitions(); i++) {
|
||||||
|
Transition t = p.transition(i);
|
||||||
|
boolean evalPreds = !config.traversedAction;
|
||||||
|
ATNConfig c = getEpsilonTarget(config, t, evalPreds);
|
||||||
|
if ( c!=null ) closure(c, configs, closureBusy);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
public ATNConfig getEpsilonTarget(ATNConfig config, Transition t, boolean evalPreds) {
|
||||||
|
ATNConfig c = null;
|
||||||
|
if ( t instanceof RuleTransition ) {
|
||||||
|
ATNState p = config.state;
|
||||||
|
RuleContext newContext;
|
||||||
|
if ( parser != null ) {
|
||||||
|
System.out.println("rule trans to rule "+parser.getRuleNames()[t.target.ruleIndex]);
|
||||||
|
newContext = parser.newContext(config.context, t.target.stateNumber, t.target.ruleIndex, -999);
|
||||||
|
newContext.invokingState = p.stateNumber;
|
||||||
|
System.out.println("new ctx type is "+newContext.getClass().getSimpleName());
|
||||||
|
}
|
||||||
|
else {
|
||||||
|
newContext = new RuleContext(config.context, p.stateNumber, t.target.stateNumber);
|
||||||
|
}
|
||||||
|
c = new ATNConfig(config, t.target, newContext);
|
||||||
|
}
|
||||||
|
else if ( t instanceof PredicateTransition ) {
|
||||||
|
PredicateTransition pt = (PredicateTransition)t;
|
||||||
|
if ( debug ) System.out.println("PRED (eval="+evalPreds+") "+pt.ruleIndex+":"+pt.predIndex);
|
||||||
|
if ( parser != null ) System.out.println("rule surrounding pred is "+
|
||||||
|
parser.getRuleNames()[pt.ruleIndex]);
|
||||||
|
// preds are epsilon if we're not doing preds.
|
||||||
|
// if we are doing preds, pred must eval to true
|
||||||
|
/* COMMENTED OUT TO COMPILE
|
||||||
|
if ( !evalPreds ||
|
||||||
|
(evalPreds && parser.sempred(config.outerContext, pt.ruleIndex, pt.predIndex)) ) {
|
||||||
|
c = new ATNConfig(config, t.target);
|
||||||
|
c.traversedPredicate = true;
|
||||||
|
}
|
||||||
|
*/
|
||||||
|
}
|
||||||
|
else if ( t instanceof ActionTransition ) {
|
||||||
|
c = new ATNConfig(config, t.target);
|
||||||
|
ActionTransition at = (ActionTransition)t;
|
||||||
|
if ( debug ) System.out.println("ACTION edge "+at.ruleIndex+":"+at.actionIndex);
|
||||||
|
if ( at.actionIndex>=0 ) {
|
||||||
|
if ( debug ) System.out.println("DO ACTION "+at.ruleIndex+":"+at.actionIndex);
|
||||||
|
// COMMENTED OUT TO COMPILE parser.action(config.outerContext, at.ruleIndex, at.actionIndex);
|
||||||
|
}
|
||||||
|
else {
|
||||||
|
// non-forced action traversed to get to t.target
|
||||||
|
if ( debug && !config.traversedAction ) {
|
||||||
|
System.out.println("NONFORCED; pruning future pred eval derived from s"+
|
||||||
|
config.state.stateNumber);
|
||||||
|
}
|
||||||
|
c.traversedAction = true;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
else if ( t.isEpsilon() ) {
|
||||||
|
c = new ATNConfig(config, t.target);
|
||||||
|
}
|
||||||
|
return c;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void reportConflict(int startIndex, int stopIndex, Set<Integer> alts, OrderedHashSet<ATNConfig> configs) {
|
||||||
|
if ( parser!=null ) parser.reportConflict(startIndex, stopIndex, alts, configs);
|
||||||
|
}
|
||||||
|
|
||||||
|
public void reportContextSensitivity(int startIndex, int stopIndex, Set<Integer> alts, OrderedHashSet<ATNConfig> configs) {
|
||||||
|
if ( parser!=null ) parser.reportContextSensitivity(startIndex, stopIndex, alts, configs);
|
||||||
|
}
|
||||||
|
|
||||||
|
/** If context sensitive parsing, we know it's ambiguity not conflict */
|
||||||
|
public void reportAmbiguity(int startIndex, int stopIndex, Set<Integer> alts, OrderedHashSet<ATNConfig> configs) {
|
||||||
|
if ( parser!=null ) parser.reportAmbiguity(startIndex, stopIndex, alts, configs);
|
||||||
|
}
|
||||||
|
|
||||||
|
public static int getUniqueAlt(Collection<ATNConfig> configs) {
|
||||||
|
int alt = ATN.INVALID_ALT_NUMBER;
|
||||||
|
for (ATNConfig c : configs) {
|
||||||
|
if ( alt == ATN.INVALID_ALT_NUMBER ) {
|
||||||
|
alt = c.alt; // found first alt
|
||||||
|
}
|
||||||
|
else if ( c.alt!=alt ) {
|
||||||
|
return ATN.INVALID_ALT_NUMBER;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return alt;
|
||||||
|
}
|
||||||
|
|
||||||
|
public Set<Integer> getAmbiguousAlts(OrderedHashSet<ATNConfig> configs) {
|
||||||
|
// System.err.println("check ambiguous "+configs);
|
||||||
|
Set<Integer> ambigAlts = null;
|
||||||
|
int numConfigs = configs.size();
|
||||||
|
// First get a list of configurations for each state.
|
||||||
|
// Most of the time, each state will have one associated configuration.
|
||||||
|
MultiMap<Integer, ATNConfig> stateToConfigListMap =
|
||||||
|
new MultiMap<Integer, ATNConfig>();
|
||||||
|
for (ATNConfig c : configs) {
|
||||||
|
stateToConfigListMap.map(c.state.stateNumber, c);
|
||||||
|
}
|
||||||
|
// potential conflicts are states with > 1 configuration and diff alts
|
||||||
|
for (List<ATNConfig> configsPerAlt : stateToConfigListMap.values()) {
|
||||||
|
ATNConfig goal = configsPerAlt.get(0);
|
||||||
|
int size = configsPerAlt.size();
|
||||||
|
for (int i=1; i< size; i++) {
|
||||||
|
ATNConfig c = configsPerAlt.get(i);
|
||||||
|
if ( c.alt!=goal.alt ) {
|
||||||
|
//System.out.println("chk stack "+goal+", "+c);
|
||||||
|
boolean sameCtx =
|
||||||
|
(goal.context==null&&c.context==null) ||
|
||||||
|
goal.context.equals(c.context) ||
|
||||||
|
c.context.conflictsWith(goal.context);
|
||||||
|
if ( sameCtx ) {
|
||||||
|
if ( debug ) {
|
||||||
|
System.out.println("we reach state "+c.state.stateNumber+
|
||||||
|
" in rule "+
|
||||||
|
(parser !=null ? parser.getRuleNames()[c.state.ruleIndex]:"n/a")+
|
||||||
|
" alts "+goal.alt+","+c.alt+" from ctx "+goal.context.toString((BaseRecognizer) parser)
|
||||||
|
+" and "+
|
||||||
|
c.context.toString((BaseRecognizer) parser));
|
||||||
|
}
|
||||||
|
if ( ambigAlts==null ) ambigAlts = new HashSet<Integer>();
|
||||||
|
ambigAlts.add(goal.alt);
|
||||||
|
ambigAlts.add(c.alt);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if ( ambigAlts!=null ) {
|
||||||
|
//System.err.println("ambig upon "+input.toString(startIndex, input.index()));
|
||||||
|
}
|
||||||
|
return ambigAlts;
|
||||||
|
}
|
||||||
|
|
||||||
|
public static int getMinAlt(Set<Integer> ambigAlts) {
|
||||||
|
int min = Integer.MAX_VALUE;
|
||||||
|
for (int alt : ambigAlts) {
|
||||||
|
if ( alt < min ) min = alt;
|
||||||
|
}
|
||||||
|
return min;
|
||||||
|
}
|
||||||
|
|
||||||
|
public static void killAlts(Set<Integer> alts, OrderedHashSet<ATNConfig> configs) {
|
||||||
|
int i = 0;
|
||||||
|
while ( i<configs.size() ) {
|
||||||
|
ATNConfig c = configs.get(i);
|
||||||
|
if ( alts.contains(c.alt) ) {
|
||||||
|
configs.remove(i);
|
||||||
|
}
|
||||||
|
else i++;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
protected DFAState addDFAEdge(DFA dfa,
|
||||||
|
OrderedHashSet<ATNConfig> p,
|
||||||
|
int t,
|
||||||
|
OrderedHashSet<ATNConfig> q)
|
||||||
|
{
|
||||||
|
System.out.println("MOVE "+p+" -> "+q+" upon "+getTokenName(t));
|
||||||
|
DFAState from = addDFAState(dfa, p);
|
||||||
|
DFAState to = addDFAState(dfa, q);
|
||||||
|
addDFAEdge(from, t, to);
|
||||||
|
return to;
|
||||||
|
}
|
||||||
|
|
||||||
|
protected void addDFAEdge(DFAState p, int t, DFAState q) {
|
||||||
|
if ( p==null ) return;
|
||||||
|
if ( p.edges==null ) {
|
||||||
|
p.edges = new DFAState[atn.maxTokenType+1+1]; // TODO: make adaptive
|
||||||
|
}
|
||||||
|
p.edges[t+1] = q; // connect
|
||||||
|
}
|
||||||
|
|
||||||
|
/** See comment on LexerInterpreter.addDFAState. */
|
||||||
|
protected DFAState addDFAState(DFA dfa, OrderedHashSet<ATNConfig> configs) {
|
||||||
|
DFAState proposed = new DFAState(configs);
|
||||||
|
DFAState existing = dfa.states.get(proposed);
|
||||||
|
if ( existing!=null ) return existing;
|
||||||
|
|
||||||
|
DFAState newState = proposed;
|
||||||
|
|
||||||
|
boolean traversedPredicate = false;
|
||||||
|
for (ATNConfig c : configs) {
|
||||||
|
if ( c.traversedPredicate ) {traversedPredicate = true; break;}
|
||||||
|
}
|
||||||
|
|
||||||
|
if ( traversedPredicate ) return null; // cannot cache
|
||||||
|
|
||||||
|
newState.stateNumber = dfa.states.size();
|
||||||
|
newState.configs = new OrderedHashSet<ATNConfig>();
|
||||||
|
newState.configs.addAll(configs);
|
||||||
|
dfa.states.put(newState, newState);
|
||||||
|
return newState;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void makeAcceptState(DFA dfa, OrderedHashSet<ATNConfig> reach, int uniqueAlt) {
|
||||||
|
DFAState accept = dfa.states.get(new DFAState(reach));
|
||||||
|
if ( accept==null ) return;
|
||||||
|
|
||||||
|
boolean usesOuterContext;
|
||||||
|
RuleContext outerContext = null;
|
||||||
|
for (ATNConfig c : accept.configs) {
|
||||||
|
/* COMMENTED OUT TO COMPILE
|
||||||
|
if ( c.outerContext!=originalContext ) {
|
||||||
|
System.out.println("### we used context "+
|
||||||
|
originalContext.toString(parser, c.outerContext)+" from "+
|
||||||
|
originalContext.toString(parser));
|
||||||
|
// TODO:Will they ever be different context sizes for the same except state?
|
||||||
|
// seems like they can only one context depth otherwise we'd
|
||||||
|
// be a different except states for different input
|
||||||
|
usesOuterContext = true;
|
||||||
|
outerContext = c.outerContext;
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
*/
|
||||||
|
}
|
||||||
|
if ( outerContext!=null ) {
|
||||||
|
// accept.setContextSensitivePrediction(originalContext, uniqueAlt);
|
||||||
|
}
|
||||||
|
else {
|
||||||
|
accept.isAcceptState = true;
|
||||||
|
accept.prediction = uniqueAlt;
|
||||||
|
accept.complete = true;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
public String getTokenName(int t) {
|
||||||
|
if ( t==-1 ) return "EOF";
|
||||||
|
if ( parser!=null && parser.getTokenNames()!=null ) return parser.getTokenNames()[t]+"<"+t+">";
|
||||||
|
return String.valueOf(t);
|
||||||
|
}
|
||||||
|
|
||||||
|
public void setContextSensitive(boolean ctxSensitive) {
|
||||||
|
this.userWantsCtxSensitive = ctxSensitive;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void dumpDeadEndConfigs(NoViableAltException nvae) {
|
||||||
|
System.err.println("dead end configs: ");
|
||||||
|
for (ATNConfig c : nvae.deadEndConfigs) {
|
||||||
|
Transition t = c.state.transition(0);
|
||||||
|
String trans = "";
|
||||||
|
if ( t instanceof AtomTransition) {
|
||||||
|
AtomTransition at = (AtomTransition)t;
|
||||||
|
trans = "Atom "+getTokenName(at.label);
|
||||||
|
}
|
||||||
|
else if ( t instanceof SetTransition ) {
|
||||||
|
SetTransition st = (SetTransition)t;
|
||||||
|
boolean not = st instanceof NotSetTransition;
|
||||||
|
trans = (not?"~":"")+"Set "+st.set.toString();
|
||||||
|
}
|
||||||
|
System.err.println(c.toString(parser, true)+":"+trans);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
|
@ -100,6 +100,16 @@ public class DFAState {
|
||||||
return alts;
|
return alts;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
public void setContextSensitivePrediction(RuleContext ctx, int predictedAlt) {
|
||||||
|
isCtxSensitive = true;
|
||||||
|
if ( ctxToPrediction==null ) {
|
||||||
|
ctxToPrediction = new LinkedHashMap<RuleContext, Integer>();
|
||||||
|
}
|
||||||
|
ctxToPrediction.put(ctx, predictedAlt);
|
||||||
|
}
|
||||||
|
*/
|
||||||
|
|
||||||
/** A decent hash for a DFA state is the sum of the ATN state/alt pairs. */
|
/** A decent hash for a DFA state is the sum of the ATN state/alt pairs. */
|
||||||
public int hashCode() {
|
public int hashCode() {
|
||||||
int h = 0;
|
int h = 0;
|
||||||
|
|
|
@ -2,7 +2,6 @@ import org.antlr.v4.Tool;
|
||||||
import org.antlr.v4.automata.ParserATNFactory;
|
import org.antlr.v4.automata.ParserATNFactory;
|
||||||
import org.antlr.v4.runtime.*;
|
import org.antlr.v4.runtime.*;
|
||||||
import org.antlr.v4.runtime.atn.ATN;
|
import org.antlr.v4.runtime.atn.ATN;
|
||||||
import org.antlr.v4.runtime.tree.Tree;
|
|
||||||
import org.antlr.v4.semantics.SemanticPipeline;
|
import org.antlr.v4.semantics.SemanticPipeline;
|
||||||
import org.antlr.v4.tool.*;
|
import org.antlr.v4.tool.*;
|
||||||
|
|
||||||
|
@ -14,7 +13,7 @@ public class TestT {
|
||||||
CommonTokenStream tokens = new CommonTokenStream(t);
|
CommonTokenStream tokens = new CommonTokenStream(t);
|
||||||
TParser p = new TParser(tokens);
|
TParser p = new TParser(tokens);
|
||||||
ParserRuleContext ret = p.s();
|
ParserRuleContext ret = p.s();
|
||||||
System.out.println(((Tree)ret.tree).toStringTree());
|
// System.out.println(((Tree)ret.tree).toStringTree());
|
||||||
}
|
}
|
||||||
|
|
||||||
public static void dump() throws Exception {
|
public static void dump() throws Exception {
|
||||||
|
|
Loading…
Reference in New Issue