mirror of
https://github.com/apache/datafusion-sqlparser-rs.git
synced 2025-08-04 06:18:17 +00:00
Allow calling prev_token()
after EOF
Before this `next_token()` would only increment the index when returning `Some(token)`. This means that the caller wishing to rewind must be careful not to call `prev_token()` on EOF (`None`), while not forgetting to call it for `Some`. Not doing this resulted in bugs in the undertested code that does error handling. After making this mistake several times, I'm changing `next_token()` / `prev_token()` so that calling `next_token(); prev_token(); next_token()` returns the same token in the first and the last invocation.
This commit is contained in:
parent
1227fddd48
commit
e02625719e
1 changed files with 37 additions and 39 deletions
|
@ -66,6 +66,7 @@ impl Error for ParserError {}
|
|||
/// SQL Parser
|
||||
pub struct Parser {
|
||||
tokens: Vec<Token>,
|
||||
/// The index of the first unprocessed token in `self.tokens`
|
||||
index: usize,
|
||||
}
|
||||
|
||||
|
@ -558,7 +559,8 @@ impl Parser {
|
|||
}
|
||||
}
|
||||
|
||||
/// Return first non-whitespace token that has not yet been processed
|
||||
/// Return the first non-whitespace token that has not yet been processed
|
||||
/// (or None if reached end-of-file)
|
||||
pub fn peek_token(&self) -> Option<Token> {
|
||||
self.peek_nth_token(0)
|
||||
}
|
||||
|
@ -567,53 +569,48 @@ impl Parser {
|
|||
pub fn peek_nth_token(&self, mut n: usize) -> Option<Token> {
|
||||
let mut index = self.index;
|
||||
loop {
|
||||
match self.tokens.get(index) {
|
||||
Some(Token::Whitespace(_)) => {
|
||||
index += 1;
|
||||
}
|
||||
Some(token) => {
|
||||
index += 1;
|
||||
match self.tokens.get(index - 1) {
|
||||
Some(Token::Whitespace(_)) => continue,
|
||||
non_whitespace => {
|
||||
if n == 0 {
|
||||
return Some(token.clone());
|
||||
return non_whitespace.cloned();
|
||||
}
|
||||
index += 1;
|
||||
n -= 1;
|
||||
}
|
||||
None => {
|
||||
return None;
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// Get the next token skipping whitespace and increment the token index
|
||||
/// Return the first non-whitespace token that has not yet been processed
|
||||
/// (or None if reached end-of-file) and mark it as processed. OK to call
|
||||
/// repeatedly after reaching EOF.
|
||||
pub fn next_token(&mut self) -> Option<Token> {
|
||||
loop {
|
||||
match self.next_token_no_skip() {
|
||||
self.index += 1;
|
||||
match self.tokens.get(self.index - 1) {
|
||||
Some(Token::Whitespace(_)) => continue,
|
||||
token => return token.cloned(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// Return the first unprocessed token, possibly whitespace.
|
||||
pub fn next_token_no_skip(&mut self) -> Option<&Token> {
|
||||
if self.index < self.tokens.len() {
|
||||
self.index += 1;
|
||||
Some(&self.tokens[self.index - 1])
|
||||
} else {
|
||||
None
|
||||
}
|
||||
self.index += 1;
|
||||
self.tokens.get(self.index - 1)
|
||||
}
|
||||
|
||||
/// Push back the last one non-whitespace token
|
||||
/// Push back the last one non-whitespace token. Must be called after
|
||||
/// `next_token()`, otherwise might panic. OK to call after
|
||||
/// `next_token()` indicates an EOF.
|
||||
pub fn prev_token(&mut self) {
|
||||
loop {
|
||||
assert!(self.index > 0);
|
||||
if self.index > 0 {
|
||||
self.index -= 1;
|
||||
if let Token::Whitespace(_) = &self.tokens[self.index] {
|
||||
continue;
|
||||
}
|
||||
};
|
||||
self.index -= 1;
|
||||
if let Some(Token::Whitespace(_)) = self.tokens.get(self.index) {
|
||||
continue;
|
||||
}
|
||||
return;
|
||||
}
|
||||
}
|
||||
|
@ -929,9 +926,7 @@ impl Parser {
|
|||
if name.is_some() {
|
||||
self.expected("PRIMARY, UNIQUE, FOREIGN, or CHECK", unexpected)
|
||||
} else {
|
||||
if unexpected.is_some() {
|
||||
self.prev_token();
|
||||
}
|
||||
self.prev_token();
|
||||
Ok(None)
|
||||
}
|
||||
}
|
||||
|
@ -1149,8 +1144,7 @@ impl Parser {
|
|||
reserved_kwds: &[&str],
|
||||
) -> Result<Option<SQLIdent>, ParserError> {
|
||||
let after_as = self.parse_keyword("AS");
|
||||
let maybe_alias = self.next_token();
|
||||
match maybe_alias {
|
||||
match self.next_token() {
|
||||
// Accept any identifier after `AS` (though many dialects have restrictions on
|
||||
// keywords that may appear here). If there's no `AS`: don't parse keywords,
|
||||
// which may start a construct allowed in this position, to be parsed as aliases.
|
||||
|
@ -1168,9 +1162,7 @@ impl Parser {
|
|||
if after_as {
|
||||
return self.expected("an identifier after AS", not_an_ident);
|
||||
}
|
||||
if not_an_ident.is_some() {
|
||||
self.prev_token();
|
||||
}
|
||||
self.prev_token();
|
||||
Ok(None) // no alias found
|
||||
}
|
||||
}
|
||||
|
@ -1192,9 +1184,7 @@ impl Parser {
|
|||
continue;
|
||||
}
|
||||
_ => {
|
||||
if token.is_some() {
|
||||
self.prev_token();
|
||||
}
|
||||
self.prev_token();
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
@ -1750,14 +1740,22 @@ mod tests {
|
|||
|
||||
#[test]
|
||||
fn test_prev_index() {
|
||||
let sql = "SELECT version()";
|
||||
let sql = "SELECT version";
|
||||
all_dialects().run_parser_method(sql, |parser| {
|
||||
assert_eq!(parser.peek_token(), Some(Token::make_keyword("SELECT")));
|
||||
assert_eq!(parser.next_token(), Some(Token::make_keyword("SELECT")));
|
||||
parser.prev_token();
|
||||
assert_eq!(parser.next_token(), Some(Token::make_keyword("SELECT")));
|
||||
assert_eq!(parser.next_token(), Some(Token::make_word("version", None)));
|
||||
parser.prev_token();
|
||||
assert_eq!(parser.peek_token(), Some(Token::make_word("version", None)));
|
||||
assert_eq!(parser.next_token(), Some(Token::make_word("version", None)));
|
||||
assert_eq!(parser.peek_token(), None);
|
||||
parser.prev_token();
|
||||
assert_eq!(parser.next_token(), Some(Token::make_word("version", None)));
|
||||
assert_eq!(parser.next_token(), None);
|
||||
assert_eq!(parser.next_token(), None);
|
||||
parser.prev_token();
|
||||
assert_eq!(parser.peek_token(), Some(Token::make_keyword("SELECT")));
|
||||
});
|
||||
}
|
||||
}
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue