mirror of
https://github.com/rust-lang/rust-analyzer.git
synced 2025-11-02 12:59:12 +00:00
parser tests work
This commit is contained in:
parent
26bfd6023f
commit
6ce587ba5a
8 changed files with 92 additions and 140 deletions
|
|
@ -24,40 +24,11 @@ mod tokens;
|
||||||
|
|
||||||
pub(crate) use token_set::TokenSet;
|
pub(crate) use token_set::TokenSet;
|
||||||
|
|
||||||
pub use syntax_kind::SyntaxKind;
|
pub use crate::{syntax_kind::SyntaxKind, tokens::Tokens};
|
||||||
|
|
||||||
use crate::tokens::Tokens;
|
|
||||||
|
|
||||||
#[derive(Debug, Clone, PartialEq, Eq, Hash)]
|
#[derive(Debug, Clone, PartialEq, Eq, Hash)]
|
||||||
pub struct ParseError(pub Box<String>);
|
pub struct ParseError(pub Box<String>);
|
||||||
|
|
||||||
/// `TokenSource` abstracts the source of the tokens parser operates on.
|
|
||||||
///
|
|
||||||
/// Hopefully this will allow us to treat text and token trees in the same way!
|
|
||||||
pub trait TokenSource {
|
|
||||||
fn current(&self) -> Token;
|
|
||||||
|
|
||||||
/// Lookahead n token
|
|
||||||
fn lookahead_nth(&self, n: usize) -> Token;
|
|
||||||
|
|
||||||
/// bump cursor to next token
|
|
||||||
fn bump(&mut self);
|
|
||||||
|
|
||||||
/// Is the current token a specified keyword?
|
|
||||||
fn is_keyword(&self, kw: &str) -> bool;
|
|
||||||
}
|
|
||||||
|
|
||||||
/// `Token` abstracts the cursor of `TokenSource` operates on.
|
|
||||||
#[derive(Debug, Copy, Clone, Eq, PartialEq)]
|
|
||||||
pub struct Token {
|
|
||||||
/// What is the current token?
|
|
||||||
pub kind: SyntaxKind,
|
|
||||||
|
|
||||||
/// Is the current token joined to the next one (`> >` vs `>>`).
|
|
||||||
pub is_jointed_to_next: bool,
|
|
||||||
pub contextual_kw: SyntaxKind,
|
|
||||||
}
|
|
||||||
|
|
||||||
/// `TreeSink` abstracts details of a particular syntax tree implementation.
|
/// `TreeSink` abstracts details of a particular syntax tree implementation.
|
||||||
pub trait TreeSink {
|
pub trait TreeSink {
|
||||||
/// Adds new token to the current branch.
|
/// Adds new token to the current branch.
|
||||||
|
|
|
||||||
|
|
@ -334,6 +334,18 @@ impl SyntaxKind {
|
||||||
};
|
};
|
||||||
Some(kw)
|
Some(kw)
|
||||||
}
|
}
|
||||||
|
pub fn from_contextual_keyword(ident: &str) -> Option<SyntaxKind> {
|
||||||
|
let kw = match ident {
|
||||||
|
"auto" => AUTO_KW,
|
||||||
|
"default" => DEFAULT_KW,
|
||||||
|
"existential" => EXISTENTIAL_KW,
|
||||||
|
"union" => UNION_KW,
|
||||||
|
"raw" => RAW_KW,
|
||||||
|
"macro_rules" => MACRO_RULES_KW,
|
||||||
|
_ => return None,
|
||||||
|
};
|
||||||
|
Some(kw)
|
||||||
|
}
|
||||||
pub fn from_char(c: char) -> Option<SyntaxKind> {
|
pub fn from_char(c: char) -> Option<SyntaxKind> {
|
||||||
let tok = match c {
|
let tok = match c {
|
||||||
';' => SEMICOLON,
|
';' => SEMICOLON,
|
||||||
|
|
|
||||||
|
|
@ -1,8 +1,19 @@
|
||||||
use crate::{SyntaxKind, Token};
|
use crate::SyntaxKind;
|
||||||
|
|
||||||
#[allow(non_camel_case_types)]
|
#[allow(non_camel_case_types)]
|
||||||
type bits = u64;
|
type bits = u64;
|
||||||
|
|
||||||
|
/// `Token` abstracts the cursor of `TokenSource` operates on.
|
||||||
|
#[derive(Debug, Copy, Clone, Eq, PartialEq)]
|
||||||
|
pub(crate) struct Token {
|
||||||
|
/// What is the current token?
|
||||||
|
pub(crate) kind: SyntaxKind,
|
||||||
|
|
||||||
|
/// Is the current token joined to the next one (`> >` vs `>>`).
|
||||||
|
pub(crate) is_jointed_to_next: bool,
|
||||||
|
pub(crate) contextual_kw: SyntaxKind,
|
||||||
|
}
|
||||||
|
|
||||||
/// Main input to the parser.
|
/// Main input to the parser.
|
||||||
///
|
///
|
||||||
/// A sequence of tokens represented internally as a struct of arrays.
|
/// A sequence of tokens represented internally as a struct of arrays.
|
||||||
|
|
@ -49,13 +60,14 @@ impl Tokens {
|
||||||
self.kind.len()
|
self.kind.len()
|
||||||
}
|
}
|
||||||
pub(crate) fn get(&self, idx: usize) -> Token {
|
pub(crate) fn get(&self, idx: usize) -> Token {
|
||||||
if idx > self.len() {
|
if idx < self.len() {
|
||||||
return self.eof();
|
let kind = self.kind[idx];
|
||||||
|
let is_jointed_to_next = self.get_joint(idx);
|
||||||
|
let contextual_kw = self.contextual_kw[idx];
|
||||||
|
Token { kind, is_jointed_to_next, contextual_kw }
|
||||||
|
} else {
|
||||||
|
self.eof()
|
||||||
}
|
}
|
||||||
let kind = self.kind[idx];
|
|
||||||
let is_jointed_to_next = self.get_joint(idx);
|
|
||||||
let contextual_kw = self.contextual_kw[idx];
|
|
||||||
Token { kind, is_jointed_to_next, contextual_kw }
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cold]
|
#[cold]
|
||||||
|
|
|
||||||
|
|
@ -2,12 +2,10 @@
|
||||||
//! incremental reparsing.
|
//! incremental reparsing.
|
||||||
|
|
||||||
pub(crate) mod lexer;
|
pub(crate) mod lexer;
|
||||||
mod text_token_source;
|
|
||||||
mod text_tree_sink;
|
mod text_tree_sink;
|
||||||
mod reparsing;
|
mod reparsing;
|
||||||
|
|
||||||
use parser::SyntaxKind;
|
use parser::SyntaxKind;
|
||||||
use text_token_source::TextTokenSource;
|
|
||||||
use text_tree_sink::TextTreeSink;
|
use text_tree_sink::TextTreeSink;
|
||||||
|
|
||||||
use crate::{syntax_node::GreenNode, AstNode, SyntaxError, SyntaxNode};
|
use crate::{syntax_node::GreenNode, AstNode, SyntaxError, SyntaxNode};
|
||||||
|
|
@ -15,12 +13,12 @@ use crate::{syntax_node::GreenNode, AstNode, SyntaxError, SyntaxNode};
|
||||||
pub(crate) use crate::parsing::{lexer::*, reparsing::incremental_reparse};
|
pub(crate) use crate::parsing::{lexer::*, reparsing::incremental_reparse};
|
||||||
|
|
||||||
pub(crate) fn parse_text(text: &str) -> (GreenNode, Vec<SyntaxError>) {
|
pub(crate) fn parse_text(text: &str) -> (GreenNode, Vec<SyntaxError>) {
|
||||||
let (tokens, lexer_errors) = tokenize(text);
|
let (lexer_tokens, lexer_errors) = tokenize(text);
|
||||||
|
let parser_tokens = to_parser_tokens(text, &lexer_tokens);
|
||||||
|
|
||||||
let mut token_source = TextTokenSource::new(text, &tokens);
|
let mut tree_sink = TextTreeSink::new(text, &lexer_tokens);
|
||||||
let mut tree_sink = TextTreeSink::new(text, &tokens);
|
|
||||||
|
|
||||||
parser::parse_source_file(&mut token_source, &mut tree_sink);
|
parser::parse_source_file(&parser_tokens, &mut tree_sink);
|
||||||
|
|
||||||
let (tree, mut parser_errors) = tree_sink.finish();
|
let (tree, mut parser_errors) = tree_sink.finish();
|
||||||
parser_errors.extend(lexer_errors);
|
parser_errors.extend(lexer_errors);
|
||||||
|
|
@ -33,26 +31,47 @@ pub(crate) fn parse_text_as<T: AstNode>(
|
||||||
text: &str,
|
text: &str,
|
||||||
entry_point: parser::ParserEntryPoint,
|
entry_point: parser::ParserEntryPoint,
|
||||||
) -> Result<T, ()> {
|
) -> Result<T, ()> {
|
||||||
let (tokens, lexer_errors) = tokenize(text);
|
let (lexer_tokens, lexer_errors) = tokenize(text);
|
||||||
if !lexer_errors.is_empty() {
|
if !lexer_errors.is_empty() {
|
||||||
return Err(());
|
return Err(());
|
||||||
}
|
}
|
||||||
|
|
||||||
let mut token_source = TextTokenSource::new(text, &tokens);
|
let parser_tokens = to_parser_tokens(text, &lexer_tokens);
|
||||||
let mut tree_sink = TextTreeSink::new(text, &tokens);
|
|
||||||
|
let mut tree_sink = TextTreeSink::new(text, &lexer_tokens);
|
||||||
|
|
||||||
// TextTreeSink assumes that there's at least some root node to which it can attach errors and
|
// TextTreeSink assumes that there's at least some root node to which it can attach errors and
|
||||||
// tokens. We arbitrarily give it a SourceFile.
|
// tokens. We arbitrarily give it a SourceFile.
|
||||||
use parser::TreeSink;
|
use parser::TreeSink;
|
||||||
tree_sink.start_node(SyntaxKind::SOURCE_FILE);
|
tree_sink.start_node(SyntaxKind::SOURCE_FILE);
|
||||||
parser::parse(&mut token_source, &mut tree_sink, entry_point);
|
parser::parse(&parser_tokens, &mut tree_sink, entry_point);
|
||||||
tree_sink.finish_node();
|
tree_sink.finish_node();
|
||||||
|
|
||||||
let (tree, parser_errors) = tree_sink.finish();
|
let (tree, parser_errors, eof) = tree_sink.finish_eof();
|
||||||
use parser::TokenSource;
|
if !parser_errors.is_empty() || !eof {
|
||||||
if !parser_errors.is_empty() || token_source.current().kind != SyntaxKind::EOF {
|
|
||||||
return Err(());
|
return Err(());
|
||||||
}
|
}
|
||||||
|
|
||||||
SyntaxNode::new_root(tree).first_child().and_then(T::cast).ok_or(())
|
SyntaxNode::new_root(tree).first_child().and_then(T::cast).ok_or(())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pub(crate) fn to_parser_tokens(text: &str, lexer_tokens: &[lexer::Token]) -> ::parser::Tokens {
|
||||||
|
let mut off = 0;
|
||||||
|
let mut res = parser::Tokens::default();
|
||||||
|
let mut was_joint = true;
|
||||||
|
for t in lexer_tokens {
|
||||||
|
if t.kind.is_trivia() {
|
||||||
|
was_joint = false;
|
||||||
|
} else if t.kind == SyntaxKind::IDENT {
|
||||||
|
let token_text = &text[off..][..usize::from(t.len)];
|
||||||
|
let contextual_kw =
|
||||||
|
SyntaxKind::from_contextual_keyword(token_text).unwrap_or(SyntaxKind::IDENT);
|
||||||
|
res.push_ident(contextual_kw);
|
||||||
|
} else {
|
||||||
|
res.push(was_joint, t.kind);
|
||||||
|
was_joint = true;
|
||||||
|
}
|
||||||
|
off += usize::from(t.len);
|
||||||
|
}
|
||||||
|
res
|
||||||
|
}
|
||||||
|
|
|
||||||
|
|
@ -12,8 +12,8 @@ use text_edit::Indel;
|
||||||
use crate::{
|
use crate::{
|
||||||
parsing::{
|
parsing::{
|
||||||
lexer::{lex_single_syntax_kind, tokenize, Token},
|
lexer::{lex_single_syntax_kind, tokenize, Token},
|
||||||
text_token_source::TextTokenSource,
|
|
||||||
text_tree_sink::TextTreeSink,
|
text_tree_sink::TextTreeSink,
|
||||||
|
to_parser_tokens,
|
||||||
},
|
},
|
||||||
syntax_node::{GreenNode, GreenToken, NodeOrToken, SyntaxElement, SyntaxNode},
|
syntax_node::{GreenNode, GreenToken, NodeOrToken, SyntaxElement, SyntaxNode},
|
||||||
SyntaxError,
|
SyntaxError,
|
||||||
|
|
@ -91,14 +91,14 @@ fn reparse_block(
|
||||||
let (node, reparser) = find_reparsable_node(root, edit.delete)?;
|
let (node, reparser) = find_reparsable_node(root, edit.delete)?;
|
||||||
let text = get_text_after_edit(node.clone().into(), edit);
|
let text = get_text_after_edit(node.clone().into(), edit);
|
||||||
|
|
||||||
let (tokens, new_lexer_errors) = tokenize(&text);
|
let (lexer_tokens, new_lexer_errors) = tokenize(&text);
|
||||||
if !is_balanced(&tokens) {
|
if !is_balanced(&lexer_tokens) {
|
||||||
return None;
|
return None;
|
||||||
}
|
}
|
||||||
|
let parser_tokens = to_parser_tokens(&text, &lexer_tokens);
|
||||||
|
|
||||||
let mut token_source = TextTokenSource::new(&text, &tokens);
|
let mut tree_sink = TextTreeSink::new(&text, &lexer_tokens);
|
||||||
let mut tree_sink = TextTreeSink::new(&text, &tokens);
|
reparser.parse(&parser_tokens, &mut tree_sink);
|
||||||
reparser.parse(&mut token_source, &mut tree_sink);
|
|
||||||
|
|
||||||
let (green, mut new_parser_errors) = tree_sink.finish();
|
let (green, mut new_parser_errors) = tree_sink.finish();
|
||||||
new_parser_errors.extend(new_lexer_errors);
|
new_parser_errors.extend(new_lexer_errors);
|
||||||
|
|
|
||||||
|
|
@ -1,82 +0,0 @@
|
||||||
//! See `TextTokenSource` docs.
|
|
||||||
|
|
||||||
use parser::TokenSource;
|
|
||||||
|
|
||||||
use crate::{parsing::lexer::Token, SyntaxKind::EOF, TextRange, TextSize};
|
|
||||||
|
|
||||||
/// Implementation of `parser::TokenSource` that takes tokens from source code text.
|
|
||||||
pub(crate) struct TextTokenSource<'t> {
|
|
||||||
text: &'t str,
|
|
||||||
/// token and its start position (non-whitespace/comment tokens)
|
|
||||||
/// ```non-rust
|
|
||||||
/// struct Foo;
|
|
||||||
/// ^------^--^-
|
|
||||||
/// | | \________
|
|
||||||
/// | \____ \
|
|
||||||
/// | \ |
|
|
||||||
/// (struct, 0) (Foo, 7) (;, 10)
|
|
||||||
/// ```
|
|
||||||
/// `[(struct, 0), (Foo, 7), (;, 10)]`
|
|
||||||
token_offset_pairs: Vec<(Token, TextSize)>,
|
|
||||||
|
|
||||||
/// Current token and position
|
|
||||||
curr: (parser::Token, usize),
|
|
||||||
}
|
|
||||||
|
|
||||||
impl<'t> TokenSource for TextTokenSource<'t> {
|
|
||||||
fn current(&self) -> parser::Token {
|
|
||||||
self.curr.0
|
|
||||||
}
|
|
||||||
|
|
||||||
fn lookahead_nth(&self, n: usize) -> parser::Token {
|
|
||||||
mk_token(self.curr.1 + n, &self.token_offset_pairs)
|
|
||||||
}
|
|
||||||
|
|
||||||
fn bump(&mut self) {
|
|
||||||
if self.curr.0.kind == EOF {
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
let pos = self.curr.1 + 1;
|
|
||||||
self.curr = (mk_token(pos, &self.token_offset_pairs), pos);
|
|
||||||
}
|
|
||||||
|
|
||||||
fn is_keyword(&self, kw: &str) -> bool {
|
|
||||||
self.token_offset_pairs
|
|
||||||
.get(self.curr.1)
|
|
||||||
.map_or(false, |(token, offset)| &self.text[TextRange::at(*offset, token.len)] == kw)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
fn mk_token(pos: usize, token_offset_pairs: &[(Token, TextSize)]) -> parser::Token {
|
|
||||||
let (kind, is_jointed_to_next) = match token_offset_pairs.get(pos) {
|
|
||||||
Some((token, offset)) => (
|
|
||||||
token.kind,
|
|
||||||
token_offset_pairs
|
|
||||||
.get(pos + 1)
|
|
||||||
.map_or(false, |(_, next_offset)| offset + token.len == *next_offset),
|
|
||||||
),
|
|
||||||
None => (EOF, false),
|
|
||||||
};
|
|
||||||
parser::Token { kind, is_jointed_to_next }
|
|
||||||
}
|
|
||||||
|
|
||||||
impl<'t> TextTokenSource<'t> {
|
|
||||||
/// Generate input from tokens(expect comment and whitespace).
|
|
||||||
pub(crate) fn new(text: &'t str, raw_tokens: &'t [Token]) -> TextTokenSource<'t> {
|
|
||||||
let token_offset_pairs: Vec<_> = raw_tokens
|
|
||||||
.iter()
|
|
||||||
.filter_map({
|
|
||||||
let mut len = 0.into();
|
|
||||||
move |token| {
|
|
||||||
let pair = if token.kind.is_trivia() { None } else { Some((*token, len)) };
|
|
||||||
len += token.len;
|
|
||||||
pair
|
|
||||||
}
|
|
||||||
})
|
|
||||||
.collect();
|
|
||||||
|
|
||||||
let first = mk_token(0, &token_offset_pairs);
|
|
||||||
TextTokenSource { text, token_offset_pairs, curr: (first, 0) }
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
@ -104,7 +104,7 @@ impl<'a> TextTreeSink<'a> {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub(super) fn finish(mut self) -> (GreenNode, Vec<SyntaxError>) {
|
pub(super) fn finish_eof(mut self) -> (GreenNode, Vec<SyntaxError>, bool) {
|
||||||
match mem::replace(&mut self.state, State::Normal) {
|
match mem::replace(&mut self.state, State::Normal) {
|
||||||
State::PendingFinish => {
|
State::PendingFinish => {
|
||||||
self.eat_trivias();
|
self.eat_trivias();
|
||||||
|
|
@ -113,7 +113,15 @@ impl<'a> TextTreeSink<'a> {
|
||||||
State::PendingStart | State::Normal => unreachable!(),
|
State::PendingStart | State::Normal => unreachable!(),
|
||||||
}
|
}
|
||||||
|
|
||||||
self.inner.finish_raw()
|
let (node, errors) = self.inner.finish_raw();
|
||||||
|
let is_eof = self.token_pos == self.tokens.len();
|
||||||
|
|
||||||
|
(node, errors, is_eof)
|
||||||
|
}
|
||||||
|
|
||||||
|
pub(super) fn finish(self) -> (GreenNode, Vec<SyntaxError>) {
|
||||||
|
let (node, errors, _eof) = self.finish_eof();
|
||||||
|
(node, errors)
|
||||||
}
|
}
|
||||||
|
|
||||||
fn eat_trivias(&mut self) {
|
fn eat_trivias(&mut self) {
|
||||||
|
|
|
||||||
|
|
@ -359,6 +359,10 @@ fn generate_syntax_kinds(grammar: KindsSrc<'_>) -> String {
|
||||||
let full_keywords =
|
let full_keywords =
|
||||||
full_keywords_values.iter().map(|kw| format_ident!("{}_KW", to_upper_snake_case(kw)));
|
full_keywords_values.iter().map(|kw| format_ident!("{}_KW", to_upper_snake_case(kw)));
|
||||||
|
|
||||||
|
let contextual_keywords_values = &grammar.contextual_keywords;
|
||||||
|
let contextual_keywords =
|
||||||
|
contextual_keywords_values.iter().map(|kw| format_ident!("{}_KW", to_upper_snake_case(kw)));
|
||||||
|
|
||||||
let all_keywords_values =
|
let all_keywords_values =
|
||||||
grammar.keywords.iter().chain(grammar.contextual_keywords.iter()).collect::<Vec<_>>();
|
grammar.keywords.iter().chain(grammar.contextual_keywords.iter()).collect::<Vec<_>>();
|
||||||
let all_keywords_idents = all_keywords_values.iter().map(|kw| format_ident!("{}", kw));
|
let all_keywords_idents = all_keywords_values.iter().map(|kw| format_ident!("{}", kw));
|
||||||
|
|
@ -428,6 +432,14 @@ fn generate_syntax_kinds(grammar: KindsSrc<'_>) -> String {
|
||||||
Some(kw)
|
Some(kw)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pub fn from_contextual_keyword(ident: &str) -> Option<SyntaxKind> {
|
||||||
|
let kw = match ident {
|
||||||
|
#(#contextual_keywords_values => #contextual_keywords,)*
|
||||||
|
_ => return None,
|
||||||
|
};
|
||||||
|
Some(kw)
|
||||||
|
}
|
||||||
|
|
||||||
pub fn from_char(c: char) -> Option<SyntaxKind> {
|
pub fn from_char(c: char) -> Option<SyntaxKind> {
|
||||||
let tok = match c {
|
let tok = match c {
|
||||||
#(#single_byte_tokens_values => #single_byte_tokens,)*
|
#(#single_byte_tokens_values => #single_byte_tokens,)*
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue