mirror of
https://github.com/rust-lang/rust-analyzer.git
synced 2025-08-22 11:24:24 +00:00
parser tests work
This commit is contained in:
parent
26bfd6023f
commit
6ce587ba5a
8 changed files with 92 additions and 140 deletions
|
@ -12,8 +12,8 @@ use text_edit::Indel;
|
|||
use crate::{
|
||||
parsing::{
|
||||
lexer::{lex_single_syntax_kind, tokenize, Token},
|
||||
text_token_source::TextTokenSource,
|
||||
text_tree_sink::TextTreeSink,
|
||||
to_parser_tokens,
|
||||
},
|
||||
syntax_node::{GreenNode, GreenToken, NodeOrToken, SyntaxElement, SyntaxNode},
|
||||
SyntaxError,
|
||||
|
@ -91,14 +91,14 @@ fn reparse_block(
|
|||
let (node, reparser) = find_reparsable_node(root, edit.delete)?;
|
||||
let text = get_text_after_edit(node.clone().into(), edit);
|
||||
|
||||
let (tokens, new_lexer_errors) = tokenize(&text);
|
||||
if !is_balanced(&tokens) {
|
||||
let (lexer_tokens, new_lexer_errors) = tokenize(&text);
|
||||
if !is_balanced(&lexer_tokens) {
|
||||
return None;
|
||||
}
|
||||
let parser_tokens = to_parser_tokens(&text, &lexer_tokens);
|
||||
|
||||
let mut token_source = TextTokenSource::new(&text, &tokens);
|
||||
let mut tree_sink = TextTreeSink::new(&text, &tokens);
|
||||
reparser.parse(&mut token_source, &mut tree_sink);
|
||||
let mut tree_sink = TextTreeSink::new(&text, &lexer_tokens);
|
||||
reparser.parse(&parser_tokens, &mut tree_sink);
|
||||
|
||||
let (green, mut new_parser_errors) = tree_sink.finish();
|
||||
new_parser_errors.extend(new_lexer_errors);
|
||||
|
|
|
@ -1,82 +0,0 @@
|
|||
//! See `TextTokenSource` docs.
|
||||
|
||||
use parser::TokenSource;
|
||||
|
||||
use crate::{parsing::lexer::Token, SyntaxKind::EOF, TextRange, TextSize};
|
||||
|
||||
/// Implementation of `parser::TokenSource` that takes tokens from source code text.
|
||||
pub(crate) struct TextTokenSource<'t> {
|
||||
text: &'t str,
|
||||
/// token and its start position (non-whitespace/comment tokens)
|
||||
/// ```non-rust
|
||||
/// struct Foo;
|
||||
/// ^------^--^-
|
||||
/// | | \________
|
||||
/// | \____ \
|
||||
/// | \ |
|
||||
/// (struct, 0) (Foo, 7) (;, 10)
|
||||
/// ```
|
||||
/// `[(struct, 0), (Foo, 7), (;, 10)]`
|
||||
token_offset_pairs: Vec<(Token, TextSize)>,
|
||||
|
||||
/// Current token and position
|
||||
curr: (parser::Token, usize),
|
||||
}
|
||||
|
||||
impl<'t> TokenSource for TextTokenSource<'t> {
|
||||
fn current(&self) -> parser::Token {
|
||||
self.curr.0
|
||||
}
|
||||
|
||||
fn lookahead_nth(&self, n: usize) -> parser::Token {
|
||||
mk_token(self.curr.1 + n, &self.token_offset_pairs)
|
||||
}
|
||||
|
||||
fn bump(&mut self) {
|
||||
if self.curr.0.kind == EOF {
|
||||
return;
|
||||
}
|
||||
|
||||
let pos = self.curr.1 + 1;
|
||||
self.curr = (mk_token(pos, &self.token_offset_pairs), pos);
|
||||
}
|
||||
|
||||
fn is_keyword(&self, kw: &str) -> bool {
|
||||
self.token_offset_pairs
|
||||
.get(self.curr.1)
|
||||
.map_or(false, |(token, offset)| &self.text[TextRange::at(*offset, token.len)] == kw)
|
||||
}
|
||||
}
|
||||
|
||||
fn mk_token(pos: usize, token_offset_pairs: &[(Token, TextSize)]) -> parser::Token {
|
||||
let (kind, is_jointed_to_next) = match token_offset_pairs.get(pos) {
|
||||
Some((token, offset)) => (
|
||||
token.kind,
|
||||
token_offset_pairs
|
||||
.get(pos + 1)
|
||||
.map_or(false, |(_, next_offset)| offset + token.len == *next_offset),
|
||||
),
|
||||
None => (EOF, false),
|
||||
};
|
||||
parser::Token { kind, is_jointed_to_next }
|
||||
}
|
||||
|
||||
impl<'t> TextTokenSource<'t> {
|
||||
/// Generate input from tokens(expect comment and whitespace).
|
||||
pub(crate) fn new(text: &'t str, raw_tokens: &'t [Token]) -> TextTokenSource<'t> {
|
||||
let token_offset_pairs: Vec<_> = raw_tokens
|
||||
.iter()
|
||||
.filter_map({
|
||||
let mut len = 0.into();
|
||||
move |token| {
|
||||
let pair = if token.kind.is_trivia() { None } else { Some((*token, len)) };
|
||||
len += token.len;
|
||||
pair
|
||||
}
|
||||
})
|
||||
.collect();
|
||||
|
||||
let first = mk_token(0, &token_offset_pairs);
|
||||
TextTokenSource { text, token_offset_pairs, curr: (first, 0) }
|
||||
}
|
||||
}
|
|
@ -104,7 +104,7 @@ impl<'a> TextTreeSink<'a> {
|
|||
}
|
||||
}
|
||||
|
||||
pub(super) fn finish(mut self) -> (GreenNode, Vec<SyntaxError>) {
|
||||
pub(super) fn finish_eof(mut self) -> (GreenNode, Vec<SyntaxError>, bool) {
|
||||
match mem::replace(&mut self.state, State::Normal) {
|
||||
State::PendingFinish => {
|
||||
self.eat_trivias();
|
||||
|
@ -113,7 +113,15 @@ impl<'a> TextTreeSink<'a> {
|
|||
State::PendingStart | State::Normal => unreachable!(),
|
||||
}
|
||||
|
||||
self.inner.finish_raw()
|
||||
let (node, errors) = self.inner.finish_raw();
|
||||
let is_eof = self.token_pos == self.tokens.len();
|
||||
|
||||
(node, errors, is_eof)
|
||||
}
|
||||
|
||||
pub(super) fn finish(self) -> (GreenNode, Vec<SyntaxError>) {
|
||||
let (node, errors, _eof) = self.finish_eof();
|
||||
(node, errors)
|
||||
}
|
||||
|
||||
fn eat_trivias(&mut self) {
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue