mirror of
https://github.com/rust-lang/rust-analyzer.git
synced 2025-09-27 04:19:13 +00:00
Derive kinds information from ungrammar file
This commit is contained in:
parent
8f044d9681
commit
983c9c122e
12 changed files with 448 additions and 718 deletions
|
@ -14,6 +14,8 @@ pub struct Abi {
|
|||
impl Abi {
|
||||
#[inline]
|
||||
pub fn extern_token(&self) -> Option<SyntaxToken> { support::token(&self.syntax, T![extern]) }
|
||||
#[inline]
|
||||
pub fn string_token(&self) -> Option<SyntaxToken> { support::token(&self.syntax, T![string]) }
|
||||
}
|
||||
|
||||
#[derive(Debug, Clone, PartialEq, Eq, Hash)]
|
||||
|
@ -833,27 +835,6 @@ impl MacroDef {
|
|||
pub fn macro_token(&self) -> Option<SyntaxToken> { support::token(&self.syntax, T![macro]) }
|
||||
}
|
||||
|
||||
#[derive(Debug, Clone, PartialEq, Eq, Hash)]
|
||||
pub struct MacroEagerInput {
|
||||
pub(crate) syntax: SyntaxNode,
|
||||
}
|
||||
impl MacroEagerInput {
|
||||
#[inline]
|
||||
pub fn exprs(&self) -> AstChildren<Expr> { support::children(&self.syntax) }
|
||||
#[inline]
|
||||
pub fn l_paren_token(&self) -> Option<SyntaxToken> { support::token(&self.syntax, T!['(']) }
|
||||
#[inline]
|
||||
pub fn r_paren_token(&self) -> Option<SyntaxToken> { support::token(&self.syntax, T![')']) }
|
||||
#[inline]
|
||||
pub fn l_brack_token(&self) -> Option<SyntaxToken> { support::token(&self.syntax, T!['[']) }
|
||||
#[inline]
|
||||
pub fn r_brack_token(&self) -> Option<SyntaxToken> { support::token(&self.syntax, T![']']) }
|
||||
#[inline]
|
||||
pub fn l_curly_token(&self) -> Option<SyntaxToken> { support::token(&self.syntax, T!['{']) }
|
||||
#[inline]
|
||||
pub fn r_curly_token(&self) -> Option<SyntaxToken> { support::token(&self.syntax, T!['}']) }
|
||||
}
|
||||
|
||||
#[derive(Debug, Clone, PartialEq, Eq, Hash)]
|
||||
pub struct MacroExpr {
|
||||
pub(crate) syntax: SyntaxNode,
|
||||
|
@ -1050,6 +1031,10 @@ impl NameRef {
|
|||
#[inline]
|
||||
pub fn ident_token(&self) -> Option<SyntaxToken> { support::token(&self.syntax, T![ident]) }
|
||||
#[inline]
|
||||
pub fn int_number_token(&self) -> Option<SyntaxToken> {
|
||||
support::token(&self.syntax, T![int_number])
|
||||
}
|
||||
#[inline]
|
||||
pub fn self_token(&self) -> Option<SyntaxToken> { support::token(&self.syntax, T![self]) }
|
||||
#[inline]
|
||||
pub fn super_token(&self) -> Option<SyntaxToken> { support::token(&self.syntax, T![super]) }
|
||||
|
@ -3021,20 +3006,6 @@ impl AstNode for MacroDef {
|
|||
#[inline]
|
||||
fn syntax(&self) -> &SyntaxNode { &self.syntax }
|
||||
}
|
||||
impl AstNode for MacroEagerInput {
|
||||
#[inline]
|
||||
fn can_cast(kind: SyntaxKind) -> bool { kind == MACRO_EAGER_INPUT }
|
||||
#[inline]
|
||||
fn cast(syntax: SyntaxNode) -> Option<Self> {
|
||||
if Self::can_cast(syntax.kind()) {
|
||||
Some(Self { syntax })
|
||||
} else {
|
||||
None
|
||||
}
|
||||
}
|
||||
#[inline]
|
||||
fn syntax(&self) -> &SyntaxNode { &self.syntax }
|
||||
}
|
||||
impl AstNode for MacroExpr {
|
||||
#[inline]
|
||||
fn can_cast(kind: SyntaxKind) -> bool { kind == MACRO_EXPR }
|
||||
|
@ -5741,11 +5712,6 @@ impl std::fmt::Display for MacroDef {
|
|||
std::fmt::Display::fmt(self.syntax(), f)
|
||||
}
|
||||
}
|
||||
impl std::fmt::Display for MacroEagerInput {
|
||||
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
||||
std::fmt::Display::fmt(self.syntax(), f)
|
||||
}
|
||||
}
|
||||
impl std::fmt::Display for MacroExpr {
|
||||
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
||||
std::fmt::Display::fmt(self.syntax(), f)
|
||||
|
|
|
@ -211,115 +211,6 @@ impl SourceFile {
|
|||
}
|
||||
}
|
||||
|
||||
impl ast::TokenTree {
|
||||
pub fn reparse_as_comma_separated_expr(
|
||||
self,
|
||||
edition: parser::Edition,
|
||||
) -> Parse<ast::MacroEagerInput> {
|
||||
let tokens = self.syntax().descendants_with_tokens().filter_map(NodeOrToken::into_token);
|
||||
|
||||
let mut parser_input = parser::Input::default();
|
||||
let mut was_joint = false;
|
||||
for t in tokens {
|
||||
let kind = t.kind();
|
||||
if kind.is_trivia() {
|
||||
was_joint = false
|
||||
} else if kind == SyntaxKind::IDENT {
|
||||
let token_text = t.text();
|
||||
let contextual_kw =
|
||||
SyntaxKind::from_contextual_keyword(token_text).unwrap_or(SyntaxKind::IDENT);
|
||||
parser_input.push_ident(contextual_kw);
|
||||
} else {
|
||||
if was_joint {
|
||||
parser_input.was_joint();
|
||||
}
|
||||
parser_input.push(kind);
|
||||
// Tag the token as joint if it is float with a fractional part
|
||||
// we use this jointness to inform the parser about what token split
|
||||
// event to emit when we encounter a float literal in a field access
|
||||
if kind == SyntaxKind::FLOAT_NUMBER {
|
||||
if !t.text().ends_with('.') {
|
||||
parser_input.was_joint();
|
||||
} else {
|
||||
was_joint = false;
|
||||
}
|
||||
} else {
|
||||
was_joint = true;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
let parser_output = parser::TopEntryPoint::MacroEagerInput.parse(&parser_input, edition);
|
||||
|
||||
let mut tokens =
|
||||
self.syntax().descendants_with_tokens().filter_map(NodeOrToken::into_token);
|
||||
let mut text = String::new();
|
||||
let mut pos = TextSize::from(0);
|
||||
let mut builder = SyntaxTreeBuilder::default();
|
||||
for event in parser_output.iter() {
|
||||
match event {
|
||||
parser::Step::Token { kind, n_input_tokens } => {
|
||||
let mut token = tokens.next().unwrap();
|
||||
while token.kind().is_trivia() {
|
||||
let text = token.text();
|
||||
pos += TextSize::from(text.len() as u32);
|
||||
builder.token(token.kind(), text);
|
||||
|
||||
token = tokens.next().unwrap();
|
||||
}
|
||||
text.push_str(token.text());
|
||||
for _ in 1..n_input_tokens {
|
||||
let token = tokens.next().unwrap();
|
||||
text.push_str(token.text());
|
||||
}
|
||||
|
||||
pos += TextSize::from(text.len() as u32);
|
||||
builder.token(kind, &text);
|
||||
text.clear();
|
||||
}
|
||||
parser::Step::FloatSplit { ends_in_dot: has_pseudo_dot } => {
|
||||
let token = tokens.next().unwrap();
|
||||
let text = token.text();
|
||||
|
||||
match text.split_once('.') {
|
||||
Some((left, right)) => {
|
||||
assert!(!left.is_empty());
|
||||
builder.start_node(SyntaxKind::NAME_REF);
|
||||
builder.token(SyntaxKind::INT_NUMBER, left);
|
||||
builder.finish_node();
|
||||
|
||||
// here we move the exit up, the original exit has been deleted in process
|
||||
builder.finish_node();
|
||||
|
||||
builder.token(SyntaxKind::DOT, ".");
|
||||
|
||||
if has_pseudo_dot {
|
||||
assert!(right.is_empty(), "{left}.{right}");
|
||||
} else {
|
||||
assert!(!right.is_empty(), "{left}.{right}");
|
||||
builder.start_node(SyntaxKind::NAME_REF);
|
||||
builder.token(SyntaxKind::INT_NUMBER, right);
|
||||
builder.finish_node();
|
||||
|
||||
// the parser creates an unbalanced start node, we are required to close it here
|
||||
builder.finish_node();
|
||||
}
|
||||
}
|
||||
None => unreachable!(),
|
||||
}
|
||||
pos += TextSize::from(text.len() as u32);
|
||||
}
|
||||
parser::Step::Enter { kind } => builder.start_node(kind),
|
||||
parser::Step::Exit => builder.finish_node(),
|
||||
parser::Step::Error { msg } => builder.error(msg.to_owned(), pos),
|
||||
}
|
||||
}
|
||||
|
||||
let (green, errors) = builder.finish_raw();
|
||||
Parse::new(green, errors)
|
||||
}
|
||||
}
|
||||
|
||||
/// Matches a `SyntaxNode` against an `ast` type.
|
||||
///
|
||||
/// # Example:
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue