mirror of
https://github.com/rust-lang/rust-analyzer.git
synced 2025-09-27 04:19:13 +00:00
organizize
This commit is contained in:
parent
26262aaf05
commit
7c67612b8a
376 changed files with 27 additions and 145 deletions
55
crates/libsyntax2/src/lib.rs
Normal file
55
crates/libsyntax2/src/lib.rs
Normal file
|
@ -0,0 +1,55 @@
|
|||
//! An experimental implementation of [Rust RFC#2256 libsyntax2.0][rfc#2256].
|
||||
//!
|
||||
//! The intent is to be an IDE-ready parser, i.e. one that offers
|
||||
//!
|
||||
//! - easy and fast incremental re-parsing,
|
||||
//! - graceful handling of errors, and
|
||||
//! - maintains all information in the source file.
|
||||
//!
|
||||
//! For more information, see [the RFC][rfc#2265], or [the working draft][RFC.md].
|
||||
//!
|
||||
//! [rfc#2256]: <https://github.com/rust-lang/rfcs/pull/2256>
|
||||
//! [RFC.md]: <https://github.com/matklad/libsyntax2/blob/master/docs/RFC.md>
|
||||
|
||||
#![forbid(
|
||||
missing_debug_implementations,
|
||||
unconditional_recursion,
|
||||
future_incompatible
|
||||
)]
|
||||
#![deny(bad_style, missing_docs)]
|
||||
#![allow(missing_docs)]
|
||||
//#![warn(unreachable_pub)] // rust-lang/rust#47816
|
||||
|
||||
extern crate itertools;
|
||||
extern crate text_unit;
|
||||
extern crate unicode_xid;
|
||||
extern crate drop_bomb;
|
||||
extern crate parking_lot;
|
||||
|
||||
pub mod algo;
|
||||
pub mod ast;
|
||||
mod lexer;
|
||||
#[macro_use]
|
||||
mod parser_api;
|
||||
mod grammar;
|
||||
mod parser_impl;
|
||||
|
||||
mod syntax_kinds;
|
||||
mod smol_str;
|
||||
mod yellow;
|
||||
/// Utilities for simple uses of the parser.
|
||||
pub mod utils;
|
||||
|
||||
pub use {
|
||||
ast::{AstNode, File},
|
||||
lexer::{tokenize, Token},
|
||||
syntax_kinds::SyntaxKind,
|
||||
text_unit::{TextRange, TextUnit},
|
||||
yellow::{SyntaxNode, SyntaxNodeRef, SyntaxRoot, TreeRoot, SyntaxError},
|
||||
};
|
||||
|
||||
|
||||
pub fn parse(text: &str) -> SyntaxNode {
|
||||
let tokens = tokenize(&text);
|
||||
parser_impl::parse::<yellow::GreenBuilder>(text, &tokens)
|
||||
}
|
Loading…
Add table
Add a link
Reference in a new issue