mirror of
https://github.com/Myriad-Dreamin/tinymist.git
synced 2025-07-19 18:55:01 +00:00
refactor: refactor hover, analysis/global, and docs crates (#755)
* dev: refactor hover.rs * refactor refactor AnalysisContext * refactor: refactor docs crate
This commit is contained in:
parent
1c1bc19caf
commit
8f3566366e
44 changed files with 694 additions and 790 deletions
|
@ -202,7 +202,7 @@ mod module_tests {
|
|||
ids
|
||||
}
|
||||
|
||||
let dependencies = construct_module_dependencies(&mut ctx.local);
|
||||
let dependencies = construct_module_dependencies(ctx);
|
||||
|
||||
let mut dependencies = dependencies
|
||||
.into_iter()
|
||||
|
|
|
@ -39,7 +39,7 @@ pub struct CallInfo {
|
|||
// todo: cache call
|
||||
/// Analyzes a function call.
|
||||
pub fn analyze_call(
|
||||
ctx: &mut AnalysisContext,
|
||||
ctx: &mut LocalContext,
|
||||
source: Source,
|
||||
node: LinkedNode,
|
||||
) -> Option<Arc<CallInfo>> {
|
||||
|
@ -64,7 +64,7 @@ pub fn analyze_call(
|
|||
/// Analyzes a function call without caching the result.
|
||||
// todo: testing
|
||||
pub fn analyze_call_no_cache(
|
||||
ctx: &mut AnalysisContext,
|
||||
ctx: &mut LocalContext,
|
||||
source: Source,
|
||||
callee_node: LinkedNode,
|
||||
args: ast::Args<'_>,
|
||||
|
|
|
@ -7,7 +7,7 @@ use typst::visualize::Color;
|
|||
use super::prelude::*;
|
||||
|
||||
/// Get color expressions from a source.
|
||||
pub fn get_color_exprs(ctx: &mut AnalysisContext, src: &Source) -> Option<Vec<ColorInformation>> {
|
||||
pub fn get_color_exprs(ctx: &mut LocalContext, src: &Source) -> Option<Vec<ColorInformation>> {
|
||||
let mut worker = ColorExprWorker {
|
||||
ctx,
|
||||
source: src.clone(),
|
||||
|
@ -18,13 +18,13 @@ pub fn get_color_exprs(ctx: &mut AnalysisContext, src: &Source) -> Option<Vec<Co
|
|||
Some(worker.colors)
|
||||
}
|
||||
|
||||
struct ColorExprWorker<'a, 'w> {
|
||||
ctx: &'a mut AnalysisContext<'w>,
|
||||
struct ColorExprWorker<'a> {
|
||||
ctx: &'a mut LocalContext,
|
||||
source: Source,
|
||||
colors: Vec<ColorInformation>,
|
||||
}
|
||||
|
||||
impl<'a, 'w> ColorExprWorker<'a, 'w> {
|
||||
impl<'a> ColorExprWorker<'a> {
|
||||
fn collect_colors(&mut self, node: LinkedNode) -> Option<()> {
|
||||
match node.kind() {
|
||||
SyntaxKind::FuncCall => {
|
||||
|
|
|
@ -24,8 +24,9 @@ use typst::syntax::{package::PackageSpec, Span, VirtualPath};
|
|||
|
||||
use crate::analysis::prelude::*;
|
||||
use crate::analysis::{
|
||||
analyze_bib, analyze_import_, analyze_signature, post_type_check, BibInfo, PathPreference,
|
||||
Signature, SignatureTarget, Ty, TypeScheme,
|
||||
analyze_bib, analyze_expr_, analyze_import_, analyze_signature, definition, post_type_check,
|
||||
AllocStats, AnalysisStats, BibInfo, Definition, PathPreference, QueryStatGuard, Signature,
|
||||
SignatureTarget, Ty, TypeScheme,
|
||||
};
|
||||
use crate::docs::{DefDocs, TidyModuleDocs};
|
||||
use crate::syntax::{
|
||||
|
@ -39,46 +40,44 @@ use crate::{
|
|||
SemanticTokenContext, TypstRange, VersionedDocument,
|
||||
};
|
||||
|
||||
use super::{analyze_expr_, definition, AllocStats, AnalysisStats, Definition, QueryStatGuard};
|
||||
|
||||
/// The analysis data holds globally.
|
||||
#[derive(Default, Clone)]
|
||||
pub struct Analysis {
|
||||
/// The position encoding for the workspace.
|
||||
pub position_encoding: PositionEncoding,
|
||||
/// The position encoding for the workspace.
|
||||
pub enable_periscope: bool,
|
||||
/// The global caches for analysis.
|
||||
pub caches: AnalysisGlobalCaches,
|
||||
/// The global caches for analysis.
|
||||
pub workers: Arc<AnalysisGlobalWorkers>,
|
||||
/// The global cache grid for analysis.
|
||||
pub cache_grid: Arc<Mutex<AnalysisGlobalCacheGrid>>,
|
||||
/// The periscope provider.
|
||||
pub periscope: Option<Arc<dyn PeriscopeProvider + Send + Sync>>,
|
||||
/// The semantic token context.
|
||||
pub tokens_ctx: Arc<SemanticTokenContext>,
|
||||
/// The global worker resources for analysis.
|
||||
pub workers: Arc<AnalysisGlobalWorkers>,
|
||||
/// The global caches for analysis.
|
||||
pub caches: AnalysisGlobalCaches,
|
||||
/// The global cache grid for analysis.
|
||||
pub cache_grid: Arc<Mutex<AnalysisGlobalCacheGrid>>,
|
||||
/// The statistics about the analyzers.
|
||||
pub analysis_stats: Arc<AnalysisStats>,
|
||||
pub stats: Arc<AnalysisStats>,
|
||||
}
|
||||
|
||||
impl Analysis {
|
||||
/// Get a snapshot of the analysis data.
|
||||
pub fn snapshot<'a>(
|
||||
&self,
|
||||
world: LspWorld,
|
||||
resources: &'a dyn AnalysisResources,
|
||||
) -> AnalysisContext<'a> {
|
||||
AnalysisContext::new(world, resources, self.clone())
|
||||
pub fn snapshot(&self, world: LspWorld) -> LocalContextGuard {
|
||||
let lifetime = self.caches.lifetime.fetch_add(1, Ordering::SeqCst);
|
||||
let slot = self.cache_grid.lock().find_revision(world.revision());
|
||||
LocalContextGuard {
|
||||
local: LocalContext {
|
||||
caches: AnalysisCaches::default(),
|
||||
shared: Arc::new(SharedContext {
|
||||
slot,
|
||||
lifetime,
|
||||
world,
|
||||
analysis: self.clone(),
|
||||
}),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
/// Clear all cached resources.
|
||||
pub fn clear_cache(&self) {
|
||||
self.caches.signatures.clear();
|
||||
self.caches.static_signatures.clear();
|
||||
self.caches.terms.clear();
|
||||
self.cache_grid.lock().clear();
|
||||
}
|
||||
|
||||
/// Lock the revision in main thread.
|
||||
/// Lock the revision in *main thread*.
|
||||
#[must_use]
|
||||
pub fn lock_revision(&self) -> RevisionLock {
|
||||
let mut grid = self.cache_grid.lock();
|
||||
|
@ -90,9 +89,17 @@ impl Analysis {
|
|||
}
|
||||
}
|
||||
|
||||
/// Clear all cached resources.
|
||||
pub fn clear_cache(&self) {
|
||||
self.caches.signatures.clear();
|
||||
self.caches.static_signatures.clear();
|
||||
self.caches.terms.clear();
|
||||
self.cache_grid.lock().clear();
|
||||
}
|
||||
|
||||
/// Report the statistics of the analysis.
|
||||
pub fn report_query_stats(&self) -> String {
|
||||
self.analysis_stats.report()
|
||||
self.stats.report()
|
||||
}
|
||||
|
||||
/// Report the statistics of the allocation.
|
||||
|
@ -101,12 +108,12 @@ impl Analysis {
|
|||
}
|
||||
}
|
||||
|
||||
/// The resources for analysis.
|
||||
pub trait AnalysisResources {
|
||||
/// The periscope provider.
|
||||
pub trait PeriscopeProvider {
|
||||
/// Resolve telescope image at the given position.
|
||||
fn periscope_at(
|
||||
&self,
|
||||
_ctx: &mut AnalysisContext,
|
||||
_ctx: &mut LocalContext,
|
||||
_doc: VersionedDocument,
|
||||
_pos: Position,
|
||||
) -> Option<String> {
|
||||
|
@ -125,15 +132,13 @@ pub struct AnalysisGlobalWorkers {
|
|||
tooltip: RateLimiter,
|
||||
}
|
||||
|
||||
/// The context for analyzers.
|
||||
pub struct AnalysisContext<'a> {
|
||||
/// The world surface for Typst compiler
|
||||
pub resources: &'a dyn AnalysisResources,
|
||||
/// Constructed shared context
|
||||
/// The local context guard that performs gc once dropped.
|
||||
pub struct LocalContextGuard {
|
||||
/// Constructed local context
|
||||
pub local: LocalContext,
|
||||
}
|
||||
|
||||
impl Deref for AnalysisContext<'_> {
|
||||
impl Deref for LocalContextGuard {
|
||||
type Target = LocalContext;
|
||||
|
||||
fn deref(&self) -> &Self::Target {
|
||||
|
@ -141,118 +146,20 @@ impl Deref for AnalysisContext<'_> {
|
|||
}
|
||||
}
|
||||
|
||||
impl DerefMut for AnalysisContext<'_> {
|
||||
impl DerefMut for LocalContextGuard {
|
||||
fn deref_mut(&mut self) -> &mut Self::Target {
|
||||
&mut self.local
|
||||
}
|
||||
}
|
||||
|
||||
// todo: gc in new thread
|
||||
impl<'w> Drop for AnalysisContext<'w> {
|
||||
impl Drop for LocalContextGuard {
|
||||
fn drop(&mut self) {
|
||||
self.gc();
|
||||
}
|
||||
}
|
||||
|
||||
impl<'w> AnalysisContext<'w> {
|
||||
/// Create a new analysis context.
|
||||
pub fn new(world: LspWorld, resources: &'w dyn AnalysisResources, a: Analysis) -> Self {
|
||||
let lifetime = a.caches.lifetime.fetch_add(1, Ordering::SeqCst);
|
||||
let slot = a.cache_grid.lock().find_revision(world.revision());
|
||||
Self {
|
||||
resources,
|
||||
local: LocalContext {
|
||||
caches: AnalysisCaches::default(),
|
||||
shared: Arc::new(SharedContext {
|
||||
slot,
|
||||
lifetime,
|
||||
world,
|
||||
analysis: a,
|
||||
}),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
/// Resolve extra font information.
|
||||
pub fn font_info(&self, font: typst::text::Font) -> Option<Arc<DataSource>> {
|
||||
self.world().font_resolver.describe_font(&font)
|
||||
}
|
||||
|
||||
/// Get the world surface for Typst compiler.
|
||||
pub fn world(&self) -> &LspWorld {
|
||||
&self.shared.world
|
||||
}
|
||||
|
||||
/// Get the shared context.
|
||||
pub fn shared(&self) -> &Arc<SharedContext> {
|
||||
&self.local.shared
|
||||
}
|
||||
|
||||
/// Get the shared context.
|
||||
pub fn shared_(&self) -> Arc<SharedContext> {
|
||||
self.local.shared.clone()
|
||||
}
|
||||
|
||||
/// Fork a new context for searching in the workspace.
|
||||
pub fn fork_for_search<'s>(&'s mut self) -> SearchCtx<'s, 'w> {
|
||||
SearchCtx {
|
||||
ctx: self,
|
||||
searched: Default::default(),
|
||||
worklist: Default::default(),
|
||||
}
|
||||
}
|
||||
|
||||
pub(crate) fn preload_package(&self, entry_point: TypstFileId) {
|
||||
self.shared_().preload_package(entry_point);
|
||||
}
|
||||
|
||||
pub(crate) fn with_vm<T>(&self, f: impl FnOnce(&mut typst::eval::Vm) -> T) -> T {
|
||||
crate::upstream::with_vm((self.world() as &dyn World).track(), f)
|
||||
}
|
||||
|
||||
pub(crate) fn const_eval(&self, rr: ast::Expr<'_>) -> Option<Value> {
|
||||
SharedContext::const_eval(rr)
|
||||
}
|
||||
|
||||
pub(crate) fn mini_eval(&self, rr: ast::Expr<'_>) -> Option<Value> {
|
||||
self.const_eval(rr)
|
||||
.or_else(|| self.with_vm(|vm| rr.eval(vm).ok()))
|
||||
}
|
||||
|
||||
pub(crate) fn type_of(&mut self, rr: &SyntaxNode) -> Option<Ty> {
|
||||
self.type_of_span(rr.span())
|
||||
}
|
||||
|
||||
pub(crate) fn type_of_span(&mut self, s: Span) -> Option<Ty> {
|
||||
let id = s.id()?;
|
||||
let source = self.source_by_id(id).ok()?;
|
||||
self.type_of_span_(&source, s)
|
||||
}
|
||||
|
||||
pub(crate) fn type_of_span_(&mut self, source: &Source, s: Span) -> Option<Ty> {
|
||||
self.type_check(source).type_of_span(s)
|
||||
}
|
||||
|
||||
pub(crate) fn literal_type_of_node(&mut self, k: LinkedNode) -> Option<Ty> {
|
||||
let id = k.span().id()?;
|
||||
let source = self.source_by_id(id).ok()?;
|
||||
let ty_chk = self.type_check(&source);
|
||||
|
||||
let ty = post_type_check(self.shared_(), &ty_chk, k.clone())
|
||||
.or_else(|| ty_chk.type_of_span(k.span()))?;
|
||||
Some(ty_chk.simplify(ty, false))
|
||||
}
|
||||
|
||||
/// Get module import at location.
|
||||
pub fn module_ins_at(&mut self, def_fid: TypstFileId, cursor: usize) -> Option<Value> {
|
||||
let def_src = self.source_by_id(def_fid).ok()?;
|
||||
let def_root = LinkedNode::new(def_src.root());
|
||||
let mod_exp = find_expr_in_import(def_root.leaf_at_compat(cursor)?)?;
|
||||
let mod_import = mod_exp.parent()?.clone();
|
||||
let mod_import_node = mod_import.cast::<ast::ModuleImport>()?;
|
||||
self.analyze_import(mod_import_node.source().to_untyped()).1
|
||||
}
|
||||
|
||||
impl LocalContextGuard {
|
||||
fn gc(&self) {
|
||||
let lifetime = self.lifetime;
|
||||
loop {
|
||||
|
@ -297,7 +204,7 @@ impl<'w> AnalysisContext<'w> {
|
|||
pub struct LocalContext {
|
||||
/// Local caches for analysis.
|
||||
pub caches: AnalysisCaches,
|
||||
/// Constructed shared context
|
||||
/// The shared context
|
||||
pub shared: Arc<SharedContext>,
|
||||
}
|
||||
|
||||
|
@ -374,6 +281,81 @@ impl LocalContext {
|
|||
}
|
||||
}
|
||||
|
||||
/// Get the world surface for Typst compiler.
|
||||
pub fn world(&self) -> &LspWorld {
|
||||
&self.shared.world
|
||||
}
|
||||
|
||||
/// Get the shared context.
|
||||
pub fn shared(&self) -> &Arc<SharedContext> {
|
||||
&self.shared
|
||||
}
|
||||
|
||||
/// Get the shared context.
|
||||
pub fn shared_(&self) -> Arc<SharedContext> {
|
||||
self.shared.clone()
|
||||
}
|
||||
|
||||
/// Fork a new context for searching in the workspace.
|
||||
pub fn fork_for_search(&mut self) -> SearchCtx {
|
||||
SearchCtx {
|
||||
ctx: self,
|
||||
searched: Default::default(),
|
||||
worklist: Default::default(),
|
||||
}
|
||||
}
|
||||
|
||||
pub(crate) fn preload_package(&self, entry_point: TypstFileId) {
|
||||
self.shared_().preload_package(entry_point);
|
||||
}
|
||||
|
||||
pub(crate) fn with_vm<T>(&self, f: impl FnOnce(&mut typst::eval::Vm) -> T) -> T {
|
||||
crate::upstream::with_vm((self.world() as &dyn World).track(), f)
|
||||
}
|
||||
|
||||
pub(crate) fn const_eval(&self, rr: ast::Expr<'_>) -> Option<Value> {
|
||||
SharedContext::const_eval(rr)
|
||||
}
|
||||
|
||||
pub(crate) fn mini_eval(&self, rr: ast::Expr<'_>) -> Option<Value> {
|
||||
self.const_eval(rr)
|
||||
.or_else(|| self.with_vm(|vm| rr.eval(vm).ok()))
|
||||
}
|
||||
|
||||
pub(crate) fn type_of(&mut self, rr: &SyntaxNode) -> Option<Ty> {
|
||||
self.type_of_span(rr.span())
|
||||
}
|
||||
|
||||
pub(crate) fn type_of_span(&mut self, s: Span) -> Option<Ty> {
|
||||
let id = s.id()?;
|
||||
let source = self.source_by_id(id).ok()?;
|
||||
self.type_of_span_(&source, s)
|
||||
}
|
||||
|
||||
pub(crate) fn type_of_span_(&mut self, source: &Source, s: Span) -> Option<Ty> {
|
||||
self.type_check(source).type_of_span(s)
|
||||
}
|
||||
|
||||
pub(crate) fn literal_type_of_node(&mut self, k: LinkedNode) -> Option<Ty> {
|
||||
let id = k.span().id()?;
|
||||
let source = self.source_by_id(id).ok()?;
|
||||
let ty_chk = self.type_check(&source);
|
||||
|
||||
let ty = post_type_check(self.shared_(), &ty_chk, k.clone())
|
||||
.or_else(|| ty_chk.type_of_span(k.span()))?;
|
||||
Some(ty_chk.simplify(ty, false))
|
||||
}
|
||||
|
||||
/// Get module import at location.
|
||||
pub fn module_ins_at(&mut self, def_fid: TypstFileId, cursor: usize) -> Option<Value> {
|
||||
let def_src = self.source_by_id(def_fid).ok()?;
|
||||
let def_root = LinkedNode::new(def_src.root());
|
||||
let mod_exp = find_expr_in_import(def_root.leaf_at_compat(cursor)?)?;
|
||||
let mod_import = mod_exp.parent()?.clone();
|
||||
let mod_import_node = mod_import.cast::<ast::ModuleImport>()?;
|
||||
self.analyze_import(mod_import_node.source().to_untyped()).1
|
||||
}
|
||||
|
||||
/// Get the expression information of a source file.
|
||||
pub(crate) fn expr_stage_by_id(&mut self, fid: TypstFileId) -> Option<Arc<ExprInfo>> {
|
||||
Some(self.expr_stage(&self.source_by_id(fid).ok()?))
|
||||
|
@ -399,11 +381,11 @@ impl LocalContext {
|
|||
match def.decl.kind() {
|
||||
DefKind::Function => {
|
||||
let sig = self.sig_of_def(def.clone())?;
|
||||
let docs = crate::docs::signature_docs(&sig, None)?;
|
||||
let docs = crate::docs::sig_docs(&sig, None)?;
|
||||
Some(DefDocs::Function(Box::new(docs)))
|
||||
}
|
||||
DefKind::Struct | DefKind::Constant | DefKind::Variable => {
|
||||
let docs = crate::docs::variable_docs(self, def.decl.span())?;
|
||||
let docs = crate::docs::var_docs(self, def.decl.span())?;
|
||||
Some(DefDocs::Variable(docs))
|
||||
}
|
||||
DefKind::Module => {
|
||||
|
@ -421,11 +403,11 @@ impl LocalContext {
|
|||
pub struct SharedContext {
|
||||
/// The caches lifetime tick for analysis.
|
||||
pub lifetime: u64,
|
||||
/// Get the world surface for Typst compiler.
|
||||
/// The world surface for Typst compiler.
|
||||
pub world: LspWorld,
|
||||
/// The analysis data
|
||||
pub analysis: Analysis,
|
||||
/// The revision slot
|
||||
/// The using revision slot
|
||||
slot: Arc<RevisionSlot>,
|
||||
}
|
||||
|
||||
|
@ -435,16 +417,6 @@ impl SharedContext {
|
|||
self.slot.revision
|
||||
}
|
||||
|
||||
fn query_stat(&self, id: TypstFileId, query: &'static str) -> QueryStatGuard {
|
||||
let stats = &self.analysis.analysis_stats.query_stats;
|
||||
let entry = stats.entry(id).or_default();
|
||||
let entry = entry.entry(query).or_default();
|
||||
QueryStatGuard {
|
||||
bucket: entry.clone(),
|
||||
since: std::time::SystemTime::now(),
|
||||
}
|
||||
}
|
||||
|
||||
/// Get the position encoding during session.
|
||||
pub(crate) fn position_encoding(&self) -> PositionEncoding {
|
||||
self.analysis.position_encoding
|
||||
|
@ -481,7 +453,7 @@ impl SharedContext {
|
|||
if matches!(w, Some("yaml" | "yml" | "bib")) {
|
||||
let bytes = self.file_by_id(fid).ok()?;
|
||||
let bytes_len = bytes.len();
|
||||
let loc = get_loc_info(bytes)?;
|
||||
let loc = loc_info(bytes)?;
|
||||
// binary search
|
||||
let start = find_loc(bytes_len, &loc, position.start, self.position_encoding())?;
|
||||
let end = find_loc(bytes_len, &loc, position.end, self.position_encoding())?;
|
||||
|
@ -579,6 +551,12 @@ impl SharedContext {
|
|||
|
||||
v
|
||||
}
|
||||
|
||||
/// Resolve extra font information.
|
||||
pub fn font_info(&self, font: typst::text::Font) -> Option<Arc<DataSource>> {
|
||||
self.world.font_resolver.describe_font(&font)
|
||||
}
|
||||
|
||||
/// Get the local packages and their descriptions.
|
||||
pub fn local_packages(&self) -> EcoVec<PackageSpec> {
|
||||
crate::package::list_package_by_namespace(&self.world.registry, eco_format!("local"))
|
||||
|
@ -587,37 +565,60 @@ impl SharedContext {
|
|||
.collect()
|
||||
}
|
||||
|
||||
pub(crate) fn type_of_func(self: &Arc<Self>, func: Func) -> Signature {
|
||||
log::debug!("convert runtime func {func:?}");
|
||||
analyze_signature(self, SignatureTarget::Convert(func)).unwrap()
|
||||
pub(crate) fn const_eval(rr: ast::Expr<'_>) -> Option<Value> {
|
||||
Some(match rr {
|
||||
ast::Expr::None(_) => Value::None,
|
||||
ast::Expr::Auto(_) => Value::Auto,
|
||||
ast::Expr::Bool(v) => Value::Bool(v.get()),
|
||||
ast::Expr::Int(v) => Value::Int(v.get()),
|
||||
ast::Expr::Float(v) => Value::Float(v.get()),
|
||||
ast::Expr::Numeric(v) => Value::numeric(v.get()),
|
||||
ast::Expr::Str(v) => Value::Str(v.get().into()),
|
||||
_ => return None,
|
||||
})
|
||||
}
|
||||
|
||||
pub(crate) fn type_of_value(self: &Arc<Self>, val: &Value) -> Ty {
|
||||
log::debug!("convert runtime value {val:?}");
|
||||
/// Get a module by file id.
|
||||
pub fn module_by_id(&self, fid: TypstFileId) -> SourceResult<Module> {
|
||||
let source = self.source_by_id(fid).at(Span::detached())?;
|
||||
self.module_by_src(source)
|
||||
}
|
||||
|
||||
// todo: check performance on peeking signature source frequently
|
||||
let cache_key = val;
|
||||
let cached = self
|
||||
.analysis
|
||||
.caches
|
||||
.terms
|
||||
.m
|
||||
.get(&hash128(&cache_key))
|
||||
.and_then(|slot| (cache_key == &slot.1 .0).then_some(slot.1 .1.clone()));
|
||||
if let Some(cached) = cached {
|
||||
return cached;
|
||||
/// Get a module by string.
|
||||
pub fn module_by_str(&self, rr: String) -> Option<Module> {
|
||||
let src = Source::new(*DETACHED_ENTRY, rr);
|
||||
self.module_by_src(src).ok()
|
||||
}
|
||||
|
||||
/// Get (Create) a module by source.
|
||||
pub fn module_by_src(&self, source: Source) -> SourceResult<Module> {
|
||||
let route = Route::default();
|
||||
let traced = Traced::default();
|
||||
let mut sink = Sink::default();
|
||||
|
||||
typst::eval::eval(
|
||||
((&self.world) as &dyn World).track(),
|
||||
traced.track(),
|
||||
sink.track_mut(),
|
||||
route.track(),
|
||||
&source,
|
||||
)
|
||||
}
|
||||
|
||||
/// Try to load a module from the current source file.
|
||||
pub fn module_by_syntax(&self, source: &SyntaxNode) -> Option<Value> {
|
||||
let (src, scope) = self.analyze_import(source);
|
||||
if let Some(scope) = scope {
|
||||
return Some(scope);
|
||||
}
|
||||
|
||||
let res = crate::analysis::term_value(self, val);
|
||||
|
||||
self.analysis
|
||||
.caches
|
||||
.terms
|
||||
.m
|
||||
.entry(hash128(&cache_key))
|
||||
.or_insert_with(|| (self.lifetime, (cache_key.clone(), res.clone())));
|
||||
|
||||
res
|
||||
match src {
|
||||
Some(Value::Str(s)) => {
|
||||
let id = resolve_id_by_path(&self.world, source.span().id()?, s.as_str())?;
|
||||
self.module_by_id(id).ok().map(Value::Module)
|
||||
}
|
||||
_ => None,
|
||||
}
|
||||
}
|
||||
|
||||
/// Get the expression information of a source file.
|
||||
|
@ -691,6 +692,39 @@ impl SharedContext {
|
|||
})
|
||||
}
|
||||
|
||||
pub(crate) fn type_of_func(self: &Arc<Self>, func: Func) -> Signature {
|
||||
log::debug!("convert runtime func {func:?}");
|
||||
analyze_signature(self, SignatureTarget::Convert(func)).unwrap()
|
||||
}
|
||||
|
||||
pub(crate) fn type_of_value(self: &Arc<Self>, val: &Value) -> Ty {
|
||||
log::debug!("convert runtime value {val:?}");
|
||||
|
||||
// todo: check performance on peeking signature source frequently
|
||||
let cache_key = val;
|
||||
let cached = self
|
||||
.analysis
|
||||
.caches
|
||||
.terms
|
||||
.m
|
||||
.get(&hash128(&cache_key))
|
||||
.and_then(|slot| (cache_key == &slot.1 .0).then_some(slot.1 .1.clone()));
|
||||
if let Some(cached) = cached {
|
||||
return cached;
|
||||
}
|
||||
|
||||
let res = crate::analysis::term_value(self, val);
|
||||
|
||||
self.analysis
|
||||
.caches
|
||||
.terms
|
||||
.m
|
||||
.entry(hash128(&cache_key))
|
||||
.or_insert_with(|| (self.lifetime, (cache_key.clone(), res.clone())));
|
||||
|
||||
res
|
||||
}
|
||||
|
||||
pub(crate) fn def_of_span(
|
||||
self: &Arc<Self>,
|
||||
source: &Source,
|
||||
|
@ -777,62 +811,6 @@ impl SharedContext {
|
|||
token.enter(|| tooltip_(&self.world, document, source, cursor))
|
||||
}
|
||||
|
||||
pub(crate) fn const_eval(rr: ast::Expr<'_>) -> Option<Value> {
|
||||
Some(match rr {
|
||||
ast::Expr::None(_) => Value::None,
|
||||
ast::Expr::Auto(_) => Value::Auto,
|
||||
ast::Expr::Bool(v) => Value::Bool(v.get()),
|
||||
ast::Expr::Int(v) => Value::Int(v.get()),
|
||||
ast::Expr::Float(v) => Value::Float(v.get()),
|
||||
ast::Expr::Numeric(v) => Value::numeric(v.get()),
|
||||
ast::Expr::Str(v) => Value::Str(v.get().into()),
|
||||
_ => return None,
|
||||
})
|
||||
}
|
||||
|
||||
/// Get a module by file id.
|
||||
pub fn module_by_id(&self, fid: TypstFileId) -> SourceResult<Module> {
|
||||
let source = self.source_by_id(fid).at(Span::detached())?;
|
||||
self.module_by_src(source)
|
||||
}
|
||||
|
||||
/// Get a module by string.
|
||||
pub fn module_by_str(&self, rr: String) -> Option<Module> {
|
||||
let src = Source::new(*DETACHED_ENTRY, rr);
|
||||
self.module_by_src(src).ok()
|
||||
}
|
||||
|
||||
/// Get (Create) a module by source.
|
||||
pub fn module_by_src(&self, source: Source) -> SourceResult<Module> {
|
||||
let route = Route::default();
|
||||
let traced = Traced::default();
|
||||
let mut sink = Sink::default();
|
||||
|
||||
typst::eval::eval(
|
||||
((&self.world) as &dyn World).track(),
|
||||
traced.track(),
|
||||
sink.track_mut(),
|
||||
route.track(),
|
||||
&source,
|
||||
)
|
||||
}
|
||||
|
||||
/// Try to load a module from the current source file.
|
||||
pub fn module_by_syntax(&self, source: &SyntaxNode) -> Option<Value> {
|
||||
let (src, scope) = self.analyze_import(source);
|
||||
if let Some(scope) = scope {
|
||||
return Some(scope);
|
||||
}
|
||||
|
||||
match src {
|
||||
Some(Value::Str(s)) => {
|
||||
let id = resolve_id_by_path(&self.world, source.span().id()?, s.as_str())?;
|
||||
self.module_by_id(id).ok().map(Value::Module)
|
||||
}
|
||||
_ => None,
|
||||
}
|
||||
}
|
||||
|
||||
/// Get the manifest of a package by file id.
|
||||
pub fn get_manifest(&self, toml_id: TypstFileId) -> StrResult<PackageManifest> {
|
||||
crate::docs::get_manifest(&self.world, toml_id)
|
||||
|
@ -878,6 +856,16 @@ impl SharedContext {
|
|||
res.get_or_init(|| compute(self)).clone()
|
||||
}
|
||||
|
||||
fn query_stat(&self, id: TypstFileId, query: &'static str) -> QueryStatGuard {
|
||||
let stats = &self.analysis.stats.query_stats;
|
||||
let entry = stats.entry(id).or_default();
|
||||
let entry = entry.entry(query).or_default();
|
||||
QueryStatGuard {
|
||||
bucket: entry.clone(),
|
||||
since: std::time::SystemTime::now(),
|
||||
}
|
||||
}
|
||||
|
||||
/// Check on a module before really needing them. But we likely use them
|
||||
/// after a while.
|
||||
pub(crate) fn prefetch_type_check(self: &Arc<Self>, _fid: TypstFileId) {
|
||||
|
@ -1189,7 +1177,7 @@ fn bib_info(
|
|||
}
|
||||
|
||||
#[comemo::memoize]
|
||||
fn get_loc_info(bytes: Bytes) -> Option<EcoVec<(usize, String)>> {
|
||||
fn loc_info(bytes: Bytes) -> Option<EcoVec<(usize, String)>> {
|
||||
let mut loc = EcoVec::new();
|
||||
let mut offset = 0;
|
||||
for line in bytes.split(|e| *e == b'\n') {
|
||||
|
@ -1234,16 +1222,16 @@ fn find_loc(
|
|||
}
|
||||
|
||||
/// The context for searching in the workspace.
|
||||
pub struct SearchCtx<'a, 'w> {
|
||||
pub struct SearchCtx<'a> {
|
||||
/// The inner analysis context.
|
||||
pub ctx: &'a mut AnalysisContext<'w>,
|
||||
pub ctx: &'a mut LocalContext,
|
||||
/// The set of files that have been searched.
|
||||
pub searched: HashSet<TypstFileId>,
|
||||
/// The files that need to be searched.
|
||||
pub worklist: Vec<TypstFileId>,
|
||||
}
|
||||
|
||||
impl SearchCtx<'_, '_> {
|
||||
impl SearchCtx<'_> {
|
||||
/// Push a file to the worklist.
|
||||
pub fn push(&mut self, id: TypstFileId) -> bool {
|
||||
if self.searched.insert(id) {
|
||||
|
|
|
@ -6,14 +6,14 @@ use super::prelude::*;
|
|||
use crate::path_to_url;
|
||||
|
||||
/// Get link expressions from a source.
|
||||
pub fn get_link_exprs(ctx: &mut AnalysisContext, src: &Source) -> Option<Vec<(Range<usize>, Url)>> {
|
||||
pub fn get_link_exprs(ctx: &mut LocalContext, src: &Source) -> Option<Vec<(Range<usize>, Url)>> {
|
||||
let root = LinkedNode::new(src.root());
|
||||
get_link_exprs_in(ctx, &root)
|
||||
}
|
||||
|
||||
/// Get link expressions in a source node.
|
||||
pub fn get_link_exprs_in(
|
||||
ctx: &mut AnalysisContext,
|
||||
ctx: &mut LocalContext,
|
||||
node: &LinkedNode,
|
||||
) -> Option<Vec<(Range<usize>, Url)>> {
|
||||
let mut worker = LinkStrWorker { ctx, links: vec![] };
|
||||
|
@ -21,12 +21,12 @@ pub fn get_link_exprs_in(
|
|||
Some(worker.links)
|
||||
}
|
||||
|
||||
struct LinkStrWorker<'a, 'w> {
|
||||
ctx: &'a mut AnalysisContext<'w>,
|
||||
struct LinkStrWorker<'a> {
|
||||
ctx: &'a mut LocalContext,
|
||||
links: Vec<(Range<usize>, Url)>,
|
||||
}
|
||||
|
||||
impl<'a, 'w> LinkStrWorker<'a, 'w> {
|
||||
impl<'a> LinkStrWorker<'a> {
|
||||
fn collect_links(&mut self, node: &LinkedNode) -> Option<()> {
|
||||
match node.kind() {
|
||||
// SyntaxKind::Link => { }
|
||||
|
|
|
@ -16,6 +16,6 @@ pub use typst_shim::syntax::LinkedNodeExt;
|
|||
pub use typst_shim::utils::LazyHash;
|
||||
|
||||
pub(crate) use super::StrRef;
|
||||
pub(crate) use super::{AnalysisContext, ToFunc};
|
||||
pub(crate) use super::{LocalContext, ToFunc};
|
||||
pub(crate) use crate::adt::interner::Interned;
|
||||
pub use crate::ty::Ty;
|
||||
|
|
|
@ -74,7 +74,7 @@ pub struct CodeActionRequest {
|
|||
impl SemanticRequest for CodeActionRequest {
|
||||
type Response = Vec<CodeActionOrCommand>;
|
||||
|
||||
fn request(self, ctx: &mut AnalysisContext) -> Option<Self::Response> {
|
||||
fn request(self, ctx: &mut LocalContext) -> Option<Self::Response> {
|
||||
let source = ctx.source_by_path(&self.path).ok()?;
|
||||
let range = ctx.to_typst_range(self.range, &source)?;
|
||||
let cursor = (range.start + 1).min(source.text().len());
|
||||
|
@ -89,15 +89,15 @@ impl SemanticRequest for CodeActionRequest {
|
|||
}
|
||||
}
|
||||
|
||||
struct CodeActionWorker<'a, 'w> {
|
||||
ctx: &'a mut AnalysisContext<'w>,
|
||||
struct CodeActionWorker<'a> {
|
||||
ctx: &'a mut LocalContext,
|
||||
actions: Vec<CodeActionOrCommand>,
|
||||
local_url: OnceCell<Option<Url>>,
|
||||
current: Source,
|
||||
}
|
||||
|
||||
impl<'a, 'w> CodeActionWorker<'a, 'w> {
|
||||
fn new(ctx: &'a mut AnalysisContext<'w>, current: Source) -> Self {
|
||||
impl<'a> CodeActionWorker<'a> {
|
||||
fn new(ctx: &'a mut LocalContext, current: Source) -> Self {
|
||||
Self {
|
||||
ctx,
|
||||
actions: Vec::new(),
|
||||
|
|
|
@ -15,7 +15,7 @@ pub struct CodeLensRequest {
|
|||
impl SemanticRequest for CodeLensRequest {
|
||||
type Response = Vec<CodeLens>;
|
||||
|
||||
fn request(self, ctx: &mut AnalysisContext) -> Option<Self::Response> {
|
||||
fn request(self, ctx: &mut LocalContext) -> Option<Self::Response> {
|
||||
let source = ctx.source_by_path(&self.path).ok()?;
|
||||
|
||||
let doc_start = ctx.to_lsp_range(0..0, &source);
|
||||
|
|
|
@ -61,7 +61,7 @@ impl StatefulRequest for CompletionRequest {
|
|||
|
||||
fn request(
|
||||
self,
|
||||
ctx: &mut AnalysisContext,
|
||||
ctx: &mut LocalContext,
|
||||
doc: Option<VersionedDocument>,
|
||||
) -> Option<Self::Response> {
|
||||
let doc = doc.as_ref().map(|doc| doc.document.as_ref());
|
||||
|
@ -331,8 +331,8 @@ mod tests {
|
|||
pkg_mode: bool,
|
||||
}
|
||||
|
||||
fn run(c: TestConfig) -> impl Fn(&mut AnalysisContext, PathBuf) {
|
||||
fn test(ctx: &mut AnalysisContext, id: TypstFileId) {
|
||||
fn run(c: TestConfig) -> impl Fn(&mut LocalContext, PathBuf) {
|
||||
fn test(ctx: &mut LocalContext, id: TypstFileId) {
|
||||
let source = ctx.source_by_id(id).unwrap();
|
||||
let rng = find_test_range(&source);
|
||||
let text = source.text()[rng.clone()].to_string();
|
||||
|
|
|
@ -7,7 +7,7 @@ pub type DiagnosticsMap = HashMap<Url, Vec<LspDiagnostic>>;
|
|||
|
||||
/// Converts a list of Typst diagnostics to LSP diagnostics.
|
||||
pub fn convert_diagnostics<'a>(
|
||||
ctx: &AnalysisContext,
|
||||
ctx: &LocalContext,
|
||||
errors: impl IntoIterator<Item = &'a TypstDiagnostic>,
|
||||
) -> DiagnosticsMap {
|
||||
errors
|
||||
|
@ -15,7 +15,7 @@ pub fn convert_diagnostics<'a>(
|
|||
.flat_map(|error| {
|
||||
convert_diagnostic(ctx, error)
|
||||
.map_err(move |conversion_err| {
|
||||
error!("could not convert Typst error to diagnostic: {conversion_err:?} error to convert: {error:?}");
|
||||
log::error!("could not convert Typst error to diagnostic: {conversion_err:?} error to convert: {error:?}");
|
||||
})
|
||||
})
|
||||
.collect::<Vec<_>>()
|
||||
|
@ -24,7 +24,7 @@ pub fn convert_diagnostics<'a>(
|
|||
}
|
||||
|
||||
fn convert_diagnostic(
|
||||
ctx: &AnalysisContext,
|
||||
ctx: &LocalContext,
|
||||
typst_diagnostic: &TypstDiagnostic,
|
||||
) -> anyhow::Result<(Url, LspDiagnostic)> {
|
||||
let uri;
|
||||
|
@ -64,7 +64,7 @@ fn convert_diagnostic(
|
|||
}
|
||||
|
||||
fn tracepoint_to_relatedinformation(
|
||||
project: &AnalysisContext,
|
||||
project: &LocalContext,
|
||||
tracepoint: &Spanned<Tracepoint>,
|
||||
position_encoding: PositionEncoding,
|
||||
) -> anyhow::Result<Option<DiagnosticRelatedInformation>> {
|
||||
|
@ -89,7 +89,7 @@ fn tracepoint_to_relatedinformation(
|
|||
}
|
||||
|
||||
fn diagnostic_related_information(
|
||||
project: &AnalysisContext,
|
||||
project: &LocalContext,
|
||||
typst_diagnostic: &TypstDiagnostic,
|
||||
position_encoding: PositionEncoding,
|
||||
) -> anyhow::Result<Vec<DiagnosticRelatedInformation>> {
|
||||
|
|
62
crates/tinymist-query/src/docs/convert.rs
Normal file
62
crates/tinymist-query/src/docs/convert.rs
Normal file
|
@ -0,0 +1,62 @@
|
|||
use std::sync::{Arc, LazyLock};
|
||||
|
||||
use ecow::{eco_format, EcoString};
|
||||
use parking_lot::Mutex;
|
||||
use tinymist_world::base::{EntryState, ShadowApi, TaskInputs};
|
||||
use tinymist_world::LspWorld;
|
||||
use typlite::scopes::Scopes;
|
||||
use typlite::value::{Value, *};
|
||||
use typst::foundations::Bytes;
|
||||
use typst::{
|
||||
diag::StrResult,
|
||||
syntax::{FileId, VirtualPath},
|
||||
};
|
||||
|
||||
// Unfortunately, we have only 65536 possible file ids and we cannot revoke
|
||||
// them. So we share a global file id for all docs conversion.
|
||||
static DOCS_CONVERT_ID: LazyLock<Mutex<FileId>> =
|
||||
LazyLock::new(|| Mutex::new(FileId::new(None, VirtualPath::new("__tinymist_docs__.typ"))));
|
||||
|
||||
pub(crate) fn convert_docs(world: &LspWorld, content: &str) -> StrResult<EcoString> {
|
||||
static DOCS_LIB: LazyLock<Arc<Scopes<Value>>> = LazyLock::new(lib);
|
||||
|
||||
let conv_id = DOCS_CONVERT_ID.lock();
|
||||
let entry = EntryState::new_rootless(conv_id.vpath().as_rooted_path().into()).unwrap();
|
||||
let entry = entry.select_in_workspace(*conv_id);
|
||||
|
||||
let mut w = world.task(TaskInputs {
|
||||
entry: Some(entry),
|
||||
inputs: None,
|
||||
});
|
||||
w.map_shadow_by_id(*conv_id, Bytes::from(content.as_bytes().to_owned()))?;
|
||||
// todo: bad performance
|
||||
w.source_db.take_state();
|
||||
|
||||
let conv = typlite::Typlite::new(Arc::new(w))
|
||||
.with_library(DOCS_LIB.clone())
|
||||
.annotate_elements(true)
|
||||
.convert()
|
||||
.map_err(|e| eco_format!("failed to convert to markdown: {e}"))?;
|
||||
|
||||
Ok(conv.replace("```example", "```typ"))
|
||||
}
|
||||
|
||||
pub(super) fn lib() -> Arc<Scopes<Value>> {
|
||||
let mut scopes = typlite::library::library();
|
||||
|
||||
// todo: how to import this function correctly?
|
||||
scopes.define("example", example as RawFunc);
|
||||
|
||||
Arc::new(scopes)
|
||||
}
|
||||
|
||||
/// Evaluate a `example`.
|
||||
pub fn example(mut args: Args) -> typlite::Result<Value> {
|
||||
let body = get_pos_named!(args, body: Content).0;
|
||||
let body = body.trim();
|
||||
let ticks = body.chars().take_while(|t| *t == '`').collect::<String>();
|
||||
let body = &body[ticks.len()..];
|
||||
let body = eco_format!("{ticks}typ{body}");
|
||||
|
||||
Ok(Value::Content(body))
|
||||
}
|
|
@ -1,21 +1,13 @@
|
|||
use core::fmt;
|
||||
use std::collections::BTreeMap;
|
||||
use std::sync::{Arc, OnceLock};
|
||||
use std::sync::OnceLock;
|
||||
|
||||
use ecow::{eco_format, EcoString};
|
||||
use parking_lot::Mutex;
|
||||
use ecow::EcoString;
|
||||
use serde::{Deserialize, Serialize};
|
||||
use tinymist_world::base::{EntryState, ShadowApi, TaskInputs};
|
||||
use tinymist_world::LspWorld;
|
||||
use typst::foundations::Bytes;
|
||||
use typst::{
|
||||
diag::StrResult,
|
||||
syntax::{FileId, Span, VirtualPath},
|
||||
};
|
||||
use typst::syntax::Span;
|
||||
|
||||
use super::tidy::*;
|
||||
use crate::analysis::{ParamAttrs, ParamSpec, Signature};
|
||||
use crate::docs::library;
|
||||
use crate::prelude::*;
|
||||
use crate::ty::Ty;
|
||||
use crate::ty::{DocSource, Interned};
|
||||
|
@ -24,12 +16,12 @@ use crate::upstream::plain_docs_sentence;
|
|||
type TypeRepr = Option<(/* short */ String, /* long */ String)>;
|
||||
type ShowTypeRepr<'a> = &'a mut dyn FnMut(Option<&Ty>) -> TypeRepr;
|
||||
|
||||
/// Documentation about a symbol (without type information).
|
||||
/// Documentation about a definition (without type information).
|
||||
pub type UntypedDefDocs = DefDocsT<()>;
|
||||
/// Documentation about a symbol.
|
||||
/// Documentation about a definition.
|
||||
pub type DefDocs = DefDocsT<TypeRepr>;
|
||||
|
||||
/// Documentation about a symbol.
|
||||
/// Documentation about a definition.
|
||||
#[derive(Debug, Clone, Serialize, Deserialize)]
|
||||
#[serde(tag = "kind")]
|
||||
pub enum DefDocsT<T> {
|
||||
|
@ -263,7 +255,7 @@ fn format_ty(ty: Option<&Ty>, doc_ty: Option<&mut ShowTypeRepr>) -> TypeRepr {
|
|||
}
|
||||
}
|
||||
|
||||
pub(crate) fn variable_docs(ctx: &mut LocalContext, pos: Span) -> Option<VarDocs> {
|
||||
pub(crate) fn var_docs(ctx: &mut LocalContext, pos: Span) -> Option<VarDocs> {
|
||||
let source = ctx.source_by_id(pos.id()?).ok()?;
|
||||
let type_info = ctx.type_check(&source);
|
||||
let ty = type_info.type_of_span(pos)?;
|
||||
|
@ -299,10 +291,7 @@ pub(crate) fn variable_docs(ctx: &mut LocalContext, pos: Span) -> Option<VarDocs
|
|||
}
|
||||
}
|
||||
|
||||
pub(crate) fn signature_docs(
|
||||
sig: &Signature,
|
||||
mut doc_ty: Option<ShowTypeRepr>,
|
||||
) -> Option<SignatureDocs> {
|
||||
pub(crate) fn sig_docs(sig: &Signature, mut doc_ty: Option<ShowTypeRepr>) -> Option<SignatureDocs> {
|
||||
let type_sig = sig.type_sig().clone();
|
||||
|
||||
let pos_in = sig
|
||||
|
@ -344,34 +333,3 @@ pub(crate) fn signature_docs(
|
|||
hover_docs: OnceLock::new(),
|
||||
})
|
||||
}
|
||||
|
||||
// Unfortunately, we have only 65536 possible file ids and we cannot revoke
|
||||
// them. So we share a global file id for all docs conversion.
|
||||
static DOCS_CONVERT_ID: std::sync::LazyLock<Mutex<FileId>> = std::sync::LazyLock::new(|| {
|
||||
Mutex::new(FileId::new(None, VirtualPath::new("__tinymist_docs__.typ")))
|
||||
});
|
||||
|
||||
pub(crate) fn convert_docs(world: &LspWorld, content: &str) -> StrResult<EcoString> {
|
||||
static DOCS_LIB: std::sync::LazyLock<Arc<typlite::scopes::Scopes<typlite::value::Value>>> =
|
||||
std::sync::LazyLock::new(library::lib);
|
||||
|
||||
let conv_id = DOCS_CONVERT_ID.lock();
|
||||
let entry = EntryState::new_rootless(conv_id.vpath().as_rooted_path().into()).unwrap();
|
||||
let entry = entry.select_in_workspace(*conv_id);
|
||||
|
||||
let mut w = world.task(TaskInputs {
|
||||
entry: Some(entry),
|
||||
inputs: None,
|
||||
});
|
||||
w.map_shadow_by_id(*conv_id, Bytes::from(content.as_bytes().to_owned()))?;
|
||||
// todo: bad performance
|
||||
w.source_db.take_state();
|
||||
|
||||
let conv = typlite::Typlite::new(Arc::new(w))
|
||||
.with_library(DOCS_LIB.clone())
|
||||
.annotate_elements(true)
|
||||
.convert()
|
||||
.map_err(|e| eco_format!("failed to convert to markdown: {e}"))?;
|
||||
|
||||
Ok(conv.replace("```example", "```typ"))
|
||||
}
|
|
@ -1,24 +0,0 @@
|
|||
use std::sync::Arc;
|
||||
|
||||
use ecow::eco_format;
|
||||
use typlite::value::*;
|
||||
|
||||
pub(super) fn lib() -> Arc<typlite::scopes::Scopes<Value>> {
|
||||
let mut scopes = typlite::library::library();
|
||||
|
||||
// todo: how to import this function correctly?
|
||||
scopes.define("example", example as RawFunc);
|
||||
|
||||
Arc::new(scopes)
|
||||
}
|
||||
|
||||
/// Evaluate a `example`.
|
||||
pub fn example(mut args: Args) -> typlite::Result<Value> {
|
||||
let body = get_pos_named!(args, body: Content).0;
|
||||
let body = body.trim();
|
||||
let ticks = body.chars().take_while(|t| *t == '`').collect::<String>();
|
||||
let body = &body[ticks.len()..];
|
||||
let body = eco_format!("{ticks}typ{body}");
|
||||
|
||||
Ok(Value::Content(body))
|
||||
}
|
|
@ -1,17 +1,18 @@
|
|||
//! Documentation generation utilities.
|
||||
//! Documentation utilities.
|
||||
|
||||
mod library;
|
||||
mod convert;
|
||||
mod def;
|
||||
mod module;
|
||||
mod package;
|
||||
mod symbol;
|
||||
mod tidy;
|
||||
|
||||
use reflexo::path::unix_slash;
|
||||
use typst::syntax::FileId;
|
||||
|
||||
pub(crate) use convert::convert_docs;
|
||||
pub use def::*;
|
||||
pub use module::*;
|
||||
pub use package::*;
|
||||
pub use symbol::*;
|
||||
pub(crate) use tidy::*;
|
||||
|
||||
fn file_id_repr(k: FileId) -> String {
|
||||
|
|
|
@ -13,12 +13,12 @@ use typst::syntax::FileId;
|
|||
use crate::docs::file_id_repr;
|
||||
use crate::syntax::{Decl, DefKind, Expr, ExprInfo};
|
||||
use crate::ty::Interned;
|
||||
use crate::AnalysisContext;
|
||||
use crate::LocalContext;
|
||||
|
||||
use super::{get_manifest_id, DefDocs, PackageInfo};
|
||||
|
||||
/// Get documentation of symbols in a package.
|
||||
pub fn package_module_docs(ctx: &mut AnalysisContext, pkg: &PackageInfo) -> StrResult<SymbolsInfo> {
|
||||
/// Get documentation of definitions in a package.
|
||||
pub fn package_module_docs(ctx: &mut LocalContext, pkg: &PackageInfo) -> StrResult<PackageDefInfo> {
|
||||
let toml_id = get_manifest_id(pkg)?;
|
||||
let manifest = ctx.get_manifest(toml_id)?;
|
||||
|
||||
|
@ -26,12 +26,12 @@ pub fn package_module_docs(ctx: &mut AnalysisContext, pkg: &PackageInfo) -> StrR
|
|||
module_docs(ctx, entry_point)
|
||||
}
|
||||
|
||||
/// Get documentation of symbols in a module.
|
||||
pub fn module_docs(ctx: &mut AnalysisContext, entry_point: FileId) -> StrResult<SymbolsInfo> {
|
||||
/// Get documentation of definitions in a module.
|
||||
pub fn module_docs(ctx: &mut LocalContext, entry_point: FileId) -> StrResult<PackageDefInfo> {
|
||||
let mut aliases = HashMap::new();
|
||||
let mut extras = vec![];
|
||||
|
||||
let mut scan_ctx = ScanSymbolCtx {
|
||||
let mut scan_ctx = ScanDefCtx {
|
||||
ctx,
|
||||
root: entry_point,
|
||||
for_spec: entry_point.package(),
|
||||
|
@ -43,7 +43,7 @@ pub fn module_docs(ctx: &mut AnalysisContext, entry_point: FileId) -> StrResult<
|
|||
.ctx
|
||||
.expr_stage_by_id(entry_point)
|
||||
.ok_or("entry point not found")?;
|
||||
let mut symbols = scan_ctx.module_sym(eco_vec![], ei);
|
||||
let mut defs = scan_ctx.defs(eco_vec![], ei);
|
||||
|
||||
let module_uses = aliases
|
||||
.into_iter()
|
||||
|
@ -55,72 +55,64 @@ pub fn module_docs(ctx: &mut AnalysisContext, entry_point: FileId) -> StrResult<
|
|||
|
||||
log::debug!("module_uses: {module_uses:#?}",);
|
||||
|
||||
symbols.children.extend(extras);
|
||||
defs.children.extend(extras);
|
||||
|
||||
Ok(SymbolsInfo {
|
||||
root: symbols,
|
||||
Ok(PackageDefInfo {
|
||||
root: defs,
|
||||
module_uses,
|
||||
})
|
||||
}
|
||||
|
||||
/// Information about a symbol.
|
||||
/// Information about a definition.
|
||||
#[derive(Debug, Clone, Default, Serialize, Deserialize)]
|
||||
pub struct SymbolInfoHead {
|
||||
/// The name of the symbol.
|
||||
pub struct DefInfo {
|
||||
/// The name of the definition.
|
||||
pub name: EcoString,
|
||||
/// The kind of the symbol.
|
||||
/// The kind of the definition.
|
||||
pub kind: DefKind,
|
||||
/// The location (file, start, end) of the symbol.
|
||||
/// The location (file, start, end) of the definition.
|
||||
pub loc: Option<(usize, usize, usize)>,
|
||||
/// Is the symbol reexport
|
||||
pub export_again: bool,
|
||||
/// Is the symbol reexport
|
||||
/// Whether the definition external to the module.
|
||||
pub is_external: bool,
|
||||
/// The link to the definition if it is external.
|
||||
pub external_link: Option<String>,
|
||||
/// The one-line documentation of the symbol.
|
||||
/// The one-line documentation of the definition.
|
||||
pub oneliner: Option<String>,
|
||||
/// The raw documentation of the symbol.
|
||||
/// The raw documentation of the definition.
|
||||
pub docs: Option<EcoString>,
|
||||
/// The parsed documentation of the symbol.
|
||||
/// The parsed documentation of the definition.
|
||||
pub parsed_docs: Option<DefDocs>,
|
||||
/// The value of the symbol.
|
||||
/// The value of the definition.
|
||||
#[serde(skip)]
|
||||
pub constant: Option<EcoString>,
|
||||
/// The name range of the symbol.
|
||||
/// The value of the symbol.
|
||||
/// The name range of the definition.
|
||||
/// The value of the definition.
|
||||
#[serde(skip)]
|
||||
pub decl: Option<Interned<Decl>>,
|
||||
/// The children of the definition.
|
||||
pub children: EcoVec<DefInfo>,
|
||||
}
|
||||
|
||||
/// Information about a symbol.
|
||||
/// Information about the definitions in a package.
|
||||
#[derive(Debug, Clone, Serialize, Deserialize)]
|
||||
pub struct SymbolInfo {
|
||||
/// The primary information about the symbol.
|
||||
#[serde(flatten)]
|
||||
pub head: SymbolInfoHead,
|
||||
/// The children of the symbol.
|
||||
pub children: EcoVec<SymbolInfo>,
|
||||
}
|
||||
|
||||
/// Information about the symbols in a package.
|
||||
#[derive(Debug, Clone, Serialize, Deserialize)]
|
||||
pub struct SymbolsInfo {
|
||||
pub struct PackageDefInfo {
|
||||
/// The root module information.
|
||||
#[serde(flatten)]
|
||||
pub root: SymbolInfo,
|
||||
pub root: DefInfo,
|
||||
/// The module accessible paths.
|
||||
pub module_uses: HashMap<String, EcoVec<String>>,
|
||||
}
|
||||
|
||||
struct ScanSymbolCtx<'a, 'w> {
|
||||
ctx: &'a mut AnalysisContext<'w>,
|
||||
struct ScanDefCtx<'a> {
|
||||
ctx: &'a mut LocalContext,
|
||||
for_spec: Option<&'a PackageSpec>,
|
||||
aliases: &'a mut HashMap<FileId, Vec<String>>,
|
||||
extras: &'a mut Vec<SymbolInfo>,
|
||||
extras: &'a mut Vec<DefInfo>,
|
||||
root: FileId,
|
||||
}
|
||||
|
||||
impl ScanSymbolCtx<'_, '_> {
|
||||
fn module_sym(&mut self, path: EcoVec<&str>, ei: Arc<ExprInfo>) -> SymbolInfo {
|
||||
impl ScanDefCtx<'_> {
|
||||
fn defs(&mut self, path: EcoVec<&str>, ei: Arc<ExprInfo>) -> DefInfo {
|
||||
let name = {
|
||||
let stem = ei.fid.vpath().as_rooted_path().file_stem();
|
||||
stem.and_then(|s| Some(Interned::new_str(s.to_str()?)))
|
||||
|
@ -129,7 +121,7 @@ impl ScanSymbolCtx<'_, '_> {
|
|||
let module_decl = Decl::module(name.clone(), ei.fid).into();
|
||||
let site = Some(self.root);
|
||||
let p = path.clone();
|
||||
self.sym(&name, p, site.as_ref(), &module_decl, None)
|
||||
self.def(&name, p, site.as_ref(), &module_decl, None)
|
||||
}
|
||||
|
||||
fn expr(
|
||||
|
@ -138,9 +130,9 @@ impl ScanSymbolCtx<'_, '_> {
|
|||
path: EcoVec<&str>,
|
||||
site: Option<&FileId>,
|
||||
val: &Expr,
|
||||
) -> SymbolInfo {
|
||||
) -> DefInfo {
|
||||
match val {
|
||||
Expr::Decl(d) => self.sym(key, path, site, d, Some(val)),
|
||||
Expr::Decl(d) => self.def(key, path, site, d, Some(val)),
|
||||
Expr::Ref(r) if r.root.is_some() => {
|
||||
self.expr(key, path, site, r.root.as_ref().unwrap())
|
||||
}
|
||||
|
@ -148,55 +140,38 @@ impl ScanSymbolCtx<'_, '_> {
|
|||
Expr::Select(..) => {
|
||||
let mut path = path.clone();
|
||||
path.push(key);
|
||||
let head = SymbolInfoHead {
|
||||
DefInfo {
|
||||
name: key.to_string().into(),
|
||||
kind: DefKind::Module,
|
||||
..Default::default()
|
||||
};
|
||||
SymbolInfo {
|
||||
head,
|
||||
children: eco_vec![],
|
||||
}
|
||||
}
|
||||
// v => panic!("unexpected export: {key} -> {v}"),
|
||||
_ => {
|
||||
let mut path = path.clone();
|
||||
path.push(key);
|
||||
let head = SymbolInfoHead {
|
||||
DefInfo {
|
||||
name: key.to_string().into(),
|
||||
kind: DefKind::Constant,
|
||||
..Default::default()
|
||||
};
|
||||
SymbolInfo {
|
||||
head,
|
||||
children: eco_vec![],
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
fn sym(
|
||||
fn def(
|
||||
&mut self,
|
||||
key: &str,
|
||||
path: EcoVec<&str>,
|
||||
site: Option<&FileId>,
|
||||
val: &Interned<Decl>,
|
||||
decl: &Interned<Decl>,
|
||||
expr: Option<&Expr>,
|
||||
) -> SymbolInfo {
|
||||
let mut head = create_head(self.ctx, key, val, expr);
|
||||
|
||||
if !matches!(val.as_ref(), Decl::Module(..)) {
|
||||
if let Some((span, mod_fid)) = head.decl.as_ref().and_then(|d| d.file_id()).zip(site) {
|
||||
if span != *mod_fid {
|
||||
head.export_again = true;
|
||||
head.oneliner = head.docs.as_deref().map(oneliner).map(|e| e.to_owned());
|
||||
head.docs = None;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
let children = match val.as_ref() {
|
||||
Decl::Module(..) => val.file_id().and_then(|fid| {
|
||||
) -> DefInfo {
|
||||
let def = self.ctx.def_of_decl(decl);
|
||||
let def_docs = def.and_then(|def| self.ctx.def_docs(&def));
|
||||
let docs = def_docs.as_ref().map(|d| d.docs().clone());
|
||||
let children = match decl.as_ref() {
|
||||
Decl::Module(..) => decl.file_id().and_then(|fid| {
|
||||
// only generate docs for the same package
|
||||
if fid.package() != self.for_spec {
|
||||
return None;
|
||||
|
@ -230,6 +205,28 @@ impl ScanSymbolCtx<'_, '_> {
|
|||
_ => None,
|
||||
};
|
||||
|
||||
let mut head = DefInfo {
|
||||
name: key.to_string().into(),
|
||||
kind: decl.kind(),
|
||||
constant: expr.map(|e| e.repr()),
|
||||
docs,
|
||||
parsed_docs: def_docs,
|
||||
decl: Some(decl.clone()),
|
||||
children: children.unwrap_or_default(),
|
||||
loc: None,
|
||||
is_external: false,
|
||||
external_link: None,
|
||||
oneliner: None,
|
||||
};
|
||||
|
||||
if let Some((span, mod_fid)) = head.decl.as_ref().and_then(|d| d.file_id()).zip(site) {
|
||||
if span != *mod_fid {
|
||||
head.is_external = true;
|
||||
head.oneliner = head.docs.as_deref().map(oneliner).map(|e| e.to_owned());
|
||||
head.docs = None;
|
||||
}
|
||||
}
|
||||
|
||||
// Insert module that is not exported
|
||||
if let Some(fid) = head.decl.as_ref().and_then(|d| d.file_id()) {
|
||||
// only generate docs for the same package
|
||||
|
@ -243,37 +240,14 @@ impl ScanSymbolCtx<'_, '_> {
|
|||
|
||||
log::debug!("found internal module: {path:?}");
|
||||
if let Some(m) = src {
|
||||
let msym = self.module_sym(path, m);
|
||||
let msym = self.defs(path, m);
|
||||
self.extras.push(msym)
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
let children = children.unwrap_or_default();
|
||||
SymbolInfo { head, children }
|
||||
}
|
||||
}
|
||||
|
||||
fn create_head(
|
||||
ctx: &mut AnalysisContext,
|
||||
k: &str,
|
||||
decl: &Interned<Decl>,
|
||||
expr: Option<&Expr>,
|
||||
) -> SymbolInfoHead {
|
||||
let kind = decl.kind();
|
||||
|
||||
let parsed_docs = ctx.def_of_decl(decl).and_then(|def| ctx.def_docs(&def));
|
||||
let docs = parsed_docs.as_ref().map(|d| d.docs().clone());
|
||||
|
||||
SymbolInfoHead {
|
||||
name: k.to_string().into(),
|
||||
kind,
|
||||
constant: expr.map(|e| e.repr()),
|
||||
docs,
|
||||
parsed_docs,
|
||||
decl: Some(decl.clone()),
|
||||
..Default::default()
|
||||
head
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -10,11 +10,11 @@ use typst::syntax::package::{PackageManifest, PackageSpec};
|
|||
use typst::syntax::{FileId, Span, VirtualPath};
|
||||
use typst::World;
|
||||
|
||||
use crate::docs::{file_id_repr, module_docs, DefDocs, SymbolsInfo};
|
||||
use crate::AnalysisContext;
|
||||
use crate::docs::{file_id_repr, module_docs, DefDocs, PackageDefInfo};
|
||||
use crate::LocalContext;
|
||||
|
||||
/// Check Package.
|
||||
pub fn check_package(ctx: &mut AnalysisContext, spec: &PackageInfo) -> StrResult<()> {
|
||||
pub fn check_package(ctx: &mut LocalContext, spec: &PackageInfo) -> StrResult<()> {
|
||||
let toml_id = get_manifest_id(spec)?;
|
||||
let manifest = ctx.get_manifest(toml_id)?;
|
||||
|
||||
|
@ -25,7 +25,7 @@ pub fn check_package(ctx: &mut AnalysisContext, spec: &PackageInfo) -> StrResult
|
|||
}
|
||||
|
||||
/// Generate full documents in markdown format
|
||||
pub fn package_docs(ctx: &mut AnalysisContext, spec: &PackageInfo) -> StrResult<String> {
|
||||
pub fn package_docs(ctx: &mut LocalContext, spec: &PackageInfo) -> StrResult<String> {
|
||||
log::info!("generate_md_docs {spec:?}");
|
||||
|
||||
let mut md = String::new();
|
||||
|
@ -37,7 +37,7 @@ pub fn package_docs(ctx: &mut AnalysisContext, spec: &PackageInfo) -> StrResult<
|
|||
|
||||
ctx.preload_package(entry_point);
|
||||
|
||||
let SymbolsInfo { root, module_uses } = module_docs(ctx, entry_point)?;
|
||||
let PackageDefInfo { root, module_uses } = module_docs(ctx, entry_point)?;
|
||||
|
||||
log::debug!("module_uses: {module_uses:#?}");
|
||||
|
||||
|
@ -62,7 +62,7 @@ pub fn package_docs(ctx: &mut AnalysisContext, spec: &PackageInfo) -> StrResult<
|
|||
let package_meta = jbase64(&meta);
|
||||
let _ = writeln!(md, "<!-- begin:package {package_meta} -->");
|
||||
|
||||
let mut modules_to_generate = vec![(root.head.name.clone(), root)];
|
||||
let mut modules_to_generate = vec![(root.name.clone(), root)];
|
||||
let mut generated_modules = HashSet::new();
|
||||
let mut file_ids: IndexSet<FileId> = IndexSet::new();
|
||||
|
||||
|
@ -82,11 +82,11 @@ pub fn package_docs(ctx: &mut AnalysisContext, spec: &PackageInfo) -> StrResult<
|
|||
};
|
||||
|
||||
while !modules_to_generate.is_empty() {
|
||||
for (parent_ident, sym) in std::mem::take(&mut modules_to_generate) {
|
||||
for (parent_ident, def) in std::mem::take(&mut modules_to_generate) {
|
||||
// parent_ident, symbols
|
||||
let symbols = sym.children;
|
||||
let children = def.children;
|
||||
|
||||
let module_val = sym.head.decl.as_ref().unwrap();
|
||||
let module_val = def.decl.as_ref().unwrap();
|
||||
let fid = module_val.file_id();
|
||||
let aka = fid.map(&mut akas).unwrap_or_default();
|
||||
|
||||
|
@ -111,15 +111,15 @@ pub fn package_docs(ctx: &mut AnalysisContext, spec: &PackageInfo) -> StrResult<
|
|||
}
|
||||
let m = jbase64(&ModuleInfo {
|
||||
prefix: primary.as_str().into(),
|
||||
name: sym.head.name.clone(),
|
||||
name: def.name.clone(),
|
||||
loc: persist_fid,
|
||||
parent_ident: parent_ident.clone(),
|
||||
aka,
|
||||
});
|
||||
let _ = writeln!(md, "<!-- begin:module {primary} {m} -->");
|
||||
|
||||
for mut sym in symbols {
|
||||
let span = sym.head.decl.as_ref().map(|d| d.span());
|
||||
for mut child in children {
|
||||
let span = child.decl.as_ref().map(|d| d.span());
|
||||
let fid_range = span.and_then(|v| {
|
||||
v.id().and_then(|e| {
|
||||
let fid = file_ids.insert_full(e).0;
|
||||
|
@ -128,13 +128,13 @@ pub fn package_docs(ctx: &mut AnalysisContext, spec: &PackageInfo) -> StrResult<
|
|||
Some((fid, rng.start, rng.end))
|
||||
})
|
||||
});
|
||||
let sym_fid = sym.head.decl.as_ref().and_then(|d| d.file_id());
|
||||
let sym_fid = sym_fid.or_else(|| span.and_then(Span::id)).or(fid);
|
||||
let child_fid = child.decl.as_ref().and_then(|d| d.file_id());
|
||||
let child_fid = child_fid.or_else(|| span.and_then(Span::id)).or(fid);
|
||||
let span = fid_range.or_else(|| {
|
||||
let fid = sym_fid?;
|
||||
let fid = child_fid?;
|
||||
Some((file_ids.insert_full(fid).0, 0, 0))
|
||||
});
|
||||
sym.head.loc = span;
|
||||
child.loc = span;
|
||||
// .ok_or_else(|| {
|
||||
// let err = format!("failed to convert docs in {title}").replace(
|
||||
// "-->", "—>", // avoid markdown comment
|
||||
|
@ -142,7 +142,7 @@ pub fn package_docs(ctx: &mut AnalysisContext, spec: &PackageInfo) -> StrResult<
|
|||
// log::error!("{err}");
|
||||
// err
|
||||
// })
|
||||
let docs = sym.head.parsed_docs.clone();
|
||||
let docs = child.parsed_docs.clone();
|
||||
// Err(e) => {
|
||||
// let err = format!("failed to convert docs: {e}").replace(
|
||||
// "-->", "—>", // avoid markdown comment
|
||||
|
@ -154,8 +154,8 @@ pub fn package_docs(ctx: &mut AnalysisContext, spec: &PackageInfo) -> StrResult<
|
|||
let convert_err = None::<EcoString>;
|
||||
match &docs {
|
||||
Some(docs) => {
|
||||
sym.head.parsed_docs = Some(docs.clone());
|
||||
sym.head.docs = None;
|
||||
child.parsed_docs = Some(docs.clone());
|
||||
child.docs = None;
|
||||
}
|
||||
None => {
|
||||
// let err = format!("failed to convert docs in {title}:
|
||||
|
@ -168,53 +168,54 @@ pub fn package_docs(ctx: &mut AnalysisContext, spec: &PackageInfo) -> StrResult<
|
|||
}
|
||||
|
||||
let ident = if !primary.is_empty() {
|
||||
eco_format!("symbol-{}-{primary}.{}", sym.head.kind, sym.head.name)
|
||||
eco_format!("symbol-{}-{primary}.{}", child.kind, child.name)
|
||||
} else {
|
||||
eco_format!("symbol-{}-{}", sym.head.kind, sym.head.name)
|
||||
eco_format!("symbol-{}-{}", child.kind, child.name)
|
||||
};
|
||||
let _ = writeln!(md, "### {}: {} in {primary}", sym.head.kind, sym.head.name);
|
||||
let _ = writeln!(md, "### {}: {} in {primary}", child.kind, child.name);
|
||||
|
||||
if sym.head.export_again {
|
||||
if let Some(fid) = sym_fid {
|
||||
if child.is_external {
|
||||
if let Some(fid) = child_fid {
|
||||
let lnk = if fid.package() == Some(for_spec) {
|
||||
let sub_aka = akas(fid);
|
||||
let sub_primary = sub_aka.first().cloned().unwrap_or_default();
|
||||
sym.head.external_link = Some(format!(
|
||||
child.external_link = Some(format!(
|
||||
"#symbol-{}-{sub_primary}.{}",
|
||||
sym.head.kind, sym.head.name
|
||||
child.kind, child.name
|
||||
));
|
||||
format!("#{}-{}-in-{sub_primary}", sym.head.kind, sym.head.name)
|
||||
format!("#{}-{}-in-{sub_primary}", child.kind, child.name)
|
||||
.replace(".", "")
|
||||
} else if let Some(spec) = fid.package() {
|
||||
let lnk = format!(
|
||||
"https://typst.app/universe/package/{}/{}",
|
||||
spec.name, spec.version
|
||||
);
|
||||
sym.head.external_link = Some(lnk.clone());
|
||||
child.external_link = Some(lnk.clone());
|
||||
lnk
|
||||
} else {
|
||||
let lnk: String = "https://typst.app/docs".into();
|
||||
sym.head.external_link = Some(lnk.clone());
|
||||
child.external_link = Some(lnk.clone());
|
||||
lnk
|
||||
};
|
||||
let _ = writeln!(md, "[Symbol Docs]({lnk})\n");
|
||||
}
|
||||
}
|
||||
|
||||
let head = jbase64(&sym.head);
|
||||
let child_children = std::mem::take(&mut child.children);
|
||||
let head = jbase64(&child);
|
||||
let _ = writeln!(md, "<!-- begin:symbol {ident} {head} -->");
|
||||
|
||||
if let Some(DefDocs::Function(sig)) = &sym.head.parsed_docs {
|
||||
if let Some(DefDocs::Function(sig)) = &child.parsed_docs {
|
||||
let _ = writeln!(md, "<!-- begin:sig -->");
|
||||
let _ = writeln!(md, "```typc");
|
||||
let _ = write!(md, "let {}", sym.head.name);
|
||||
let _ = write!(md, "let {}", child.name);
|
||||
let _ = sig.print(&mut md);
|
||||
let _ = writeln!(md, ";");
|
||||
let _ = writeln!(md, "```");
|
||||
let _ = writeln!(md, "<!-- end:sig -->");
|
||||
}
|
||||
|
||||
match (&sym.head.parsed_docs, convert_err) {
|
||||
match (&child.parsed_docs, convert_err) {
|
||||
(_, Some(err)) => {
|
||||
let err = format!("failed to convert docs in {title}: {err}").replace(
|
||||
"-->", "—>", // avoid markdown comment
|
||||
|
@ -244,8 +245,8 @@ pub fn package_docs(ctx: &mut AnalysisContext, spec: &PackageInfo) -> StrResult<
|
|||
(None, None) => {}
|
||||
}
|
||||
|
||||
let plain_docs = sym.head.docs.as_deref();
|
||||
let plain_docs = plain_docs.or(sym.head.oneliner.as_deref());
|
||||
let plain_docs = child.docs.as_deref();
|
||||
let plain_docs = plain_docs.or(child.oneliner.as_deref());
|
||||
|
||||
if let Some(docs) = plain_docs {
|
||||
let contains_code = docs.contains("```");
|
||||
|
@ -258,9 +259,9 @@ pub fn package_docs(ctx: &mut AnalysisContext, spec: &PackageInfo) -> StrResult<
|
|||
}
|
||||
}
|
||||
|
||||
if !sym.children.is_empty() {
|
||||
log::debug!("sub_fid: {sym_fid:?}");
|
||||
match sym_fid {
|
||||
if !child_children.is_empty() {
|
||||
log::debug!("sub_fid: {child_fid:?}");
|
||||
match child_fid {
|
||||
Some(fid) => {
|
||||
let aka = akas(fid);
|
||||
let primary = aka.first().cloned().unwrap_or_default();
|
||||
|
@ -268,7 +269,8 @@ pub fn package_docs(ctx: &mut AnalysisContext, spec: &PackageInfo) -> StrResult<
|
|||
let _ = writeln!(md, "[Module Docs](#{link})\n");
|
||||
|
||||
if generated_modules.insert(fid) {
|
||||
modules_to_generate.push((ident.clone(), sym));
|
||||
child.children = child_children;
|
||||
modules_to_generate.push((ident.clone(), child));
|
||||
}
|
||||
}
|
||||
None => {
|
||||
|
|
|
@ -24,7 +24,7 @@ pub struct DocumentColorRequest {
|
|||
impl SemanticRequest for DocumentColorRequest {
|
||||
type Response = Vec<ColorInformation>;
|
||||
|
||||
fn request(self, ctx: &mut AnalysisContext) -> Option<Self::Response> {
|
||||
fn request(self, ctx: &mut LocalContext) -> Option<Self::Response> {
|
||||
let source = ctx.source_by_path(&self.path).ok()?;
|
||||
get_color_exprs(ctx, &source)
|
||||
}
|
||||
|
|
|
@ -16,7 +16,7 @@ pub struct DocumentHighlightRequest {
|
|||
impl SemanticRequest for DocumentHighlightRequest {
|
||||
type Response = Vec<DocumentHighlight>;
|
||||
|
||||
fn request(self, ctx: &mut AnalysisContext) -> Option<Self::Response> {
|
||||
fn request(self, ctx: &mut LocalContext) -> Option<Self::Response> {
|
||||
let source = ctx.source_by_path(&self.path).ok()?;
|
||||
let cursor = ctx.to_typst_pos(self.position, &source)?;
|
||||
|
||||
|
@ -44,15 +44,15 @@ impl SemanticRequest for DocumentHighlightRequest {
|
|||
}
|
||||
}
|
||||
|
||||
struct DocumentHighlightWorker<'a, 'w> {
|
||||
ctx: &'a mut AnalysisContext<'w>,
|
||||
struct DocumentHighlightWorker<'a> {
|
||||
ctx: &'a mut LocalContext,
|
||||
current: &'a Source,
|
||||
highlights: Vec<DocumentHighlight>,
|
||||
worklist: Vec<LinkedNode<'a>>,
|
||||
}
|
||||
|
||||
impl<'a, 'w> DocumentHighlightWorker<'a, 'w> {
|
||||
fn new(ctx: &'a mut AnalysisContext<'w>, current: &'a Source) -> Self {
|
||||
impl<'a> DocumentHighlightWorker<'a> {
|
||||
fn new(ctx: &'a mut LocalContext, current: &'a Source) -> Self {
|
||||
Self {
|
||||
ctx,
|
||||
current,
|
||||
|
@ -138,7 +138,7 @@ impl<'a, 'w> DocumentHighlightWorker<'a, 'w> {
|
|||
}
|
||||
|
||||
fn highlight_func_returns(
|
||||
ctx: &mut AnalysisContext,
|
||||
ctx: &mut LocalContext,
|
||||
node: &LinkedNode,
|
||||
) -> Option<Vec<DocumentHighlight>> {
|
||||
let _ = ctx;
|
||||
|
|
|
@ -21,7 +21,7 @@ pub struct DocumentLinkRequest {
|
|||
impl SemanticRequest for DocumentLinkRequest {
|
||||
type Response = Vec<DocumentLink>;
|
||||
|
||||
fn request(self, ctx: &mut AnalysisContext) -> Option<Self::Response> {
|
||||
fn request(self, ctx: &mut LocalContext) -> Option<Self::Response> {
|
||||
let source = ctx.source_by_path(&self.path).ok()?;
|
||||
let links = get_link_exprs(ctx, &source);
|
||||
links.map(|links| {
|
||||
|
|
|
@ -100,7 +100,7 @@ impl StatefulRequest for DocumentMetricsRequest {
|
|||
|
||||
fn request(
|
||||
self,
|
||||
ctx: &mut AnalysisContext,
|
||||
ctx: &mut LocalContext,
|
||||
doc: Option<VersionedDocument>,
|
||||
) -> Option<Self::Response> {
|
||||
let doc = doc?;
|
||||
|
@ -134,14 +134,14 @@ struct FontInfoValue {
|
|||
first_occur_column: Option<u32>,
|
||||
}
|
||||
|
||||
struct DocumentMetricsWorker<'a, 'w> {
|
||||
ctx: &'a mut AnalysisContext<'w>,
|
||||
struct DocumentMetricsWorker<'a> {
|
||||
ctx: &'a mut LocalContext,
|
||||
span_info: HashMap<Arc<DataSource>, u32>,
|
||||
span_info2: Vec<DataSource>,
|
||||
font_info: HashMap<Font, FontInfoValue>,
|
||||
}
|
||||
|
||||
impl<'a, 'w> DocumentMetricsWorker<'a, 'w> {
|
||||
impl<'a> DocumentMetricsWorker<'a> {
|
||||
fn work(&mut self, doc: &Document) -> Option<()> {
|
||||
for page in &doc.pages {
|
||||
self.work_frame(&page.frame)?;
|
||||
|
|
|
@ -30,14 +30,14 @@ pub struct GotoDeclarationRequest {
|
|||
impl SemanticRequest for GotoDeclarationRequest {
|
||||
type Response = GotoDeclarationResponse;
|
||||
|
||||
fn request(self, _ctx: &mut AnalysisContext) -> Option<Self::Response> {
|
||||
fn request(self, _ctx: &mut LocalContext) -> Option<Self::Response> {
|
||||
let _ = find_declarations;
|
||||
todo!()
|
||||
}
|
||||
}
|
||||
|
||||
fn find_declarations(
|
||||
_ctx: &AnalysisContext,
|
||||
_ctx: &LocalContext,
|
||||
_expr_info: Arc<crate::syntax::ExprInfo>,
|
||||
_deref_target: DerefTarget<'_>,
|
||||
) -> Option<Vec<Range<usize>>> {
|
||||
|
|
|
@ -28,7 +28,7 @@ impl StatefulRequest for GotoDefinitionRequest {
|
|||
|
||||
fn request(
|
||||
self,
|
||||
ctx: &mut AnalysisContext,
|
||||
ctx: &mut LocalContext,
|
||||
doc: Option<VersionedDocument>,
|
||||
) -> Option<Self::Response> {
|
||||
let source = ctx.source_by_path(&self.path).ok()?;
|
||||
|
@ -38,9 +38,7 @@ impl StatefulRequest for GotoDefinitionRequest {
|
|||
let def = ctx.def_of_syntax(&source, doc.as_ref(), deref_target)?;
|
||||
|
||||
let (fid, def_range) = def.def_at(ctx.shared())?;
|
||||
|
||||
let uri = ctx.uri_for_id(fid).ok()?;
|
||||
|
||||
let range = ctx.to_lsp_range_(def_range, fid)?;
|
||||
|
||||
let res = Some(GotoDefinitionResponse::Link(vec![LocationLink {
|
||||
|
|
|
@ -1,17 +1,12 @@
|
|||
use core::fmt::Write;
|
||||
use core::fmt::{self, Write};
|
||||
|
||||
use typst::foundations::repr::separated_list;
|
||||
use typst_shim::syntax::LinkedNodeExt;
|
||||
|
||||
use crate::{
|
||||
analysis::{get_link_exprs_in, Definition},
|
||||
docs::DefDocs,
|
||||
jump_from_cursor,
|
||||
prelude::*,
|
||||
syntax::{get_deref_target, Decl, DefKind},
|
||||
ty::PathPreference,
|
||||
upstream::{expr_tooltip, route_of_value, truncated_repr, Tooltip},
|
||||
LspHoverContents, StatefulRequest,
|
||||
};
|
||||
use crate::analysis::get_link_exprs_in;
|
||||
use crate::jump_from_cursor;
|
||||
use crate::prelude::*;
|
||||
use crate::upstream::{expr_tooltip, route_of_value, truncated_repr, Tooltip};
|
||||
|
||||
/// The [`textDocument/hover`] request asks the server for hover information at
|
||||
/// a given text document position.
|
||||
|
@ -33,7 +28,7 @@ impl StatefulRequest for HoverRequest {
|
|||
|
||||
fn request(
|
||||
self,
|
||||
ctx: &mut AnalysisContext,
|
||||
ctx: &mut LocalContext,
|
||||
doc: Option<VersionedDocument>,
|
||||
) -> Option<Self::Response> {
|
||||
let doc_ref = doc.as_ref().map(|doc| doc.document.as_ref());
|
||||
|
@ -43,23 +38,18 @@ impl StatefulRequest for HoverRequest {
|
|||
// the typst's cursor is 1-based, so we need to add 1 to the offset
|
||||
let cursor = offset + 1;
|
||||
|
||||
let node = LinkedNode::new(source.root()).leaf_at_compat(cursor)?;
|
||||
let range = ctx.to_lsp_range(node.range(), &source);
|
||||
|
||||
let contents = def_tooltip(ctx, &source, doc.as_ref(), cursor)
|
||||
.or_else(|| star_tooltip(ctx, &source, cursor))
|
||||
.or_else(|| link_tooltip(ctx, &source, cursor));
|
||||
|
||||
let contents = contents.or_else(|| {
|
||||
Some(typst_to_lsp::tooltip(
|
||||
&ctx.tooltip(doc_ref, &source, cursor)?,
|
||||
))
|
||||
})?;
|
||||
|
||||
let ast_node = LinkedNode::new(source.root()).leaf_at_compat(cursor)?;
|
||||
let range = ctx.to_lsp_range(ast_node.range(), &source);
|
||||
.or_else(|| star_tooltip(ctx, &node))
|
||||
.or_else(|| link_tooltip(ctx, &node, cursor))
|
||||
.or_else(|| Some(to_lsp_tooltip(&ctx.tooltip(doc_ref, &source, cursor)?)))?;
|
||||
|
||||
// Neovim shows ugly hover if the hover content is in array, so we join them
|
||||
// manually with divider bars.
|
||||
let mut contents = match contents {
|
||||
LspHoverContents::Array(contents) => contents
|
||||
HoverContents::Array(contents) => contents
|
||||
.into_iter()
|
||||
.map(|e| match e {
|
||||
MarkedString::LanguageString(e) => {
|
||||
|
@ -68,8 +58,8 @@ impl StatefulRequest for HoverRequest {
|
|||
MarkedString::String(e) => e,
|
||||
})
|
||||
.join("\n\n---\n"),
|
||||
LspHoverContents::Scalar(MarkedString::String(contents)) => contents,
|
||||
LspHoverContents::Scalar(MarkedString::LanguageString(contents)) => {
|
||||
HoverContents::Scalar(MarkedString::String(contents)) => contents,
|
||||
HoverContents::Scalar(MarkedString::LanguageString(contents)) => {
|
||||
format!("```{}\n{}\n```", contents.language, contents.value)
|
||||
}
|
||||
lsp_types::HoverContents::Markup(e) => {
|
||||
|
@ -81,7 +71,7 @@ impl StatefulRequest for HoverRequest {
|
|||
}
|
||||
};
|
||||
|
||||
if ctx.analysis.enable_periscope {
|
||||
if let Some(p) = ctx.analysis.periscope.clone() {
|
||||
if let Some(doc) = doc.clone() {
|
||||
let position = jump_from_cursor(&doc.document, &source, cursor);
|
||||
let position = position.or_else(|| {
|
||||
|
@ -106,7 +96,7 @@ impl StatefulRequest for HoverRequest {
|
|||
});
|
||||
|
||||
log::info!("telescope position: {:?}", position);
|
||||
let content = position.and_then(|pos| ctx.resources.periscope_at(ctx, doc, pos));
|
||||
let content = position.and_then(|pos| p.periscope_at(ctx, doc, pos));
|
||||
if let Some(preview_content) = content {
|
||||
contents = format!("{preview_content}\n---\n{contents}");
|
||||
}
|
||||
|
@ -114,114 +104,18 @@ impl StatefulRequest for HoverRequest {
|
|||
}
|
||||
|
||||
Some(Hover {
|
||||
contents: LspHoverContents::Scalar(MarkedString::String(contents)),
|
||||
contents: HoverContents::Scalar(MarkedString::String(contents)),
|
||||
range: Some(range),
|
||||
})
|
||||
}
|
||||
}
|
||||
|
||||
fn link_tooltip(
|
||||
ctx: &mut AnalysisContext<'_>,
|
||||
source: &Source,
|
||||
cursor: usize,
|
||||
) -> Option<HoverContents> {
|
||||
let mut node = LinkedNode::new(source.root()).leaf_at_compat(cursor)?;
|
||||
while !matches!(node.kind(), SyntaxKind::FuncCall) {
|
||||
node = node.parent()?.clone();
|
||||
}
|
||||
|
||||
let mut links = get_link_exprs_in(ctx, &node)?;
|
||||
links.retain(|link| link.0.contains(&cursor));
|
||||
if links.is_empty() {
|
||||
return None;
|
||||
}
|
||||
|
||||
let mut results = vec![];
|
||||
let mut actions = vec![];
|
||||
for (_, target) in links {
|
||||
// open file in tab or system application
|
||||
actions.push(CommandLink {
|
||||
title: Some("Open in Tab".to_string()),
|
||||
command_or_links: vec![CommandOrLink::Command(Command {
|
||||
id: "tinymist.openInternal".to_string(),
|
||||
args: vec![JsonValue::String(target.to_string())],
|
||||
})],
|
||||
});
|
||||
actions.push(CommandLink {
|
||||
title: Some("Open Externally".to_string()),
|
||||
command_or_links: vec![CommandOrLink::Command(Command {
|
||||
id: "tinymist.openExternal".to_string(),
|
||||
args: vec![JsonValue::String(target.to_string())],
|
||||
})],
|
||||
});
|
||||
if let Some(kind) = PathPreference::from_ext(target.path()) {
|
||||
let preview = format!("A `{kind:?}` file.");
|
||||
results.push(MarkedString::String(preview));
|
||||
}
|
||||
}
|
||||
render_actions(&mut results, actions);
|
||||
if results.is_empty() {
|
||||
return None;
|
||||
}
|
||||
|
||||
Some(LspHoverContents::Array(results))
|
||||
}
|
||||
|
||||
fn star_tooltip(
|
||||
ctx: &mut AnalysisContext,
|
||||
source: &Source,
|
||||
cursor: usize,
|
||||
) -> Option<HoverContents> {
|
||||
let leaf = LinkedNode::new(source.root()).leaf_at_compat(cursor)?;
|
||||
|
||||
if !matches!(leaf.kind(), SyntaxKind::Star) {
|
||||
return None;
|
||||
}
|
||||
|
||||
let mut leaf = &leaf;
|
||||
while !matches!(leaf.kind(), SyntaxKind::ModuleImport) {
|
||||
leaf = leaf.parent()?;
|
||||
}
|
||||
|
||||
let mi: ast::ModuleImport = leaf.cast()?;
|
||||
let source = mi.source();
|
||||
let module = ctx.analyze_import(source.to_untyped()).1;
|
||||
log::debug!("star import: {source:?} => {:?}", module.is_some());
|
||||
|
||||
let i = module?;
|
||||
let scope = i.scope()?;
|
||||
|
||||
let mut results = vec![];
|
||||
|
||||
let mut names = scope.iter().map(|(name, _, _)| name).collect::<Vec<_>>();
|
||||
names.sort();
|
||||
let items = typst::foundations::repr::separated_list(&names, "and");
|
||||
|
||||
results.push(MarkedString::String(format!("This star imports {items}")));
|
||||
Some(LspHoverContents::Array(results))
|
||||
}
|
||||
|
||||
struct Command {
|
||||
id: String,
|
||||
args: Vec<JsonValue>,
|
||||
}
|
||||
|
||||
enum CommandOrLink {
|
||||
Link(String),
|
||||
Command(Command),
|
||||
}
|
||||
|
||||
struct CommandLink {
|
||||
title: Option<String>,
|
||||
command_or_links: Vec<CommandOrLink>,
|
||||
}
|
||||
|
||||
fn def_tooltip(
|
||||
ctx: &mut AnalysisContext,
|
||||
ctx: &mut LocalContext,
|
||||
source: &Source,
|
||||
document: Option<&VersionedDocument>,
|
||||
cursor: usize,
|
||||
) -> Option<LspHoverContents> {
|
||||
) -> Option<HoverContents> {
|
||||
let leaf = LinkedNode::new(source.root()).leaf_at_compat(cursor)?;
|
||||
let deref_target = get_deref_target(leaf.clone(), cursor)?;
|
||||
let def = ctx.def_of_syntax(source, document, deref_target.clone())?;
|
||||
|
@ -238,7 +132,7 @@ fn def_tooltip(
|
|||
let c = truncated_repr(&c);
|
||||
results.push(MarkedString::String(format!("{c}")));
|
||||
}
|
||||
Some(LspHoverContents::Array(results))
|
||||
Some(HoverContents::Array(results))
|
||||
}
|
||||
BibEntry(..) => {
|
||||
results.push(MarkedString::String(format!(
|
||||
|
@ -246,7 +140,7 @@ fn def_tooltip(
|
|||
def.name()
|
||||
)));
|
||||
|
||||
Some(LspHoverContents::Array(results))
|
||||
Some(HoverContents::Array(results))
|
||||
}
|
||||
_ => {
|
||||
let sym_docs = ctx.def_docs(&def);
|
||||
|
@ -308,16 +202,82 @@ fn def_tooltip(
|
|||
// results.push(MarkedString::String(doc));
|
||||
// }
|
||||
|
||||
if let Some(link) = ExternalDocLink::get(ctx, &def) {
|
||||
if let Some(link) = ExternalDocLink::get(&def) {
|
||||
actions.push(link);
|
||||
}
|
||||
|
||||
render_actions(&mut results, actions);
|
||||
Some(LspHoverContents::Array(results))
|
||||
Some(HoverContents::Array(results))
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
fn star_tooltip(ctx: &mut LocalContext, mut node: &LinkedNode) -> Option<HoverContents> {
|
||||
if !matches!(node.kind(), SyntaxKind::Star) {
|
||||
return None;
|
||||
}
|
||||
|
||||
while !matches!(node.kind(), SyntaxKind::ModuleImport) {
|
||||
node = node.parent()?;
|
||||
}
|
||||
|
||||
let import_node = node.cast::<ast::ModuleImport>()?;
|
||||
let scope_val = ctx.analyze_import(import_node.source().to_untyped()).1?;
|
||||
|
||||
let scope_items = scope_val.scope()?.iter();
|
||||
let mut names = scope_items.map(|item| item.0.as_str()).collect::<Vec<_>>();
|
||||
names.sort();
|
||||
|
||||
let content = format!("This star imports {}", separated_list(&names, "and"));
|
||||
Some(HoverContents::Scalar(MarkedString::String(content)))
|
||||
}
|
||||
|
||||
fn link_tooltip(
|
||||
ctx: &mut LocalContext,
|
||||
mut node: &LinkedNode,
|
||||
cursor: usize,
|
||||
) -> Option<HoverContents> {
|
||||
while !matches!(node.kind(), SyntaxKind::FuncCall) {
|
||||
node = node.parent()?;
|
||||
}
|
||||
|
||||
let mut links = get_link_exprs_in(ctx, node)?;
|
||||
links.retain(|link| link.0.contains(&cursor));
|
||||
if links.is_empty() {
|
||||
return None;
|
||||
}
|
||||
|
||||
let mut results = vec![];
|
||||
let mut actions = vec![];
|
||||
for (_, target) in links {
|
||||
// open file in tab or system application
|
||||
actions.push(CommandLink {
|
||||
title: Some("Open in Tab".to_string()),
|
||||
command_or_links: vec![CommandOrLink::Command {
|
||||
id: "tinymist.openInternal".to_string(),
|
||||
args: vec![JsonValue::String(target.to_string())],
|
||||
}],
|
||||
});
|
||||
actions.push(CommandLink {
|
||||
title: Some("Open Externally".to_string()),
|
||||
command_or_links: vec![CommandOrLink::Command {
|
||||
id: "tinymist.openExternal".to_string(),
|
||||
args: vec![JsonValue::String(target.to_string())],
|
||||
}],
|
||||
});
|
||||
if let Some(kind) = PathPreference::from_ext(target.path()) {
|
||||
let preview = format!("A `{kind:?}` file.");
|
||||
results.push(MarkedString::String(preview));
|
||||
}
|
||||
}
|
||||
render_actions(&mut results, actions);
|
||||
if results.is_empty() {
|
||||
return None;
|
||||
}
|
||||
|
||||
Some(HoverContents::Array(results))
|
||||
}
|
||||
|
||||
fn push_result_ty(name: &str, ty_repr: Option<&(String, String)>, type_doc: &mut String) {
|
||||
let Some((short, _)) = ty_repr else {
|
||||
return;
|
||||
|
@ -334,48 +294,14 @@ fn render_actions(results: &mut Vec<MarkedString>, actions: Vec<CommandLink>) {
|
|||
return;
|
||||
}
|
||||
|
||||
let g = actions
|
||||
.into_iter()
|
||||
.map(|action| {
|
||||
// https://github.com/rust-lang/rust-analyzer/blob/1a5bb27c018c947dab01ab70ffe1d267b0481a17/editors/code/src/client.ts#L59
|
||||
let title = action.title.unwrap_or("".to_owned());
|
||||
let command_or_links = action
|
||||
.command_or_links
|
||||
.into_iter()
|
||||
.map(|col| match col {
|
||||
CommandOrLink::Link(link) => link,
|
||||
CommandOrLink::Command(command) => {
|
||||
let id = command.id;
|
||||
// <https://code.visualstudio.com/api/extension-guides/command#command-uris>
|
||||
if command.args.is_empty() {
|
||||
format!("command:{id}")
|
||||
} else {
|
||||
let args = serde_json::to_string(&command.args).unwrap();
|
||||
let args = percent_encoding::utf8_percent_encode(
|
||||
&args,
|
||||
percent_encoding::NON_ALPHANUMERIC,
|
||||
);
|
||||
format!("command:{id}?{args}")
|
||||
}
|
||||
}
|
||||
})
|
||||
.collect::<Vec<_>>()
|
||||
.join(" ");
|
||||
format!("[{title}]({command_or_links})")
|
||||
})
|
||||
.collect::<Vec<_>>()
|
||||
.join(" | ");
|
||||
let g = actions.into_iter().join(" | ");
|
||||
results.push(MarkedString::String(g));
|
||||
}
|
||||
|
||||
struct ExternalDocLink;
|
||||
|
||||
impl ExternalDocLink {
|
||||
fn get(ctx: &mut AnalysisContext, def: &Definition) -> Option<CommandLink> {
|
||||
self::ExternalDocLink::get_inner(ctx, def)
|
||||
}
|
||||
|
||||
fn get_inner(_ctx: &mut AnalysisContext, def: &Definition) -> Option<CommandLink> {
|
||||
fn get(def: &Definition) -> Option<CommandLink> {
|
||||
let value = def.value();
|
||||
|
||||
if matches!(value, Some(Value::Func(..))) {
|
||||
|
@ -386,9 +312,7 @@ impl ExternalDocLink {
|
|||
|
||||
value.and_then(|value| Self::builtin_value_tooltip("https://typst.app/docs/", &value))
|
||||
}
|
||||
}
|
||||
|
||||
impl ExternalDocLink {
|
||||
fn builtin_func_tooltip(base: &str, def: &Definition) -> Option<CommandLink> {
|
||||
let Some(Value::Func(func)) = def.value() else {
|
||||
return None;
|
||||
|
@ -422,6 +346,57 @@ impl ExternalDocLink {
|
|||
}
|
||||
}
|
||||
|
||||
struct CommandLink {
|
||||
title: Option<String>,
|
||||
command_or_links: Vec<CommandOrLink>,
|
||||
}
|
||||
|
||||
impl fmt::Display for CommandLink {
|
||||
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
||||
// https://github.com/rust-lang/rust-analyzer/blob/1a5bb27c018c947dab01ab70ffe1d267b0481a17/editors/code/src/client.ts#L59
|
||||
let title = self.title.as_deref().unwrap_or("");
|
||||
let command_or_links = self.command_or_links.iter().join(" ");
|
||||
write!(f, "[{title}]({command_or_links})")
|
||||
}
|
||||
}
|
||||
|
||||
enum CommandOrLink {
|
||||
Link(String),
|
||||
Command { id: String, args: Vec<JsonValue> },
|
||||
}
|
||||
|
||||
impl fmt::Display for CommandOrLink {
|
||||
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
||||
match self {
|
||||
Self::Link(link) => f.write_str(link),
|
||||
Self::Command { id, args } => {
|
||||
// <https://code.visualstudio.com/api/extension-guides/command#command-uris>
|
||||
if args.is_empty() {
|
||||
return write!(f, "command:{id}");
|
||||
}
|
||||
|
||||
let args = serde_json::to_string(&args).unwrap();
|
||||
let args = percent_encoding::utf8_percent_encode(
|
||||
&args,
|
||||
percent_encoding::NON_ALPHANUMERIC,
|
||||
);
|
||||
write!(f, "command:{id}?{args}")
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
fn to_lsp_tooltip(typst_tooltip: &Tooltip) -> HoverContents {
|
||||
let lsp_marked_string = match typst_tooltip {
|
||||
Tooltip::Text(text) => MarkedString::String(text.to_string()),
|
||||
Tooltip::Code(code) => MarkedString::LanguageString(LanguageString {
|
||||
language: "typc".to_owned(),
|
||||
value: code.to_string(),
|
||||
}),
|
||||
};
|
||||
HoverContents::Scalar(lsp_marked_string)
|
||||
}
|
||||
|
||||
#[cfg(test)]
|
||||
mod tests {
|
||||
use super::*;
|
||||
|
|
|
@ -63,7 +63,7 @@ pub struct InlayHintRequest {
|
|||
impl SemanticRequest for InlayHintRequest {
|
||||
type Response = Vec<InlayHint>;
|
||||
|
||||
fn request(self, ctx: &mut AnalysisContext) -> Option<Self::Response> {
|
||||
fn request(self, ctx: &mut LocalContext) -> Option<Self::Response> {
|
||||
let source = ctx.source_by_path(&self.path).ok()?;
|
||||
let range = ctx.to_typst_range(self.range, &source)?;
|
||||
|
||||
|
@ -83,22 +83,22 @@ impl SemanticRequest for InlayHintRequest {
|
|||
}
|
||||
|
||||
fn inlay_hint(
|
||||
ctx: &mut AnalysisContext,
|
||||
ctx: &mut LocalContext,
|
||||
source: &Source,
|
||||
range: Range<usize>,
|
||||
encoding: PositionEncoding,
|
||||
) -> FileResult<Vec<InlayHint>> {
|
||||
const SMART: InlayHintConfig = InlayHintConfig::smart();
|
||||
|
||||
struct InlayHintWorker<'a, 'w> {
|
||||
ctx: &'a mut AnalysisContext<'w>,
|
||||
struct InlayHintWorker<'a> {
|
||||
ctx: &'a mut LocalContext,
|
||||
source: &'a Source,
|
||||
range: Range<usize>,
|
||||
encoding: PositionEncoding,
|
||||
hints: Vec<InlayHint>,
|
||||
}
|
||||
|
||||
impl InlayHintWorker<'_, '_> {
|
||||
impl InlayHintWorker<'_> {
|
||||
fn analyze(&mut self, node: LinkedNode) {
|
||||
let rng = node.range();
|
||||
if rng.start >= self.range.end || rng.end <= self.range.start {
|
||||
|
|
|
@ -15,7 +15,7 @@ pub mod syntax;
|
|||
pub mod ty;
|
||||
mod upstream;
|
||||
|
||||
pub use analysis::AnalysisContext;
|
||||
pub use analysis::{LocalContext, LocalContextGuard};
|
||||
pub use upstream::with_vm;
|
||||
|
||||
mod diagnostics;
|
||||
|
@ -119,7 +119,7 @@ pub trait SemanticRequest {
|
|||
type Response;
|
||||
|
||||
/// Request the information from the given context.
|
||||
fn request(self, ctx: &mut AnalysisContext) -> Option<Self::Response>;
|
||||
fn request(self, ctx: &mut LocalContext) -> Option<Self::Response>;
|
||||
}
|
||||
|
||||
/// A request handler with given (semantic) analysis context and a versioned
|
||||
|
@ -131,7 +131,7 @@ pub trait StatefulRequest {
|
|||
/// Request the information from the given context.
|
||||
fn request(
|
||||
self,
|
||||
ctx: &mut AnalysisContext,
|
||||
ctx: &mut LocalContext,
|
||||
doc: Option<VersionedDocument>,
|
||||
) -> Option<Self::Response>;
|
||||
}
|
||||
|
|
|
@ -3,10 +3,7 @@
|
|||
// todo: remove this
|
||||
#![allow(missing_docs)]
|
||||
|
||||
use std::path::{Path, PathBuf};
|
||||
|
||||
use lsp_types::{self, Url};
|
||||
use once_cell::sync::Lazy;
|
||||
use crate::prelude::*;
|
||||
use reflexo::path::PathClean;
|
||||
|
||||
pub type LspPosition = lsp_types::Position;
|
||||
|
@ -24,9 +21,6 @@ pub type TypstSpan = typst::syntax::Span;
|
|||
pub type LspRange = lsp_types::Range;
|
||||
pub type TypstRange = std::ops::Range<usize>;
|
||||
|
||||
pub type TypstTooltip = crate::upstream::Tooltip;
|
||||
pub type LspHoverContents = lsp_types::HoverContents;
|
||||
|
||||
pub type LspDiagnostic = lsp_types::Diagnostic;
|
||||
pub type TypstDiagnostic = typst::diag::SourceDiagnostic;
|
||||
|
||||
|
@ -64,7 +58,7 @@ impl From<PositionEncoding> for lsp_types::PositionEncodingKind {
|
|||
}
|
||||
|
||||
const UNTITLED_ROOT: &str = "/untitled";
|
||||
static EMPTY_URL: Lazy<Url> = Lazy::new(|| Url::parse("file://").unwrap());
|
||||
static EMPTY_URL: LazyLock<Url> = LazyLock::new(|| Url::parse("file://").unwrap());
|
||||
|
||||
pub fn path_to_url(path: &Path) -> anyhow::Result<Url> {
|
||||
if let Ok(untitled) = path.strip_prefix(UNTITLED_ROOT) {
|
||||
|
@ -215,10 +209,6 @@ pub mod lsp_to_typst {
|
|||
}
|
||||
|
||||
pub mod typst_to_lsp {
|
||||
|
||||
use lsp_types::{LanguageString, MarkedString};
|
||||
use typst::syntax::Source;
|
||||
|
||||
use super::*;
|
||||
|
||||
pub fn offset_to_position(
|
||||
|
@ -270,23 +260,11 @@ pub mod typst_to_lsp {
|
|||
|
||||
LspRange::new(lsp_start, lsp_end)
|
||||
}
|
||||
|
||||
pub fn tooltip(typst_tooltip: &TypstTooltip) -> LspHoverContents {
|
||||
let lsp_marked_string = match typst_tooltip {
|
||||
TypstTooltip::Text(text) => MarkedString::String(text.to_string()),
|
||||
TypstTooltip::Code(code) => MarkedString::LanguageString(LanguageString {
|
||||
language: "typc".to_owned(),
|
||||
value: code.to_string(),
|
||||
}),
|
||||
};
|
||||
LspHoverContents::Scalar(lsp_marked_string)
|
||||
}
|
||||
}
|
||||
|
||||
#[cfg(test)]
|
||||
mod test {
|
||||
use lsp_types::Position;
|
||||
use typst::syntax::Source;
|
||||
|
||||
use super::*;
|
||||
|
||||
|
|
|
@ -1,15 +1,11 @@
|
|||
pub use std::{
|
||||
collections::HashMap,
|
||||
iter,
|
||||
ops::Range,
|
||||
path::{Path, PathBuf},
|
||||
sync::Arc,
|
||||
};
|
||||
pub use std::collections::HashMap;
|
||||
pub use std::iter;
|
||||
pub use std::ops::Range;
|
||||
pub use std::path::{Path, PathBuf};
|
||||
pub use std::sync::{Arc, LazyLock};
|
||||
|
||||
pub use ecow::eco_vec;
|
||||
pub use ecow::EcoVec;
|
||||
pub use ecow::{eco_vec, EcoVec};
|
||||
pub use itertools::{Format, Itertools};
|
||||
pub use log::error;
|
||||
pub use lsp_types::{
|
||||
request::GotoDeclarationResponse, CodeAction, CodeActionKind, CodeActionOrCommand, CodeLens,
|
||||
ColorInformation, ColorPresentation, CompletionResponse, DiagnosticRelatedInformation,
|
||||
|
@ -24,16 +20,18 @@ pub use reflexo::vector::ir::DefId;
|
|||
pub use serde_json::Value as JsonValue;
|
||||
pub use typst::diag::{EcoString, FileResult, Tracepoint};
|
||||
pub use typst::foundations::Value;
|
||||
pub use typst::syntax::FileId as TypstFileId;
|
||||
pub use typst::syntax::ast::{self, AstNode};
|
||||
pub use typst::syntax::{
|
||||
ast::{self, AstNode},
|
||||
LinkedNode, Source, Spanned, SyntaxKind, SyntaxNode,
|
||||
FileId as TypstFileId, LinkedNode, Source, Spanned, SyntaxKind, SyntaxNode,
|
||||
};
|
||||
pub use typst::World;
|
||||
|
||||
pub use crate::analysis::{AnalysisContext, LocalContext};
|
||||
pub use crate::analysis::{Definition, LocalContext};
|
||||
pub use crate::docs::DefDocs;
|
||||
pub use crate::lsp_typst_boundary::{
|
||||
lsp_to_typst, path_to_url, typst_to_lsp, LspDiagnostic, LspRange, LspSeverity,
|
||||
PositionEncoding, TypstDiagnostic, TypstSeverity, TypstSpan,
|
||||
};
|
||||
pub use crate::syntax::{get_deref_target, Decl, DefKind};
|
||||
pub(crate) use crate::ty::PathPreference;
|
||||
pub use crate::{SemanticRequest, StatefulRequest, VersionedDocument};
|
||||
|
|
|
@ -35,7 +35,7 @@ impl StatefulRequest for PrepareRenameRequest {
|
|||
|
||||
fn request(
|
||||
self,
|
||||
ctx: &mut AnalysisContext,
|
||||
ctx: &mut LocalContext,
|
||||
doc: Option<VersionedDocument>,
|
||||
) -> Option<Self::Response> {
|
||||
let source = ctx.source_by_path(&self.path).ok()?;
|
||||
|
@ -59,7 +59,7 @@ impl StatefulRequest for PrepareRenameRequest {
|
|||
}
|
||||
|
||||
pub(crate) fn prepare_renaming(
|
||||
ctx: &mut AnalysisContext,
|
||||
ctx: &mut LocalContext,
|
||||
deref_target: &DerefTarget,
|
||||
def: &Definition,
|
||||
) -> Option<(String, Option<LspRange>)> {
|
||||
|
|
|
@ -26,7 +26,7 @@ impl StatefulRequest for ReferencesRequest {
|
|||
|
||||
fn request(
|
||||
self,
|
||||
ctx: &mut AnalysisContext,
|
||||
ctx: &mut LocalContext,
|
||||
doc: Option<VersionedDocument>,
|
||||
) -> Option<Self::Response> {
|
||||
let source = ctx.source_by_path(&self.path).ok()?;
|
||||
|
@ -40,7 +40,7 @@ impl StatefulRequest for ReferencesRequest {
|
|||
}
|
||||
|
||||
pub(crate) fn find_references(
|
||||
ctx: &mut AnalysisContext,
|
||||
ctx: &mut LocalContext,
|
||||
source: &Source,
|
||||
doc: Option<&VersionedDocument>,
|
||||
target: DerefTarget<'_>,
|
||||
|
@ -69,13 +69,13 @@ pub(crate) fn find_references(
|
|||
}
|
||||
}
|
||||
|
||||
struct ReferencesWorker<'a, 'w> {
|
||||
ctx: SearchCtx<'a, 'w>,
|
||||
struct ReferencesWorker<'a> {
|
||||
ctx: SearchCtx<'a>,
|
||||
references: Vec<LspLocation>,
|
||||
def: Definition,
|
||||
}
|
||||
|
||||
impl<'a, 'w> ReferencesWorker<'a, 'w> {
|
||||
impl<'a> ReferencesWorker<'a> {
|
||||
fn label_root(mut self) -> Option<Vec<LspLocation>> {
|
||||
let mut ids = vec![];
|
||||
|
||||
|
|
|
@ -35,7 +35,7 @@ impl StatefulRequest for RenameRequest {
|
|||
|
||||
fn request(
|
||||
self,
|
||||
ctx: &mut AnalysisContext,
|
||||
ctx: &mut LocalContext,
|
||||
doc: Option<VersionedDocument>,
|
||||
) -> Option<Self::Response> {
|
||||
let source = ctx.source_by_path(&self.path).ok()?;
|
||||
|
@ -117,7 +117,7 @@ impl StatefulRequest for RenameRequest {
|
|||
}
|
||||
|
||||
pub(crate) fn do_rename_file(
|
||||
ctx: &mut AnalysisContext,
|
||||
ctx: &mut LocalContext,
|
||||
def_fid: TypstFileId,
|
||||
diff: PathBuf,
|
||||
edits: &mut HashMap<Url, Vec<TextEdit>>,
|
||||
|
@ -167,7 +167,7 @@ pub(crate) fn edits_to_document_changes(
|
|||
}
|
||||
|
||||
fn rename_importer(
|
||||
ctx: &AnalysisContext,
|
||||
ctx: &LocalContext,
|
||||
src: &Source,
|
||||
span: Span,
|
||||
diff: &Path,
|
||||
|
|
|
@ -25,7 +25,7 @@ impl SemanticRequest for SemanticTokensDeltaRequest {
|
|||
type Response = SemanticTokensFullDeltaResult;
|
||||
/// Handles the request to compute the semantic tokens delta for a given
|
||||
/// document.
|
||||
fn request(self, ctx: &mut AnalysisContext) -> Option<Self::Response> {
|
||||
fn request(self, ctx: &mut LocalContext) -> Option<Self::Response> {
|
||||
let source = ctx.source_by_path(&self.path).ok()?;
|
||||
let ei = ctx.expr_stage(&source);
|
||||
|
||||
|
|
|
@ -26,7 +26,7 @@ impl SemanticRequest for SemanticTokensFullRequest {
|
|||
type Response = SemanticTokensResult;
|
||||
|
||||
/// Handles the request to compute the semantic tokens for a given document.
|
||||
fn request(self, ctx: &mut AnalysisContext) -> Option<Self::Response> {
|
||||
fn request(self, ctx: &mut LocalContext) -> Option<Self::Response> {
|
||||
let source = ctx.source_by_path(&self.path).ok()?;
|
||||
let ei = ctx.expr_stage(&source);
|
||||
let token_ctx = &ctx.analysis.tokens_ctx;
|
||||
|
|
|
@ -25,7 +25,7 @@ pub struct SignatureHelpRequest {
|
|||
impl SemanticRequest for SignatureHelpRequest {
|
||||
type Response = SignatureHelp;
|
||||
|
||||
fn request(self, ctx: &mut AnalysisContext) -> Option<Self::Response> {
|
||||
fn request(self, ctx: &mut LocalContext) -> Option<Self::Response> {
|
||||
let source = ctx.source_by_path(&self.path).ok()?;
|
||||
let cursor = ctx.to_typst_pos(self.position, &source)? + 1;
|
||||
|
||||
|
@ -150,11 +150,11 @@ impl SemanticRequest for SignatureHelpRequest {
|
|||
pub(crate) struct DocTooltip;
|
||||
|
||||
impl DocTooltip {
|
||||
pub fn get(ctx: &mut AnalysisContext, def: &Definition) -> Option<String> {
|
||||
pub fn get(ctx: &mut LocalContext, def: &Definition) -> Option<String> {
|
||||
self::DocTooltip::get_inner(ctx, def).map(|s| "\n\n".to_owned() + &s)
|
||||
}
|
||||
|
||||
fn get_inner(ctx: &mut AnalysisContext, def: &Definition) -> Option<String> {
|
||||
fn get_inner(ctx: &mut LocalContext, def: &Definition) -> Option<String> {
|
||||
let value = def.value();
|
||||
if matches!(value, Some(Value::Func(..))) {
|
||||
if let Some(builtin) = Self::builtin_func_tooltip(def) {
|
||||
|
|
|
@ -31,7 +31,7 @@ pub struct SymbolRequest {
|
|||
impl SemanticRequest for SymbolRequest {
|
||||
type Response = Vec<SymbolInformation>;
|
||||
|
||||
fn request(self, ctx: &mut AnalysisContext) -> Option<Self::Response> {
|
||||
fn request(self, ctx: &mut LocalContext) -> Option<Self::Response> {
|
||||
// todo: let typst.ts expose source
|
||||
|
||||
let mut symbols = vec![];
|
||||
|
|
|
@ -22,16 +22,13 @@ pub use tinymist_world::{LspUniverse, LspUniverseBuilder};
|
|||
use typst_shim::syntax::LinkedNodeExt;
|
||||
|
||||
use crate::{
|
||||
analysis::{Analysis, AnalysisResources},
|
||||
prelude::AnalysisContext,
|
||||
typst_to_lsp, LspPosition, PositionEncoding, VersionedDocument,
|
||||
analysis::Analysis, prelude::LocalContext, typst_to_lsp, LspPosition, PositionEncoding,
|
||||
VersionedDocument,
|
||||
};
|
||||
|
||||
type CompileDriver<C> = CompileDriverImpl<C, tinymist_world::LspCompilerFeat>;
|
||||
|
||||
impl AnalysisResources for () {}
|
||||
|
||||
pub fn snapshot_testing(name: &str, f: &impl Fn(&mut AnalysisContext, PathBuf)) {
|
||||
pub fn snapshot_testing(name: &str, f: &impl Fn(&mut LocalContext, PathBuf)) {
|
||||
let mut settings = insta::Settings::new();
|
||||
settings.set_prepend_module_to_snapshot(false);
|
||||
settings.set_snapshot_path(format!("fixtures/{name}/snaps"));
|
||||
|
@ -52,7 +49,7 @@ pub fn snapshot_testing(name: &str, f: &impl Fn(&mut AnalysisContext, PathBuf))
|
|||
pub fn run_with_ctx<T>(
|
||||
w: &mut LspUniverse,
|
||||
p: PathBuf,
|
||||
f: &impl Fn(&mut AnalysisContext, PathBuf) -> T,
|
||||
f: &impl Fn(&mut LocalContext, PathBuf) -> T,
|
||||
) -> T {
|
||||
let root = w.workspace_root().unwrap();
|
||||
let paths = w
|
||||
|
@ -61,7 +58,7 @@ pub fn run_with_ctx<T>(
|
|||
.map(|p| TypstFileId::new(None, VirtualPath::new(p.strip_prefix(&root).unwrap())))
|
||||
.collect::<Vec<_>>();
|
||||
|
||||
let mut ctx = Arc::new(Analysis::default()).snapshot(w.snapshot(), &());
|
||||
let mut ctx = Arc::new(Analysis::default()).snapshot(w.snapshot());
|
||||
ctx.test_completion_files(Vec::new);
|
||||
ctx.test_files(|| paths);
|
||||
f(&mut ctx, p)
|
||||
|
@ -79,7 +76,7 @@ pub fn get_test_properties(s: &str) -> HashMap<&'_ str, &'_ str> {
|
|||
}
|
||||
|
||||
pub fn compile_doc_for_test(
|
||||
ctx: &mut AnalysisContext,
|
||||
ctx: &mut LocalContext,
|
||||
properties: &HashMap<&str, &str>,
|
||||
) -> Option<VersionedDocument> {
|
||||
let must_compile = properties
|
||||
|
|
|
@ -17,7 +17,7 @@ use unscanny::Scanner;
|
|||
use super::{plain_docs_sentence, summarize_font_family};
|
||||
use crate::adt::interner::Interned;
|
||||
use crate::analysis::{analyze_labels, DynLabel, Ty};
|
||||
use crate::AnalysisContext;
|
||||
use crate::LocalContext;
|
||||
|
||||
mod ext;
|
||||
pub use ext::complete_path;
|
||||
|
@ -551,7 +551,7 @@ fn complete_imports(ctx: &mut CompletionContext) -> bool {
|
|||
|
||||
/// Add completions for all exports of a module.
|
||||
fn import_item_completions<'a>(
|
||||
ctx: &mut CompletionContext<'a, '_>,
|
||||
ctx: &mut CompletionContext<'a>,
|
||||
existing: ast::ImportItems<'a>,
|
||||
source: &LinkedNode,
|
||||
) {
|
||||
|
@ -951,8 +951,8 @@ fn code_completions(ctx: &mut CompletionContext, hash: bool) {
|
|||
}
|
||||
|
||||
/// Context for autocompletion.
|
||||
pub struct CompletionContext<'a, 'b> {
|
||||
pub ctx: &'a mut AnalysisContext<'b>,
|
||||
pub struct CompletionContext<'a> {
|
||||
pub ctx: &'a mut LocalContext,
|
||||
pub document: Option<&'a Document>,
|
||||
pub text: &'a str,
|
||||
pub before: &'a str,
|
||||
|
@ -974,11 +974,11 @@ pub struct CompletionContext<'a, 'b> {
|
|||
pub seen_fields: HashSet<Interned<str>>,
|
||||
}
|
||||
|
||||
impl<'a, 'w> CompletionContext<'a, 'w> {
|
||||
impl<'a> CompletionContext<'a> {
|
||||
/// Create a new autocompletion context.
|
||||
#[allow(clippy::too_many_arguments)]
|
||||
pub fn new(
|
||||
ctx: &'a mut AnalysisContext<'w>,
|
||||
ctx: &'a mut LocalContext,
|
||||
document: Option<&'a Document>,
|
||||
source: &'a Source,
|
||||
cursor: usize,
|
||||
|
|
|
@ -20,7 +20,7 @@ use crate::upstream::plain_docs_sentence;
|
|||
|
||||
use crate::{completion_kind, prelude::*, LspCompletion};
|
||||
|
||||
impl<'a, 'w> CompletionContext<'a, 'w> {
|
||||
impl<'a> CompletionContext<'a> {
|
||||
pub fn world(&self) -> &LspWorld {
|
||||
self.ctx.world()
|
||||
}
|
||||
|
@ -456,7 +456,7 @@ impl<'a, 'w> CompletionContext<'a, 'w> {
|
|||
}
|
||||
}
|
||||
|
||||
fn describe_value(ctx: &mut AnalysisContext, v: &Value) -> EcoString {
|
||||
fn describe_value(ctx: &mut LocalContext, v: &Value) -> EcoString {
|
||||
match v {
|
||||
Value::Func(f) => {
|
||||
let mut f = f;
|
||||
|
@ -555,7 +555,7 @@ pub fn value_to_completion_kind(value: &Value) -> CompletionKind {
|
|||
|
||||
/// Add completions for the parameters of a function.
|
||||
pub fn param_completions<'a>(
|
||||
ctx: &mut CompletionContext<'a, '_>,
|
||||
ctx: &mut CompletionContext<'a>,
|
||||
callee: ast::Expr<'a>,
|
||||
set: bool,
|
||||
args: ast::Args<'a>,
|
||||
|
@ -695,7 +695,7 @@ pub fn param_completions<'a>(
|
|||
}
|
||||
|
||||
fn type_completion(
|
||||
ctx: &mut CompletionContext<'_, '_>,
|
||||
ctx: &mut CompletionContext<'_>,
|
||||
infer_type: &Ty,
|
||||
docs: Option<&str>,
|
||||
) -> Option<()> {
|
||||
|
@ -963,7 +963,7 @@ fn type_completion(
|
|||
|
||||
/// Add completions for the values of a named function parameter.
|
||||
pub fn named_param_value_completions<'a>(
|
||||
ctx: &mut CompletionContext<'a, '_>,
|
||||
ctx: &mut CompletionContext<'a>,
|
||||
callee: ast::Expr<'a>,
|
||||
name: &Interned<str>,
|
||||
ty: Option<&Ty>,
|
||||
|
@ -1100,7 +1100,7 @@ pub(crate) fn complete_type(ctx: &mut CompletionContext) -> Option<()> {
|
|||
}
|
||||
|
||||
pub fn complete_path(
|
||||
ctx: &AnalysisContext,
|
||||
ctx: &LocalContext,
|
||||
v: Option<LinkedNode>,
|
||||
source: &Source,
|
||||
cursor: usize,
|
||||
|
|
|
@ -17,7 +17,7 @@ impl StatefulRequest for WillRenameFilesRequest {
|
|||
|
||||
fn request(
|
||||
self,
|
||||
ctx: &mut AnalysisContext,
|
||||
ctx: &mut LocalContext,
|
||||
_doc: Option<VersionedDocument>,
|
||||
) -> Option<Self::Response> {
|
||||
let mut edits: HashMap<Url, Vec<TextEdit>> = HashMap::new();
|
||||
|
|
|
@ -16,7 +16,7 @@ pub struct WorkspaceLabelRequest {}
|
|||
impl SemanticRequest for WorkspaceLabelRequest {
|
||||
type Response = Vec<SymbolInformation>;
|
||||
|
||||
fn request(self, ctx: &mut AnalysisContext) -> Option<Self::Response> {
|
||||
fn request(self, ctx: &mut LocalContext) -> Option<Self::Response> {
|
||||
// todo: let typst.ts expose source
|
||||
|
||||
let mut symbols = vec![];
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue