Better support client completion resolve caps

This commit is contained in:
Kirill Bulatov 2020-12-02 01:02:15 +02:00
parent 47464e556c
commit 2a7be4afb0
5 changed files with 118 additions and 103 deletions

View file

@ -12,6 +12,7 @@ use crate::config::SnippetCap;
/// `CompletionItem` describes a single completion variant in the editor pop-up. /// `CompletionItem` describes a single completion variant in the editor pop-up.
/// It is basically a POD with various properties. To construct a /// It is basically a POD with various properties. To construct a
/// `CompletionItem`, use `new` method and the `Builder` struct. /// `CompletionItem`, use `new` method and the `Builder` struct.
#[derive(Clone)]
pub struct CompletionItem { pub struct CompletionItem {
/// Used only internally in tests, to check only specific kind of /// Used only internally in tests, to check only specific kind of
/// completion (postfix, keyword, reference, etc). /// completion (postfix, keyword, reference, etc).

View file

@ -97,6 +97,7 @@ pub(crate) fn check_edit_with_config(
.unwrap_or_else(|| panic!("can't find {:?} completion in {:#?}", what, completions)); .unwrap_or_else(|| panic!("can't find {:?} completion in {:#?}", what, completions));
let mut actual = db.file_text(position.file_id).to_string(); let mut actual = db.file_text(position.file_id).to_string();
completion.text_edit().apply(&mut actual); completion.text_edit().apply(&mut actual);
// TODO kb how to apply imports now?
assert_eq_text!(&ra_fixture_after, &actual) assert_eq_text!(&ra_fixture_after, &actual)
} }

View file

@ -7,7 +7,7 @@ use std::{sync::Arc, time::Instant};
use crossbeam_channel::{unbounded, Receiver, Sender}; use crossbeam_channel::{unbounded, Receiver, Sender};
use flycheck::FlycheckHandle; use flycheck::FlycheckHandle;
use ide::{Analysis, AnalysisHost, Change, FileId, ImportToAdd}; use ide::{Analysis, AnalysisHost, Change, CompletionItem, FileId};
use ide_db::base_db::{CrateId, VfsPath}; use ide_db::base_db::{CrateId, VfsPath};
use lsp_types::{SemanticTokens, Url}; use lsp_types::{SemanticTokens, Url};
use parking_lot::{Mutex, RwLock}; use parking_lot::{Mutex, RwLock};
@ -51,6 +51,11 @@ pub(crate) struct Handle<H, C> {
pub(crate) type ReqHandler = fn(&mut GlobalState, lsp_server::Response); pub(crate) type ReqHandler = fn(&mut GlobalState, lsp_server::Response);
pub(crate) type ReqQueue = lsp_server::ReqQueue<(String, Instant), ReqHandler>; pub(crate) type ReqQueue = lsp_server::ReqQueue<(String, Instant), ReqHandler>;
pub(crate) struct CompletionResolveData {
pub(crate) file_id: FileId,
pub(crate) item: CompletionItem,
}
/// `GlobalState` is the primary mutable state of the language server /// `GlobalState` is the primary mutable state of the language server
/// ///
/// The most interesting components are `vfs`, which stores a consistent /// The most interesting components are `vfs`, which stores a consistent
@ -69,7 +74,7 @@ pub(crate) struct GlobalState {
pub(crate) config: Config, pub(crate) config: Config,
pub(crate) analysis_host: AnalysisHost, pub(crate) analysis_host: AnalysisHost,
pub(crate) diagnostics: DiagnosticCollection, pub(crate) diagnostics: DiagnosticCollection,
pub(crate) additional_imports: FxHashMap<usize, ImportToAdd>, pub(crate) completion_resolve_data: FxHashMap<usize, CompletionResolveData>,
pub(crate) mem_docs: FxHashMap<VfsPath, DocumentData>, pub(crate) mem_docs: FxHashMap<VfsPath, DocumentData>,
pub(crate) semantic_tokens_cache: Arc<Mutex<FxHashMap<Url, SemanticTokens>>>, pub(crate) semantic_tokens_cache: Arc<Mutex<FxHashMap<Url, SemanticTokens>>>,
pub(crate) vfs: Arc<RwLock<(vfs::Vfs, FxHashMap<FileId, LineEndings>)>>, pub(crate) vfs: Arc<RwLock<(vfs::Vfs, FxHashMap<FileId, LineEndings>)>>,
@ -122,7 +127,7 @@ impl GlobalState {
config, config,
analysis_host, analysis_host,
diagnostics: Default::default(), diagnostics: Default::default(),
additional_imports: FxHashMap::default(), completion_resolve_data: FxHashMap::default(),
mem_docs: FxHashMap::default(), mem_docs: FxHashMap::default(),
semantic_tokens_cache: Arc::new(Default::default()), semantic_tokens_cache: Arc::new(Default::default()),
vfs: Arc::new(RwLock::new((vfs::Vfs::default(), FxHashMap::default()))), vfs: Arc::new(RwLock::new((vfs::Vfs::default(), FxHashMap::default()))),

View file

@ -5,14 +5,12 @@
use std::{ use std::{
io::Write as _, io::Write as _,
process::{self, Stdio}, process::{self, Stdio},
sync::Arc,
}; };
use ide::{ use ide::{
FileId, FilePosition, FileRange, HoverAction, HoverGotoTypeData, ImportToAdd, LineIndex, FileId, FilePosition, FileRange, HoverAction, HoverGotoTypeData, NavigationTarget, Query,
NavigationTarget, Query, RangeInfo, Runnable, RunnableKind, SearchScope, TextEdit, RangeInfo, Runnable, RunnableKind, SearchScope, TextEdit,
}; };
use ide_db::helpers::{insert_use, mod_path_to_ast};
use itertools::Itertools; use itertools::Itertools;
use lsp_server::ErrorCode; use lsp_server::ErrorCode;
use lsp_types::{ use lsp_types::{
@ -36,10 +34,10 @@ use crate::{
cargo_target_spec::CargoTargetSpec, cargo_target_spec::CargoTargetSpec,
config::RustfmtConfig, config::RustfmtConfig,
from_json, from_proto, from_json, from_proto,
global_state::{GlobalState, GlobalStateSnapshot}, global_state::{CompletionResolveData, GlobalState, GlobalStateSnapshot},
line_endings::LineEndings,
lsp_ext::{self, InlayHint, InlayHintsParams}, lsp_ext::{self, InlayHint, InlayHintsParams},
to_proto, LspError, Result, to_proto::{self, append_import_edits},
LspError, Result,
}; };
pub(crate) fn handle_analyzer_status( pub(crate) fn handle_analyzer_status(
@ -538,12 +536,6 @@ pub(crate) fn handle_runnables(
Ok(res) Ok(res)
} }
#[derive(Debug, Copy, Clone, Serialize, Deserialize)]
pub(crate) struct ResolveCompletionData {
completion_id: usize,
completion_file_id: u32,
}
pub(crate) fn handle_completion( pub(crate) fn handle_completion(
global_state: &mut GlobalState, global_state: &mut GlobalState,
params: lsp_types::CompletionParams, params: lsp_types::CompletionParams,
@ -579,38 +571,31 @@ pub(crate) fn handle_completion(
}; };
let line_index = snap.analysis.file_line_index(position.file_id)?; let line_index = snap.analysis.file_line_index(position.file_id)?;
let line_endings = snap.file_line_endings(position.file_id); let line_endings = snap.file_line_endings(position.file_id);
let mut additional_imports = FxHashMap::default(); let mut completion_resolve_data = FxHashMap::default();
let items: Vec<CompletionItem> = items let items: Vec<CompletionItem> = items
.into_iter() .into_iter()
.enumerate() .enumerate()
.flat_map(|(item_index, item)| { .flat_map(|(item_index, item)| {
let resolve_completion_data = ResolveCompletionData { let mut new_completion_items = to_proto::completion_item(
completion_id: item_index, &line_index,
completion_file_id: position.file_id.0, line_endings,
}; item.clone(),
let import_to_add = item.import_to_add().cloned(); &snap.config.completion.resolve_capabilities,
let mut new_completion_items = );
to_proto::completion_item(&line_index, line_endings, item);
if let Some(import_to_add) = import_to_add { let item_id = serde_json::to_value(&item_index)
for new_item in &mut new_completion_items { .expect(&format!("Should be able to serialize usize value {}", item_index));
match serde_json::to_value(&resolve_completion_data) { completion_resolve_data
Ok(resolve_value) => { .insert(item_index, CompletionResolveData { file_id: position.file_id, item });
new_item.data = Some(resolve_value); for new_item in &mut new_completion_items {
additional_imports.insert(item_index, import_to_add.clone()); new_item.data = Some(item_id.clone());
}
Err(e) => {
log::error!("Failed to serialize completion resolve metadata: {}", e)
}
}
}
} }
new_completion_items new_completion_items
}) })
.collect(); .collect();
global_state.additional_imports = additional_imports; global_state.completion_resolve_data = completion_resolve_data;
let completion_list = lsp_types::CompletionList { is_incomplete: true, items }; let completion_list = lsp_types::CompletionList { is_incomplete: true, items };
Ok(Some(completion_list.into())) Ok(Some(completion_list.into()))
@ -622,73 +607,40 @@ pub(crate) fn handle_resolve_completion(
) -> Result<lsp_types::CompletionItem> { ) -> Result<lsp_types::CompletionItem> {
let _p = profile::span("handle_resolve_completion"); let _p = profile::span("handle_resolve_completion");
match original_completion.data.as_ref() { let server_completion_data = match original_completion
Some(completion_data) => { .data
match serde_json::from_value::<ResolveCompletionData>(completion_data.clone()) { .as_ref()
Ok(resolve_completion_data) => { .map(|data| serde_json::from_value::<usize>(data.clone()))
if let Some(import_to_add) = .transpose()?
global_state.additional_imports.get(&resolve_completion_data.completion_id) .and_then(|server_completion_id| {
{ global_state.completion_resolve_data.get(&server_completion_id)
let snap = global_state.snapshot(); }) {
let file_id = FileId(resolve_completion_data.completion_file_id); Some(data) => data,
let line_index = snap.analysis.file_line_index(file_id)?; None => return Ok(original_completion),
let line_endings = snap.file_line_endings(file_id); };
let resolved_edits = let snap = &global_state.snapshot();
resolve_additional_edits(import_to_add, line_index, line_endings); for supported_completion_resolve_cap in &snap.config.completion.resolve_capabilities {
match supported_completion_resolve_cap {
original_completion.additional_text_edits = ide::CompletionResolveCapability::AdditionalTextEdits => {
match original_completion.additional_text_edits { // TODO kb actually add all additional edits here?
Some(mut original_additional_edits) => { if let Some(import_to_add) = server_completion_data.item.import_to_add() {
if let Some(mut new_edits) = resolved_edits { append_import_edits(
original_additional_edits.extend(new_edits.drain(..)) &mut original_completion,
} import_to_add,
Some(original_additional_edits) snap.analysis.file_line_index(server_completion_data.file_id)?.as_ref(),
} snap.file_line_endings(server_completion_data.file_id),
None => resolved_edits, );
};
} else {
log::error!(
"Got no import data for completion with label {}, id {}",
original_completion.label,
resolve_completion_data.completion_id
)
}
} }
Err(e) => log::error!("Failed to deserialize completion resolve metadata: {}", e),
} }
// TODO kb calculate the rest also?
_ => {}
} }
None => (),
} }
Ok(original_completion) Ok(original_completion)
} }
// TODO kb what to do when no resolve is available on the client?
fn resolve_additional_edits(
import_to_add: &ImportToAdd,
line_index: Arc<LineIndex>,
line_endings: LineEndings,
) -> Option<Vec<lsp_types::TextEdit>> {
let _p = profile::span("resolve_additional_edits");
let rewriter = insert_use::insert_use(
&import_to_add.import_scope,
mod_path_to_ast(&import_to_add.import_path),
import_to_add.merge_behaviour,
);
let old_ast = rewriter.rewrite_root()?;
let mut import_insert = TextEdit::builder();
algo::diff(&old_ast, &rewriter.rewrite(&old_ast)).into_text_edit(&mut import_insert);
let text_edit = import_insert.finish();
Some(
text_edit
.into_iter()
.map(|indel| to_proto::text_edit(&line_index, line_endings, indel))
.collect_vec(),
)
}
pub(crate) fn handle_folding_range( pub(crate) fn handle_folding_range(
snap: GlobalStateSnapshot, snap: GlobalStateSnapshot,
params: FoldingRangeParams, params: FoldingRangeParams,

View file

@ -5,14 +5,19 @@ use std::{
}; };
use ide::{ use ide::{
Assist, AssistKind, CallInfo, CompletionItem, CompletionItemKind, Documentation, Assist, AssistKind, CallInfo, CompletionItem, CompletionItemKind, CompletionResolveCapability,
FileSystemEdit, Fold, FoldKind, Highlight, HighlightModifier, HighlightTag, HighlightedRange, Documentation, FileSystemEdit, Fold, FoldKind, Highlight, HighlightModifier, HighlightTag,
Indel, InlayHint, InlayKind, InsertTextFormat, LineIndex, Markup, NavigationTarget, HighlightedRange, ImportToAdd, Indel, InlayHint, InlayKind, InsertTextFormat, LineIndex,
ReferenceAccess, ResolvedAssist, Runnable, Severity, SourceChange, SourceFileEdit, TextEdit, Markup, NavigationTarget, ReferenceAccess, ResolvedAssist, Runnable, Severity, SourceChange,
SourceFileEdit, TextEdit,
};
use ide_db::{
base_db::{FileId, FileRange},
helpers::{insert_use, mod_path_to_ast},
}; };
use ide_db::base_db::{FileId, FileRange};
use itertools::Itertools; use itertools::Itertools;
use syntax::{SyntaxKind, TextRange, TextSize}; use rustc_hash::FxHashSet;
use syntax::{algo, SyntaxKind, TextRange, TextSize};
use crate::{ use crate::{
cargo_target_spec::CargoTargetSpec, global_state::GlobalStateSnapshot, cargo_target_spec::CargoTargetSpec, global_state::GlobalStateSnapshot,
@ -158,6 +163,7 @@ pub(crate) fn completion_item(
line_index: &LineIndex, line_index: &LineIndex,
line_endings: LineEndings, line_endings: LineEndings,
completion_item: CompletionItem, completion_item: CompletionItem,
resolve_capabilities: &FxHashSet<CompletionResolveCapability>,
) -> Vec<lsp_types::CompletionItem> { ) -> Vec<lsp_types::CompletionItem> {
fn set_score(res: &mut lsp_types::CompletionItem, label: &str) { fn set_score(res: &mut lsp_types::CompletionItem, label: &str) {
res.preselect = Some(true); res.preselect = Some(true);
@ -231,9 +237,17 @@ pub(crate) fn completion_item(
None => vec![res], None => vec![res],
}; };
let unapplied_import_data = completion_item.import_to_add().filter(|_| {
!resolve_capabilities.contains(&CompletionResolveCapability::AdditionalTextEdits)
});
for mut r in all_results.iter_mut() { for mut r in all_results.iter_mut() {
r.insert_text_format = Some(insert_text_format(completion_item.insert_text_format())); r.insert_text_format = Some(insert_text_format(completion_item.insert_text_format()));
if let Some(unapplied_import_data) = unapplied_import_data {
append_import_edits(r, unapplied_import_data, line_index, line_endings);
}
} }
all_results all_results
} }
@ -817,6 +831,47 @@ pub(crate) fn markup_content(markup: Markup) -> lsp_types::MarkupContent {
lsp_types::MarkupContent { kind: lsp_types::MarkupKind::Markdown, value } lsp_types::MarkupContent { kind: lsp_types::MarkupKind::Markdown, value }
} }
pub(crate) fn import_into_edits(
import_to_add: &ImportToAdd,
line_index: &LineIndex,
line_endings: LineEndings,
) -> Option<Vec<lsp_types::TextEdit>> {
let _p = profile::span("add_import_edits");
let rewriter = insert_use::insert_use(
&import_to_add.import_scope,
mod_path_to_ast(&import_to_add.import_path),
import_to_add.merge_behaviour,
);
let old_ast = rewriter.rewrite_root()?;
let mut import_insert = TextEdit::builder();
algo::diff(&old_ast, &rewriter.rewrite(&old_ast)).into_text_edit(&mut import_insert);
let import_edit = import_insert.finish();
Some(
import_edit
.into_iter()
.map(|indel| text_edit(line_index, line_endings, indel))
.collect_vec(),
)
}
pub(crate) fn append_import_edits(
completion: &mut lsp_types::CompletionItem,
import_to_add: &ImportToAdd,
line_index: &LineIndex,
line_endings: LineEndings,
) {
let new_edits = import_into_edits(import_to_add, line_index, line_endings);
if let Some(original_additional_edits) = completion.additional_text_edits.as_mut() {
if let Some(mut new_edits) = new_edits {
original_additional_edits.extend(new_edits.drain(..))
}
} else {
completion.additional_text_edits = new_edits;
}
}
#[cfg(test)] #[cfg(test)]
mod tests { mod tests {
use ide::Analysis; use ide::Analysis;
@ -836,6 +891,7 @@ mod tests {
let (offset, text) = test_utils::extract_offset(fixture); let (offset, text) = test_utils::extract_offset(fixture);
let line_index = LineIndex::new(&text); let line_index = LineIndex::new(&text);
let (analysis, file_id) = Analysis::from_single_file(text); let (analysis, file_id) = Analysis::from_single_file(text);
let resolve_caps = FxHashSet::default();
let completions: Vec<(String, Option<String>)> = analysis let completions: Vec<(String, Option<String>)> = analysis
.completions( .completions(
&ide::CompletionConfig::default(), &ide::CompletionConfig::default(),
@ -845,7 +901,7 @@ mod tests {
.unwrap() .unwrap()
.into_iter() .into_iter()
.filter(|c| c.label().ends_with("arg")) .filter(|c| c.label().ends_with("arg"))
.map(|c| completion_item(&line_index, LineEndings::Unix, c)) .map(|c| completion_item(&line_index, LineEndings::Unix, c, &resolve_caps))
.flat_map(|comps| comps.into_iter().map(|c| (c.label, c.sort_text))) .flat_map(|comps| comps.into_iter().map(|c| (c.label, c.sort_text)))
.collect(); .collect();
expect_test::expect![[r#" expect_test::expect![[r#"