//! FIXME: write short doc here use hir::{Semantics, SemanticsScope}; use ra_db::SourceDatabase; use ra_ide_db::RootDatabase; use ra_syntax::{ algo::{find_covering_element, find_node_at_offset}, ast, AstNode, SyntaxKind::*, SyntaxNode, SyntaxToken, TextRange, TextUnit, }; use ra_text_edit::AtomTextEdit; use crate::{completion::CompletionConfig, FilePosition}; /// `CompletionContext` is created early during completion to figure out, where /// exactly is the cursor, syntax-wise. #[derive(Debug)] pub(crate) struct CompletionContext<'a> { pub(super) sema: Semantics<'a, RootDatabase>, pub(super) db: &'a RootDatabase, pub(super) config: &'a CompletionConfig, pub(super) offset: TextUnit, /// The token before the cursor, in the original file. pub(super) original_token: SyntaxToken, /// The token before the cursor, in the macro-expanded file. pub(super) token: SyntaxToken, pub(super) krate: Option, pub(super) name_ref_syntax: Option, pub(super) function_syntax: Option, pub(super) use_item_syntax: Option, pub(super) record_lit_syntax: Option, pub(super) record_lit_pat: Option, pub(super) impl_def: Option, pub(super) is_param: bool, /// If a name-binding or reference to a const in a pattern. /// Irrefutable patterns (like let) are excluded. pub(super) is_pat_binding_or_const: bool, /// A single-indent path, like `foo`. `::foo` should not be considered a trivial path. pub(super) is_trivial_path: bool, /// If not a trivial path, the prefix (qualifier). pub(super) path_prefix: Option, pub(super) after_if: bool, /// `true` if we are a statement or a last expr in the block. pub(super) can_be_stmt: bool, /// Something is typed at the "top" level, in module or impl/trait. pub(super) is_new_item: bool, /// The receiver if this is a field or method access, i.e. writing something.<|> pub(super) dot_receiver: Option, pub(super) dot_receiver_is_ambiguous_float_literal: bool, /// If this is a call (method or function) in particular, i.e. the () are already there. pub(super) is_call: bool, /// If this is a macro call, i.e. the () are already there. pub(super) is_macro_call: bool, pub(super) is_path_type: bool, pub(super) has_type_args: bool, } impl<'a> CompletionContext<'a> { pub(super) fn new( db: &'a RootDatabase, position: FilePosition, config: &'a CompletionConfig, ) -> Option> { let sema = Semantics::new(db); let original_file = sema.parse(position.file_id); // Insert a fake ident to get a valid parse tree. We will use this file // to determine context, though the original_file will be used for // actual completion. let file_with_fake_ident = { let parse = db.parse(position.file_id); let edit = AtomTextEdit::insert(position.offset, "intellijRulezz".to_string()); parse.reparse(&edit).tree() }; let fake_ident_token = file_with_fake_ident.syntax().token_at_offset(position.offset).right_biased().unwrap(); let krate = sema.to_module_def(position.file_id).map(|m| m.krate()); let original_token = original_file.syntax().token_at_offset(position.offset).left_biased()?; let token = sema.descend_into_macros(original_token.clone()); let mut ctx = CompletionContext { sema, db, config, original_token, token, offset: position.offset, krate, name_ref_syntax: None, function_syntax: None, use_item_syntax: None, record_lit_syntax: None, record_lit_pat: None, impl_def: None, is_param: false, is_pat_binding_or_const: false, is_trivial_path: false, path_prefix: None, after_if: false, can_be_stmt: false, is_new_item: false, dot_receiver: None, is_call: false, is_macro_call: false, is_path_type: false, has_type_args: false, dot_receiver_is_ambiguous_float_literal: false, }; let mut original_file = original_file.syntax().clone(); let mut hypothetical_file = file_with_fake_ident.syntax().clone(); let mut offset = position.offset; let mut fake_ident_token = fake_ident_token; // Are we inside a macro call? while let (Some(actual_macro_call), Some(macro_call_with_fake_ident)) = ( find_node_at_offset::(&original_file, offset), find_node_at_offset::(&hypothetical_file, offset), ) { if actual_macro_call.path().as_ref().map(|s| s.syntax().text()) != macro_call_with_fake_ident.path().as_ref().map(|s| s.syntax().text()) { break; } let hypothetical_args = match macro_call_with_fake_ident.token_tree() { Some(tt) => tt, None => break, }; if let (Some(actual_expansion), Some(hypothetical_expansion)) = ( ctx.sema.expand(&actual_macro_call), ctx.sema.expand_hypothetical( &actual_macro_call, &hypothetical_args, fake_ident_token, ), ) { let new_offset = hypothetical_expansion.1.text_range().start(); if new_offset > actual_expansion.text_range().end() { break; } original_file = actual_expansion; hypothetical_file = hypothetical_expansion.0; fake_ident_token = hypothetical_expansion.1; offset = new_offset; } else { break; } } ctx.fill(&original_file, hypothetical_file, offset); Some(ctx) } // The range of the identifier that is being completed. pub(crate) fn source_range(&self) -> TextRange { // check kind of macro-expanded token, but use range of original token match self.token.kind() { // workaroud when completion is triggered by trigger characters. IDENT => self.original_token.text_range(), _ => TextRange::offset_len(self.offset, 0.into()), } } pub(crate) fn scope(&self) -> SemanticsScope<'_, RootDatabase> { self.sema.scope_at_offset(&self.token.parent(), self.offset) } fn fill( &mut self, original_file: &SyntaxNode, file_with_fake_ident: SyntaxNode, offset: TextUnit, ) { // First, let's try to complete a reference to some declaration. if let Some(name_ref) = find_node_at_offset::(&file_with_fake_ident, offset) { // Special case, `trait T { fn foo(i_am_a_name_ref) {} }`. // See RFC#1685. if is_node::(name_ref.syntax()) { self.is_param = true; return; } self.classify_name_ref(original_file, name_ref, offset); } // Otherwise, see if this is a declaration. We can use heuristics to // suggest declaration names, see `CompletionKind::Magic`. if let Some(name) = find_node_at_offset::(&file_with_fake_ident, offset) { if let Some(bind_pat) = name.syntax().ancestors().find_map(ast::BindPat::cast) { self.is_pat_binding_or_const = true; if bind_pat.has_at() || bind_pat.is_ref() || bind_pat.is_mutable() { self.is_pat_binding_or_const = false; } if bind_pat.syntax().parent().and_then(ast::RecordFieldPatList::cast).is_some() { self.is_pat_binding_or_const = false; } if let Some(let_stmt) = bind_pat.syntax().ancestors().find_map(ast::LetStmt::cast) { if let Some(pat) = let_stmt.pat() { if bind_pat.syntax().text_range().is_subrange(&pat.syntax().text_range()) { self.is_pat_binding_or_const = false; } } } } if is_node::(name.syntax()) { self.is_param = true; return; } if name.syntax().ancestors().find_map(ast::RecordFieldPatList::cast).is_some() { self.record_lit_pat = self.sema.find_node_at_offset_with_macros(&original_file, offset); } } } fn classify_name_ref( &mut self, original_file: &SyntaxNode, name_ref: ast::NameRef, offset: TextUnit, ) { self.name_ref_syntax = find_node_at_offset(&original_file, name_ref.syntax().text_range().start()); let name_range = name_ref.syntax().text_range(); if name_ref.syntax().parent().and_then(ast::RecordField::cast).is_some() { self.record_lit_syntax = self.sema.find_node_at_offset_with_macros(&original_file, offset); } self.impl_def = self .sema .ancestors_with_macros(self.token.parent()) .take_while(|it| it.kind() != SOURCE_FILE && it.kind() != MODULE) .find_map(ast::ImplDef::cast); let top_node = name_ref .syntax() .ancestors() .take_while(|it| it.text_range() == name_range) .last() .unwrap(); match top_node.parent().map(|it| it.kind()) { Some(SOURCE_FILE) | Some(ITEM_LIST) => { self.is_new_item = true; return; } _ => (), } self.use_item_syntax = self.sema.ancestors_with_macros(self.token.parent()).find_map(ast::UseItem::cast); self.function_syntax = self .sema .ancestors_with_macros(self.token.parent()) .take_while(|it| it.kind() != SOURCE_FILE && it.kind() != MODULE) .find_map(ast::FnDef::cast); let parent = match name_ref.syntax().parent() { Some(it) => it, None => return, }; if let Some(segment) = ast::PathSegment::cast(parent.clone()) { let path = segment.parent_path(); self.is_call = path .syntax() .parent() .and_then(ast::PathExpr::cast) .and_then(|it| it.syntax().parent().and_then(ast::CallExpr::cast)) .is_some(); self.is_macro_call = path.syntax().parent().and_then(ast::MacroCall::cast).is_some(); self.is_path_type = path.syntax().parent().and_then(ast::PathType::cast).is_some(); self.has_type_args = segment.type_arg_list().is_some(); if let Some(path) = hir::Path::from_ast(path.clone()) { if let Some(path_prefix) = path.qualifier() { self.path_prefix = Some(path_prefix); return; } } if path.qualifier().is_none() { self.is_trivial_path = true; // Find either enclosing expr statement (thing with `;`) or a // block. If block, check that we are the last expr. self.can_be_stmt = name_ref .syntax() .ancestors() .find_map(|node| { if let Some(stmt) = ast::ExprStmt::cast(node.clone()) { return Some( stmt.syntax().text_range() == name_ref.syntax().text_range(), ); } if let Some(block) = ast::Block::cast(node) { return Some( block.expr().map(|e| e.syntax().text_range()) == Some(name_ref.syntax().text_range()), ); } None }) .unwrap_or(false); if let Some(off) = name_ref.syntax().text_range().start().checked_sub(2.into()) { if let Some(if_expr) = self.sema.find_node_at_offset_with_macros::(original_file, off) { if if_expr.syntax().text_range().end() < name_ref.syntax().text_range().start() { self.after_if = true; } } } } } if let Some(field_expr) = ast::FieldExpr::cast(parent.clone()) { // The receiver comes before the point of insertion of the fake // ident, so it should have the same range in the non-modified file self.dot_receiver = field_expr .expr() .map(|e| e.syntax().text_range()) .and_then(|r| find_node_with_range(original_file, r)); self.dot_receiver_is_ambiguous_float_literal = if let Some(ast::Expr::Literal(l)) = &self.dot_receiver { match l.kind() { ast::LiteralKind::FloatNumber { .. } => l.token().text().ends_with('.'), _ => false, } } else { false } } if let Some(method_call_expr) = ast::MethodCallExpr::cast(parent) { // As above self.dot_receiver = method_call_expr .expr() .map(|e| e.syntax().text_range()) .and_then(|r| find_node_with_range(original_file, r)); self.is_call = true; } } } fn find_node_with_range(syntax: &SyntaxNode, range: TextRange) -> Option { find_covering_element(syntax, range).ancestors().find_map(N::cast) } fn is_node(node: &SyntaxNode) -> bool { match node.ancestors().find_map(N::cast) { None => false, Some(n) => n.syntax().text_range() == node.text_range(), } }