//! "Recursive" Syntax highlighting for code in doctests and fixtures. use std::mem; use either::Either; use hir::{InFile, Semantics}; use ide_db::{call_info::ActiveParameter, helpers::rust_doc::is_rust_fence, SymbolKind}; use syntax::{ ast::{self, AstNode, IsString}, AstToken, NodeOrToken, SyntaxNode, SyntaxToken, TextRange, TextSize, }; use crate::{ doc_links::{doc_attributes, extract_definitions_from_markdown, resolve_doc_path_for_def}, Analysis, HlMod, HlRange, HlTag, RootDatabase, }; use super::{highlights::Highlights, injector::Injector}; pub(super) fn ra_fixture( hl: &mut Highlights, sema: &Semantics, literal: ast::String, expanded: SyntaxToken, ) -> Option<()> { let active_parameter = ActiveParameter::at_token(sema, expanded)?; if !active_parameter.ident().map_or(false, |name| name.text().starts_with("ra_fixture")) { return None; } let value = literal.value()?; if let Some(range) = literal.open_quote_text_range() { hl.add(HlRange { range, highlight: HlTag::StringLiteral.into(), binding_hash: None }) } let mut inj = Injector::default(); let mut text = &*value; let mut offset: TextSize = 0.into(); while !text.is_empty() { let marker = "$0"; let idx = text.find(marker).unwrap_or(text.len()); let (chunk, next) = text.split_at(idx); inj.add(chunk, TextRange::at(offset, TextSize::of(chunk))); text = next; offset += TextSize::of(chunk); if let Some(next) = text.strip_prefix(marker) { if let Some(range) = literal.map_range_up(TextRange::at(offset, TextSize::of(marker))) { hl.add(HlRange { range, highlight: HlTag::Keyword.into(), binding_hash: None }); } text = next; let marker_len = TextSize::of(marker); offset += marker_len; } } let (analysis, tmp_file_id) = Analysis::from_single_file(inj.text().to_string()); for mut hl_range in analysis.highlight(tmp_file_id).unwrap() { for range in inj.map_range_up(hl_range.range) { if let Some(range) = literal.map_range_up(range) { hl_range.range = range; hl.add(hl_range); } } } if let Some(range) = literal.close_quote_text_range() { hl.add(HlRange { range, highlight: HlTag::StringLiteral.into(), binding_hash: None }) } Some(()) } const RUSTDOC_FENCE: &'static str = "```"; /// Injection of syntax highlighting of doctests. pub(super) fn doc_comment( hl: &mut Highlights, sema: &Semantics, node: InFile<&SyntaxNode>, ) { let (attributes, def) = match doc_attributes(sema, node.value) { Some(it) => it, None => return, }; let mut inj = Injector::default(); inj.add_unmapped("fn doctest() {\n"); let attrs_source_map = attributes.source_map(sema.db); let mut is_codeblock = false; let mut is_doctest = false; // Replace the original, line-spanning comment ranges by new, only comment-prefix // spanning comment ranges. let mut new_comments = Vec::new(); let mut string; if let Some((docs, doc_mapping)) = attributes.docs_with_rangemap(sema.db) { extract_definitions_from_markdown(docs.as_str()) .into_iter() .filter_map(|(range, link, ns)| { let def = resolve_doc_path_for_def(sema.db, def, &link, ns)?; let InFile { file_id, value: range } = doc_mapping.map(range)?; (file_id == node.file_id).then(|| (range, def)) }) .for_each(|(range, def)| { hl.add(HlRange { range, highlight: module_def_to_hl_tag(def) | HlMod::Documentation | HlMod::Injected | HlMod::IntraDocLink, binding_hash: None, }) }); } for attr in attributes.by_key("doc").attrs() { let InFile { file_id, value: src } = attrs_source_map.source_of(attr); if file_id != node.file_id { continue; } let (line, range, prefix) = match &src { Either::Left(it) => { string = match find_doc_string_in_attr(attr, it) { Some(it) => it, None => continue, }; let text_range = string.syntax().text_range(); let text_range = TextRange::new( text_range.start() + TextSize::from(1), text_range.end() - TextSize::from(1), ); let text = string.text(); (&text[1..text.len() - 1], text_range, "") } Either::Right(comment) => { (comment.text(), comment.syntax().text_range(), comment.prefix()) } }; let mut pos = TextSize::from(prefix.len() as u32); let mut range_start = range.start(); for line in line.split('\n') { let line_len = TextSize::from(line.len() as u32); let prev_range_start = { let next_range_start = range_start + line_len + TextSize::from(1); mem::replace(&mut range_start, next_range_start) }; // only first line has the prefix so take it away for future iterations let mut pos = mem::take(&mut pos); match line.find(RUSTDOC_FENCE) { Some(idx) => { is_codeblock = !is_codeblock; // Check whether code is rust by inspecting fence guards let guards = &line[idx + RUSTDOC_FENCE.len()..]; let is_rust = is_rust_fence(guards); is_doctest = is_codeblock && is_rust; continue; } None if !is_doctest => continue, None => (), } // whitespace after comment is ignored if let Some(ws) = line[pos.into()..].chars().next().filter(|c| c.is_whitespace()) { pos += TextSize::of(ws); } // lines marked with `#` should be ignored in output, we skip the `#` char if line[pos.into()..].starts_with('#') { pos += TextSize::of('#'); } new_comments.push(TextRange::at(prev_range_start, pos)); inj.add(&line[pos.into()..], TextRange::new(pos, line_len) + prev_range_start); inj.add_unmapped("\n"); } } if new_comments.is_empty() { return; // no need to run an analysis on an empty file } inj.add_unmapped("\n}"); let (analysis, tmp_file_id) = Analysis::from_single_file(inj.text().to_string()); for HlRange { range, highlight, binding_hash } in analysis.with_db(|db| super::highlight(db, tmp_file_id, None, true)).unwrap() { for range in inj.map_range_up(range) { hl.add(HlRange { range, highlight: highlight | HlMod::Injected, binding_hash }); } } for range in new_comments { hl.add(HlRange { range, highlight: HlTag::Comment | HlMod::Documentation, binding_hash: None, }); } } fn find_doc_string_in_attr(attr: &hir::Attr, it: &ast::Attr) -> Option { match it.expr() { // #[doc = lit] Some(ast::Expr::Literal(lit)) => match lit.kind() { ast::LiteralKind::String(it) => Some(it), _ => None, }, // #[cfg_attr(..., doc = "", ...)] None => { // We gotta hunt the string token manually here let text = attr.string_value()?; // FIXME: We just pick the first string literal that has the same text as the doc attribute // This means technically we might highlight the wrong one it.syntax() .descendants_with_tokens() .filter_map(NodeOrToken::into_token) .filter_map(ast::String::cast) .find(|string| { string.text().get(1..string.text().len() - 1).map_or(false, |it| it == text) }) } _ => None, } } fn module_def_to_hl_tag(def: hir::ModuleDef) -> HlTag { let symbol = match def { hir::ModuleDef::Module(_) => SymbolKind::Module, hir::ModuleDef::Function(_) => SymbolKind::Function, hir::ModuleDef::Adt(hir::Adt::Struct(_)) => SymbolKind::Struct, hir::ModuleDef::Adt(hir::Adt::Enum(_)) => SymbolKind::Enum, hir::ModuleDef::Adt(hir::Adt::Union(_)) => SymbolKind::Union, hir::ModuleDef::Variant(_) => SymbolKind::Variant, hir::ModuleDef::Const(_) => SymbolKind::Const, hir::ModuleDef::Static(_) => SymbolKind::Static, hir::ModuleDef::Trait(_) => SymbolKind::Trait, hir::ModuleDef::TypeAlias(_) => SymbolKind::TypeAlias, hir::ModuleDef::BuiltinType(_) => return HlTag::BuiltinType, }; HlTag::Symbol(symbol) }