From 6504c3c32a9795173fc0c9c94befe7f5d0e7fe9e Mon Sep 17 00:00:00 2001 From: Jonas Schievink Date: Mon, 21 Jun 2021 22:57:54 +0200 Subject: Move subtree collection out of `TokenConvertor` --- crates/mbe/src/syntax_bridge.rs | 248 ++++++++++++++++++++-------------------- 1 file changed, 124 insertions(+), 124 deletions(-) (limited to 'crates') diff --git a/crates/mbe/src/syntax_bridge.rs b/crates/mbe/src/syntax_bridge.rs index 7526bd8e6..adf5a56ec 100644 --- a/crates/mbe/src/syntax_bridge.rs +++ b/crates/mbe/src/syntax_bridge.rs @@ -24,7 +24,7 @@ pub fn ast_to_token_tree(ast: &impl ast::AstNode) -> (tt::Subtree, TokenMap) { pub fn syntax_node_to_token_tree(node: &SyntaxNode) -> (tt::Subtree, TokenMap) { let global_offset = node.text_range().start(); let mut c = Convertor::new(node, global_offset); - let subtree = c.go(); + let subtree = convert_tokens(&mut c); c.id_alloc.map.shrink_to_fit(); (subtree, c.id_alloc.map) } @@ -80,7 +80,7 @@ pub fn parse_to_token_tree(text: &str) -> Option<(tt::Subtree, TokenMap)> { }, }; - let subtree = conv.go(); + let subtree = convert_tokens(&mut conv); Some((subtree, conv.id_alloc.map)) } @@ -121,6 +121,128 @@ pub fn parse_exprs_with_sep(tt: &tt::Subtree, sep: char) -> Vec { res } +fn convert_tokens(conv: &mut C) -> tt::Subtree { + let mut subtree = tt::Subtree { delimiter: None, ..Default::default() }; + while conv.peek().is_some() { + collect_leaf(conv, &mut subtree.token_trees); + } + if subtree.token_trees.len() == 1 { + if let tt::TokenTree::Subtree(first) = &subtree.token_trees[0] { + return first.clone(); + } + } + return subtree; + + fn collect_leaf(conv: &mut C, result: &mut Vec) { + let (token, range) = match conv.bump() { + None => return, + Some(it) => it, + }; + + let k: SyntaxKind = token.kind(); + if k == COMMENT { + if let Some(tokens) = conv.convert_doc_comment(&token) { + result.extend(tokens); + } + return; + } + + result.push(if k.is_punct() && k != UNDERSCORE { + assert_eq!(range.len(), TextSize::of('.')); + let delim = match k { + T!['('] => Some((tt::DelimiterKind::Parenthesis, T![')'])), + T!['{'] => Some((tt::DelimiterKind::Brace, T!['}'])), + T!['['] => Some((tt::DelimiterKind::Bracket, T![']'])), + _ => None, + }; + + if let Some((kind, closed)) = delim { + let mut subtree = tt::Subtree::default(); + let (id, idx) = conv.id_alloc().open_delim(range); + subtree.delimiter = Some(tt::Delimiter { id, kind }); + + while conv.peek().map_or(false, |it| it.kind() != closed) { + collect_leaf(conv, &mut subtree.token_trees); + } + let last_range = match conv.bump() { + None => { + // For error resilience, we insert an char punct for the opening delim here + conv.id_alloc().close_delim(idx, None); + let leaf: tt::Leaf = tt::Punct { + id: conv.id_alloc().alloc(range), + char: token.to_char().unwrap(), + spacing: tt::Spacing::Alone, + } + .into(); + result.push(leaf.into()); + result.extend(subtree.token_trees); + return; + } + Some(it) => it.1, + }; + conv.id_alloc().close_delim(idx, Some(last_range)); + subtree.into() + } else { + let spacing = match conv.peek() { + Some(next) + if next.kind().is_trivia() + || next.kind() == T!['['] + || next.kind() == T!['{'] + || next.kind() == T!['('] => + { + tt::Spacing::Alone + } + Some(next) if next.kind().is_punct() && next.kind() != UNDERSCORE => { + tt::Spacing::Joint + } + _ => tt::Spacing::Alone, + }; + let char = match token.to_char() { + Some(c) => c, + None => { + panic!("Token from lexer must be single char: token = {:#?}", token); + } + }; + tt::Leaf::from(tt::Punct { char, spacing, id: conv.id_alloc().alloc(range) }).into() + } + } else { + macro_rules! make_leaf { + ($i:ident) => { + tt::$i { id: conv.id_alloc().alloc(range), text: token.to_text() }.into() + }; + } + let leaf: tt::Leaf = match k { + T![true] | T![false] => make_leaf!(Ident), + IDENT => make_leaf!(Ident), + UNDERSCORE => make_leaf!(Ident), + k if k.is_keyword() => make_leaf!(Ident), + k if k.is_literal() => make_leaf!(Literal), + LIFETIME_IDENT => { + let char_unit = TextSize::of('\''); + let r = TextRange::at(range.start(), char_unit); + let apostrophe = tt::Leaf::from(tt::Punct { + char: '\'', + spacing: tt::Spacing::Joint, + id: conv.id_alloc().alloc(r), + }); + result.push(apostrophe.into()); + + let r = TextRange::at(range.start() + char_unit, range.len() - char_unit); + let ident = tt::Leaf::from(tt::Ident { + text: SmolStr::new(&token.to_text()[1..]), + id: conv.id_alloc().alloc(r), + }); + result.push(ident.into()); + return; + } + _ => return, + }; + + leaf.into() + }); + } +} + /// Returns the textual content of a doc comment block as a quoted string /// That is, strips leading `///` (or `/**`, etc) /// and strips the ending `*/` @@ -242,128 +364,6 @@ trait SrcToken: std::fmt::Debug { trait TokenConvertor { type Token: SrcToken; - fn go(&mut self) -> tt::Subtree { - let mut subtree = tt::Subtree { delimiter: None, ..Default::default() }; - while self.peek().is_some() { - self.collect_leaf(&mut subtree.token_trees); - } - if subtree.token_trees.len() == 1 { - if let tt::TokenTree::Subtree(first) = &subtree.token_trees[0] { - return first.clone(); - } - } - subtree - } - - fn collect_leaf(&mut self, result: &mut Vec) { - let (token, range) = match self.bump() { - None => return, - Some(it) => it, - }; - - let k: SyntaxKind = token.kind(); - if k == COMMENT { - if let Some(tokens) = self.convert_doc_comment(&token) { - result.extend(tokens); - } - return; - } - - result.push(if k.is_punct() && k != UNDERSCORE { - assert_eq!(range.len(), TextSize::of('.')); - let delim = match k { - T!['('] => Some((tt::DelimiterKind::Parenthesis, T![')'])), - T!['{'] => Some((tt::DelimiterKind::Brace, T!['}'])), - T!['['] => Some((tt::DelimiterKind::Bracket, T![']'])), - _ => None, - }; - - if let Some((kind, closed)) = delim { - let mut subtree = tt::Subtree::default(); - let (id, idx) = self.id_alloc().open_delim(range); - subtree.delimiter = Some(tt::Delimiter { id, kind }); - - while self.peek().map_or(false, |it| it.kind() != closed) { - self.collect_leaf(&mut subtree.token_trees); - } - let last_range = match self.bump() { - None => { - // For error resilience, we insert an char punct for the opening delim here - self.id_alloc().close_delim(idx, None); - let leaf: tt::Leaf = tt::Punct { - id: self.id_alloc().alloc(range), - char: token.to_char().unwrap(), - spacing: tt::Spacing::Alone, - } - .into(); - result.push(leaf.into()); - result.extend(subtree.token_trees); - return; - } - Some(it) => it.1, - }; - self.id_alloc().close_delim(idx, Some(last_range)); - subtree.into() - } else { - let spacing = match self.peek() { - Some(next) - if next.kind().is_trivia() - || next.kind() == T!['['] - || next.kind() == T!['{'] - || next.kind() == T!['('] => - { - tt::Spacing::Alone - } - Some(next) if next.kind().is_punct() && next.kind() != UNDERSCORE => { - tt::Spacing::Joint - } - _ => tt::Spacing::Alone, - }; - let char = match token.to_char() { - Some(c) => c, - None => { - panic!("Token from lexer must be single char: token = {:#?}", token); - } - }; - tt::Leaf::from(tt::Punct { char, spacing, id: self.id_alloc().alloc(range) }).into() - } - } else { - macro_rules! make_leaf { - ($i:ident) => { - tt::$i { id: self.id_alloc().alloc(range), text: token.to_text() }.into() - }; - } - let leaf: tt::Leaf = match k { - T![true] | T![false] => make_leaf!(Ident), - IDENT => make_leaf!(Ident), - UNDERSCORE => make_leaf!(Ident), - k if k.is_keyword() => make_leaf!(Ident), - k if k.is_literal() => make_leaf!(Literal), - LIFETIME_IDENT => { - let char_unit = TextSize::of('\''); - let r = TextRange::at(range.start(), char_unit); - let apostrophe = tt::Leaf::from(tt::Punct { - char: '\'', - spacing: tt::Spacing::Joint, - id: self.id_alloc().alloc(r), - }); - result.push(apostrophe.into()); - - let r = TextRange::at(range.start() + char_unit, range.len() - char_unit); - let ident = tt::Leaf::from(tt::Ident { - text: SmolStr::new(&token.to_text()[1..]), - id: self.id_alloc().alloc(r), - }); - result.push(ident.into()); - return; - } - _ => return, - }; - - leaf.into() - }); - } - fn convert_doc_comment(&self, token: &Self::Token) -> Option>; fn bump(&mut self) -> Option<(Self::Token, TextRange)>; -- cgit v1.2.3 From c6669776e1174170c157b685a66026d1a31ede85 Mon Sep 17 00:00:00 2001 From: Jonas Schievink Date: Tue, 22 Jun 2021 14:23:31 +0200 Subject: Rewrite `convert_tokens` to use an explicit stack --- crates/mbe/src/syntax_bridge.rs | 120 ++++++++++++++++++++++++++-------------- 1 file changed, 80 insertions(+), 40 deletions(-) (limited to 'crates') diff --git a/crates/mbe/src/syntax_bridge.rs b/crates/mbe/src/syntax_bridge.rs index adf5a56ec..ae6058cbc 100644 --- a/crates/mbe/src/syntax_bridge.rs +++ b/crates/mbe/src/syntax_bridge.rs @@ -122,20 +122,25 @@ pub fn parse_exprs_with_sep(tt: &tt::Subtree, sep: char) -> Vec { } fn convert_tokens(conv: &mut C) -> tt::Subtree { - let mut subtree = tt::Subtree { delimiter: None, ..Default::default() }; - while conv.peek().is_some() { - collect_leaf(conv, &mut subtree.token_trees); + struct StackEntry { + subtree: tt::Subtree, + idx: usize, + open_range: TextRange, } - if subtree.token_trees.len() == 1 { - if let tt::TokenTree::Subtree(first) = &subtree.token_trees[0] { - return first.clone(); - } - } - return subtree; - fn collect_leaf(conv: &mut C, result: &mut Vec) { + let entry = StackEntry { + subtree: tt::Subtree { delimiter: None, ..Default::default() }, + // never used (delimiter is `None`) + idx: !0, + open_range: TextRange::empty(TextSize::of('.')), + }; + let mut stack = vec![entry]; + + loop { + let entry = stack.last_mut().unwrap(); + let result = &mut entry.subtree.token_trees; let (token, range) = match conv.bump() { - None => return, + None => break, Some(it) => it, }; @@ -144,44 +149,40 @@ fn convert_tokens(conv: &mut C) -> tt::Subtree { if let Some(tokens) = conv.convert_doc_comment(&token) { result.extend(tokens); } - return; + continue; } result.push(if k.is_punct() && k != UNDERSCORE { assert_eq!(range.len(), TextSize::of('.')); + + if let Some(delim) = entry.subtree.delimiter { + let expected = match delim.kind { + tt::DelimiterKind::Parenthesis => T![')'], + tt::DelimiterKind::Brace => T!['}'], + tt::DelimiterKind::Bracket => T![']'], + }; + + if k == expected { + let entry = stack.pop().unwrap(); + conv.id_alloc().close_delim(entry.idx, Some(range)); + stack.last_mut().unwrap().subtree.token_trees.push(entry.subtree.into()); + continue; + } + } + let delim = match k { - T!['('] => Some((tt::DelimiterKind::Parenthesis, T![')'])), - T!['{'] => Some((tt::DelimiterKind::Brace, T!['}'])), - T!['['] => Some((tt::DelimiterKind::Bracket, T![']'])), + T!['('] => Some(tt::DelimiterKind::Parenthesis), + T!['{'] => Some(tt::DelimiterKind::Brace), + T!['['] => Some(tt::DelimiterKind::Bracket), _ => None, }; - if let Some((kind, closed)) = delim { + if let Some(kind) = delim { let mut subtree = tt::Subtree::default(); let (id, idx) = conv.id_alloc().open_delim(range); subtree.delimiter = Some(tt::Delimiter { id, kind }); - - while conv.peek().map_or(false, |it| it.kind() != closed) { - collect_leaf(conv, &mut subtree.token_trees); - } - let last_range = match conv.bump() { - None => { - // For error resilience, we insert an char punct for the opening delim here - conv.id_alloc().close_delim(idx, None); - let leaf: tt::Leaf = tt::Punct { - id: conv.id_alloc().alloc(range), - char: token.to_char().unwrap(), - spacing: tt::Spacing::Alone, - } - .into(); - result.push(leaf.into()); - result.extend(subtree.token_trees); - return; - } - Some(it) => it.1, - }; - conv.id_alloc().close_delim(idx, Some(last_range)); - subtree.into() + stack.push(StackEntry { subtree, idx, open_range: range }); + continue; } else { let spacing = match conv.peek() { Some(next) @@ -233,14 +234,44 @@ fn convert_tokens(conv: &mut C) -> tt::Subtree { id: conv.id_alloc().alloc(r), }); result.push(ident.into()); - return; + continue; } - _ => return, + _ => continue, }; leaf.into() }); } + + // If we get here, we've consumed all input tokens. + // We might have more than one subtree in the stack, if the delimiters are improperly balanced. + // Merge them so we're left with one. + while stack.len() > 1 { + let entry = stack.pop().unwrap(); + let parent = stack.last_mut().unwrap(); + + conv.id_alloc().close_delim(entry.idx, None); + let leaf: tt::Leaf = tt::Punct { + id: conv.id_alloc().alloc(entry.open_range), + char: match entry.subtree.delimiter.unwrap().kind { + tt::DelimiterKind::Parenthesis => '(', + tt::DelimiterKind::Brace => '{', + tt::DelimiterKind::Bracket => '[', + }, + spacing: tt::Spacing::Alone, + } + .into(); + parent.subtree.token_trees.push(leaf.into()); + parent.subtree.token_trees.extend(entry.subtree.token_trees); + } + + let subtree = stack.pop().unwrap().subtree; + if subtree.token_trees.len() == 1 { + if let tt::TokenTree::Subtree(first) = &subtree.token_trees[0] { + return first.clone(); + } + } + subtree } /// Returns the textual content of a doc comment block as a quoted string @@ -683,6 +714,7 @@ mod tests { algo::{insert_children, InsertPosition}, ast::AstNode, }; + use test_utils::assert_eq_text; #[test] fn convert_tt_token_source() { @@ -792,4 +824,12 @@ mod tests { let tt = ast_to_token_tree(&struct_def).0; token_tree_to_syntax_node(&tt, FragmentKind::Item).unwrap(); } + + #[test] + fn test_missing_closing_delim() { + let source_file = ast::SourceFile::parse("m!(x").tree(); + let node = source_file.syntax().descendants().find_map(ast::TokenTree::cast).unwrap(); + let tt = ast_to_token_tree(&node).0.to_string(); + assert_eq_text!(&*tt, "( x"); + } } -- cgit v1.2.3