aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
-rw-r--r--crates/ra_mbe/src/lib.rs57
-rw-r--r--crates/ra_mbe/src/subtree_parser.rs3
-rw-r--r--crates/ra_mbe/src/subtree_source.rs354
-rw-r--r--crates/ra_mbe/src/syntax_bridge.rs47
-rw-r--r--crates/ra_parser/src/parser.rs12
5 files changed, 255 insertions, 218 deletions
diff --git a/crates/ra_mbe/src/lib.rs b/crates/ra_mbe/src/lib.rs
index a21ea4dbc..4126854d1 100644
--- a/crates/ra_mbe/src/lib.rs
+++ b/crates/ra_mbe/src/lib.rs
@@ -39,7 +39,7 @@ pub enum ExpandError {
39 BindingError(String), 39 BindingError(String),
40} 40}
41 41
42pub use crate::syntax_bridge::{ast_to_token_tree, token_tree_to_ast_item_list}; 42pub use crate::syntax_bridge::{ast_to_token_tree, token_tree_to_ast_item_list, syntax_node_to_token_tree};
43 43
44/// This struct contains AST for a single `macro_rules` definition. What might 44/// This struct contains AST for a single `macro_rules` definition. What might
45/// be very confusing is that AST has almost exactly the same shape as 45/// be very confusing is that AST has almost exactly the same shape as
@@ -192,6 +192,15 @@ impl_froms!(TokenTree: Leaf, Subtree);
192 pub(crate) fn assert_expansion(rules: &MacroRules, invocation: &str, expansion: &str) { 192 pub(crate) fn assert_expansion(rules: &MacroRules, invocation: &str, expansion: &str) {
193 let expanded = expand(rules, invocation); 193 let expanded = expand(rules, invocation);
194 assert_eq!(expanded.to_string(), expansion); 194 assert_eq!(expanded.to_string(), expansion);
195
196 let tree = token_tree_to_ast_item_list(&expanded);
197
198 // Eat all white space by parse it back and forth
199 let expansion = ast::SourceFile::parse(expansion);
200 let expansion = syntax_node_to_token_tree(expansion.syntax()).unwrap().0;
201 let file = token_tree_to_ast_item_list(&expansion);
202
203 assert_eq!(tree.syntax().debug_dump().trim(), file.syntax().debug_dump().trim());
195 } 204 }
196 205
197 #[test] 206 #[test]
@@ -288,6 +297,36 @@ impl_froms!(TokenTree: Leaf, Subtree);
288 } 297 }
289 298
290 #[test] 299 #[test]
300 fn test_match_group_pattern_with_multiple_defs() {
301 let rules = create_rules(
302 r#"
303 macro_rules! foo {
304 ($ ($ i:ident),*) => ( struct Bar { $ (
305 fn $ i {}
306 )*} );
307 }
308"#,
309 );
310
311 assert_expansion(&rules, "foo! { foo, bar }", "struct Bar {fn foo {} fn bar {}}");
312 }
313
314 #[test]
315 fn test_match_group_pattern_with_multiple_statement() {
316 let rules = create_rules(
317 r#"
318 macro_rules! foo {
319 ($ ($ i:ident),*) => ( fn baz { $ (
320 $ i ();
321 )*} );
322 }
323"#,
324 );
325
326 assert_expansion(&rules, "foo! { foo, bar }", "fn baz {foo () ; bar () ;}");
327 }
328
329 #[test]
291 fn expand_to_item_list() { 330 fn expand_to_item_list() {
292 let rules = create_rules( 331 let rules = create_rules(
293 " 332 "
@@ -415,7 +454,7 @@ SOURCE_FILE@[0; 40)
415 assert_expansion( 454 assert_expansion(
416 &rules, 455 &rules,
417 "foo! { bar::<u8>::baz::<u8> }", 456 "foo! { bar::<u8>::baz::<u8> }",
418 "fn foo () {let a = bar ::< u8 > ::baz ::< u8 > ;}", 457 "fn foo () {let a = bar :: < u8 > :: baz :: < u8 > ;}",
419 ); 458 );
420 } 459 }
421 460
@@ -432,4 +471,18 @@ SOURCE_FILE@[0; 40)
432 ); 471 );
433 assert_expansion(&rules, "foo! { foo, bar }", "fn foo () {let a = foo ; let b = bar ;}"); 472 assert_expansion(&rules, "foo! { foo, bar }", "fn foo () {let a = foo ; let b = bar ;}");
434 } 473 }
474
475 #[test]
476 fn test_path_with_path() {
477 let rules = create_rules(
478 r#"
479 macro_rules! foo {
480 ($ i:path) => {
481 fn foo() { let a = $ i :: bar; }
482 }
483 }
484"#,
485 );
486 assert_expansion(&rules, "foo! { foo }", "fn foo () {let a = foo :: bar ;}");
487 }
435} 488}
diff --git a/crates/ra_mbe/src/subtree_parser.rs b/crates/ra_mbe/src/subtree_parser.rs
index ce39a40bb..164240d92 100644
--- a/crates/ra_mbe/src/subtree_parser.rs
+++ b/crates/ra_mbe/src/subtree_parser.rs
@@ -34,8 +34,7 @@ impl<'a> Parser<'a> {
34 where 34 where
35 F: FnOnce(&dyn TokenSource, &mut dyn TreeSink), 35 F: FnOnce(&dyn TokenSource, &mut dyn TreeSink),
36 { 36 {
37 let mut src = SubtreeTokenSource::new(self.subtree); 37 let mut src = SubtreeTokenSource::new(&self.subtree.token_trees[*self.cur_pos..]);
38 src.start_from_nth(*self.cur_pos);
39 let mut sink = OffsetTokenSink { token_pos: 0 }; 38 let mut sink = OffsetTokenSink { token_pos: 0 };
40 39
41 f(&src, &mut sink); 40 f(&src, &mut sink);
diff --git a/crates/ra_mbe/src/subtree_source.rs b/crates/ra_mbe/src/subtree_source.rs
index 4b37c2bda..6aa20057e 100644
--- a/crates/ra_mbe/src/subtree_source.rs
+++ b/crates/ra_mbe/src/subtree_source.rs
@@ -2,6 +2,64 @@ use ra_parser::{TokenSource};
2use ra_syntax::{classify_literal, SmolStr, SyntaxKind, SyntaxKind::*}; 2use ra_syntax::{classify_literal, SmolStr, SyntaxKind, SyntaxKind::*};
3use std::cell::{RefCell}; 3use std::cell::{RefCell};
4 4
5// A Sequece of Token,
6#[derive(Debug, Clone, Eq, PartialEq)]
7pub(super) enum TokenSeq<'a> {
8 Subtree(&'a tt::Subtree),
9 Seq(&'a [tt::TokenTree]),
10}
11
12impl<'a> From<&'a tt::Subtree> for TokenSeq<'a> {
13 fn from(s: &'a tt::Subtree) -> TokenSeq<'a> {
14 TokenSeq::Subtree(s)
15 }
16}
17
18impl<'a> From<&'a [tt::TokenTree]> for TokenSeq<'a> {
19 fn from(s: &'a [tt::TokenTree]) -> TokenSeq<'a> {
20 TokenSeq::Seq(s)
21 }
22}
23
24enum DelimToken<'a> {
25 Delim(&'a tt::Delimiter, bool),
26 Token(&'a tt::TokenTree),
27 End,
28}
29
30impl<'a> TokenSeq<'a> {
31 fn get(&self, pos: usize) -> DelimToken<'a> {
32 match self {
33 TokenSeq::Subtree(subtree) => {
34 let len = subtree.token_trees.len() + 2;
35 match pos {
36 p if p >= len => DelimToken::End,
37 p if p == len - 1 => DelimToken::Delim(&subtree.delimiter, true),
38 0 => DelimToken::Delim(&subtree.delimiter, false),
39 p => DelimToken::Token(&subtree.token_trees[p - 1]),
40 }
41 }
42 TokenSeq::Seq(tokens) => {
43 tokens.get(pos).map(DelimToken::Token).unwrap_or(DelimToken::End)
44 }
45 }
46 }
47
48 fn len(&self) -> usize {
49 match self {
50 TokenSeq::Subtree(subtree) => subtree.token_trees.len() + 2,
51 TokenSeq::Seq(tokens) => tokens.len(),
52 }
53 }
54
55 fn child_slice(&self) -> &[tt::TokenTree] {
56 match self {
57 TokenSeq::Subtree(subtree) => &subtree.token_trees,
58 TokenSeq::Seq(tokens) => &tokens,
59 }
60 }
61}
62
5#[derive(Debug, Clone, Eq, PartialEq)] 63#[derive(Debug, Clone, Eq, PartialEq)]
6struct TtToken { 64struct TtToken {
7 pub kind: SyntaxKind, 65 pub kind: SyntaxKind,
@@ -12,29 +70,27 @@ struct TtToken {
12 70
13#[derive(Debug, Clone, Eq, PartialEq)] 71#[derive(Debug, Clone, Eq, PartialEq)]
14enum WalkCursor { 72enum WalkCursor {
15 DelimiterBegin(Option<TtToken>), 73 Token(usize, TtToken),
16 Token(usize, Option<TtToken>),
17 DelimiterEnd(Option<TtToken>),
18 Eof, 74 Eof,
19} 75}
20 76
21#[derive(Debug)] 77#[derive(Debug)]
22struct SubTreeWalker<'a> { 78struct SubTreeWalker<'a> {
23 pos: usize, 79 pos: usize,
24 stack: Vec<(&'a tt::Subtree, Option<usize>)>, 80 stack: Vec<(TokenSeq<'a>, usize)>,
25 cursor: WalkCursor, 81 cursor: WalkCursor,
26 last_steps: Vec<usize>, 82 last_steps: Vec<usize>,
27 subtree: &'a tt::Subtree, 83 ts: TokenSeq<'a>,
28} 84}
29 85
30impl<'a> SubTreeWalker<'a> { 86impl<'a> SubTreeWalker<'a> {
31 fn new(subtree: &tt::Subtree) -> SubTreeWalker { 87 fn new(ts: TokenSeq<'a>) -> SubTreeWalker {
32 let mut res = SubTreeWalker { 88 let mut res = SubTreeWalker {
33 pos: 0, 89 pos: 0,
34 stack: vec![], 90 stack: vec![],
35 cursor: WalkCursor::Eof, 91 cursor: WalkCursor::Eof,
36 last_steps: vec![], 92 last_steps: vec![],
37 subtree, 93 ts,
38 }; 94 };
39 95
40 res.reset(); 96 res.reset();
@@ -47,209 +103,106 @@ impl<'a> SubTreeWalker<'a> {
47 103
48 fn reset(&mut self) { 104 fn reset(&mut self) {
49 self.pos = 0; 105 self.pos = 0;
50 self.stack = vec![(self.subtree, None)]; 106 self.stack = vec![];
51 self.cursor = WalkCursor::DelimiterBegin(convert_delim(self.subtree.delimiter, false));
52 self.last_steps = vec![]; 107 self.last_steps = vec![];
53 108
54 while self.is_empty_delimiter() { 109 self.cursor = match self.ts.get(0) {
55 self.forward_unchecked(); 110 DelimToken::Token(token) => match token {
56 } 111 tt::TokenTree::Subtree(subtree) => {
57 } 112 WalkCursor::Token(0, convert_delim(subtree.delimiter, false))
58 113 }
59 // This funciton will fast forward the cursor, 114 tt::TokenTree::Leaf(leaf) => {
60 // Such that backward will stop at `start_pos` point 115 let next_tokens = self.ts.child_slice();
61 fn start_from_nth(&mut self, start_pos: usize) { 116 WalkCursor::Token(0, convert_leaf(&next_tokens, leaf))
62 self.reset(); 117 }
63 self.pos = start_pos; 118 },
64 self.cursor = self.walk_token(start_pos, 0, false); 119 DelimToken::Delim(delim, is_end) => {
65 120 assert!(!is_end);
66 while self.is_empty_delimiter() { 121 WalkCursor::Token(0, convert_delim(*delim, false))
67 self.forward_unchecked(); 122 }
123 DelimToken::End => WalkCursor::Eof,
68 } 124 }
69 } 125 }
70 126
71 fn current(&self) -> Option<&TtToken> { 127 fn current(&self) -> Option<&TtToken> {
72 match &self.cursor { 128 match &self.cursor {
73 WalkCursor::DelimiterBegin(t) => t.as_ref(), 129 WalkCursor::Token(_, t) => Some(t),
74 WalkCursor::Token(_, t) => t.as_ref(),
75 WalkCursor::DelimiterEnd(t) => t.as_ref(),
76 WalkCursor::Eof => None, 130 WalkCursor::Eof => None,
77 } 131 }
78 } 132 }
79 133
80 fn is_empty_delimiter(&self) -> bool { 134 fn top(&self) -> &TokenSeq {
81 match &self.cursor { 135 self.stack.last().map(|(t, _)| t).unwrap_or(&self.ts)
82 WalkCursor::DelimiterBegin(None) => true,
83 WalkCursor::DelimiterEnd(None) => true,
84 _ => false,
85 }
86 } 136 }
87 137
88 /// Move cursor backward by 1 step with empty checking 138 /// Move cursor backward by 1 step
89 fn backward(&mut self) { 139 fn backward(&mut self) {
90 if self.last_steps.is_empty() { 140 if self.last_steps.is_empty() {
91 return; 141 return;
92 } 142 }
143
93 self.pos -= 1; 144 self.pos -= 1;
94 loop { 145 let last_step = self.last_steps.pop().unwrap();
95 self.backward_unchecked();
96 // Skip Empty delimiter
97 if self.last_steps.is_empty() || !self.is_empty_delimiter() {
98 break;
99 }
100 }
101 146
102 // Move forward if it is empty delimiter 147 self.cursor = match self.cursor {
103 if self.last_steps.is_empty() { 148 WalkCursor::Token(idx, _) => self.walk_token(idx, last_step, true),
104 while self.is_empty_delimiter() { 149 WalkCursor::Eof => {
105 self.forward_unchecked(); 150 let len = self.top().len();
151 self.walk_token(len, last_step, true)
106 } 152 }
107 } 153 }
108 } 154 }
109 155
110 /// Move cursor backward by 1 step without empty check 156 /// Move cursor forward by 1 step
111 ///
112 /// Depends on the current state of cursor:
113 ///
114 /// * Delimiter Begin => Pop the stack, goto last walking token (`walk_token`)
115 /// * Token => Goto prev token (`walk_token`)
116 /// * Delimiter End => Goto the last child token (`walk_token`)
117 /// * Eof => push the root subtree, and set it as Delimiter End
118 fn backward_unchecked(&mut self) {
119 if self.last_steps.is_empty() {
120 return;
121 }
122
123 let last_step = self.last_steps.pop().unwrap();
124 let do_walk_token = match self.cursor {
125 WalkCursor::DelimiterBegin(_) => None,
126 WalkCursor::Token(u, _) => Some(u),
127 WalkCursor::DelimiterEnd(_) => {
128 let (top, _) = self.stack.last().unwrap();
129 Some(top.token_trees.len())
130 }
131 WalkCursor::Eof => None,
132 };
133
134 self.cursor = match do_walk_token {
135 Some(u) => self.walk_token(u, last_step, true),
136 None => match self.cursor {
137 WalkCursor::Eof => {
138 self.stack.push((self.subtree, None));
139 WalkCursor::DelimiterEnd(convert_delim(
140 self.stack.last().unwrap().0.delimiter,
141 true,
142 ))
143 }
144 _ => {
145 let (_, last_top_cursor) = self.stack.pop().unwrap();
146 assert!(!self.stack.is_empty());
147
148 self.walk_token(last_top_cursor.unwrap(), last_step, true)
149 }
150 },
151 };
152 }
153
154 /// Move cursor forward by 1 step with empty checking
155 fn forward(&mut self) { 157 fn forward(&mut self) {
156 if self.is_eof() { 158 if self.is_eof() {
157 return; 159 return;
158 } 160 }
159
160 self.pos += 1; 161 self.pos += 1;
161 loop {
162 self.forward_unchecked();
163 if !self.is_empty_delimiter() {
164 break;
165 }
166 }
167 }
168
169 /// Move cursor forward by 1 step without empty checking
170 ///
171 /// Depends on the current state of cursor:
172 ///
173 /// * Delimiter Begin => Goto the first child token (`walk_token`)
174 /// * Token => Goto next token (`walk_token`)
175 /// * Delimiter End => Pop the stack, goto last walking token (`walk_token`)
176 ///
177 fn forward_unchecked(&mut self) {
178 if self.is_eof() {
179 return;
180 }
181 162
182 let step = self.current().map(|x| x.n_tokens).unwrap_or(1); 163 let step = self.current().map(|x| x.n_tokens).unwrap_or(1);
183 self.last_steps.push(step); 164 self.last_steps.push(step);
184 165
185 let do_walk_token = match self.cursor { 166 if let WalkCursor::Token(u, _) = self.cursor {
186 WalkCursor::DelimiterBegin(_) => Some((0, 0)), 167 self.cursor = self.walk_token(u, step, false)
187 WalkCursor::Token(u, _) => Some((u, step)), 168 }
188 WalkCursor::DelimiterEnd(_) => None,
189 _ => unreachable!(),
190 };
191
192 self.cursor = match do_walk_token {
193 Some((u, step)) => self.walk_token(u, step, false),
194 None => {
195 let (_, last_top_idx) = self.stack.pop().unwrap();
196 match self.stack.last() {
197 Some(_) => self.walk_token(last_top_idx.unwrap(), 1, false),
198 None => WalkCursor::Eof,
199 }
200 }
201 };
202 } 169 }
203 170
204 /// Traversal child token 171 /// Traversal child token
205 /// Depends on the new position, it returns:
206 ///
207 /// * new position < 0 => DelimiterBegin
208 /// * new position > token_tree.len() => DelimiterEnd
209 /// * if new position is a subtree, depends on traversal direction:
210 /// ** backward => DelimiterEnd
211 /// ** forward => DelimiterBegin
212 /// * if new psoition is a leaf, return walk_leaf()
213 fn walk_token(&mut self, pos: usize, offset: usize, backward: bool) -> WalkCursor { 172 fn walk_token(&mut self, pos: usize, offset: usize, backward: bool) -> WalkCursor {
214 let (top, _) = self.stack.last().unwrap(); 173 let top = self.stack.last().map(|(t, _)| t).unwrap_or(&self.ts);
215 174
216 if backward && pos < offset { 175 if backward && pos < offset {
217 return WalkCursor::DelimiterBegin(convert_delim( 176 let (_, last_idx) = self.stack.pop().unwrap();
218 self.stack.last().unwrap().0.delimiter, 177 return self.walk_token(last_idx, offset, backward);
219 false,
220 ));
221 }
222
223 if !backward && pos + offset >= top.token_trees.len() {
224 return WalkCursor::DelimiterEnd(convert_delim(
225 self.stack.last().unwrap().0.delimiter,
226 true,
227 ));
228 } 178 }
229 179
230 let pos = if backward { pos - offset } else { pos + offset }; 180 let pos = if backward { pos - offset } else { pos + offset };
231 181
232 match &top.token_trees[pos] { 182 match top.get(pos) {
233 tt::TokenTree::Subtree(subtree) => { 183 DelimToken::Token(token) => match token {
234 self.stack.push((subtree, Some(pos))); 184 tt::TokenTree::Subtree(subtree) => {
235 let delim = convert_delim(self.stack.last().unwrap().0.delimiter, backward); 185 let ts = TokenSeq::from(subtree);
236 if backward { 186 let new_idx = if backward { ts.len() - 1 } else { 0 };
237 WalkCursor::DelimiterEnd(delim) 187 self.stack.push((ts, pos));
238 } else { 188 WalkCursor::Token(new_idx, convert_delim(subtree.delimiter, backward))
239 WalkCursor::DelimiterBegin(delim) 189 }
190 tt::TokenTree::Leaf(leaf) => {
191 let next_tokens = top.child_slice();
192 WalkCursor::Token(pos, convert_leaf(&next_tokens[pos..], leaf))
240 } 193 }
194 },
195 DelimToken::Delim(delim, is_end) => {
196 WalkCursor::Token(pos, convert_delim(*delim, is_end))
241 } 197 }
242 tt::TokenTree::Leaf(leaf) => WalkCursor::Token(pos, Some(self.walk_leaf(leaf, pos))), 198 DelimToken::End => {
243 } 199 // it is the top level
244 } 200 if let Some((_, last_idx)) = self.stack.pop() {
245 201 assert!(!backward);
246 fn walk_leaf(&mut self, leaf: &tt::Leaf, pos: usize) -> TtToken { 202 self.walk_token(last_idx, offset, backward)
247 match leaf { 203 } else {
248 tt::Leaf::Literal(l) => convert_literal(l), 204 WalkCursor::Eof
249 tt::Leaf::Ident(ident) => convert_ident(ident), 205 }
250 tt::Leaf::Punct(punct) => {
251 let (top, _) = self.stack.last().unwrap();
252 convert_punct(punct, top, pos)
253 } 206 }
254 } 207 }
255 } 208 }
@@ -263,27 +216,20 @@ pub(crate) trait Querier {
263#[derive(Debug)] 216#[derive(Debug)]
264pub(crate) struct WalkerOwner<'a> { 217pub(crate) struct WalkerOwner<'a> {
265 walker: RefCell<SubTreeWalker<'a>>, 218 walker: RefCell<SubTreeWalker<'a>>,
266 offset: usize,
267} 219}
268 220
269impl<'a> WalkerOwner<'a> { 221impl<'a> WalkerOwner<'a> {
270 fn new(subtree: &'a tt::Subtree) -> Self { 222 fn new<I: Into<TokenSeq<'a>>>(ts: I) -> Self {
271 WalkerOwner { walker: RefCell::new(SubTreeWalker::new(subtree)), offset: 0 } 223 WalkerOwner { walker: RefCell::new(SubTreeWalker::new(ts.into())) }
272 } 224 }
273 225
274 fn get<'b>(&self, pos: usize) -> Option<TtToken> { 226 fn get<'b>(&self, pos: usize) -> Option<TtToken> {
275 self.set_walker_pos(pos); 227 self.set_pos(pos);
276 let walker = self.walker.borrow(); 228 let walker = self.walker.borrow();
277 walker.current().cloned() 229 walker.current().cloned()
278 } 230 }
279 231
280 fn start_from_nth(&mut self, pos: usize) { 232 fn set_pos(&self, pos: usize) {
281 self.offset = pos;
282 self.walker.borrow_mut().start_from_nth(pos);
283 }
284
285 fn set_walker_pos(&self, mut pos: usize) {
286 pos += self.offset;
287 let mut walker = self.walker.borrow_mut(); 233 let mut walker = self.walker.borrow_mut();
288 while pos > walker.pos && !walker.is_eof() { 234 while pos > walker.pos && !walker.is_eof() {
289 walker.forward(); 235 walker.forward();
@@ -294,19 +240,26 @@ impl<'a> WalkerOwner<'a> {
294 } 240 }
295 241
296 fn collect_token_trees(&mut self, n: usize) -> Vec<&tt::TokenTree> { 242 fn collect_token_trees(&mut self, n: usize) -> Vec<&tt::TokenTree> {
297 self.start_from_nth(self.offset);
298
299 let mut res = vec![]; 243 let mut res = vec![];
300 let mut walker = self.walker.borrow_mut(); 244 let mut walker = self.walker.borrow_mut();
245 walker.reset();
301 246
302 while walker.pos - self.offset < n { 247 while walker.pos < n {
303 if let WalkCursor::Token(u, tt) = &walker.cursor { 248 if let WalkCursor::Token(u, tt) = &walker.cursor {
304 if walker.stack.len() == 1 { 249 // We only collect the topmost child
305 // We only collect the topmost child 250 if walker.stack.len() == 0 {
306 res.push(&walker.stack[0].0.token_trees[*u]); 251 for i in 0..tt.n_tokens {
307 if let Some(tt) = tt { 252 if let DelimToken::Token(token) = walker.ts.get(u + i) {
308 for i in 0..tt.n_tokens - 1 { 253 res.push(token);
309 res.push(&walker.stack[0].0.token_trees[u + i]); 254 }
255 }
256 } else if walker.stack.len() == 1 {
257 if let DelimToken::Delim(_, is_end) = walker.ts.get(*u) {
258 if !is_end {
259 let (_, last_idx) = &walker.stack[0];
260 if let DelimToken::Token(token) = walker.ts.get(*last_idx) {
261 res.push(token);
262 }
310 } 263 }
311 } 264 }
312 } 265 }
@@ -331,12 +284,8 @@ pub(crate) struct SubtreeTokenSource<'a> {
331} 284}
332 285
333impl<'a> SubtreeTokenSource<'a> { 286impl<'a> SubtreeTokenSource<'a> {
334 pub fn new(subtree: &tt::Subtree) -> SubtreeTokenSource { 287 pub fn new<I: Into<TokenSeq<'a>>>(ts: I) -> SubtreeTokenSource<'a> {
335 SubtreeTokenSource { walker: WalkerOwner::new(subtree) } 288 SubtreeTokenSource { walker: WalkerOwner::new(ts) }
336 }
337
338 pub fn start_from_nth(&mut self, n: usize) {
339 self.walker.start_from_nth(n);
340 } 289 }
341 290
342 pub fn querier<'b>(&'a self) -> &'b WalkerOwner<'a> 291 pub fn querier<'b>(&'a self) -> &'b WalkerOwner<'a>
@@ -467,18 +416,18 @@ where
467 None 416 None
468} 417}
469 418
470fn convert_delim(d: tt::Delimiter, closing: bool) -> Option<TtToken> { 419fn convert_delim(d: tt::Delimiter, closing: bool) -> TtToken {
471 let (kinds, texts) = match d { 420 let (kinds, texts) = match d {
472 tt::Delimiter::Parenthesis => ([L_PAREN, R_PAREN], "()"), 421 tt::Delimiter::Parenthesis => ([L_PAREN, R_PAREN], "()"),
473 tt::Delimiter::Brace => ([L_CURLY, R_CURLY], "{}"), 422 tt::Delimiter::Brace => ([L_CURLY, R_CURLY], "{}"),
474 tt::Delimiter::Bracket => ([L_BRACK, R_BRACK], "[]"), 423 tt::Delimiter::Bracket => ([L_BRACK, R_BRACK], "[]"),
475 tt::Delimiter::None => return None, 424 tt::Delimiter::None => ([L_DOLLAR, R_DOLLAR], ""),
476 }; 425 };
477 426
478 let idx = closing as usize; 427 let idx = closing as usize;
479 let kind = kinds[idx]; 428 let kind = kinds[idx];
480 let text = &texts[idx..texts.len() - (1 - idx)]; 429 let text = if texts.len() > 0 { &texts[idx..texts.len() - (1 - idx)] } else { "" };
481 Some(TtToken { kind, is_joint_to_next: false, text: SmolStr::new(text), n_tokens: 1 }) 430 TtToken { kind, is_joint_to_next: false, text: SmolStr::new(text), n_tokens: 1 }
482} 431}
483 432
484fn convert_literal(l: &tt::Literal) -> TtToken { 433fn convert_literal(l: &tt::Literal) -> TtToken {
@@ -495,8 +444,9 @@ fn convert_ident(ident: &tt::Ident) -> TtToken {
495 TtToken { kind, is_joint_to_next: false, text: ident.text.clone(), n_tokens: 1 } 444 TtToken { kind, is_joint_to_next: false, text: ident.text.clone(), n_tokens: 1 }
496} 445}
497 446
498fn convert_punct(p: &tt::Punct, parent: &tt::Subtree, next: usize) -> TtToken { 447fn convert_punct(p: &tt::Punct, next_tokens: &[tt::TokenTree]) -> TtToken {
499 let iter = parent.token_trees[next + 1..].iter(); 448 let mut iter = next_tokens.iter();
449 iter.next();
500 let mut peek = TokenPeek::new(iter); 450 let mut peek = TokenPeek::new(iter);
501 451
502 if let Some((kind, is_joint_to_next, text, size)) = convert_multi_char_punct(p, &mut peek) { 452 if let Some((kind, is_joint_to_next, text, size)) = convert_multi_char_punct(p, &mut peek) {
@@ -519,3 +469,11 @@ fn convert_punct(p: &tt::Punct, parent: &tt::Subtree, next: usize) -> TtToken {
519 TtToken { kind, is_joint_to_next: p.spacing == tt::Spacing::Joint, text, n_tokens: 1 } 469 TtToken { kind, is_joint_to_next: p.spacing == tt::Spacing::Joint, text, n_tokens: 1 }
520 } 470 }
521} 471}
472
473fn convert_leaf(tokens: &[tt::TokenTree], leaf: &tt::Leaf) -> TtToken {
474 match leaf {
475 tt::Leaf::Literal(l) => convert_literal(l),
476 tt::Leaf::Ident(ident) => convert_ident(ident),
477 tt::Leaf::Punct(punct) => convert_punct(punct, tokens),
478 }
479}
diff --git a/crates/ra_mbe/src/syntax_bridge.rs b/crates/ra_mbe/src/syntax_bridge.rs
index 19c17bd55..28ded7870 100644
--- a/crates/ra_mbe/src/syntax_bridge.rs
+++ b/crates/ra_mbe/src/syntax_bridge.rs
@@ -22,6 +22,14 @@ pub fn ast_to_token_tree(ast: &ast::TokenTree) -> Option<(tt::Subtree, TokenMap)
22 Some((tt, token_map)) 22 Some((tt, token_map))
23} 23}
24 24
25/// Convert the syntax node to a `TokenTree` (what macro
26/// will consume).
27pub fn syntax_node_to_token_tree(node: &SyntaxNode) -> Option<(tt::Subtree, TokenMap)> {
28 let mut token_map = TokenMap::default();
29 let tt = convert_tt(&mut token_map, node.range().start(), node)?;
30 Some((tt, token_map))
31}
32
25/// Parses the token tree (result of macro expansion) as a sequence of items 33/// Parses the token tree (result of macro expansion) as a sequence of items
26pub fn token_tree_to_ast_item_list(tt: &tt::Subtree) -> TreeArc<ast::SourceFile> { 34pub fn token_tree_to_ast_item_list(tt: &tt::Subtree) -> TreeArc<ast::SourceFile> {
27 let token_source = SubtreeTokenSource::new(tt); 35 let token_source = SubtreeTokenSource::new(tt);
@@ -51,15 +59,17 @@ fn convert_tt(
51) -> Option<tt::Subtree> { 59) -> Option<tt::Subtree> {
52 let first_child = tt.first_child_or_token()?; 60 let first_child = tt.first_child_or_token()?;
53 let last_child = tt.last_child_or_token()?; 61 let last_child = tt.last_child_or_token()?;
54 let delimiter = match (first_child.kind(), last_child.kind()) { 62 let (delimiter, skip_first) = match (first_child.kind(), last_child.kind()) {
55 (L_PAREN, R_PAREN) => tt::Delimiter::Parenthesis, 63 (L_PAREN, R_PAREN) => (tt::Delimiter::Parenthesis, true),
56 (L_CURLY, R_CURLY) => tt::Delimiter::Brace, 64 (L_CURLY, R_CURLY) => (tt::Delimiter::Brace, true),
57 (L_BRACK, R_BRACK) => tt::Delimiter::Bracket, 65 (L_BRACK, R_BRACK) => (tt::Delimiter::Bracket, true),
58 _ => return None, 66 _ => (tt::Delimiter::None, false),
59 }; 67 };
68
60 let mut token_trees = Vec::new(); 69 let mut token_trees = Vec::new();
61 for child in tt.children_with_tokens().skip(1) { 70 for child in tt.children_with_tokens().skip(skip_first as usize) {
62 if child == first_child || child == last_child || child.kind().is_trivia() { 71 if (skip_first && (child == first_child || child == last_child)) || child.kind().is_trivia()
72 {
63 continue; 73 continue;
64 } 74 }
65 match child { 75 match child {
@@ -127,6 +137,11 @@ impl<'a, Q: Querier> TtTreeSink<'a, Q> {
127 137
128impl<'a, Q: Querier> TreeSink for TtTreeSink<'a, Q> { 138impl<'a, Q: Querier> TreeSink for TtTreeSink<'a, Q> {
129 fn token(&mut self, kind: SyntaxKind, n_tokens: u8) { 139 fn token(&mut self, kind: SyntaxKind, n_tokens: u8) {
140 if kind == L_DOLLAR || kind == R_DOLLAR {
141 self.token_pos += n_tokens as usize;
142 return;
143 }
144
130 for _ in 0..n_tokens { 145 for _ in 0..n_tokens {
131 self.buf += &self.src_querier.token(self.token_pos).1; 146 self.buf += &self.src_querier.token(self.token_pos).1;
132 self.token_pos += 1; 147 self.token_pos += 1;
@@ -176,19 +191,19 @@ mod tests {
176 191
177 let query = tt_src.querier(); 192 let query = tt_src.querier();
178 193
179 // [{] 194 // [${]
180 // [let] [a] [=] ['c'] [;] 195 // [let] [a] [=] ['c'] [;]
181 assert_eq!(query.token(1 + 3).1, "'c'"); 196 assert_eq!(query.token(2 + 3).1, "'c'");
182 assert_eq!(query.token(1 + 3).0, CHAR); 197 assert_eq!(query.token(2 + 3).0, CHAR);
183 // [let] [c] [=] [1000] [;] 198 // [let] [c] [=] [1000] [;]
184 assert_eq!(query.token(1 + 5 + 3).1, "1000"); 199 assert_eq!(query.token(2 + 5 + 3).1, "1000");
185 assert_eq!(query.token(1 + 5 + 3).0, INT_NUMBER); 200 assert_eq!(query.token(2 + 5 + 3).0, INT_NUMBER);
186 // [let] [f] [=] [12E+99_f64] [;] 201 // [let] [f] [=] [12E+99_f64] [;]
187 assert_eq!(query.token(1 + 10 + 3).1, "12E+99_f64"); 202 assert_eq!(query.token(2 + 10 + 3).1, "12E+99_f64");
188 assert_eq!(query.token(1 + 10 + 3).0, FLOAT_NUMBER); 203 assert_eq!(query.token(2 + 10 + 3).0, FLOAT_NUMBER);
189 204
190 // [let] [s] [=] ["rust1"] [;] 205 // [let] [s] [=] ["rust1"] [;]
191 assert_eq!(query.token(1 + 15 + 3).1, "\"rust1\""); 206 assert_eq!(query.token(2 + 15 + 3).1, "\"rust1\"");
192 assert_eq!(query.token(1 + 15 + 3).0, STRING); 207 assert_eq!(query.token(2 + 15 + 3).0, STRING);
193 } 208 }
194} 209}
diff --git a/crates/ra_parser/src/parser.rs b/crates/ra_parser/src/parser.rs
index 56f8b7126..3cb57ed9c 100644
--- a/crates/ra_parser/src/parser.rs
+++ b/crates/ra_parser/src/parser.rs
@@ -99,6 +99,8 @@ impl<'t> Parser<'t> {
99 /// consumed between the `start` and the corresponding `Marker::complete` 99 /// consumed between the `start` and the corresponding `Marker::complete`
100 /// belong to the same node. 100 /// belong to the same node.
101 pub(crate) fn start(&mut self) -> Marker { 101 pub(crate) fn start(&mut self) -> Marker {
102 self.eat_dollars();
103
102 let pos = self.events.len() as u32; 104 let pos = self.events.len() as u32;
103 self.push_event(Event::tombstone()); 105 self.push_event(Event::tombstone());
104 Marker::new(pos) 106 Marker::new(pos)
@@ -180,13 +182,23 @@ impl<'t> Parser<'t> {
180 } 182 }
181 183
182 fn do_bump(&mut self, kind: SyntaxKind, n_raw_tokens: u8) { 184 fn do_bump(&mut self, kind: SyntaxKind, n_raw_tokens: u8) {
185 self.eat_dollars();
183 self.token_pos += usize::from(n_raw_tokens); 186 self.token_pos += usize::from(n_raw_tokens);
184 self.push_event(Event::Token { kind, n_raw_tokens }); 187 self.push_event(Event::Token { kind, n_raw_tokens });
188 self.eat_dollars();
185 } 189 }
186 190
187 fn push_event(&mut self, event: Event) { 191 fn push_event(&mut self, event: Event) {
188 self.events.push(event) 192 self.events.push(event)
189 } 193 }
194
195 fn eat_dollars(&mut self) {
196 while self.nth(0) == SyntaxKind::L_DOLLAR || self.nth(0) == SyntaxKind::R_DOLLAR {
197 let kind = self.nth(0);
198 self.token_pos += 1;
199 self.push_event(Event::Token { kind, n_raw_tokens: 1 });
200 }
201 }
190} 202}
191 203
192/// See `Parser::start`. 204/// See `Parser::start`.