aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorbors[bot] <bors[bot]@users.noreply.github.com>2019-04-05 09:55:50 +0100
committerbors[bot] <bors[bot]@users.noreply.github.com>2019-04-05 09:55:50 +0100
commitbe9a44e9bad262ac5e615730e540fd434f846a0e (patch)
tree7082f2b1398f8481a5a583a8c499d9b931c5e590
parent7713416477fd59348ad60d44f0ec3a3aebcf4b9f (diff)
parent6ff16c7ad9c1b72c4f04bb0d6c7fc0fdc2cc00cb (diff)
Merge #1111
1111: Add multi-byte token support in token tree to ast item list r=matklad a=edwin0cheng As discusion in https://github.com/rust-analyzer/rust-analyzer/pull/1105 , this PR add implement all multi-byte tokens in `ra_mbe` crate. Co-authored-by: Edwin Cheng <[email protected]>
-rw-r--r--Cargo.lock1
-rw-r--r--crates/ra_mbe/Cargo.toml2
-rw-r--r--crates/ra_mbe/src/syntax_bridge.rs156
3 files changed, 139 insertions, 20 deletions
diff --git a/Cargo.lock b/Cargo.lock
index 603abe058..a14312821 100644
--- a/Cargo.lock
+++ b/Cargo.lock
@@ -1040,6 +1040,7 @@ dependencies = [
1040name = "ra_mbe" 1040name = "ra_mbe"
1041version = "0.1.0" 1041version = "0.1.0"
1042dependencies = [ 1042dependencies = [
1043 "itertools 0.8.0 (registry+https://github.com/rust-lang/crates.io-index)",
1043 "ra_parser 0.1.0", 1044 "ra_parser 0.1.0",
1044 "ra_syntax 0.1.0", 1045 "ra_syntax 0.1.0",
1045 "ra_tt 0.1.0", 1046 "ra_tt 0.1.0",
diff --git a/crates/ra_mbe/Cargo.toml b/crates/ra_mbe/Cargo.toml
index 6e785f570..1d0c2a340 100644
--- a/crates/ra_mbe/Cargo.toml
+++ b/crates/ra_mbe/Cargo.toml
@@ -8,5 +8,5 @@ authors = ["rust-analyzer developers"]
8ra_syntax = { path = "../ra_syntax" } 8ra_syntax = { path = "../ra_syntax" }
9ra_parser = { path = "../ra_parser" } 9ra_parser = { path = "../ra_parser" }
10tt = { path = "../ra_tt", package = "ra_tt" } 10tt = { path = "../ra_tt", package = "ra_tt" }
11 11itertools = "0.8.0"
12rustc-hash = "1.0.0" 12rustc-hash = "1.0.0"
diff --git a/crates/ra_mbe/src/syntax_bridge.rs b/crates/ra_mbe/src/syntax_bridge.rs
index 05f9817da..257503de8 100644
--- a/crates/ra_mbe/src/syntax_bridge.rs
+++ b/crates/ra_mbe/src/syntax_bridge.rs
@@ -113,6 +113,51 @@ struct TtToken {
113 text: SmolStr, 113 text: SmolStr,
114} 114}
115 115
116// Some helper functions
117fn to_punct(tt: &tt::TokenTree) -> Option<&tt::Punct> {
118 if let tt::TokenTree::Leaf(tt::Leaf::Punct(pp)) = tt {
119 return Some(pp);
120 }
121 None
122}
123
124struct TokenPeek<'a, I>
125where
126 I: Iterator<Item = &'a tt::TokenTree>,
127{
128 iter: itertools::MultiPeek<I>,
129}
130
131impl<'a, I> TokenPeek<'a, I>
132where
133 I: Iterator<Item = &'a tt::TokenTree>,
134{
135 fn next(&mut self) -> Option<&tt::TokenTree> {
136 self.iter.next()
137 }
138
139 fn current_punct2(&mut self, p: &tt::Punct) -> Option<((char, char), bool)> {
140 if p.spacing != tt::Spacing::Joint {
141 return None;
142 }
143
144 self.iter.reset_peek();
145 let p1 = to_punct(self.iter.peek()?)?;
146 Some(((p.char, p1.char), p1.spacing == tt::Spacing::Joint))
147 }
148
149 fn current_punct3(&mut self, p: &tt::Punct) -> Option<((char, char, char), bool)> {
150 self.current_punct2(p).and_then(|((p0, p1), last_joint)| {
151 if !last_joint {
152 None
153 } else {
154 let p2 = to_punct(*self.iter.peek()?)?;
155 Some(((p0, p1, p2.char), p2.spacing == tt::Spacing::Joint))
156 }
157 })
158 }
159}
160
116impl TtTokenSource { 161impl TtTokenSource {
117 fn new(tt: &tt::Subtree) -> TtTokenSource { 162 fn new(tt: &tt::Subtree) -> TtTokenSource {
118 let mut res = TtTokenSource { tokens: Vec::new() }; 163 let mut res = TtTokenSource { tokens: Vec::new() };
@@ -121,16 +166,27 @@ impl TtTokenSource {
121 } 166 }
122 fn convert_subtree(&mut self, sub: &tt::Subtree) { 167 fn convert_subtree(&mut self, sub: &tt::Subtree) {
123 self.push_delim(sub.delimiter, false); 168 self.push_delim(sub.delimiter, false);
124 sub.token_trees.iter().for_each(|tt| self.convert_tt(tt)); 169 let mut peek = TokenPeek { iter: itertools::multipeek(sub.token_trees.iter()) };
170 while let Some(tt) = peek.iter.next() {
171 self.convert_tt(tt, &mut peek);
172 }
125 self.push_delim(sub.delimiter, true) 173 self.push_delim(sub.delimiter, true)
126 } 174 }
127 fn convert_tt(&mut self, tt: &tt::TokenTree) { 175
176 fn convert_tt<'a, I>(&mut self, tt: &tt::TokenTree, iter: &mut TokenPeek<'a, I>)
177 where
178 I: Iterator<Item = &'a tt::TokenTree>,
179 {
128 match tt { 180 match tt {
129 tt::TokenTree::Leaf(token) => self.convert_token(token), 181 tt::TokenTree::Leaf(token) => self.convert_token(token, iter),
130 tt::TokenTree::Subtree(sub) => self.convert_subtree(sub), 182 tt::TokenTree::Subtree(sub) => self.convert_subtree(sub),
131 } 183 }
132 } 184 }
133 fn convert_token(&mut self, token: &tt::Leaf) { 185
186 fn convert_token<'a, I>(&mut self, token: &tt::Leaf, iter: &mut TokenPeek<'a, I>)
187 where
188 I: Iterator<Item = &'a tt::TokenTree>,
189 {
134 let tok = match token { 190 let tok = match token {
135 tt::Leaf::Literal(l) => TtToken { 191 tt::Leaf::Literal(l) => TtToken {
136 kind: SyntaxKind::INT_NUMBER, // FIXME 192 kind: SyntaxKind::INT_NUMBER, // FIXME
@@ -138,21 +194,25 @@ impl TtTokenSource {
138 text: l.text.clone(), 194 text: l.text.clone(),
139 }, 195 },
140 tt::Leaf::Punct(p) => { 196 tt::Leaf::Punct(p) => {
141 let kind = match p.char { 197 if let Some(tt) = Self::convert_multi_char_punct(p, iter) {
142 // lexer may produce combpund tokens for these ones 198 tt
143 '.' => DOT, 199 } else {
144 ':' => COLON, 200 let kind = match p.char {
145 '=' => EQ, 201 // lexer may produce combpund tokens for these ones
146 '!' => EXCL, 202 '.' => DOT,
147 '-' => MINUS, 203 ':' => COLON,
148 c => SyntaxKind::from_char(c).unwrap(), 204 '=' => EQ,
149 }; 205 '!' => EXCL,
150 let text = { 206 '-' => MINUS,
151 let mut buf = [0u8; 4]; 207 c => SyntaxKind::from_char(c).unwrap(),
152 let s: &str = p.char.encode_utf8(&mut buf); 208 };
153 SmolStr::new(s) 209 let text = {
154 }; 210 let mut buf = [0u8; 4];
155 TtToken { kind, is_joint_to_next: p.spacing == tt::Spacing::Joint, text } 211 let s: &str = p.char.encode_utf8(&mut buf);
212 SmolStr::new(s)
213 };
214 TtToken { kind, is_joint_to_next: p.spacing == tt::Spacing::Joint, text }
215 }
156 } 216 }
157 tt::Leaf::Ident(ident) => { 217 tt::Leaf::Ident(ident) => {
158 let kind = SyntaxKind::from_keyword(ident.text.as_str()).unwrap_or(IDENT); 218 let kind = SyntaxKind::from_keyword(ident.text.as_str()).unwrap_or(IDENT);
@@ -161,6 +221,64 @@ impl TtTokenSource {
161 }; 221 };
162 self.tokens.push(tok) 222 self.tokens.push(tok)
163 } 223 }
224
225 fn convert_multi_char_punct<'a, I>(
226 p: &tt::Punct,
227 iter: &mut TokenPeek<'a, I>,
228 ) -> Option<TtToken>
229 where
230 I: Iterator<Item = &'a tt::TokenTree>,
231 {
232 if let Some((m, is_joint_to_next)) = iter.current_punct3(p) {
233 if let Some((kind, text)) = match m {
234 ('<', '<', '=') => Some((SHLEQ, "<<=")),
235 ('>', '>', '=') => Some((SHREQ, ">>=")),
236 ('.', '.', '.') => Some((DOTDOTDOT, "...")),
237 ('.', '.', '=') => Some((DOTDOTEQ, "..=")),
238 _ => None,
239 } {
240 iter.next();
241 iter.next();
242 return Some(TtToken { kind, is_joint_to_next, text: text.into() });
243 }
244 }
245
246 if let Some((m, is_joint_to_next)) = iter.current_punct2(p) {
247 if let Some((kind, text)) = match m {
248 ('<', '<') => Some((SHL, "<<")),
249 ('>', '>') => Some((SHR, ">>")),
250
251 ('|', '|') => Some((PIPEPIPE, "||")),
252 ('&', '&') => Some((AMPAMP, "&&")),
253 ('%', '=') => Some((PERCENTEQ, "%=")),
254 ('*', '=') => Some((STAREQ, "*=")),
255 ('/', '=') => Some((SLASHEQ, "/=")),
256 ('^', '=') => Some((CARETEQ, "^=")),
257
258 ('&', '=') => Some((AMPEQ, "&=")),
259 ('|', '=') => Some((PIPEEQ, "|=")),
260 ('-', '=') => Some((MINUSEQ, "-=")),
261 ('+', '=') => Some((PLUSEQ, "+=")),
262 ('>', '=') => Some((GTEQ, ">=")),
263 ('<', '=') => Some((LTEQ, "<=")),
264
265 ('-', '>') => Some((THIN_ARROW, "->")),
266 ('!', '=') => Some((NEQ, "!=")),
267 ('=', '>') => Some((FAT_ARROW, "=>")),
268 ('=', '=') => Some((EQEQ, "==")),
269 ('.', '.') => Some((DOTDOT, "..")),
270 (':', ':') => Some((COLONCOLON, "::")),
271
272 _ => None,
273 } {
274 iter.next();
275 return Some(TtToken { kind, is_joint_to_next, text: text.into() });
276 }
277 }
278
279 None
280 }
281
164 fn push_delim(&mut self, d: tt::Delimiter, closing: bool) { 282 fn push_delim(&mut self, d: tt::Delimiter, closing: bool) {
165 let (kinds, texts) = match d { 283 let (kinds, texts) = match d {
166 tt::Delimiter::Parenthesis => ([L_PAREN, R_PAREN], "()"), 284 tt::Delimiter::Parenthesis => ([L_PAREN, R_PAREN], "()"),