|
1 | | -use proc_macro2::{Ident, Literal, Spacing, Span, TokenStream, TokenTree}; |
| 1 | +use proc_macro2::{Delimiter, Ident, Literal, Spacing, Span, TokenStream, TokenTree}; |
2 | 2 | use std::str::{self, FromStr}; |
3 | 3 |
|
4 | 4 | #[test] |
@@ -171,8 +171,6 @@ fn fail() { |
171 | 171 | #[cfg(span_locations)] |
172 | 172 | #[test] |
173 | 173 | fn span_test() { |
174 | | - use proc_macro2::TokenTree; |
175 | | - |
176 | 174 | fn check_spans(p: &str, mut lines: &[(usize, usize, usize, usize)]) { |
177 | 175 | let ts = p.parse::<TokenStream>().unwrap(); |
178 | 176 | check_spans_internal(ts, &mut lines); |
@@ -279,47 +277,47 @@ fn span_join() { |
279 | 277 | #[test] |
280 | 278 | fn no_panic() { |
281 | 279 | let s = str::from_utf8(b"b\'\xc2\x86 \x00\x00\x00^\"").unwrap(); |
282 | | - assert!(s.parse::<proc_macro2::TokenStream>().is_err()); |
| 280 | + assert!(s.parse::<TokenStream>().is_err()); |
283 | 281 | } |
284 | 282 |
|
285 | 283 | #[test] |
286 | 284 | fn tricky_doc_comment() { |
287 | | - let stream = "/**/".parse::<proc_macro2::TokenStream>().unwrap(); |
| 285 | + let stream = "/**/".parse::<TokenStream>().unwrap(); |
288 | 286 | let tokens = stream.into_iter().collect::<Vec<_>>(); |
289 | 287 | assert!(tokens.is_empty(), "not empty -- {:?}", tokens); |
290 | 288 |
|
291 | | - let stream = "/// doc".parse::<proc_macro2::TokenStream>().unwrap(); |
| 289 | + let stream = "/// doc".parse::<TokenStream>().unwrap(); |
292 | 290 | let tokens = stream.into_iter().collect::<Vec<_>>(); |
293 | 291 | assert!(tokens.len() == 2, "not length 2 -- {:?}", tokens); |
294 | 292 | match tokens[0] { |
295 | | - proc_macro2::TokenTree::Punct(ref tt) => assert_eq!(tt.as_char(), '#'), |
| 293 | + TokenTree::Punct(ref tt) => assert_eq!(tt.as_char(), '#'), |
296 | 294 | _ => panic!("wrong token {:?}", tokens[0]), |
297 | 295 | } |
298 | 296 | let mut tokens = match tokens[1] { |
299 | | - proc_macro2::TokenTree::Group(ref tt) => { |
300 | | - assert_eq!(tt.delimiter(), proc_macro2::Delimiter::Bracket); |
| 297 | + TokenTree::Group(ref tt) => { |
| 298 | + assert_eq!(tt.delimiter(), Delimiter::Bracket); |
301 | 299 | tt.stream().into_iter() |
302 | 300 | } |
303 | 301 | _ => panic!("wrong token {:?}", tokens[0]), |
304 | 302 | }; |
305 | 303 |
|
306 | 304 | match tokens.next().unwrap() { |
307 | | - proc_macro2::TokenTree::Ident(ref tt) => assert_eq!(tt.to_string(), "doc"), |
| 305 | + TokenTree::Ident(ref tt) => assert_eq!(tt.to_string(), "doc"), |
308 | 306 | t => panic!("wrong token {:?}", t), |
309 | 307 | } |
310 | 308 | match tokens.next().unwrap() { |
311 | | - proc_macro2::TokenTree::Punct(ref tt) => assert_eq!(tt.as_char(), '='), |
| 309 | + TokenTree::Punct(ref tt) => assert_eq!(tt.as_char(), '='), |
312 | 310 | t => panic!("wrong token {:?}", t), |
313 | 311 | } |
314 | 312 | match tokens.next().unwrap() { |
315 | | - proc_macro2::TokenTree::Literal(ref tt) => { |
| 313 | + TokenTree::Literal(ref tt) => { |
316 | 314 | assert_eq!(tt.to_string(), "\" doc\""); |
317 | 315 | } |
318 | 316 | t => panic!("wrong token {:?}", t), |
319 | 317 | } |
320 | 318 | assert!(tokens.next().is_none()); |
321 | 319 |
|
322 | | - let stream = "//! doc".parse::<proc_macro2::TokenStream>().unwrap(); |
| 320 | + let stream = "//! doc".parse::<TokenStream>().unwrap(); |
323 | 321 | let tokens = stream.into_iter().collect::<Vec<_>>(); |
324 | 322 | assert!(tokens.len() == 3, "not length 3 -- {:?}", tokens); |
325 | 323 | } |
|
0 commit comments