|
1 |
| -use proc_macro2::{Ident, Literal, Spacing, Span, TokenStream, TokenTree}; |
| 1 | +use proc_macro2::{Delimiter, Ident, Literal, Spacing, Span, TokenStream, TokenTree}; |
2 | 2 | use std::str::{self, FromStr};
|
3 | 3 |
|
4 | 4 | #[test]
|
@@ -171,8 +171,6 @@ fn fail() {
|
171 | 171 | #[cfg(span_locations)]
|
172 | 172 | #[test]
|
173 | 173 | fn span_test() {
|
174 |
| - use proc_macro2::TokenTree; |
175 |
| - |
176 | 174 | fn check_spans(p: &str, mut lines: &[(usize, usize, usize, usize)]) {
|
177 | 175 | let ts = p.parse::<TokenStream>().unwrap();
|
178 | 176 | check_spans_internal(ts, &mut lines);
|
@@ -279,47 +277,47 @@ fn span_join() {
|
279 | 277 | #[test]
|
280 | 278 | fn no_panic() {
|
281 | 279 | let s = str::from_utf8(b"b\'\xc2\x86 \x00\x00\x00^\"").unwrap();
|
282 |
| - assert!(s.parse::<proc_macro2::TokenStream>().is_err()); |
| 280 | + assert!(s.parse::<TokenStream>().is_err()); |
283 | 281 | }
|
284 | 282 |
|
285 | 283 | #[test]
|
286 | 284 | fn tricky_doc_comment() {
|
287 |
| - let stream = "/**/".parse::<proc_macro2::TokenStream>().unwrap(); |
| 285 | + let stream = "/**/".parse::<TokenStream>().unwrap(); |
288 | 286 | let tokens = stream.into_iter().collect::<Vec<_>>();
|
289 | 287 | assert!(tokens.is_empty(), "not empty -- {:?}", tokens);
|
290 | 288 |
|
291 |
| - let stream = "/// doc".parse::<proc_macro2::TokenStream>().unwrap(); |
| 289 | + let stream = "/// doc".parse::<TokenStream>().unwrap(); |
292 | 290 | let tokens = stream.into_iter().collect::<Vec<_>>();
|
293 | 291 | assert!(tokens.len() == 2, "not length 2 -- {:?}", tokens);
|
294 | 292 | match tokens[0] {
|
295 |
| - proc_macro2::TokenTree::Punct(ref tt) => assert_eq!(tt.as_char(), '#'), |
| 293 | + TokenTree::Punct(ref tt) => assert_eq!(tt.as_char(), '#'), |
296 | 294 | _ => panic!("wrong token {:?}", tokens[0]),
|
297 | 295 | }
|
298 | 296 | let mut tokens = match tokens[1] {
|
299 |
| - proc_macro2::TokenTree::Group(ref tt) => { |
300 |
| - assert_eq!(tt.delimiter(), proc_macro2::Delimiter::Bracket); |
| 297 | + TokenTree::Group(ref tt) => { |
| 298 | + assert_eq!(tt.delimiter(), Delimiter::Bracket); |
301 | 299 | tt.stream().into_iter()
|
302 | 300 | }
|
303 | 301 | _ => panic!("wrong token {:?}", tokens[0]),
|
304 | 302 | };
|
305 | 303 |
|
306 | 304 | match tokens.next().unwrap() {
|
307 |
| - proc_macro2::TokenTree::Ident(ref tt) => assert_eq!(tt.to_string(), "doc"), |
| 305 | + TokenTree::Ident(ref tt) => assert_eq!(tt.to_string(), "doc"), |
308 | 306 | t => panic!("wrong token {:?}", t),
|
309 | 307 | }
|
310 | 308 | match tokens.next().unwrap() {
|
311 |
| - proc_macro2::TokenTree::Punct(ref tt) => assert_eq!(tt.as_char(), '='), |
| 309 | + TokenTree::Punct(ref tt) => assert_eq!(tt.as_char(), '='), |
312 | 310 | t => panic!("wrong token {:?}", t),
|
313 | 311 | }
|
314 | 312 | match tokens.next().unwrap() {
|
315 |
| - proc_macro2::TokenTree::Literal(ref tt) => { |
| 313 | + TokenTree::Literal(ref tt) => { |
316 | 314 | assert_eq!(tt.to_string(), "\" doc\"");
|
317 | 315 | }
|
318 | 316 | t => panic!("wrong token {:?}", t),
|
319 | 317 | }
|
320 | 318 | assert!(tokens.next().is_none());
|
321 | 319 |
|
322 |
| - let stream = "//! doc".parse::<proc_macro2::TokenStream>().unwrap(); |
| 320 | + let stream = "//! doc".parse::<TokenStream>().unwrap(); |
323 | 321 | let tokens = stream.into_iter().collect::<Vec<_>>();
|
324 | 322 | assert!(tokens.len() == 3, "not length 3 -- {:?}", tokens);
|
325 | 323 | }
|
|
0 commit comments