Skip to content

Commit 74c36c6

Browse files
committed
Remove unnecessary paths from test suite
1 parent 8cccfe2 commit 74c36c6

File tree

1 file changed

+11
-13
lines changed

1 file changed

+11
-13
lines changed

tests/test.rs

Lines changed: 11 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
1-
use proc_macro2::{Ident, Literal, Spacing, Span, TokenStream, TokenTree};
1+
use proc_macro2::{Delimiter, Ident, Literal, Spacing, Span, TokenStream, TokenTree};
22
use std::str::{self, FromStr};
33

44
#[test]
@@ -171,8 +171,6 @@ fn fail() {
171171
#[cfg(span_locations)]
172172
#[test]
173173
fn span_test() {
174-
use proc_macro2::TokenTree;
175-
176174
fn check_spans(p: &str, mut lines: &[(usize, usize, usize, usize)]) {
177175
let ts = p.parse::<TokenStream>().unwrap();
178176
check_spans_internal(ts, &mut lines);
@@ -279,47 +277,47 @@ fn span_join() {
279277
#[test]
280278
fn no_panic() {
281279
let s = str::from_utf8(b"b\'\xc2\x86 \x00\x00\x00^\"").unwrap();
282-
assert!(s.parse::<proc_macro2::TokenStream>().is_err());
280+
assert!(s.parse::<TokenStream>().is_err());
283281
}
284282

285283
#[test]
286284
fn tricky_doc_comment() {
287-
let stream = "/**/".parse::<proc_macro2::TokenStream>().unwrap();
285+
let stream = "/**/".parse::<TokenStream>().unwrap();
288286
let tokens = stream.into_iter().collect::<Vec<_>>();
289287
assert!(tokens.is_empty(), "not empty -- {:?}", tokens);
290288

291-
let stream = "/// doc".parse::<proc_macro2::TokenStream>().unwrap();
289+
let stream = "/// doc".parse::<TokenStream>().unwrap();
292290
let tokens = stream.into_iter().collect::<Vec<_>>();
293291
assert!(tokens.len() == 2, "not length 2 -- {:?}", tokens);
294292
match tokens[0] {
295-
proc_macro2::TokenTree::Punct(ref tt) => assert_eq!(tt.as_char(), '#'),
293+
TokenTree::Punct(ref tt) => assert_eq!(tt.as_char(), '#'),
296294
_ => panic!("wrong token {:?}", tokens[0]),
297295
}
298296
let mut tokens = match tokens[1] {
299-
proc_macro2::TokenTree::Group(ref tt) => {
300-
assert_eq!(tt.delimiter(), proc_macro2::Delimiter::Bracket);
297+
TokenTree::Group(ref tt) => {
298+
assert_eq!(tt.delimiter(), Delimiter::Bracket);
301299
tt.stream().into_iter()
302300
}
303301
_ => panic!("wrong token {:?}", tokens[0]),
304302
};
305303

306304
match tokens.next().unwrap() {
307-
proc_macro2::TokenTree::Ident(ref tt) => assert_eq!(tt.to_string(), "doc"),
305+
TokenTree::Ident(ref tt) => assert_eq!(tt.to_string(), "doc"),
308306
t => panic!("wrong token {:?}", t),
309307
}
310308
match tokens.next().unwrap() {
311-
proc_macro2::TokenTree::Punct(ref tt) => assert_eq!(tt.as_char(), '='),
309+
TokenTree::Punct(ref tt) => assert_eq!(tt.as_char(), '='),
312310
t => panic!("wrong token {:?}", t),
313311
}
314312
match tokens.next().unwrap() {
315-
proc_macro2::TokenTree::Literal(ref tt) => {
313+
TokenTree::Literal(ref tt) => {
316314
assert_eq!(tt.to_string(), "\" doc\"");
317315
}
318316
t => panic!("wrong token {:?}", t),
319317
}
320318
assert!(tokens.next().is_none());
321319

322-
let stream = "//! doc".parse::<proc_macro2::TokenStream>().unwrap();
320+
let stream = "//! doc".parse::<TokenStream>().unwrap();
323321
let tokens = stream.into_iter().collect::<Vec<_>>();
324322
assert!(tokens.len() == 3, "not length 3 -- {:?}", tokens);
325323
}

0 commit comments

Comments
 (0)