@@ -21,11 +21,10 @@ use util::runner::{run_all, Test};
2121
2222use markup5ever:: buffer_queue:: BufferQueue ;
2323use xml5ever:: tendril:: { SliceExt , StrTendril } ;
24- use xml5ever:: tokenizer:: { CharacterTokens , Token , TokenSink } ;
25- use xml5ever:: tokenizer:: { CommentToken , EmptyTag , EndTag , ShortTag , StartTag , Tag } ;
26- use xml5ever:: tokenizer:: { Doctype , DoctypeToken , PIToken , Pi } ;
27- use xml5ever:: tokenizer:: { EOFToken , XmlTokenizer , XmlTokenizerOpts } ;
28- use xml5ever:: tokenizer:: { NullCharacterToken , ParseError , TagToken } ;
24+ use xml5ever:: tokenizer:: {
25+ Doctype , EmptyTag , EndTag , Pi , ShortTag , StartTag , Tag , Token , TokenSink , XmlTokenizer ,
26+ XmlTokenizerOpts ,
27+ } ;
2928use xml5ever:: { ns, Attribute , LocalName , QualName } ;
3029
3130mod util {
@@ -81,7 +80,7 @@ impl TokenLogger {
8180 fn finish_str ( & self ) {
8281 if !self . current_str . borrow ( ) . is_empty ( ) {
8382 let s = self . current_str . take ( ) ;
84- self . tokens . borrow_mut ( ) . push ( CharacterTokens ( s) ) ;
83+ self . tokens . borrow_mut ( ) . push ( Token :: Characters ( s) ) ;
8584 }
8685 }
8786
@@ -96,21 +95,20 @@ impl TokenSink for TokenLogger {
9695
9796 fn process_token ( & self , token : Token ) -> ProcessResult < ( ) > {
9897 match token {
99- CharacterTokens ( b ) => {
100- self . current_str . borrow_mut ( ) . push_slice ( & b ) ;
98+ Token :: Characters ( characters ) => {
99+ self . current_str . borrow_mut ( ) . push_slice ( & characters ) ;
101100 } ,
102101
103- NullCharacterToken => {
102+ Token :: NullCharacter => {
104103 self . current_str . borrow_mut ( ) . push_char ( '\0' ) ;
105104 } ,
106105
107- ParseError ( _) => {
106+ Token :: ParseError ( _) => {
108107 if self . exact_errors {
109- self . push ( ParseError ( Borrowed ( "" ) ) ) ;
108+ self . push ( Token :: ParseError ( Borrowed ( "" ) ) ) ;
110109 }
111110 } ,
112-
113- TagToken ( mut t) => {
111+ Token :: Tag ( mut t) => {
114112 // The spec seems to indicate that one can emit
115113 // erroneous end tags with attrs, but the test
116114 // cases don't contain them.
@@ -120,11 +118,9 @@ impl TokenSink for TokenLogger {
120118 } ,
121119 _ => t. attrs . sort_by ( |a1, a2| a1. name . cmp ( & a2. name ) ) ,
122120 }
123- self . push ( TagToken ( t) ) ;
121+ self . push ( Token :: Tag ( t) ) ;
124122 } ,
125-
126- EOFToken => ( ) ,
127-
123+ Token :: EndOfFile => ( ) ,
128124 _ => self . push ( token) ,
129125 } ;
130126 ProcessResult :: Continue
@@ -211,7 +207,7 @@ fn json_to_token(js: &Value) -> Token {
211207 // Collect refs here so we don't have to use "ref" in all the patterns below.
212208 let args: Vec < & Value > = parts[ 1 ..] . iter ( ) . collect ( ) ;
213209 match & * parts[ 0 ] . get_str ( ) {
214- "StartTag" => TagToken ( Tag {
210+ "StartTag" => Token :: Tag ( Tag {
215211 kind : StartTag ,
216212 name : QualName :: new ( None , ns ! ( ) , LocalName :: from ( args[ 0 ] . get_str ( ) ) ) ,
217213 attrs : args[ 1 ]
@@ -224,19 +220,19 @@ fn json_to_token(js: &Value) -> Token {
224220 . collect ( ) ,
225221 } ) ,
226222
227- "EndTag" => TagToken ( Tag {
223+ "EndTag" => Token :: Tag ( Tag {
228224 kind : EndTag ,
229225 name : QualName :: new ( None , ns ! ( ) , LocalName :: from ( args[ 0 ] . get_str ( ) ) ) ,
230226 attrs : vec ! [ ] ,
231227 } ) ,
232228
233- "ShortTag" => TagToken ( Tag {
229+ "ShortTag" => Token :: Tag ( Tag {
234230 kind : ShortTag ,
235231 name : QualName :: new ( None , ns ! ( ) , LocalName :: from ( args[ 0 ] . get_str ( ) ) ) ,
236232 attrs : vec ! [ ] ,
237233 } ) ,
238234
239- "EmptyTag" => TagToken ( Tag {
235+ "EmptyTag" => Token :: Tag ( Tag {
240236 kind : EmptyTag ,
241237 name : QualName :: new ( None , ns ! ( ) , LocalName :: from ( args[ 0 ] . get_str ( ) ) ) ,
242238 attrs : args[ 1 ]
@@ -249,16 +245,16 @@ fn json_to_token(js: &Value) -> Token {
249245 . collect ( ) ,
250246 } ) ,
251247
252- "Comment" => CommentToken ( args[ 0 ] . get_tendril ( ) ) ,
248+ "Comment" => Token :: Comment ( args[ 0 ] . get_tendril ( ) ) ,
253249
254- "Character" => CharacterTokens ( args[ 0 ] . get_tendril ( ) ) ,
250+ "Character" => Token :: Characters ( args[ 0 ] . get_tendril ( ) ) ,
255251
256- "PI" => PIToken ( Pi {
252+ "PI" => Token :: ProcessingInstruction ( Pi {
257253 target : args[ 0 ] . get_tendril ( ) ,
258254 data : args[ 1 ] . get_tendril ( ) ,
259255 } ) ,
260256
261- "DOCTYPE" => DoctypeToken ( Doctype {
257+ "DOCTYPE" => Token :: Doctype ( Doctype {
262258 name : args[ 0 ] . get_nullable_tendril ( ) ,
263259 public_id : args[ 1 ] . get_nullable_tendril ( ) ,
264260 system_id : args[ 2 ] . get_nullable_tendril ( ) ,
@@ -278,7 +274,7 @@ fn json_to_tokens(js: &Value, exact_errors: bool) -> Vec<Token> {
278274 for tok in js. as_array ( ) . unwrap ( ) . iter ( ) {
279275 match * tok {
280276 Value :: String ( ref s) if & s[ ..] == "ParseError" => {
281- let _ = sink. process_token ( ParseError ( Borrowed ( "" ) ) ) ;
277+ let _ = sink. process_token ( Token :: ParseError ( Borrowed ( "" ) ) ) ;
282278 } ,
283279 _ => {
284280 let _ = sink. process_token ( json_to_token ( tok) ) ;
0 commit comments