1
1
import sys
2
2
import os
3
- import unittest
4
3
import cStringIO
5
4
import warnings
6
5
import re
@@ -122,7 +121,7 @@ def tokensMatch(expectedTokens, receivedTokens, ignoreErrorOrder,
122
121
tokens [tokenType ][1 ].append (token )
123
122
return tokens ["expected" ] == tokens ["received" ]
124
123
125
- def unescape_test (test ):
124
+ def unescape (test ):
126
125
def decode (inp ):
127
126
return inp .decode ("unicode-escape" )
128
127
@@ -138,7 +137,6 @@ def decode(inp):
138
137
token [2 ][decode (key )] = decode (value )
139
138
return test
140
139
141
-
142
140
def runTokenizerTest (test ):
143
141
#XXX - move this out into the setup function
144
142
#concatenate all consecutive character tokens into a single token
@@ -163,8 +161,7 @@ def runTokenizerTest(test):
163
161
"\n received:" , unicode (tokens )])
164
162
errorMsg = errorMsg .encode ("utf-8" )
165
163
ignoreErrorOrder = test .get ('ignoreErrorOrder' , False )
166
- assert tokensMatch (expected , received , ignoreErrorOrder ), errorMsg
167
-
164
+ assert tokensMatch (expected , received , ignoreErrorOrder , True ), errorMsg
168
165
169
166
def _doCapitalize (match ):
170
167
return match .group (1 ).upper ()
@@ -176,8 +173,7 @@ def capitalize(s):
176
173
s = _capitalizeRe (_doCapitalize , s )
177
174
return s
178
175
179
-
180
- def test_tokenizer ():
176
+ def testTokenizer ():
181
177
for filename in html5lib_test_files ('tokenizer' , '*.test' ):
182
178
tests = json .load (file (filename ))
183
179
testName = os .path .basename (filename ).replace (".test" ,"" )
@@ -190,4 +186,3 @@ def test_tokenizer():
190
186
for initialState in test ["initialStates" ]:
191
187
test ["initialState" ] = capitalize (initialState )
192
188
yield runTokenizerTest , test
193
-
0 commit comments