diff --git a/html5lib/html5parser.py b/html5lib/html5parser.py index 9be3e134..5b9ce7d7 100644 --- a/html5lib/html5parser.py +++ b/html5lib/html5parser.py @@ -1216,8 +1216,7 @@ def startTagIsIndex(self, token): attributes["name"] = "isindex" self.processStartTag(impliedTagToken("input", "StartTag", attributes=attributes, - selfClosing= - token["selfClosing"])) + selfClosing=token["selfClosing"])) self.processEndTag(impliedTagToken("label")) self.processStartTag(impliedTagToken("hr", "StartTag")) self.processEndTag(impliedTagToken("form")) diff --git a/html5lib/sanitizer.py b/html5lib/sanitizer.py index 71dc5212..469d9b40 100644 --- a/html5lib/sanitizer.py +++ b/html5lib/sanitizer.py @@ -245,7 +245,7 @@ def sanitize_css(self, style): elif prop.split('-')[0].lower() in ['background', 'border', 'margin', 'padding']: for keyword in value.split(): - if not keyword in self.acceptable_css_keywords and \ + if keyword not in self.acceptable_css_keywords and \ not re.match("^(#[0-9a-f]+|rgb\(\d+%?,\d*%?,?\d*%?\)?|\d{0,2}\.?\d{0,2}(cm|em|ex|in|mm|pc|pt|px|%|,|\))?)$", keyword): break else: diff --git a/html5lib/serializer/htmlserializer.py b/html5lib/serializer/htmlserializer.py index 412a5a22..4a891ff5 100644 --- a/html5lib/serializer/htmlserializer.py +++ b/html5lib/serializer/htmlserializer.py @@ -35,7 +35,7 @@ v = utils.surrogatePairToCodepoint(v) else: v = ord(v) - if not v in encode_entity_map or k.islower(): + if v not in encode_entity_map or k.islower(): # prefer < over < and similarly for &, >, etc. encode_entity_map[v] = k @@ -291,7 +291,7 @@ def serialize(self, treewalker, encoding=None): elif type == "Entity": name = token["name"] key = name + ";" - if not key in entities: + if key not in entities: self.serializeError(_("Entity %s not recognized" % name)) if self.resolve_entities and key not in xmlEntities: data = entities[key] diff --git a/html5lib/tests/test_tokenizer.py b/html5lib/tests/test_tokenizer.py index 7d7b5258..90315ab3 100644 --- a/html5lib/tests/test_tokenizer.py +++ b/html5lib/tests/test_tokenizer.py @@ -68,8 +68,8 @@ def processParseError(self, token): def concatenateCharacterTokens(tokens): outputTokens = [] for token in tokens: - if not "ParseError" in token and token[0] == "Character": - if (outputTokens and not "ParseError" in outputTokens[-1] and + if "ParseError" not in token and token[0] == "Character": + if (outputTokens and "ParseError" not in outputTokens[-1] and outputTokens[-1][0] == "Character"): outputTokens[-1][1] += token[1] else: @@ -112,7 +112,7 @@ def tokensMatch(expectedTokens, receivedTokens, ignoreErrorOrder, # Sort the tokens into two groups; non-parse errors and parse errors tokens = {"expected": [[], []], "received": [[], []]} for tokenType, tokenList in zip(list(tokens.keys()), - (expectedTokens, receivedTokens)): + (expectedTokens, receivedTokens)): for token in tokenList: if token != "ParseError": tokens[tokenType][0].append(token) diff --git a/html5lib/treebuilders/dom.py b/html5lib/treebuilders/dom.py index 61e5ed79..234233b7 100644 --- a/html5lib/treebuilders/dom.py +++ b/html5lib/treebuilders/dom.py @@ -158,7 +158,7 @@ def insertText(self, data, parent=None): else: # HACK: allow text nodes as children of the document node if hasattr(self.dom, '_child_node_types'): - if not Node.TEXT_NODE in self.dom._child_node_types: + if Node.TEXT_NODE not in self.dom._child_node_types: self.dom._child_node_types = list(self.dom._child_node_types) self.dom._child_node_types.append(Node.TEXT_NODE) self.dom.appendChild(self.dom.createTextNode(data))
Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.
Alternative Proxies: