mirror of
https://github.com/python/cpython.git
synced 2024-11-24 18:34:43 +08:00
934 lines
34 KiB
Python
934 lines
34 KiB
Python
"""A parser for XML, using the derived class as static DTD."""
|
|
|
|
# Author: Sjoerd Mullender.
|
|
|
|
import re
|
|
import string
|
|
|
|
import warnings
|
|
warnings.warn("The xmllib module is obsolete. Use xml.sax instead.",
|
|
DeprecationWarning)
|
|
del warnings
|
|
|
|
version = '0.3'
|
|
|
|
class Error(RuntimeError):
|
|
pass
|
|
|
|
# Regular expressions used for parsing
|
|
|
|
_S = '[ \t\r\n]+' # white space
|
|
_opS = '[ \t\r\n]*' # optional white space
|
|
_Name = '[a-zA-Z_:][-a-zA-Z0-9._:]*' # valid XML name
|
|
_QStr = "(?:'[^']*'|\"[^\"]*\")" # quoted XML string
|
|
illegal = re.compile('[^\t\r\n -\176\240-\377]') # illegal chars in content
|
|
interesting = re.compile('[]&<]')
|
|
|
|
amp = re.compile('&')
|
|
ref = re.compile('&(' + _Name + '|#[0-9]+|#x[0-9a-fA-F]+)[^-a-zA-Z0-9._:]')
|
|
entityref = re.compile('&(?P<name>' + _Name + ')[^-a-zA-Z0-9._:]')
|
|
charref = re.compile('&#(?P<char>[0-9]+[^0-9]|x[0-9a-fA-F]+[^0-9a-fA-F])')
|
|
space = re.compile(_S + '$')
|
|
newline = re.compile('\n')
|
|
|
|
attrfind = re.compile(
|
|
_S + '(?P<name>' + _Name + ')'
|
|
'(' + _opS + '=' + _opS +
|
|
'(?P<value>'+_QStr+'|[-a-zA-Z0-9.:+*%?!\(\)_#=~]+))?')
|
|
starttagopen = re.compile('<' + _Name)
|
|
starttagend = re.compile(_opS + '(?P<slash>/?)>')
|
|
starttagmatch = re.compile('<(?P<tagname>'+_Name+')'
|
|
'(?P<attrs>(?:'+attrfind.pattern+')*)'+
|
|
starttagend.pattern)
|
|
endtagopen = re.compile('</')
|
|
endbracket = re.compile(_opS + '>')
|
|
endbracketfind = re.compile('(?:[^>\'"]|'+_QStr+')*>')
|
|
tagfind = re.compile(_Name)
|
|
cdataopen = re.compile(r'<!\[CDATA\[')
|
|
cdataclose = re.compile(r'\]\]>')
|
|
# this matches one of the following:
|
|
# SYSTEM SystemLiteral
|
|
# PUBLIC PubidLiteral SystemLiteral
|
|
_SystemLiteral = '(?P<%s>'+_QStr+')'
|
|
_PublicLiteral = '(?P<%s>"[-\'\(\)+,./:=?;!*#@$_%% \n\ra-zA-Z0-9]*"|' \
|
|
"'[-\(\)+,./:=?;!*#@$_%% \n\ra-zA-Z0-9]*')"
|
|
_ExternalId = '(?:SYSTEM|' \
|
|
'PUBLIC'+_S+_PublicLiteral%'pubid'+ \
|
|
')'+_S+_SystemLiteral%'syslit'
|
|
doctype = re.compile('<!DOCTYPE'+_S+'(?P<name>'+_Name+')'
|
|
'(?:'+_S+_ExternalId+')?'+_opS)
|
|
xmldecl = re.compile('<\?xml'+_S+
|
|
'version'+_opS+'='+_opS+'(?P<version>'+_QStr+')'+
|
|
'(?:'+_S+'encoding'+_opS+'='+_opS+
|
|
"(?P<encoding>'[A-Za-z][-A-Za-z0-9._]*'|"
|
|
'"[A-Za-z][-A-Za-z0-9._]*"))?'
|
|
'(?:'+_S+'standalone'+_opS+'='+_opS+
|
|
'(?P<standalone>\'(?:yes|no)\'|"(?:yes|no)"))?'+
|
|
_opS+'\?>')
|
|
procopen = re.compile(r'<\?(?P<proc>' + _Name + ')' + _opS)
|
|
procclose = re.compile(_opS + r'\?>')
|
|
commentopen = re.compile('<!--')
|
|
commentclose = re.compile('-->')
|
|
doubledash = re.compile('--')
|
|
attrtrans = string.maketrans(' \r\n\t', ' ')
|
|
|
|
# definitions for XML namespaces
|
|
_NCName = '[a-zA-Z_][-a-zA-Z0-9._]*' # XML Name, minus the ":"
|
|
ncname = re.compile(_NCName + '$')
|
|
qname = re.compile('(?:(?P<prefix>' + _NCName + '):)?' # optional prefix
|
|
'(?P<local>' + _NCName + ')$')
|
|
|
|
xmlns = re.compile('xmlns(?::(?P<ncname>'+_NCName+'))?$')
|
|
|
|
# XML parser base class -- find tags and call handler functions.
|
|
# Usage: p = XMLParser(); p.feed(data); ...; p.close().
|
|
# The dtd is defined by deriving a class which defines methods with
|
|
# special names to handle tags: start_foo and end_foo to handle <foo>
|
|
# and </foo>, respectively. The data between tags is passed to the
|
|
# parser by calling self.handle_data() with some data as argument (the
|
|
# data may be split up in arbitrary chunks).
|
|
|
|
class XMLParser:
|
|
attributes = {} # default, to be overridden
|
|
elements = {} # default, to be overridden
|
|
|
|
# parsing options, settable using keyword args in __init__
|
|
__accept_unquoted_attributes = 0
|
|
__accept_missing_endtag_name = 0
|
|
__map_case = 0
|
|
__accept_utf8 = 0
|
|
__translate_attribute_references = 1
|
|
|
|
# Interface -- initialize and reset this instance
|
|
def __init__(self, **kw):
|
|
self.__fixed = 0
|
|
if 'accept_unquoted_attributes' in kw:
|
|
self.__accept_unquoted_attributes = kw['accept_unquoted_attributes']
|
|
if 'accept_missing_endtag_name' in kw:
|
|
self.__accept_missing_endtag_name = kw['accept_missing_endtag_name']
|
|
if 'map_case' in kw:
|
|
self.__map_case = kw['map_case']
|
|
if 'accept_utf8' in kw:
|
|
self.__accept_utf8 = kw['accept_utf8']
|
|
if 'translate_attribute_references' in kw:
|
|
self.__translate_attribute_references = kw['translate_attribute_references']
|
|
self.reset()
|
|
|
|
def __fixelements(self):
|
|
self.__fixed = 1
|
|
self.elements = {}
|
|
self.__fixdict(self.__dict__)
|
|
self.__fixclass(self.__class__)
|
|
|
|
def __fixclass(self, kl):
|
|
self.__fixdict(kl.__dict__)
|
|
for k in kl.__bases__:
|
|
self.__fixclass(k)
|
|
|
|
def __fixdict(self, dict):
|
|
for key in dict.keys():
|
|
if key[:6] == 'start_':
|
|
tag = key[6:]
|
|
start, end = self.elements.get(tag, (None, None))
|
|
if start is None:
|
|
self.elements[tag] = getattr(self, key), end
|
|
elif key[:4] == 'end_':
|
|
tag = key[4:]
|
|
start, end = self.elements.get(tag, (None, None))
|
|
if end is None:
|
|
self.elements[tag] = start, getattr(self, key)
|
|
|
|
# Interface -- reset this instance. Loses all unprocessed data
|
|
def reset(self):
|
|
self.rawdata = ''
|
|
self.stack = []
|
|
self.nomoretags = 0
|
|
self.literal = 0
|
|
self.lineno = 1
|
|
self.__at_start = 1
|
|
self.__seen_doctype = None
|
|
self.__seen_starttag = 0
|
|
self.__use_namespaces = 0
|
|
self.__namespaces = {'xml':None} # xml is implicitly declared
|
|
# backward compatibility hack: if elements not overridden,
|
|
# fill it in ourselves
|
|
if self.elements is XMLParser.elements:
|
|
self.__fixelements()
|
|
|
|
# For derived classes only -- enter literal mode (CDATA) till EOF
|
|
def setnomoretags(self):
|
|
self.nomoretags = self.literal = 1
|
|
|
|
# For derived classes only -- enter literal mode (CDATA)
|
|
def setliteral(self, *args):
|
|
self.literal = 1
|
|
|
|
# Interface -- feed some data to the parser. Call this as
|
|
# often as you want, with as little or as much text as you
|
|
# want (may include '\n'). (This just saves the text, all the
|
|
# processing is done by goahead().)
|
|
def feed(self, data):
|
|
self.rawdata = self.rawdata + data
|
|
self.goahead(0)
|
|
|
|
# Interface -- handle the remaining data
|
|
def close(self):
|
|
self.goahead(1)
|
|
if self.__fixed:
|
|
self.__fixed = 0
|
|
# remove self.elements so that we don't leak
|
|
del self.elements
|
|
|
|
# Interface -- translate references
|
|
def translate_references(self, data, all = 1):
|
|
if not self.__translate_attribute_references:
|
|
return data
|
|
i = 0
|
|
while 1:
|
|
res = amp.search(data, i)
|
|
if res is None:
|
|
return data
|
|
s = res.start(0)
|
|
res = ref.match(data, s)
|
|
if res is None:
|
|
self.syntax_error("bogus `&'")
|
|
i = s+1
|
|
continue
|
|
i = res.end(0)
|
|
str = res.group(1)
|
|
rescan = 0
|
|
if str[0] == '#':
|
|
if str[1] == 'x':
|
|
str = chr(int(str[2:], 16))
|
|
else:
|
|
str = chr(int(str[1:]))
|
|
if data[i - 1] != ';':
|
|
self.syntax_error("`;' missing after char reference")
|
|
i = i-1
|
|
elif all:
|
|
if str in self.entitydefs:
|
|
str = self.entitydefs[str]
|
|
rescan = 1
|
|
elif data[i - 1] != ';':
|
|
self.syntax_error("bogus `&'")
|
|
i = s + 1 # just past the &
|
|
continue
|
|
else:
|
|
self.syntax_error("reference to unknown entity `&%s;'" % str)
|
|
str = '&' + str + ';'
|
|
elif data[i - 1] != ';':
|
|
self.syntax_error("bogus `&'")
|
|
i = s + 1 # just past the &
|
|
continue
|
|
|
|
# when we get here, str contains the translated text and i points
|
|
# to the end of the string that is to be replaced
|
|
data = data[:s] + str + data[i:]
|
|
if rescan:
|
|
i = s
|
|
else:
|
|
i = s + len(str)
|
|
|
|
# Interface - return a dictionary of all namespaces currently valid
|
|
def getnamespace(self):
|
|
nsdict = {}
|
|
for t, d, nst in self.stack:
|
|
nsdict.update(d)
|
|
return nsdict
|
|
|
|
# Internal -- handle data as far as reasonable. May leave state
|
|
# and data to be processed by a subsequent call. If 'end' is
|
|
# true, force handling all data as if followed by EOF marker.
|
|
def goahead(self, end):
|
|
rawdata = self.rawdata
|
|
i = 0
|
|
n = len(rawdata)
|
|
while i < n:
|
|
if i > 0:
|
|
self.__at_start = 0
|
|
if self.nomoretags:
|
|
data = rawdata[i:n]
|
|
self.handle_data(data)
|
|
self.lineno = self.lineno + data.count('\n')
|
|
i = n
|
|
break
|
|
res = interesting.search(rawdata, i)
|
|
if res:
|
|
j = res.start(0)
|
|
else:
|
|
j = n
|
|
if i < j:
|
|
data = rawdata[i:j]
|
|
if self.__at_start and space.match(data) is None:
|
|
self.syntax_error('illegal data at start of file')
|
|
self.__at_start = 0
|
|
if not self.stack and space.match(data) is None:
|
|
self.syntax_error('data not in content')
|
|
if not self.__accept_utf8 and illegal.search(data):
|
|
self.syntax_error('illegal character in content')
|
|
self.handle_data(data)
|
|
self.lineno = self.lineno + data.count('\n')
|
|
i = j
|
|
if i == n: break
|
|
if rawdata[i] == '<':
|
|
if starttagopen.match(rawdata, i):
|
|
if self.literal:
|
|
data = rawdata[i]
|
|
self.handle_data(data)
|
|
self.lineno = self.lineno + data.count('\n')
|
|
i = i+1
|
|
continue
|
|
k = self.parse_starttag(i)
|
|
if k < 0: break
|
|
self.__seen_starttag = 1
|
|
self.lineno = self.lineno + rawdata[i:k].count('\n')
|
|
i = k
|
|
continue
|
|
if endtagopen.match(rawdata, i):
|
|
k = self.parse_endtag(i)
|
|
if k < 0: break
|
|
self.lineno = self.lineno + rawdata[i:k].count('\n')
|
|
i = k
|
|
continue
|
|
if commentopen.match(rawdata, i):
|
|
if self.literal:
|
|
data = rawdata[i]
|
|
self.handle_data(data)
|
|
self.lineno = self.lineno + data.count('\n')
|
|
i = i+1
|
|
continue
|
|
k = self.parse_comment(i)
|
|
if k < 0: break
|
|
self.lineno = self.lineno + rawdata[i:k].count('\n')
|
|
i = k
|
|
continue
|
|
if cdataopen.match(rawdata, i):
|
|
k = self.parse_cdata(i)
|
|
if k < 0: break
|
|
self.lineno = self.lineno + rawdata[i:k].count('\n')
|
|
i = k
|
|
continue
|
|
res = xmldecl.match(rawdata, i)
|
|
if res:
|
|
if not self.__at_start:
|
|
self.syntax_error("<?xml?> declaration not at start of document")
|
|
version, encoding, standalone = res.group('version',
|
|
'encoding',
|
|
'standalone')
|
|
if version[1:-1] != '1.0':
|
|
raise Error('only XML version 1.0 supported')
|
|
if encoding: encoding = encoding[1:-1]
|
|
if standalone: standalone = standalone[1:-1]
|
|
self.handle_xml(encoding, standalone)
|
|
i = res.end(0)
|
|
continue
|
|
res = procopen.match(rawdata, i)
|
|
if res:
|
|
k = self.parse_proc(i)
|
|
if k < 0: break
|
|
self.lineno = self.lineno + rawdata[i:k].count('\n')
|
|
i = k
|
|
continue
|
|
res = doctype.match(rawdata, i)
|
|
if res:
|
|
if self.literal:
|
|
data = rawdata[i]
|
|
self.handle_data(data)
|
|
self.lineno = self.lineno + data.count('\n')
|
|
i = i+1
|
|
continue
|
|
if self.__seen_doctype:
|
|
self.syntax_error('multiple DOCTYPE elements')
|
|
if self.__seen_starttag:
|
|
self.syntax_error('DOCTYPE not at beginning of document')
|
|
k = self.parse_doctype(res)
|
|
if k < 0: break
|
|
self.__seen_doctype = res.group('name')
|
|
if self.__map_case:
|
|
self.__seen_doctype = self.__seen_doctype.lower()
|
|
self.lineno = self.lineno + rawdata[i:k].count('\n')
|
|
i = k
|
|
continue
|
|
elif rawdata[i] == '&':
|
|
if self.literal:
|
|
data = rawdata[i]
|
|
self.handle_data(data)
|
|
i = i+1
|
|
continue
|
|
res = charref.match(rawdata, i)
|
|
if res is not None:
|
|
i = res.end(0)
|
|
if rawdata[i-1] != ';':
|
|
self.syntax_error("`;' missing in charref")
|
|
i = i-1
|
|
if not self.stack:
|
|
self.syntax_error('data not in content')
|
|
self.handle_charref(res.group('char')[:-1])
|
|
self.lineno = self.lineno + res.group(0).count('\n')
|
|
continue
|
|
res = entityref.match(rawdata, i)
|
|
if res is not None:
|
|
i = res.end(0)
|
|
if rawdata[i-1] != ';':
|
|
self.syntax_error("`;' missing in entityref")
|
|
i = i-1
|
|
name = res.group('name')
|
|
if self.__map_case:
|
|
name = name.lower()
|
|
if name in self.entitydefs:
|
|
self.rawdata = rawdata = rawdata[:res.start(0)] + self.entitydefs[name] + rawdata[i:]
|
|
n = len(rawdata)
|
|
i = res.start(0)
|
|
else:
|
|
self.unknown_entityref(name)
|
|
self.lineno = self.lineno + res.group(0).count('\n')
|
|
continue
|
|
elif rawdata[i] == ']':
|
|
if self.literal:
|
|
data = rawdata[i]
|
|
self.handle_data(data)
|
|
i = i+1
|
|
continue
|
|
if n-i < 3:
|
|
break
|
|
if cdataclose.match(rawdata, i):
|
|
self.syntax_error("bogus `]]>'")
|
|
self.handle_data(rawdata[i])
|
|
i = i+1
|
|
continue
|
|
else:
|
|
raise Error('neither < nor & ??')
|
|
# We get here only if incomplete matches but
|
|
# nothing else
|
|
break
|
|
# end while
|
|
if i > 0:
|
|
self.__at_start = 0
|
|
if end and i < n:
|
|
data = rawdata[i]
|
|
self.syntax_error("bogus `%s'" % data)
|
|
if not self.__accept_utf8 and illegal.search(data):
|
|
self.syntax_error('illegal character in content')
|
|
self.handle_data(data)
|
|
self.lineno = self.lineno + data.count('\n')
|
|
self.rawdata = rawdata[i+1:]
|
|
return self.goahead(end)
|
|
self.rawdata = rawdata[i:]
|
|
if end:
|
|
if not self.__seen_starttag:
|
|
self.syntax_error('no elements in file')
|
|
if self.stack:
|
|
self.syntax_error('missing end tags')
|
|
while self.stack:
|
|
self.finish_endtag(self.stack[-1][0])
|
|
|
|
# Internal -- parse comment, return length or -1 if not terminated
|
|
def parse_comment(self, i):
|
|
rawdata = self.rawdata
|
|
if rawdata[i:i+4] != '<!--':
|
|
raise Error('unexpected call to handle_comment')
|
|
res = commentclose.search(rawdata, i+4)
|
|
if res is None:
|
|
return -1
|
|
if doubledash.search(rawdata, i+4, res.start(0)):
|
|
self.syntax_error("`--' inside comment")
|
|
if rawdata[res.start(0)-1] == '-':
|
|
self.syntax_error('comment cannot end in three dashes')
|
|
if not self.__accept_utf8 and \
|
|
illegal.search(rawdata, i+4, res.start(0)):
|
|
self.syntax_error('illegal character in comment')
|
|
self.handle_comment(rawdata[i+4: res.start(0)])
|
|
return res.end(0)
|
|
|
|
# Internal -- handle DOCTYPE tag, return length or -1 if not terminated
|
|
def parse_doctype(self, res):
|
|
rawdata = self.rawdata
|
|
n = len(rawdata)
|
|
name = res.group('name')
|
|
if self.__map_case:
|
|
name = name.lower()
|
|
pubid, syslit = res.group('pubid', 'syslit')
|
|
if pubid is not None:
|
|
pubid = pubid[1:-1] # remove quotes
|
|
pubid = ' '.join(pubid.split()) # normalize
|
|
if syslit is not None: syslit = syslit[1:-1] # remove quotes
|
|
j = k = res.end(0)
|
|
if k >= n:
|
|
return -1
|
|
if rawdata[k] == '[':
|
|
level = 0
|
|
k = k+1
|
|
dq = sq = 0
|
|
while k < n:
|
|
c = rawdata[k]
|
|
if not sq and c == '"':
|
|
dq = not dq
|
|
elif not dq and c == "'":
|
|
sq = not sq
|
|
elif sq or dq:
|
|
pass
|
|
elif level <= 0 and c == ']':
|
|
res = endbracket.match(rawdata, k+1)
|
|
if res is None:
|
|
return -1
|
|
self.handle_doctype(name, pubid, syslit, rawdata[j+1:k])
|
|
return res.end(0)
|
|
elif c == '<':
|
|
level = level + 1
|
|
elif c == '>':
|
|
level = level - 1
|
|
if level < 0:
|
|
self.syntax_error("bogus `>' in DOCTYPE")
|
|
k = k+1
|
|
res = endbracketfind.match(rawdata, k)
|
|
if res is None:
|
|
return -1
|
|
if endbracket.match(rawdata, k) is None:
|
|
self.syntax_error('garbage in DOCTYPE')
|
|
self.handle_doctype(name, pubid, syslit, None)
|
|
return res.end(0)
|
|
|
|
# Internal -- handle CDATA tag, return length or -1 if not terminated
|
|
def parse_cdata(self, i):
|
|
rawdata = self.rawdata
|
|
if rawdata[i:i+9] != '<![CDATA[':
|
|
raise Error('unexpected call to parse_cdata')
|
|
res = cdataclose.search(rawdata, i+9)
|
|
if res is None:
|
|
return -1
|
|
if not self.__accept_utf8 and \
|
|
illegal.search(rawdata, i+9, res.start(0)):
|
|
self.syntax_error('illegal character in CDATA')
|
|
if not self.stack:
|
|
self.syntax_error('CDATA not in content')
|
|
self.handle_cdata(rawdata[i+9:res.start(0)])
|
|
return res.end(0)
|
|
|
|
__xml_namespace_attributes = {'ns':None, 'src':None, 'prefix':None}
|
|
# Internal -- handle a processing instruction tag
|
|
def parse_proc(self, i):
|
|
rawdata = self.rawdata
|
|
end = procclose.search(rawdata, i)
|
|
if end is None:
|
|
return -1
|
|
j = end.start(0)
|
|
if not self.__accept_utf8 and illegal.search(rawdata, i+2, j):
|
|
self.syntax_error('illegal character in processing instruction')
|
|
res = tagfind.match(rawdata, i+2)
|
|
if res is None:
|
|
raise Error('unexpected call to parse_proc')
|
|
k = res.end(0)
|
|
name = res.group(0)
|
|
if self.__map_case:
|
|
name = name.lower()
|
|
if name == 'xml:namespace':
|
|
self.syntax_error('old-fashioned namespace declaration')
|
|
self.__use_namespaces = -1
|
|
# namespace declaration
|
|
# this must come after the <?xml?> declaration (if any)
|
|
# and before the <!DOCTYPE> (if any).
|
|
if self.__seen_doctype or self.__seen_starttag:
|
|
self.syntax_error('xml:namespace declaration too late in document')
|
|
attrdict, namespace, k = self.parse_attributes(name, k, j)
|
|
if namespace:
|
|
self.syntax_error('namespace declaration inside namespace declaration')
|
|
for attrname in attrdict.keys():
|
|
if not attrname in self.__xml_namespace_attributes:
|
|
self.syntax_error("unknown attribute `%s' in xml:namespace tag" % attrname)
|
|
if not 'ns' in attrdict or not 'prefix' in attrdict:
|
|
self.syntax_error('xml:namespace without required attributes')
|
|
prefix = attrdict.get('prefix')
|
|
if ncname.match(prefix) is None:
|
|
self.syntax_error('xml:namespace illegal prefix value')
|
|
return end.end(0)
|
|
if prefix in self.__namespaces:
|
|
self.syntax_error('xml:namespace prefix not unique')
|
|
self.__namespaces[prefix] = attrdict['ns']
|
|
else:
|
|
if name.lower() == 'xml':
|
|
self.syntax_error('illegal processing instruction target name')
|
|
self.handle_proc(name, rawdata[k:j])
|
|
return end.end(0)
|
|
|
|
# Internal -- parse attributes between i and j
|
|
def parse_attributes(self, tag, i, j):
|
|
rawdata = self.rawdata
|
|
attrdict = {}
|
|
namespace = {}
|
|
while i < j:
|
|
res = attrfind.match(rawdata, i)
|
|
if res is None:
|
|
break
|
|
attrname, attrvalue = res.group('name', 'value')
|
|
if self.__map_case:
|
|
attrname = attrname.lower()
|
|
i = res.end(0)
|
|
if attrvalue is None:
|
|
self.syntax_error("no value specified for attribute `%s'" % attrname)
|
|
attrvalue = attrname
|
|
elif attrvalue[:1] == "'" == attrvalue[-1:] or \
|
|
attrvalue[:1] == '"' == attrvalue[-1:]:
|
|
attrvalue = attrvalue[1:-1]
|
|
elif not self.__accept_unquoted_attributes:
|
|
self.syntax_error("attribute `%s' value not quoted" % attrname)
|
|
res = xmlns.match(attrname)
|
|
if res is not None:
|
|
# namespace declaration
|
|
ncname = res.group('ncname')
|
|
namespace[ncname or ''] = attrvalue or None
|
|
if not self.__use_namespaces:
|
|
self.__use_namespaces = len(self.stack)+1
|
|
continue
|
|
if '<' in attrvalue:
|
|
self.syntax_error("`<' illegal in attribute value")
|
|
if attrname in attrdict:
|
|
self.syntax_error("attribute `%s' specified twice" % attrname)
|
|
attrvalue = attrvalue.translate(attrtrans)
|
|
attrdict[attrname] = self.translate_references(attrvalue)
|
|
return attrdict, namespace, i
|
|
|
|
# Internal -- handle starttag, return length or -1 if not terminated
|
|
def parse_starttag(self, i):
|
|
rawdata = self.rawdata
|
|
# i points to start of tag
|
|
end = endbracketfind.match(rawdata, i+1)
|
|
if end is None:
|
|
return -1
|
|
tag = starttagmatch.match(rawdata, i)
|
|
if tag is None or tag.end(0) != end.end(0):
|
|
self.syntax_error('garbage in starttag')
|
|
return end.end(0)
|
|
nstag = tagname = tag.group('tagname')
|
|
if self.__map_case:
|
|
nstag = tagname = nstag.lower()
|
|
if not self.__seen_starttag and self.__seen_doctype and \
|
|
tagname != self.__seen_doctype:
|
|
self.syntax_error('starttag does not match DOCTYPE')
|
|
if self.__seen_starttag and not self.stack:
|
|
self.syntax_error('multiple elements on top level')
|
|
k, j = tag.span('attrs')
|
|
attrdict, nsdict, k = self.parse_attributes(tagname, k, j)
|
|
self.stack.append((tagname, nsdict, nstag))
|
|
if self.__use_namespaces:
|
|
res = qname.match(tagname)
|
|
else:
|
|
res = None
|
|
if res is not None:
|
|
prefix, nstag = res.group('prefix', 'local')
|
|
if prefix is None:
|
|
prefix = ''
|
|
ns = None
|
|
for t, d, nst in self.stack:
|
|
if prefix in d:
|
|
ns = d[prefix]
|
|
if ns is None and prefix != '':
|
|
ns = self.__namespaces.get(prefix)
|
|
if ns is not None:
|
|
nstag = ns + ' ' + nstag
|
|
elif prefix != '':
|
|
nstag = prefix + ':' + nstag # undo split
|
|
self.stack[-1] = tagname, nsdict, nstag
|
|
# translate namespace of attributes
|
|
attrnamemap = {} # map from new name to old name (used for error reporting)
|
|
for key in attrdict.keys():
|
|
attrnamemap[key] = key
|
|
if self.__use_namespaces:
|
|
nattrdict = {}
|
|
for key, val in attrdict.items():
|
|
okey = key
|
|
res = qname.match(key)
|
|
if res is not None:
|
|
aprefix, key = res.group('prefix', 'local')
|
|
if self.__map_case:
|
|
key = key.lower()
|
|
if aprefix is None:
|
|
aprefix = ''
|
|
ans = None
|
|
for t, d, nst in self.stack:
|
|
if aprefix in d:
|
|
ans = d[aprefix]
|
|
if ans is None and aprefix != '':
|
|
ans = self.__namespaces.get(aprefix)
|
|
if ans is not None:
|
|
key = ans + ' ' + key
|
|
elif aprefix != '':
|
|
key = aprefix + ':' + key
|
|
elif ns is not None:
|
|
key = ns + ' ' + key
|
|
nattrdict[key] = val
|
|
attrnamemap[key] = okey
|
|
attrdict = nattrdict
|
|
attributes = self.attributes.get(nstag)
|
|
if attributes is not None:
|
|
for key in attrdict.keys():
|
|
if not key in attributes:
|
|
self.syntax_error("unknown attribute `%s' in tag `%s'" % (attrnamemap[key], tagname))
|
|
for key, val in attributes.items():
|
|
if val is not None and not key in attrdict:
|
|
attrdict[key] = val
|
|
method = self.elements.get(nstag, (None, None))[0]
|
|
self.finish_starttag(nstag, attrdict, method)
|
|
if tag.group('slash') == '/':
|
|
self.finish_endtag(tagname)
|
|
return tag.end(0)
|
|
|
|
# Internal -- parse endtag
|
|
def parse_endtag(self, i):
|
|
rawdata = self.rawdata
|
|
end = endbracketfind.match(rawdata, i+1)
|
|
if end is None:
|
|
return -1
|
|
res = tagfind.match(rawdata, i+2)
|
|
if res is None:
|
|
if self.literal:
|
|
self.handle_data(rawdata[i])
|
|
return i+1
|
|
if not self.__accept_missing_endtag_name:
|
|
self.syntax_error('no name specified in end tag')
|
|
tag = self.stack[-1][0]
|
|
k = i+2
|
|
else:
|
|
tag = res.group(0)
|
|
if self.__map_case:
|
|
tag = tag.lower()
|
|
if self.literal:
|
|
if not self.stack or tag != self.stack[-1][0]:
|
|
self.handle_data(rawdata[i])
|
|
return i+1
|
|
k = res.end(0)
|
|
if endbracket.match(rawdata, k) is None:
|
|
self.syntax_error('garbage in end tag')
|
|
self.finish_endtag(tag)
|
|
return end.end(0)
|
|
|
|
# Internal -- finish processing of start tag
|
|
def finish_starttag(self, tagname, attrdict, method):
|
|
if method is not None:
|
|
self.handle_starttag(tagname, method, attrdict)
|
|
else:
|
|
self.unknown_starttag(tagname, attrdict)
|
|
|
|
# Internal -- finish processing of end tag
|
|
def finish_endtag(self, tag):
|
|
self.literal = 0
|
|
if not tag:
|
|
self.syntax_error('name-less end tag')
|
|
found = len(self.stack) - 1
|
|
if found < 0:
|
|
self.unknown_endtag(tag)
|
|
return
|
|
else:
|
|
found = -1
|
|
for i in range(len(self.stack)):
|
|
if tag == self.stack[i][0]:
|
|
found = i
|
|
if found == -1:
|
|
self.syntax_error('unopened end tag')
|
|
return
|
|
while len(self.stack) > found:
|
|
if found < len(self.stack) - 1:
|
|
self.syntax_error('missing close tag for %s' % self.stack[-1][2])
|
|
nstag = self.stack[-1][2]
|
|
method = self.elements.get(nstag, (None, None))[1]
|
|
if method is not None:
|
|
self.handle_endtag(nstag, method)
|
|
else:
|
|
self.unknown_endtag(nstag)
|
|
if self.__use_namespaces == len(self.stack):
|
|
self.__use_namespaces = 0
|
|
del self.stack[-1]
|
|
|
|
# Overridable -- handle xml processing instruction
|
|
def handle_xml(self, encoding, standalone):
|
|
pass
|
|
|
|
# Overridable -- handle DOCTYPE
|
|
def handle_doctype(self, tag, pubid, syslit, data):
|
|
pass
|
|
|
|
# Overridable -- handle start tag
|
|
def handle_starttag(self, tag, method, attrs):
|
|
method(attrs)
|
|
|
|
# Overridable -- handle end tag
|
|
def handle_endtag(self, tag, method):
|
|
method()
|
|
|
|
# Example -- handle character reference, no need to override
|
|
def handle_charref(self, name):
|
|
try:
|
|
if name[0] == 'x':
|
|
n = int(name[1:], 16)
|
|
else:
|
|
n = int(name)
|
|
except ValueError:
|
|
self.unknown_charref(name)
|
|
return
|
|
if not 0 <= n <= 255:
|
|
self.unknown_charref(name)
|
|
return
|
|
self.handle_data(chr(n))
|
|
|
|
# Definition of entities -- derived classes may override
|
|
entitydefs = {'lt': '<', # must use charref
|
|
'gt': '>',
|
|
'amp': '&', # must use charref
|
|
'quot': '"',
|
|
'apos': ''',
|
|
}
|
|
|
|
# Example -- handle data, should be overridden
|
|
def handle_data(self, data):
|
|
pass
|
|
|
|
# Example -- handle cdata, could be overridden
|
|
def handle_cdata(self, data):
|
|
pass
|
|
|
|
# Example -- handle comment, could be overridden
|
|
def handle_comment(self, data):
|
|
pass
|
|
|
|
# Example -- handle processing instructions, could be overridden
|
|
def handle_proc(self, name, data):
|
|
pass
|
|
|
|
# Example -- handle relatively harmless syntax errors, could be overridden
|
|
def syntax_error(self, message):
|
|
raise Error('Syntax error at line %d: %s' % (self.lineno, message))
|
|
|
|
# To be overridden -- handlers for unknown objects
|
|
def unknown_starttag(self, tag, attrs): pass
|
|
def unknown_endtag(self, tag): pass
|
|
def unknown_charref(self, ref): pass
|
|
def unknown_entityref(self, name):
|
|
self.syntax_error("reference to unknown entity `&%s;'" % name)
|
|
|
|
|
|
class TestXMLParser(XMLParser):
|
|
|
|
def __init__(self, **kw):
|
|
self.testdata = ""
|
|
apply(XMLParser.__init__, (self,), kw)
|
|
|
|
def handle_xml(self, encoding, standalone):
|
|
self.flush()
|
|
print 'xml: encoding =',encoding,'standalone =',standalone
|
|
|
|
def handle_doctype(self, tag, pubid, syslit, data):
|
|
self.flush()
|
|
print 'DOCTYPE:',tag, `data`
|
|
|
|
def handle_data(self, data):
|
|
self.testdata = self.testdata + data
|
|
if len(`self.testdata`) >= 70:
|
|
self.flush()
|
|
|
|
def flush(self):
|
|
data = self.testdata
|
|
if data:
|
|
self.testdata = ""
|
|
print 'data:', `data`
|
|
|
|
def handle_cdata(self, data):
|
|
self.flush()
|
|
print 'cdata:', `data`
|
|
|
|
def handle_proc(self, name, data):
|
|
self.flush()
|
|
print 'processing:',name,`data`
|
|
|
|
def handle_comment(self, data):
|
|
self.flush()
|
|
r = `data`
|
|
if len(r) > 68:
|
|
r = r[:32] + '...' + r[-32:]
|
|
print 'comment:', r
|
|
|
|
def syntax_error(self, message):
|
|
print 'error at line %d:' % self.lineno, message
|
|
|
|
def unknown_starttag(self, tag, attrs):
|
|
self.flush()
|
|
if not attrs:
|
|
print 'start tag: <' + tag + '>'
|
|
else:
|
|
print 'start tag: <' + tag,
|
|
for name, value in attrs.items():
|
|
print name + '=' + '"' + value + '"',
|
|
print '>'
|
|
|
|
def unknown_endtag(self, tag):
|
|
self.flush()
|
|
print 'end tag: </' + tag + '>'
|
|
|
|
def unknown_entityref(self, ref):
|
|
self.flush()
|
|
print '*** unknown entity ref: &' + ref + ';'
|
|
|
|
def unknown_charref(self, ref):
|
|
self.flush()
|
|
print '*** unknown char ref: &#' + ref + ';'
|
|
|
|
def close(self):
|
|
XMLParser.close(self)
|
|
self.flush()
|
|
|
|
def test(args = None):
|
|
import sys, getopt
|
|
from time import time
|
|
|
|
if not args:
|
|
args = sys.argv[1:]
|
|
|
|
opts, args = getopt.getopt(args, 'st')
|
|
klass = TestXMLParser
|
|
do_time = 0
|
|
for o, a in opts:
|
|
if o == '-s':
|
|
klass = XMLParser
|
|
elif o == '-t':
|
|
do_time = 1
|
|
|
|
if args:
|
|
file = args[0]
|
|
else:
|
|
file = 'test.xml'
|
|
|
|
if file == '-':
|
|
f = sys.stdin
|
|
else:
|
|
try:
|
|
f = open(file, 'r')
|
|
except IOError, msg:
|
|
print file, ":", msg
|
|
sys.exit(1)
|
|
|
|
data = f.read()
|
|
if f is not sys.stdin:
|
|
f.close()
|
|
|
|
x = klass()
|
|
t0 = time()
|
|
try:
|
|
if do_time:
|
|
x.feed(data)
|
|
x.close()
|
|
else:
|
|
for c in data:
|
|
x.feed(c)
|
|
x.close()
|
|
except Error, msg:
|
|
t1 = time()
|
|
print msg
|
|
if do_time:
|
|
print 'total time: %g' % (t1-t0)
|
|
sys.exit(1)
|
|
t1 = time()
|
|
if do_time:
|
|
print 'total time: %g' % (t1-t0)
|
|
|
|
|
|
if __name__ == '__main__':
|
|
test()
|