xsl-list
[Top] [All Lists]

Re: [xsl] Processing large XML Documents [> 50MB]

2010-02-22 17:51:34
Ramkumar Menon schrieb am 22.02.2010 um 15:40:47 (-0800):
We have a need to process XML Documents that could be over 50 megs in
size.

Due to the huge size of the document, XSLT is getting tough, with the
environment we are running in.

Basically, the nature of the data procesing are

a) assemble around 30-40 XML documents [each with a common header and
its own lines] into one single XML document, with the common header
and all the lines b) Update the assembled document in specific
locations c) generate multiple XML document fragments from the huge
XML document based on query criteria. Each XML frgment is created by
mapping specific fields in the big document. Each document is created
for a specific key element value in the huge document.

Am puzzled how to handle this one efficiently.

* Saxon streaming extension
  http://www.saxonica.com/documentation/sourcedocs/serial.html

* an XML database and XQuery (Berkeley DB XML, eXist, MarkLogic, others)

* SAX filters (might easily get way too complicated, or even impossible)

From your desciprion, it sounds like SAX isn't an option. So I would
explore the other two options.

-- 
Michael Ludwig

--~------------------------------------------------------------------
XSL-List info and archive:  http://www.mulberrytech.com/xsl/xsl-list
To unsubscribe, go to: http://lists.mulberrytech.com/xsl-list/
or e-mail: <mailto:xsl-list-unsubscribe(_at_)lists(_dot_)mulberrytech(_dot_)com>
--~--

<Prev in Thread] Current Thread [Next in Thread>