It is not a database problem, it seems to me something about the reading of
the XML file. This is my code about this:
$data = implode(file("example.xml"), "");
xml_parse_into_struct($parser, $data, &$structure, &$index);
xml_parser_free($parser);
it could be the array $data creation fails because the example.xml is too
big, infact if I decrease the size of xml file the script works (even if it
lasts several minutes), unfortunately the xml_parse_into_struct function
needs the $data array as argument.
My PHP version is 4.0.6.
"Raditha Dissanayake" <[EMAIL PROTECTED]> ha scritto nel messaggio
news:[EMAIL PROTECTED]
> Ago wrote:
>
> >I have a PHP script who reads data from a XML file uploaded via http
form.
> >This file is about 15 MB. The script builds a query of thousands inserts
as
> >many as the products the XML file contains (more or less 4 inserts).
> >After 4 minutes the httpd process aborts by itself. What can it depend
on? I
> >have also increased the memory_limit parameter in php.ini to 32 MB and
> >set_time_limit to 0. What can I do anymore?
> >
> >
> Use a single insert statement instead of 40,000 - or -
> Drop the primary key before the insert and recreate it after wards. - or -
> save data into a csv file and use 'load data in file'
>
>
> Having said that the page shouldn't stop loading after 4 minutes if you
> use set_time_limit please see if you have some other setting that may be
> killing it.
>
>
> >Thanks.
> >
> >
> >
>
>
> --
> Raditha Dissanayake.
> -
> http://www.raditha.com/megaupload/upload.php
> Sneak past the PHP file upload limits.
--
PHP General Mailing List (http://www.php.net/)
To unsubscribe, visit: http://www.php.net/unsub.php