• Welcome to Sitemap Generator Forum.
 

Doesn't complete the crawl

Started by bloach, March 07, 2008, 02:20:00 PM

Previous topic - Next topic

bloach

Hi,

I have a site with around 200,000 pages. Sitemap crawler only crawls around 5000 and then stops or times out.

Is there any workaround for large sites?

XML-Sitemaps Support

Hello,

it looks like your server configuration doesn't allow to run the script long enough to create full sitemap. Please try to increase memory_limit and max_execution_time settings in php configuration at your host (php.ini file) or contact hosting support regarding this.

bloach

Hi Oleg,

I have increased all three resource variables but it would still stop after 500 products.

Any other ideas?

bloach

OK . I tried running the runcrawl.php via PuTTy SSH client and it indexed around 15,000 products and then gave following error:

PHP Fatal error:  Allowed memory size of 67108864 bytes exhausted (tried to allocate 14580497 bytes) in /var/www/vhosts/ZZZZ.com/subdomains/comparison/httpdocs/generator/pages/class.utils.inc.php(2) : eval()'d code on line 6
PHP Warning:  Unknown: open(/var/lib/php/session/sess_d4633d7ba9cdb38103129a155e390286, O_RDWR) failed: Permission denied (13) in Unknown on line 0
PHP Warning:  Unknown: Failed to write session data (files). Please verify that the current setting of [ External links are visible to forum administrators only ]_path is correct (/var/lib/php/session) in Unknown on line 0


da_lyman

I'm not real familiar with SSH.  Can you explain how I can access your software with SSH.  My site is currently hosted.

XML-Sitemaps Support

I mean if you have SSH access to your server (not to sitemap generator). In this case you can execute it in command line with:
/usr/local/bin/php /path/to/generator/runcrawl.php