Thank you for your suggestion.
I examine that page carefully and I found that the page is quite bad
cuz there around 2000 lines of <li>, but without an closing tag.
like this:
<li><a href="
http://www.swcruise.com/store/?
furie=grassiness">Grassiness</a>
<li><a href="
http://www.kiop.ru/addfav.php?furie=glossiest">Glossiest</
a>
So the recursion is really to deep. And I set the recursion limit to
3000 and still can only parse half of this page, there's no memory
left for me to add the limit any more. But I think this page is too
bad to be prettified
---
Btw, after handling 30,000 pages, 320 of them cannot be souped, which
is 1% of them. The performance is good but I still hope some
improvement can be made. I will check out the reasons of these pages
after I run my program again.
On Nov 20, 11:38 am, Andrew Spiers <
7and...@gmail.com> wrote: