Hi Alex,
no, it does not.
It would complicate the crawling, not only because of the execution of Javascript,
but also to hold all required libraries and dependencies to avoid that they
are refetched again with every HTML page.
The same problem appears on the other side when a page is read from the WARC archives.
The WARC simply archives the HTTP traffic between crawler and server(s),
but a dynamic page is composed of multiple requests and responses which may
not necessarily end up in the same WARC file.
Of course, there are many good reasons why a crawler should take dynamic content into
account...
Best,
Sebastian
> --
> You received this message because you are subscribed to the Google Groups "Common Crawl" group.
> To unsubscribe from this group and stop receiving emails from it, send an email to
>
common-crawl...@googlegroups.com <mailto:
common-crawl...@googlegroups.com>.
> To post to this group, send email to
common...@googlegroups.com
> <mailto:
common...@googlegroups.com>.
> Visit this group at
https://groups.google.com/group/common-crawl.
> For more options, visit
https://groups.google.com/d/optout.