< Back to Blog
May 31, 2019

May 2019 crawl archive now available

Note: this post has been marked as obsolete.
The crawl archive for May 2019 is now available! It contains 2.65 billion web pages or 220 TiB of uncompressed content, crawled between May 19th and 27th.
Sebastian Nagel
Sebastian Nagel
Sebastian is a Distinguished Engineer with Common Crawl.

The crawl archive for May 2019 is now available! It contains 2.65 billion web pages or 220 TiB of uncompressed content, crawled between May 19th and 27th.

Data Type File List #Files Total Size
Compressed (TiB)
Segments segment.paths.gz 100
WARC warc.paths.gz 56000 51.48
WAT wat.paths.gz 56000 17.75
WET wet.paths.gz 56000 7.70
Robots.txt files robotstxt.paths.gz 56000 0.17
Non-200 responses non200responses.paths.gz 56000 1.84
URL index files cc-index.paths.gz 302 0.20
Columnar URL index files cc-index-table.paths.gz 900 0.25

The May crawl contains page captures of 825 million URLs not contained in any crawl archive before. New URLs are sampled based on the host and domain ranks (harmonic centrality) published as part of the Feb/Mar/Apr 2019 webgraph data set from the following sources:

  • sitemaps, RSS and Atom feeds
  • a breadth-first side crawl within a maximum of 4 links (“hops”) away from the homepages of the top 60 million hosts and domains and a random sample of 1 million human-readable sitemap pages (HTML format)
  • a random sample of 1.6 billion outlinks taken from WAT files of the April crawl

Archive Location and Download

The May crawl archive is located in the commoncrawl bucket at crawl-data/CC-MAIN-2019-22/. To assist with exploring and using the dataset, we provide gzipped files which list all segments, WARC, WAT and WET files.

By simply adding either s3://commoncrawl/ or https://data.commoncrawl.org/ to each line, you end up with the S3 and HTTP paths respectively.

The Common Crawl URL Index for this crawl is available at: https://index.commoncrawl.org/CC-MAIN-2019-22/. Also the columnar index has been updated to contain this crawl.

Please donate to Common Crawl if you appreciate our free datasets! We’re also seeking corporate sponsors to partner with Common Crawl for our non-profit work in open data. Please contact info@commoncrawl.org for sponsorship information.

This release was authored by:
No items found.

Erratum: 

WAT data: repeated WARC and HTTP headers are not preserved

Originally reported by: 
Permalink

Repeated HTTP and WARC headers were not represented in the JSON data in WAT files. When a header was repeated adding a further value of that header, only the last value was stored and other values were lost. This issues was fixed with CC-MAIN-2024-51, see ia-web-commons#18. All WAT files from CC-MAIN-2013-20 until CC-MAIN-2024-46 are affected.

Erratum: 

WARC revisit metadata records

Originally reported by: 
Permalink

The revisit records in the Common Crawl WARC archives in all crawls from CC-MAIN-2018-34 to CC-MAIN-2024-46 (since Aug 2018) lack the metadata record which is attached to all response records. Fixed with CC-MAIN-2024-51, see commoncrawl/nutch#33. Note: before CC-MAIN-2018-34, WARC revisit records were not stored at all.

Erratum: 

Erroneous title field in WAT records

Originally reported by: 
Robert Waksmunski
Permalink

The "Title" extracted in WAT records to the JSON path `Envelope > Payload-Metadata > HTTP-Response-Metadata > HTML-Metadata > Head > Title` is not the content included in the <title> element in the HTML header (<head> element) if the page contains further <title> elements in the page body. The content of the last <title> element is written to the WAT "Title". This bug was observed if the HTML page includes embedded SVG graphics.

The issue was reported by the user Robert Waksmunski:

...and was fixed for CC-MAIN-2024-42 by commoncrawl/ia-web-commons#37.

This erratum affects all crawls from CC-MAIN-2013-20 until CC-MAIN-2024-38.

Erratum: 

Incorrect fetch_time metadata

Originally reported by: 
Permalink

In crawls CC-MAIN-2016-36 to CC-MAIN-2016-50, and CC-MAIN-2018-34 to CC-MAIN-2019-47 the fetch_time metadata for robots.txt might be incorrect. The correct times can be found in collinfo.json. See the related issue (commoncrawl/nutch#14) for more information.