< Back to Blog
December 16, 2016

December 2016 Crawl Archive Now Available

Note: this post has been marked as obsolete.
The crawl archive for December 2016 is now available! The archive contains more than 2.85 billion web pages.
Sebastian Nagel
Sebastian Nagel
Sebastian is a Distinguished Engineer with Common Crawl.

The crawl archive for December 2016 is now available! The archive is located in the commoncrawl bucket at crawl-data/CC-MAIN-2016-50/. It contains more than 2.85 billion web pages. Similar to the preceding September and October crawls, we used sitemaps to find new URLs for already known hosts shortly before the crawl was launched. In addition to the top-million domains from Alexa, sitemaps were mined for a list of multi-lingual sites. Thanks to the Statistical Machine Translation Group at the University of Edinburgh, which created this resource and made it available. We hope to have greater coverage of multi-lingual content in this and future crawls.

We are also grateful to webxtrakt for the continued donation of verified, DNS-resolvable domain names of European country-code TLDs (.eu, .fr, .be, .de, .ch, .nl, .pl, .ru, .dk).

Data Type File List #Files Total Size
Compressed (TiB)
Segments segment.paths.gz 100
WARC warc.paths.gz 51200 49.35
WAT wat.paths.gz 51200 18.72
WET wet.paths.gz 51200 8.19
Robots.txt files robotstxt.paths.gz 51200 0.09
Non-200 responses non200responses.paths.gz 51200 0.60
URL index files cc-index.paths.gz 302 0.19
Columnar URL index files cc-index-table.paths.gz 900 0.23

To assist with exploring and using the dataset, we provide gzipped files that list:

By simply adding either s3://commoncrawl/ or https://data.commoncrawl.org/ to each line, you end up with the S3 and HTTP paths respectively.

The Common Crawl URL Index for this crawl is available at: https://index.commoncrawl.org/CC-MAIN-2016-50/. Note that the index now contains robots.txt responses and URLs that have not been successfully fetched (i.e. 404's and pages with HTTP status codes other than 200). For more information on working with the URL index, please refer to the previous blog post or the Index Server API. There is also a command-line tool client for common use cases of the URL index. Please donate to Common Crawl if you appreciate our free datasets! We’re also seeking corporate sponsors to partner with Common Crawl for our non-profit work in open data. Please contact info@commoncrawl.org for sponsorship information and packages.

This release was authored by:
No items found.

Erratum: 

Erroneous title field in WAT records

Originally reported by: 
Robert Waksmunski
Permalink

The "Title" extracted in WAT records to the JSON path `Envelope > Payload-Metadata > HTTP-Response-Metadata > HTML-Metadata > Head > Title` is not the content included in the <title> element in the HTML header (<head> element) if the page contains further <title> elements in the page body. The content of the last <title> element is written to the WAT "Title". This bug was observed if the HTML page includes embedded SVG graphics.

The issue was reported by the user Robert Waksmunski:

...and was fixed for CC-MAIN-2024-42 by commoncrawl/ia-web-commons#37.

This erratum affects all crawls from CC-MAIN-2013-20 until CC-MAIN-2024-38.

Erratum: 

Incorrect fetch_time metadata

Originally reported by: 
Permalink

In crawls CC-MAIN-2016-36 to CC-MAIN-2016-50, the fetch_time metadata for robots.txt is incorrect. The correct times are as follows:

earliest_fetch_timelatest_fetch_timecrawl
2016-08-23 20:56:23.0002016-09-01 07:28:38.000CC-MAIN-2016-36
2016-09-24 20:47:41.0002016-10-02 00:13:21.000CC-MAIN-2016-40
2016-10-20 19:25:54.0002016-10-29 01:20:47.000CC-MAIN-2016-44
2016-12-02 17:51:44.0002016-12-11 15:40:44.000CC-MAIN-2016-50

Erratum: 

Missing Language Classification

Originally reported by: 
Permalink

Starting with crawl CC-MAIN-2018-39 we added a language classification field (‘content-languages’) to the columnar indexes, WAT files, and WARC metadata for all subsequent crawls. The CLD2 classifier was used, and includes up to three languages per document. We use the ISO-639-3 (three-character) language codes.