Indian Flag
Government Of India
A-
A
A+
Common Crawl

Common Crawl

Provides petabytes of web page data collected over time, offering a vast resource for training language models on diverse internet text.

About Dataset

Common Crawl is a large-scale open web archive that provides petabytes of web page data collected through regular web crawls. The dataset includes raw HTML pages, extracted text, metadata, and link graphs from billions of publicly accessible web pages. Maintained by the Common Crawl Foundation, it is one of the largest openly available sources of web text. The data spans diverse domains, languages, and content types, reflecting the structure and diversity of the public internet over time.

Purpose of Dataset

Common Crawl Is Widely Used For Pretraining Large Language Models And Conducting Large-scale Natural Language Processing Research. Its Size And Diversity Make It Suitable For Learning General Language Patterns, World Knowledge, And Multilingual Representations. Researchers Also Use It For Web Mining, Information Retrieval, And Linguistic Analysis. Due To Its Scale, It Plays A Foundational Role In Building General-purpose Llms And Studying The Impact Of Web-scale Data On Model Performance.

Activity Overview Activity Overview

  • Downloads0
  • Redirect 1
  • Views 9
  • File Size 0

Tags Tags

  • Multilingual
  • LLM Pretraining
  • Web Scale Data
  • Raw Text

License Control License Control

Other