a, to Futurology
@a@paperbay.org avatar

Common Crawl September/October 2023 Crawl Archive (CC-MAIN-2023-40) is out and release.

100TiB compressed of fresh web crawled which can used in your next data mining project.

🔗 https://data.commoncrawl.org/crawl-data/CC-MAIN-2023-40/index.html

ButterflyOfFire, to random French
@ButterflyOfFire@mstdn.fr avatar

Est-ce que cette page fonctionne chez-vous ? 👀

http://urlsearch.commoncrawl.org/

#CommonCrawl

pluralistic, to random
@pluralistic@mamot.fr avatar

The crybabies who freak out about The Communist Manifesto appearing on university curriculum clearly never read it - chapter one is basically a long hymn to capitalism's flexibility and inventiveness, its ability to change form and adapt itself to everything the world throws at it and come out on top:

https://www.marxists.org/archive/marx/works/1848/communist-manifesto/ch01.htm#007

1/

pluralistic,
@pluralistic@mamot.fr avatar

Many of the biggest "open AI" companies are totally opaque when it comes to training data. Google and OpenAI won't even say how many pieces of data went into their models' training - let alone which data they used.

Other "open AI" companies use publicly available datasets like #ThePile and #CommonCrawl. But you can't replicate their models by shoveling these datasets into an algorithm. Each one has to be groomed - labeled, sorted, de-duplicated, and otherwise filtered.

28/

tallison, to infosec
@tallison@mastodon.social avatar

I've gotten a bunch of followers over the last coupla days.

For those interested in and especially PDFs, please take a look at our fairly newly released 8 million/8TB PDF corpus, derived from and then augmented by our team at

https://digitalcorpora.org/corpora/file-corpora/cc-main-2021-31-pdf-untruncated/

tallison, to random
@tallison@mastodon.social avatar
AmyDentata, to random

deleted_by_author

  • Loading...
  • bornach,
    @bornach@masto.ai avatar

    @morganmay @AmyDentata
    Well there is a thought process but it is unfortunately a very human one that is

    • contaminated by the biases perpetuated in subreddits hoovered up by ,
    • cleaned, labelled and fine tuned by low paid offshore workers in the gig economy,
    • edited by the user's own vulnerability to simple confidence tricks perpetrated by pushers
    willoremus, to random
    @willoremus@mastodon.social avatar

    This visual deep dive into one of the largest AI language datasets is nonstop fascinating, jaw-dropping, and troubling, and anyone who is remotely interested in how LLMs really work, their biases, or intellectual property should read it. https://www.washingtonpost.com/technology/interactive/2023/ai-chatbot-learning/

    tallison,
    @tallison@mastodon.social avatar

    @willoremus I ❤️ that Google uses #CommonCrawl and thereby the fruits of #ApacheTika and #ApacheNutch.

  • All
  • Subscribed
  • Moderated
  • Favorites
  • provamag3
  • InstantRegret
  • magazineikmin
  • modclub
  • khanakhh
  • Youngstown
  • rosin
  • mdbf
  • slotface
  • Durango
  • ngwrru68w68
  • thenastyranch
  • kavyap
  • DreamBathrooms
  • JUstTest
  • cubers
  • osvaldo12
  • Leos
  • anitta
  • everett
  • ethstaker
  • GTA5RPClips
  • tester
  • cisconetworking
  • megavids
  • tacticalgear
  • normalnudes
  • lostlight
  • All magazines