Kaydol

Flood göndermek, insanların floodlarını okumak ve diğer insanlarla bağlantı kurmak için sosyal Floodlar ve Flood Yanıtları Motorumuza kaydolun.

Oturum aç

Flood göndermek, insanların floodlarını okumak ve diğer insanlarla bağlantı kurmak için sosyal Floodlar ve Flood Yanıtları Motorumuza giriş yapın.

Şifremi hatırlamıyorum

Şifreni mi unuttun? Lütfen e-mail adresinizi giriniz. Bir bağlantı alacaksınız ve e-posta yoluyla yeni bir şifre oluşturacaksınız.

3 ve kadim dostu 1 olan sj'yi rakamla giriniz. ( 31 )

Üzgünüz, Flood yazma yetkiniz yok, Flood girmek için giriş yapmalısınız.

Lütfen bu Floodun neden bildirilmesi gerektiğini düşündüğünüzü kısaca açıklayın.

Lütfen bu cevabın neden bildirilmesi gerektiğini kısaca açıklayın.

Please briefly explain why you feel this user should be reported.

600 GB corpus of all paywalled scholarly sources of Wikipedia

[Scholarly publishing is a racket](https://www.monbiot.com/2018/09/17/honourable-theft/) where research entities pay all the costs for the outputs but publishers seize them for free and get the same research entities to pay exorbitant prices to rent them back. In an open publication like Wikipedia, [two thirds](https://wikimediafoundation.org/2018/08/20/how-many-wikipedia-references-are-available-to-read/) of the academic sources are paywalled, although many are green open access (archived by open repositories).

A torrent has been compiled from various sources to provide a corpus with nearly all such paywalled papers for Wikipedia in all languages. It contains about 561k PDFs for as many [DOIs](https://en.wikipedia.org/wiki/Digital_object_identifier), divided in 1529 ZIP files for a total of 600 GB. The archives are sorted by publisher (DOI prefix), so you can download just what you can or are most interested in. There’s a list of DOIs inside.

The corpus is only for personal and research use, for instance to help [text and data mining](https://libereurope.eu/text-data-mining/) and any other effort which requires fast or bulk download.

Please preserve a copy and seed! The info_hash and magnet link are:

>f968e22840df243befc6dd123e7d37613002fca8
>
>magnet:?xt=urn:btih:f968e22840df243befc6dd123e7d37613002fca8&tr=udp%3A%2F%2Ftracker.coppersurfer.tk%3A6969%2Fannounce

Benzer Yazılar

Yorum eklemek için giriş yapmalısınız.

28 Yorumları

  1. Anyone have an up to date magnet?

    *Asking for a friend*

  2. It seems Springer has been doing some homework: [https://www.lumendatabase.org/notices/17452283](https://www.lumendatabase.org/notices/17452283#)

  3. Been hosting this for 20 days, here’s the stats on the seedbox:

    Downloaded:
    558.6 GiB (558.6 GiB)
    Uploaded:
    3179.7 GiB (3179.7 GiB)
    Share Ratio:
    5.692
    Next Announce:
    26m 12s
    Tracker Status:
    coppersurfer.tk: Announce OK

    Speed:
    0.0 KiB/s
    Speed:
    284.9 KiB/s
    ETA:

    Pieces:
    142998 (4.0 MiB)

    Seeders:
    0 (58)
    Peers:
    6 (16)
    Availability:
    0.000
    Auto Managed:
    True

    Active Time:
    20d 10h
    Seeding Time:
    19d 14h
    Seed Rank:
    293
    Date Added:
    24/09/2018 07:49:20

  4. Just added it to my seed box, will seed for as long as I can.

  5. Ratio of 8 so far! Lots of people grabbing it, long may it seed.

  6. Started this morning, just finished ~18h later, will seed for as long as I can spare the b/w.

    Thanks to all those involved in putting this together. If only grabbing all of Sci-Mag/Libgen was this straightforward….

  7. Does anybody have a list of paywalled academic sources cited on Wikipedia which don’t appear in this torrent? I’d love to help put a dent in it if possible.

  8. I used to have close to 4tb of textbooks and scientific journals but the hard drive went kaboom. There was a 500gb or so torrent called “chemical journals” and a 1.3tb, I want to say newsgroup, that was all scientific journals. I’d like to rebuild all that I can. Anybody here either of these or something else they’d be willing to share or point in the right direction? That chemical journals torrent is still up but without seeds and I can no longer find the massive newsgroup.

  9. Just started downloading. Getting around 60MB/s down and uploading at 40KB/s (I have no idea what’s going on there, it’s a symmetrical internet connection).

  10. Hey, how do can i check the content of these ZIP files? this isn’t written in these text files which are in the bundle

  11. Is there a guide in there on how to free up 600GB?

  12. magnet:?xt=urn:btih:f968e22840df243befc6dd123e7d37613002fca8&tr=udp%3A%2F%2Ftracker.coppersurfer.tk%3A6969%2Fannounce gives error

    .

    f968e22840df243befc6dd123e7d37613002fca8 does not start

    .

    magnet:?xt=urn:btih:f968e22840df243befc6dd123e7d37613002fca8&tr=udp://tracker.coppersurfer.tk:6969/announce also nothing

  13. Downloading now. getting around 26MB/sec, uploading at 17MB/sec so far.. only 2% complete so far!

  14. I don’t mean to ruin the fun, but isn’t it technically illegal to do this? At least seeding it would be distribution of stolen, copyrighted material, right?

    Don’t get me wrong, I’m all for open and free research. But I don’t necessarily want to go to jail or pay huge fines for it…

    Does anyone have information on this?

  15. Very nice, I can expand my library again 🙂

    grabbing it, currently only 700 kiB/s but I’ll let it run!

  16. I’ll definitely be downloading this. It’s a good start to mirroring the vast library that sci-hub has.

    Any downloads for studies not cited by wikipedia? Does sci-hub share their set?

  17. I’m assuming that this has already been submitted to [Sci-Hub](https://whereisscihub.now.sh/go)/LibGen?

  18. I’m a bot, *bleep*, *bloop*. Someone has linked to this thread from another place on reddit:

    – [/r/datasets] [600 GB corpus of all paywalled scholarly sources of Wikipedia](https://www.reddit.com/r/datasets/comments/9ih486/600_gb_corpus_of_all_paywalled_scholarly_sources/)

    – [/r/datasets] [600 GB corpus of all paywalled scholarly sources of Wikipedia](https://www.reddit.com/r/datasets/comments/9iii3m/600_gb_corpus_of_all_paywalled_scholarly_sources/)

    – [/r/machineslearn] [600 GB corpus of all paywalled scholarly sources of Wikipedia](https://www.reddit.com/r/MachinesLearn/comments/9ihvdz/600_gb_corpus_of_all_paywalled_scholarly_sources/)

     *^(If you follow any of the above links, please respect the rules of reddit and don’t vote in the other threads.) ^([Info](/r/TotesMessenger) ^/ ^[Contact](/message/compose?to=/r/TotesMessenger))*

  19. Maybe a stupid question, but I’m using rtorrent on a host where I don’t really have access to upgrade it. As far as I know it doesn’t support magnet files. Any way to get a torrent file of this? It’s got like 40 TB of storage so I can seed indefinitely.

  20. Is there a text file with what’s in the file?

  21. I’m trying to grab all of it and seed but my download speed is tragic

  22. I wish I had the free space. That’s a gold mine!

    Although if I want a paper, I just mail the author. 90% of them, they’ll send it to you free of charge!