| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | """PMC Open Access Subset.""" |
| |
|
| | import datetime |
| | from functools import lru_cache |
| |
|
| | import fsspec |
| | import pandas as pd |
| |
|
| | import datasets |
| |
|
| |
|
| | _CITATION = """\ |
| | PMC Open Access Subset [Internet]. Bethesda (MD): National Library of Medicine. 2003 - [cited YEAR MONTH DAY]. Available from https://www.ncbi.nlm.nih.gov/pmc/tools/openftlist/ |
| | """ |
| |
|
| | _DESCRIPTION = """\ |
| | The PMC Open Access Subset includes more than 3.4 million journal articles and preprints that are made available under |
| | license terms that allow reuse. |
| | |
| | Not all articles in PMC are available for text mining and other reuse, many have copyright protection, however articles |
| | in the PMC Open Access Subset are made available under Creative Commons or similar licenses that generally allow more |
| | liberal redistribution and reuse than a traditional copyrighted work. |
| | |
| | The PMC Open Access Subset is one part of the PMC Article Datasets |
| | """ |
| |
|
| | _HOMEPAGE = "https://www.ncbi.nlm.nih.gov/pmc/tools/openftlist/" |
| |
|
| | _LICENSE = """\ |
| | Within the PMC Open Access Subset, there are three groupings based on available license terms: |
| | - Commercial Use Allowed - CC0, CC BY, CC BY-SA, CC BY-ND licenses; |
| | - Non-Commercial Use Only - CC BY-NC, CC BY-NC-SA, CC BY-NC-ND licenses; and |
| | - Other - no machine-readable Creative Commons license, no license, or a custom license. |
| | """ |
| |
|
| | _URL = "https://ftp.ncbi.nlm.nih.gov/pub/pmc/oa_bulk/{subset}/txt/" |
| | _SUBSETS = { |
| | "commercial": "oa_comm", |
| | "non_commercial": "oa_noncomm", |
| | "other": "oa_other", |
| | } |
| |
|
| |
|
| | @lru_cache(maxsize=None) |
| | def request_data_urls(): |
| | fs = fsspec.filesystem("https") |
| | result = {} |
| | for subset, subset_url in _SUBSETS.items(): |
| | urls = fs.ls(_URL.format(subset=subset_url), detail=False) |
| | baseline_urls = [ |
| | url for url in urls for filename in url.split("/")[-1:] if filename.startswith(f"{subset_url}_txt.PMC") |
| | ] |
| | baseline_date = parse_date(baseline_urls[0]) |
| | baseline_file_list_urls = [url for url in baseline_urls if url.endswith(".csv")] |
| | baseline_archive_urls = [url for url in baseline_urls if url.endswith(".tar.gz")] |
| | incremental_urls = [ |
| | url for url in urls for filename in url.split("/")[-1:] if filename.startswith(f"{subset_url}_txt.incr.") |
| | ] |
| | incremental_file_list_urls = [url for url in incremental_urls if url.endswith(".csv")] |
| | incremental_archive_urls = [url for url in incremental_urls if url.endswith(".tar.gz")] |
| | result["baseline_date"] = baseline_date |
| | result[subset] = { |
| | "baseline_urls": list(zip(baseline_file_list_urls, baseline_archive_urls)), |
| | "incremental_urls": list(zip(incremental_file_list_urls, incremental_archive_urls)), |
| | } |
| | return result |
| |
|
| |
|
| | def parse_date(url): |
| | return url.split("/")[-1].split(".")[-3] |
| |
|
| |
|
| | class OpenAccessConfig(datasets.BuilderConfig): |
| | """BuilderConfig for the PMC Open Access Subset.""" |
| |
|
| | def __init__(self, date=None, subsets="all", **kwargs): |
| | """BuilderConfig for the PMC Open Access Subset. |
| | |
| | Args: |
| | date (`str`, default BASELINE_DATE) : Up to date, in ISO format. Pass 'latest' for latest date. |
| | subsets (`str` or `list[str]`, default 'all'): List of subsets to load. Possible values are 'all' or any combination |
| | of {'commercial', 'non_commercial', 'other'}. |
| | **kwargs: Keyword arguments forwarded to `BuilderConfig`. |
| | """ |
| | if date is None: |
| | date = request_data_urls()["baseline_date"] |
| | date = datetime.date.today().isoformat() if date == "latest" else date |
| | subsets = [subsets] if isinstance(subsets, str) else subsets |
| | subsets_name = "+".join(subsets) |
| | name = f"{date}.{subsets_name}" |
| | super().__init__(name=name, **kwargs) |
| | self.subsets = subsets if subsets_name != "all" else list(_SUBSETS.keys()) |
| | self.date = date |
| |
|
| |
|
| | class OpenAccess(datasets.GeneratorBasedBuilder): |
| | """PMC Open Access Subset.""" |
| |
|
| | VERSION = datasets.Version("1.0.0") |
| | BUILDER_CONFIG_CLASS = OpenAccessConfig |
| | BUILDER_CONFIGS = [OpenAccessConfig(subsets="all")] + [OpenAccessConfig(subsets=subset) for subset in _SUBSETS] |
| | DEFAULT_CONFIG_NAME = f"{request_data_urls()['baseline_date']}.all" |
| |
|
| | def _info(self): |
| | return datasets.DatasetInfo( |
| | description=_DESCRIPTION, |
| | features=datasets.Features( |
| | { |
| | "text": datasets.Value("string"), |
| | "pmid": datasets.Value("string"), |
| | "accession_id": datasets.Value("string"), |
| | "license": datasets.Value("string"), |
| | "last_updated": datasets.Value("string"), |
| | "retracted": datasets.Value("string"), |
| | "citation": datasets.Value("string"), |
| | } |
| | ), |
| | homepage=_HOMEPAGE, |
| | license=_LICENSE, |
| | citation=_CITATION, |
| | ) |
| |
|
| | def _split_generators(self, dl_manager): |
| | urls = request_data_urls() |
| | date = datetime.date.fromisoformat(self.config.date) |
| | paths = [] |
| | for subset in self.config.subsets: |
| | |
| | baseline_urls = urls[subset]["baseline_urls"] |
| | |
| | incremental_urls = [ |
| | url_pair |
| | for url_pair in urls[subset]["incremental_urls"] |
| | if datetime.date.fromisoformat(parse_date(url_pair[0])) <= date |
| | ] |
| | paths += dl_manager.download(baseline_urls + incremental_urls) |
| |
|
| | return [ |
| | datasets.SplitGenerator( |
| | name=datasets.Split.TRAIN, |
| | gen_kwargs={ |
| | "paths": [(file_list, dl_manager.iter_archive(archive)) for file_list, archive in paths], |
| | }, |
| | ), |
| | ] |
| |
|
| | def _generate_examples(self, paths): |
| | key = 0 |
| | for file_list, archive in paths: |
| | file_list_data = pd.read_csv(file_list, index_col="Article File").to_dict(orient="index") |
| | for path, file in archive: |
| | data = file_list_data.pop(path) |
| | content = file.read() |
| | try: |
| | text = content.decode("utf-8").strip() |
| | except UnicodeDecodeError as e: |
| | text = content.decode("latin-1").strip() |
| | data = { |
| | "text": text, |
| | "pmid": data["PMID"], |
| | "accession_id": data["AccessionID"], |
| | "license": data["License"], |
| | "last_updated": data["LastUpdated (YYYY-MM-DD HH:MM:SS)"], |
| | "retracted": data["Retracted"], |
| | "citation": data["Article Citation"], |
| | } |
| | yield key, data |
| | key += 1 |
| |
|