| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | import os |
| |
|
| | import datasets |
| | import numpy as np |
| | import pandas as pd |
| |
|
| | from .bigbiohub import text_features |
| | from .bigbiohub import kb_features |
| | from .bigbiohub import BigBioConfig |
| | from .bigbiohub import Tasks |
| |
|
| | _LANGUAGES = ['French'] |
| | _PUBMED = False |
| | _LOCAL = True |
| | _CITATION = """\ |
| | @inproceedings{grabar-etal-2018-cas, |
| | title = {{CAS}: {F}rench Corpus with Clinical Cases}, |
| | author = {Grabar, Natalia and Claveau, Vincent and Dalloux, Cl{\'e}ment}, |
| | year = 2018, |
| | month = oct, |
| | booktitle = { |
| | Proceedings of the Ninth International Workshop on Health Text Mining and |
| | Information Analysis |
| | }, |
| | publisher = {Association for Computational Linguistics}, |
| | address = {Brussels, Belgium}, |
| | pages = {122--128}, |
| | doi = {10.18653/v1/W18-5614}, |
| | url = {https://aclanthology.org/W18-5614}, |
| | abstract = { |
| | Textual corpora are extremely important for various NLP applications as |
| | they provide information necessary for creating, setting and testing these |
| | applications and the corresponding tools. They are also crucial for |
| | designing reliable methods and reproducible results. Yet, in some areas, |
| | such as the medical area, due to confidentiality or to ethical reasons, it |
| | is complicated and even impossible to access textual data representative of |
| | those produced in these areas. We propose the CAS corpus built with |
| | clinical cases, such as they are reported in the published scientific |
| | literature in French. We describe this corpus, currently containing over |
| | 397,000 word occurrences, and the existing linguistic and semantic |
| | annotations. |
| | } |
| | }""" |
| |
|
| | _DATASETNAME = "cas" |
| | _DISPLAYNAME = "CAS" |
| |
|
| | _DESCRIPTION = """\ |
| | We manually annotated two corpora from the biomedical field. The ESSAI corpus \ |
| | contains clinical trial protocols in French. They were mainly obtained from the \ |
| | National Cancer Institute The typical protocol consists of two parts: the \ |
| | summary of the trial, which indicates the purpose of the trial and the methods \ |
| | applied; and a detailed description of the trial with the inclusion and \ |
| | exclusion criteria. The CAS corpus contains clinical cases published in \ |
| | scientific literature and training material. They are published in different \ |
| | journals from French-speaking countries (France, Belgium, Switzerland, Canada, \ |
| | African countries, tropical countries) and are related to various medical \ |
| | specialties (cardiology, urology, oncology, obstetrics, pulmonology, \ |
| | gastro-enterology). The purpose of clinical cases is to describe clinical \ |
| | situations of patients. Hence, their content is close to the content of clinical \ |
| | narratives (description of diagnoses, treatments or procedures, evolution, \ |
| | family history, expected audience, etc.). In clinical cases, the negation is \ |
| | frequently used for describing the patient signs, symptoms, and diagnosis. \ |
| | Speculation is present as well but less frequently. |
| | |
| | This version only contain the annotated CAS corpus |
| | """ |
| |
|
| | _HOMEPAGE = "https://clementdalloux.fr/?page_id=28" |
| |
|
| | _LICENSE = 'Data User Agreement' |
| |
|
| | _URLS = { |
| | "cas_source": "", |
| | "cas_bigbio_text": "", |
| | "cas_bigbio_kb": "", |
| | } |
| |
|
| | _SOURCE_VERSION = "1.0.0" |
| | _BIGBIO_VERSION = "1.0.0" |
| |
|
| | _SUPPORTED_TASKS = [Tasks.TEXT_CLASSIFICATION] |
| |
|
| |
|
| | class CAS(datasets.GeneratorBasedBuilder): |
| | SOURCE_VERSION = datasets.Version(_SOURCE_VERSION) |
| | BIGBIO_VERSION = datasets.Version(_BIGBIO_VERSION) |
| |
|
| | DEFAULT_CONFIG_NAME = "cas_source" |
| |
|
| | BUILDER_CONFIGS = [ |
| | BigBioConfig( |
| | name="cas_source", |
| | version=SOURCE_VERSION, |
| | description="CAS source schema", |
| | schema="source", |
| | subset_id="cas", |
| | ), |
| | BigBioConfig( |
| | name="cas_bigbio_text", |
| | version=BIGBIO_VERSION, |
| | description="CAS simplified BigBio schema for negation/speculation classification", |
| | schema="bigbio_text", |
| | subset_id="cas", |
| | ), |
| | BigBioConfig( |
| | name="cas_bigbio_kb", |
| | version=BIGBIO_VERSION, |
| | description="CAS simplified BigBio schema for part-of-speech-tagging", |
| | schema="bigbio_kb", |
| | subset_id="cas", |
| | ), |
| | ] |
| |
|
| | def _info(self): |
| | if self.config.schema == "source": |
| | features = datasets.Features( |
| | { |
| | "id": datasets.Value("string"), |
| | "document_id": datasets.Value("string"), |
| | "text": [datasets.Value("string")], |
| | "lemmas": [datasets.Value("string")], |
| | "POS_tags": [datasets.Value("string")], |
| | "labels": [datasets.Value("string")], |
| | } |
| | ) |
| | elif self.config.schema == "bigbio_text": |
| | features = text_features |
| | elif self.config.schema == "bigbio_kb": |
| | features = kb_features |
| |
|
| | return datasets.DatasetInfo( |
| | description=_DESCRIPTION, |
| | features=features, |
| | supervised_keys=None, |
| | homepage=_HOMEPAGE, |
| | license=str(_LICENSE), |
| | citation=_CITATION, |
| | ) |
| |
|
| | def _split_generators(self, dl_manager): |
| | if self.config.data_dir is None: |
| | raise ValueError( |
| | "This is a local dataset. Please pass the data_dir kwarg to load_dataset." |
| | ) |
| | else: |
| | data_dir = self.config.data_dir |
| | return [ |
| | datasets.SplitGenerator( |
| | name=datasets.Split.TRAIN, |
| | gen_kwargs={"datadir": data_dir}, |
| | ), |
| | ] |
| |
|
| | def _generate_examples(self, datadir): |
| | key = 0 |
| | for file in ["CAS_neg.txt", "CAS_spec.txt"]: |
| | filepath = os.path.join(datadir, file) |
| | label = "negation" if "neg" in file else "speculation" |
| | id_docs = [] |
| | id_words = [] |
| | words = [] |
| | lemmas = [] |
| | POS_tags = [] |
| |
|
| | with open(filepath) as f: |
| | for line in f.readlines(): |
| | line_content = line.split("\t") |
| | if len(line_content) > 1: |
| | id_docs.append(line_content[0]) |
| | id_words.append(line_content[1]) |
| | words.append(line_content[2]) |
| | lemmas.append(line_content[3]) |
| | POS_tags.append(line_content[4]) |
| |
|
| | dic = { |
| | "id_docs": np.array(list(map(int, id_docs))), |
| | "id_words": id_words, |
| | "words": words, |
| | "lemmas": lemmas, |
| | "POS_tags": POS_tags, |
| | } |
| | if self.config.schema == "source": |
| | for doc_id in set(dic["id_docs"]): |
| | idces = np.argwhere(dic["id_docs"] == doc_id)[:, 0] |
| | text = [dic["words"][id] for id in idces] |
| | text_lemmas = [dic["lemmas"][id] for id in idces] |
| | POS_tags_ = [dic["POS_tags"][id] for id in idces] |
| | yield key, { |
| | "id": key, |
| | "document_id": doc_id, |
| | "text": text, |
| | "lemmas": text_lemmas, |
| | "POS_tags": POS_tags_, |
| | "labels": [label], |
| | } |
| | key += 1 |
| | elif self.config.schema == "bigbio_text": |
| | for doc_id in set(dic["id_docs"]): |
| | idces = np.argwhere(dic["id_docs"] == doc_id)[:, 0] |
| | text = " ".join([dic["words"][id] for id in idces]) |
| | yield key, { |
| | "id": key, |
| | "document_id": doc_id, |
| | "text": text, |
| | "labels": [label], |
| | } |
| | key += 1 |
| | elif self.config.schema == "bigbio_kb": |
| | for doc_id in set(dic["id_docs"]): |
| | idces = np.argwhere(dic["id_docs"] == doc_id)[:, 0] |
| | text = [dic["words"][id] for id in idces] |
| | POS_tags_ = [dic["POS_tags"][id] for id in idces] |
| |
|
| | data = { |
| | "id": str(key), |
| | "document_id": doc_id, |
| | "passages": [], |
| | "entities": [], |
| | "relations": [], |
| | "events": [], |
| | "coreferences": [], |
| | } |
| | key += 1 |
| |
|
| | data["passages"] = [ |
| | { |
| | "id": str(key + i), |
| | "type": "sentence", |
| | "text": [text[i]], |
| | "offsets": [[i, i + 1]], |
| | } |
| | for i in range(len(text)) |
| | ] |
| | key += len(text) |
| |
|
| | for i in range(len(text)): |
| | entity = { |
| | "id": key, |
| | "type": "POS_tag", |
| | "text": [POS_tags_[i]], |
| | "offsets": [[i, i + 1]], |
| | "normalized": [], |
| | } |
| | data["entities"].append(entity) |
| | key += 1 |
| |
|
| | yield key, data |
| |
|