Dataset Viewer
Full Screen Viewer
Full Screen
The dataset viewer is not available for this dataset.
Cannot get the config names for the dataset.
Error code: ConfigNamesError Exception: ImportError Message: To be able to use thbndi/Mimic4Dataset, you need to install the following dependency: git. Please install it using 'pip install git' for instance. Traceback: Traceback (most recent call last): File "/src/services/worker/src/worker/job_runners/dataset/config_names.py", line 66, in compute_config_names_response config_names = get_dataset_config_names( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/inspect.py", line 347, in get_dataset_config_names dataset_module = dataset_module_factory( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/load.py", line 1914, in dataset_module_factory raise e1 from None File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/load.py", line 1880, in dataset_module_factory return HubDatasetModuleFactoryWithScript( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/load.py", line 1504, in get_module local_imports = _download_additional_modules( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/load.py", line 354, in _download_additional_modules raise ImportError( ImportError: To be able to use thbndi/Mimic4Dataset, you need to install the following dependency: git. Please install it using 'pip install git' for instance.
Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
Dataset Usage
Description
The Mimic-IV dataset generate data by executing the Pipeline available on https://github.com/healthylaife/MIMIC-IV-Data-Pipeline.
Function Signature
load_dataset('thbndi/Mimic4Dataset', task, mimic_path=mimic_data, config_path=config_file, encoding=encod, generate_cohort=gen_cohort, val_size=size, cache_dir=cache)
Arguments
task
(string) :- Description: Specifies the task you want to perform with the dataset.
- Default: "Mortality"
- Note: Possible Values : 'Phenotype', 'Length of Stay', 'Readmission', 'Mortality'
mimic_path
(string) :- Description: Complete path to the Mimic-IV raw data on user's machine.
- Note: You need to provide the appropriate path where the Mimic-IV data is stored. The path should end with the version of mimic (eg : mimiciv/2.2). Supported version : 2.2 and 1.0 as provided by the authors of the pipeline.
config_path
(string) optionnal :- Description: Path to the configuration file for the cohort generation choices (more infos in '/config/readme.md').
- Default: Configuration file provided in the 'config' folder.
encoding
(string) optionnal :- Description: Data encoding option for the features.
- Options: "concat", "aggreg", "tensor", "raw", "text"
- Default: "concat"
- Note: Choose one of the following options for data encoding:
- "concat": Concatenates the one-hot encoded diagnoses, demographic data vector, and dynamic features at each measured time instant, resulting in a high-dimensional feature vector.
- "aggreg": Concatenates the one-hot encoded diagnoses, demographic data vector, and dynamic features, where each item_id is replaced by the average of the measured time instants, resulting in a reduced-dimensional feature vector.
- "tensor": Represents each feature as an 2D array. There are separate arrays for labels, demographic data ('DEMO'), diagnosis ('COND'), medications ('MEDS'), procedures ('PROC'), chart/lab events ('CHART/LAB'), and output events data ('OUT'). Dynamic features are represented as 2D arrays where each row contains values at a specific time instant.
- "raw": Provide cohort from the pipeline without any encoding for custom data processing.
- "text": Represents diagnoses as text suitable for BERT or other similar text-based models.
- For 'concat' and 'aggreg' the composition of the vector is given in './data/dict/"task"/features_aggreg.csv' or './data/dict/"task"/features_concat.csv' file and in 'features_names' column of the dataset.
generate_cohort
(bool) optionnal :- Description: Determines whether to generate a new cohort from Mimic-IV data.
- Default: True
- Note: Set it to True to generate a cohort, or False to skip cohort generation.
val_size
, 'test_size' (float) optionnal :- Description: Proportion of the dataset used for validation during training.
- Default: 0.1 for validation size and 0.2 for testing size.
- Note: Can be set to 0.
cache_dir
(string) optionnal :- Description: Directory where the processed dataset will be cached.
- Note: Providing a cache directory for each encoding type can avoid errors when changing the encoding type.
Example Usage
import datasets
from datasets import load_dataset
# Example 1: Load dataset with default settings
dataset = load_dataset('thbndi/Mimic4Dataset', task="Mortality", mimic_path="/path/to/mimic_data")
# Example 2: Load dataset with custom settings
dataset = load_dataset('thbndi/Mimic4Dataset', task="Phenotype", mimic_path="/path/to/mimic_data", config_path="/path/to/config_file", encoding="aggreg", generate_cohort=False, val_size=0.2, cache_dir="/path/to/cache_dir")
Please note that the provided examples are for illustrative purposes only, and you should adjust the paths and settings based on your actual dataset and specific use case.
- Downloads last month
- 379