|
import numpy as np |
|
import pandas as pd |
|
from tqdm import tqdm |
|
from datetime import datetime |
|
from sklearn.preprocessing import LabelEncoder |
|
import pickle |
|
import datetime |
|
import os |
|
import sys |
|
from pathlib import Path |
|
sys.path.append(os.path.dirname(os.path.abspath(__file__)) + './../..') |
|
if not os.path.exists("./data/dict"): |
|
os.makedirs("./data/dict") |
|
if not os.path.exists("./data/csv"): |
|
os.makedirs("./data/csv") |
|
|
|
class Generator(): |
|
def __init__(self,task,cohort_output,if_mort,if_admn,if_los,feat_cond,feat_proc,feat_out,feat_chart,feat_med,impute,include_time=24,bucket=1,predW=6): |
|
self.feat_cond,self.feat_proc,self.feat_out,self.feat_chart,self.feat_med = feat_cond,feat_proc,feat_out,feat_chart,feat_med |
|
self.cohort_output=cohort_output |
|
self.impute=impute |
|
self.task = task |
|
self.data = self.generate_adm() |
|
if not os.path.exists("./data/dict/"+self.task): |
|
os.makedirs("./data/dict/"+self.task) |
|
print("[ READ COHORT ]") |
|
|
|
self.generate_feat() |
|
print("[ READ ALL FEATURES ]") |
|
|
|
if if_mort: |
|
self.mortality_length(include_time,predW) |
|
print("[ PROCESSED TIME SERIES TO EQUAL LENGTH ]") |
|
elif if_admn: |
|
self.readmission_length(include_time) |
|
print("[ PROCESSED TIME SERIES TO EQUAL LENGTH ]") |
|
elif if_los: |
|
self.los_length(include_time) |
|
print("[ PROCESSED TIME SERIES TO EQUAL LENGTH ]") |
|
|
|
self.smooth_meds(bucket) |
|
print("[ SUCCESSFULLY SAVED DATA DICTIONARIES ]") |
|
|
|
def generate_feat(self): |
|
if(self.feat_cond): |
|
print("[ ======READING DIAGNOSIS ]") |
|
self.generate_cond() |
|
if(self.feat_proc): |
|
print("[ ======READING PROCEDURES ]") |
|
self.generate_proc() |
|
if(self.feat_out): |
|
print("[ ======READING OUT EVENTS ]") |
|
self.generate_out() |
|
if(self.feat_chart): |
|
print("[ ======READING CHART EVENTS ]") |
|
self.generate_chart() |
|
if(self.feat_med): |
|
print("[ ======READING MEDICATIONS ]") |
|
self.generate_meds() |
|
|
|
def generate_adm(self): |
|
data=pd.read_csv(f"./data/cohort/{self.cohort_output}.csv.gz", compression='gzip', header=0, index_col=None) |
|
data['intime'] = pd.to_datetime(data['intime']) |
|
data['outtime'] = pd.to_datetime(data['outtime']) |
|
data['los']=pd.to_timedelta(data['outtime']-data['intime'],unit='h') |
|
data['los']=data['los'].astype(str) |
|
data[['days', 'dummy','hours']] = data['los'].str.split(' ', expand=True) |
|
data[['hours','min','sec']] = data['hours'].str.split(':', expand=True) |
|
data['los']=pd.to_numeric(data['days'])*24+pd.to_numeric(data['hours']) |
|
data=data.drop(columns=['days', 'dummy','hours','min','sec']) |
|
data=data[data['los']>0] |
|
data['Age']=data['Age'].astype(int) |
|
|
|
|
|
return data |
|
|
|
def generate_cond(self): |
|
cond=pd.read_csv("./data/features/preproc_diag_icu.csv.gz", compression='gzip', header=0, index_col=None) |
|
cond=cond[cond['stay_id'].isin(self.data['stay_id'])] |
|
cond_per_adm = cond.groupby('stay_id').size().max() |
|
self.cond, self.cond_per_adm = cond, cond_per_adm |
|
|
|
def generate_proc(self): |
|
proc=pd.read_csv("./data/features/preproc_proc_icu.csv.gz", compression='gzip', header=0, index_col=None) |
|
proc=proc[proc['stay_id'].isin(self.data['stay_id'])] |
|
proc[['start_days', 'dummy','start_hours']] = proc['event_time_from_admit'].str.split(' ', -1, expand=True) |
|
proc[['start_hours','min','sec']] = proc['start_hours'].str.split(':', -1, expand=True) |
|
proc['start_time']=pd.to_numeric(proc['start_days'])*24+pd.to_numeric(proc['start_hours']) |
|
proc=proc.drop(columns=['start_days', 'dummy','start_hours','min','sec']) |
|
proc=proc[proc['start_time']>=0] |
|
|
|
|
|
proc=pd.merge(proc,self.data[['stay_id','los']],on='stay_id',how='left') |
|
proc['sanity']=proc['los']-proc['start_time'] |
|
proc=proc[proc['sanity']>0] |
|
del proc['sanity'] |
|
|
|
self.proc=proc |
|
|
|
def generate_out(self): |
|
out=pd.read_csv("./data/features/preproc_out_icu.csv.gz", compression='gzip', header=0, index_col=None) |
|
out=out[out['stay_id'].isin(self.data['stay_id'])] |
|
out[['start_days', 'dummy','start_hours']] = out['event_time_from_admit'].str.split(' ', -1, expand=True) |
|
out[['start_hours','min','sec']] = out['start_hours'].str.split(':', -1, expand=True) |
|
out['start_time']=pd.to_numeric(out['start_days'])*24+pd.to_numeric(out['start_hours']) |
|
out=out.drop(columns=['start_days', 'dummy','start_hours','min','sec']) |
|
out=out[out['start_time']>=0] |
|
|
|
|
|
out=pd.merge(out,self.data[['stay_id','los']],on='stay_id',how='left') |
|
out['sanity']=out['los']-out['start_time'] |
|
out=out[out['sanity']>0] |
|
del out['sanity'] |
|
|
|
self.out=out |
|
|
|
|
|
def generate_chart(self): |
|
chunksize = 5000000 |
|
final=pd.DataFrame() |
|
for chart in tqdm(pd.read_csv("./data/features/preproc_chart_icu.csv.gz", compression='gzip', header=0, index_col=None,chunksize=chunksize)): |
|
chart=chart[chart['stay_id'].isin(self.data['stay_id'])] |
|
chart[['start_days', 'dummy','start_hours']] = chart['event_time_from_admit'].str.split(' ', -1, expand=True) |
|
chart[['start_hours','min','sec']] = chart['start_hours'].str.split(':', -1, expand=True) |
|
chart['start_time']=pd.to_numeric(chart['start_days'])*24+pd.to_numeric(chart['start_hours']) |
|
chart=chart.drop(columns=['start_days', 'dummy','start_hours','min','sec','event_time_from_admit']) |
|
chart=chart[chart['start_time']>=0] |
|
|
|
|
|
chart=pd.merge(chart,self.data[['stay_id','los']],on='stay_id',how='left') |
|
chart['sanity']=chart['los']-chart['start_time'] |
|
chart=chart[chart['sanity']>0] |
|
del chart['sanity'] |
|
del chart['los'] |
|
|
|
if final.empty: |
|
final=chart |
|
else: |
|
final=final.append(chart, ignore_index=True) |
|
|
|
self.chart=final |
|
|
|
|
|
|
|
def generate_meds(self): |
|
meds=pd.read_csv("./data/features/preproc_med_icu.csv.gz", compression='gzip', header=0, index_col=None) |
|
meds[['start_days', 'dummy','start_hours']] = meds['start_hours_from_admit'].str.split(' ', -1, expand=True) |
|
meds[['start_hours','min','sec']] = meds['start_hours'].str.split(':', -1, expand=True) |
|
meds['start_time']=pd.to_numeric(meds['start_days'])*24+pd.to_numeric(meds['start_hours']) |
|
meds[['start_days', 'dummy','start_hours']] = meds['stop_hours_from_admit'].str.split(' ', -1, expand=True) |
|
meds[['start_hours','min','sec']] = meds['start_hours'].str.split(':', -1, expand=True) |
|
meds['stop_time']=pd.to_numeric(meds['start_days'])*24+pd.to_numeric(meds['start_hours']) |
|
meds=meds.drop(columns=['start_days', 'dummy','start_hours','min','sec']) |
|
|
|
meds['sanity']=meds['stop_time']-meds['start_time'] |
|
meds=meds[meds['sanity']>0] |
|
del meds['sanity'] |
|
|
|
meds=meds[meds['stay_id'].isin(self.data['stay_id'])] |
|
meds=pd.merge(meds,self.data[['stay_id','los']],on='stay_id',how='left') |
|
|
|
|
|
meds['sanity']=meds['los']-meds['start_time'] |
|
meds=meds[meds['sanity']>0] |
|
del meds['sanity'] |
|
|
|
meds.loc[meds['stop_time'] > meds['los'],'stop_time']=meds.loc[meds['stop_time'] > meds['los'],'los'] |
|
del meds['los'] |
|
|
|
meds['rate']=meds['rate'].apply(pd.to_numeric, errors='coerce') |
|
meds['amount']=meds['amount'].apply(pd.to_numeric, errors='coerce') |
|
|
|
self.meds=meds |
|
|
|
def mortality_length(self,include_time,predW): |
|
print("include_time",include_time) |
|
self.los=include_time |
|
self.data=self.data[(self.data['los']>=include_time+predW)] |
|
self.hids=self.data['stay_id'].unique() |
|
|
|
if(self.feat_cond): |
|
self.cond=self.cond[self.cond['stay_id'].isin(self.data['stay_id'])] |
|
|
|
self.data['los']=include_time |
|
|
|
|
|
|
|
|
|
if(self.feat_med): |
|
self.meds=self.meds[self.meds['stay_id'].isin(self.data['stay_id'])] |
|
self.meds=self.meds[self.meds['start_time']<=include_time] |
|
self.meds.loc[self.meds.stop_time >include_time, 'stop_time']=include_time |
|
|
|
|
|
|
|
if(self.feat_proc): |
|
self.proc=self.proc[self.proc['stay_id'].isin(self.data['stay_id'])] |
|
self.proc=self.proc[self.proc['start_time']<=include_time] |
|
|
|
|
|
if(self.feat_out): |
|
self.out=self.out[self.out['stay_id'].isin(self.data['stay_id'])] |
|
self.out=self.out[self.out['start_time']<=include_time] |
|
|
|
|
|
if(self.feat_chart): |
|
self.chart=self.chart[self.chart['stay_id'].isin(self.data['stay_id'])] |
|
self.chart=self.chart[self.chart['start_time']<=include_time] |
|
|
|
|
|
def los_length(self,include_time): |
|
print("include_time",include_time) |
|
self.los=include_time |
|
self.data=self.data[(self.data['los']>=include_time)] |
|
self.hids=self.data['stay_id'].unique() |
|
|
|
if(self.feat_cond): |
|
self.cond=self.cond[self.cond['stay_id'].isin(self.data['stay_id'])] |
|
|
|
self.data['los']=include_time |
|
|
|
|
|
|
|
|
|
if(self.feat_med): |
|
self.meds=self.meds[self.meds['stay_id'].isin(self.data['stay_id'])] |
|
self.meds=self.meds[self.meds['start_time']<=include_time] |
|
self.meds.loc[self.meds.stop_time >include_time, 'stop_time']=include_time |
|
|
|
|
|
|
|
if(self.feat_proc): |
|
self.proc=self.proc[self.proc['stay_id'].isin(self.data['stay_id'])] |
|
self.proc=self.proc[self.proc['start_time']<=include_time] |
|
|
|
|
|
if(self.feat_out): |
|
self.out=self.out[self.out['stay_id'].isin(self.data['stay_id'])] |
|
self.out=self.out[self.out['start_time']<=include_time] |
|
|
|
|
|
if(self.feat_chart): |
|
self.chart=self.chart[self.chart['stay_id'].isin(self.data['stay_id'])] |
|
self.chart=self.chart[self.chart['start_time']<=include_time] |
|
|
|
def readmission_length(self,include_time): |
|
self.los=include_time |
|
self.data=self.data[(self.data['los']>=include_time)] |
|
self.hids=self.data['stay_id'].unique() |
|
|
|
if(self.feat_cond): |
|
self.cond=self.cond[self.cond['stay_id'].isin(self.data['stay_id'])] |
|
self.data['select_time']=self.data['los']-include_time |
|
self.data['los']=include_time |
|
|
|
|
|
|
|
|
|
if(self.feat_med): |
|
self.meds=self.meds[self.meds['stay_id'].isin(self.data['stay_id'])] |
|
self.meds=pd.merge(self.meds,self.data[['stay_id','select_time']],on='stay_id',how='left') |
|
self.meds['stop_time']=self.meds['stop_time']-self.meds['select_time'] |
|
self.meds['start_time']=self.meds['start_time']-self.meds['select_time'] |
|
self.meds=self.meds[self.meds['stop_time']>=0] |
|
self.meds.loc[self.meds.start_time <0, 'start_time']=0 |
|
|
|
|
|
if(self.feat_proc): |
|
self.proc=self.proc[self.proc['stay_id'].isin(self.data['stay_id'])] |
|
self.proc=pd.merge(self.proc,self.data[['stay_id','select_time']],on='stay_id',how='left') |
|
self.proc['start_time']=self.proc['start_time']-self.proc['select_time'] |
|
self.proc=self.proc[self.proc['start_time']>=0] |
|
|
|
|
|
if(self.feat_out): |
|
self.out=self.out[self.out['stay_id'].isin(self.data['stay_id'])] |
|
self.out=pd.merge(self.out,self.data[['stay_id','select_time']],on='stay_id',how='left') |
|
self.out['start_time']=self.out['start_time']-self.out['select_time'] |
|
self.out=self.out[self.out['start_time']>=0] |
|
|
|
|
|
if(self.feat_chart): |
|
self.chart=self.chart[self.chart['stay_id'].isin(self.data['stay_id'])] |
|
self.chart=pd.merge(self.chart,self.data[['stay_id','select_time']],on='stay_id',how='left') |
|
self.chart['start_time']=self.chart['start_time']-self.chart['select_time'] |
|
self.chart=self.chart[self.chart['start_time']>=0] |
|
|
|
|
|
def smooth_meds(self,bucket): |
|
final_meds=pd.DataFrame() |
|
final_proc=pd.DataFrame() |
|
final_out=pd.DataFrame() |
|
final_chart=pd.DataFrame() |
|
|
|
if(self.feat_med): |
|
self.meds=self.meds.sort_values(by=['start_time']) |
|
if(self.feat_proc): |
|
self.proc=self.proc.sort_values(by=['start_time']) |
|
if(self.feat_out): |
|
self.out=self.out.sort_values(by=['start_time']) |
|
if(self.feat_chart): |
|
self.chart=self.chart.sort_values(by=['start_time']) |
|
|
|
t=0 |
|
for i in tqdm(range(0,self.los,bucket)): |
|
|
|
if(self.feat_med): |
|
sub_meds=self.meds[(self.meds['start_time']>=i) & (self.meds['start_time']<i+bucket)].groupby(['stay_id','itemid','orderid']).agg({'stop_time':'max','subject_id':'max','rate':np.nanmean,'amount':np.nanmean}) |
|
sub_meds=sub_meds.reset_index() |
|
sub_meds['start_time']=t |
|
sub_meds['stop_time']=sub_meds['stop_time']/bucket |
|
if final_meds.empty: |
|
final_meds=sub_meds |
|
else: |
|
final_meds=final_meds.append(sub_meds) |
|
|
|
|
|
if(self.feat_proc): |
|
sub_proc=self.proc[(self.proc['start_time']>=i) & (self.proc['start_time']<i+bucket)].groupby(['stay_id','itemid']).agg({'subject_id':'max'}) |
|
sub_proc=sub_proc.reset_index() |
|
sub_proc['start_time']=t |
|
if final_proc.empty: |
|
final_proc=sub_proc |
|
else: |
|
final_proc=final_proc.append(sub_proc) |
|
|
|
|
|
if(self.feat_out): |
|
sub_out=self.out[(self.out['start_time']>=i) & (self.out['start_time']<i+bucket)].groupby(['stay_id','itemid']).agg({'subject_id':'max'}) |
|
sub_out=sub_out.reset_index() |
|
sub_out['start_time']=t |
|
if final_out.empty: |
|
final_out=sub_out |
|
else: |
|
final_out=final_out.append(sub_out) |
|
|
|
|
|
|
|
if(self.feat_chart): |
|
sub_chart=self.chart[(self.chart['start_time']>=i) & (self.chart['start_time']<i+bucket)].groupby(['stay_id','itemid']).agg({'valuenum':np.nanmean}) |
|
sub_chart=sub_chart.reset_index() |
|
sub_chart['start_time']=t |
|
if final_chart.empty: |
|
final_chart=sub_chart |
|
else: |
|
final_chart=final_chart.append(sub_chart) |
|
|
|
t=t+1 |
|
print("bucket",bucket) |
|
los=int(self.los/bucket) |
|
|
|
|
|
|
|
if(self.feat_med): |
|
f2_meds=final_meds.groupby(['stay_id','itemid','orderid']).size() |
|
self.med_per_adm=f2_meds.groupby('stay_id').sum().reset_index()[0].max() |
|
self.medlength_per_adm=final_meds.groupby('stay_id').size().max() |
|
|
|
|
|
if(self.feat_proc): |
|
f2_proc=final_proc.groupby(['stay_id','itemid']).size() |
|
self.proc_per_adm=f2_proc.groupby('stay_id').sum().reset_index()[0].max() |
|
self.proclength_per_adm=final_proc.groupby('stay_id').size().max() |
|
|
|
|
|
if(self.feat_out): |
|
f2_out=final_out.groupby(['stay_id','itemid']).size() |
|
self.out_per_adm=f2_out.groupby('stay_id').sum().reset_index()[0].max() |
|
self.outlength_per_adm=final_out.groupby('stay_id').size().max() |
|
|
|
|
|
|
|
if(self.feat_chart): |
|
f2_chart=final_chart.groupby(['stay_id','itemid']).size() |
|
self.chart_per_adm=f2_chart.groupby('stay_id').sum().reset_index()[0].max() |
|
self.chartlength_per_adm=final_chart.groupby('stay_id').size().max() |
|
|
|
print("[ PROCESSED TIME SERIES TO EQUAL TIME INTERVAL ]") |
|
|
|
|
|
|
|
|
|
self.create_Dict(final_meds,final_proc,final_out,final_chart,los) |
|
|
|
|
|
def create_Dict(self,meds,proc,out,chart,los): |
|
dataDic={} |
|
|
|
for hid in self.hids: |
|
grp=self.data[self.data['stay_id']==hid] |
|
dataDic[hid]={'Cond':{},'Proc':{},'Med':{},'Out':{},'Chart':{},'ethnicity':grp['ethnicity'].iloc[0],'age':int(grp['Age']),'gender':grp['gender'].iloc[0],'label':int(grp['label']),'insurance':grp['insurance'].iloc[0]} |
|
|
|
|
|
for hid in tqdm(self.hids): |
|
grp=self.data[self.data['stay_id']==hid] |
|
|
|
|
|
if(self.feat_med): |
|
feat=meds['itemid'].unique() |
|
df2=meds[meds['stay_id']==hid] |
|
if df2.shape[0]==0: |
|
amount=pd.DataFrame(np.zeros([los,len(feat)]),columns=feat) |
|
amount=amount.fillna(0) |
|
amount.columns=pd.MultiIndex.from_product([["MEDS"], amount.columns]) |
|
else: |
|
rate=df2.pivot_table(index='start_time',columns='itemid',values='rate') |
|
|
|
amount=df2.pivot_table(index='start_time',columns='itemid',values='amount') |
|
df2=df2.pivot_table(index='start_time',columns='itemid',values='stop_time') |
|
|
|
add_indices = pd.Index(range(los)).difference(df2.index) |
|
add_df = pd.DataFrame(index=add_indices, columns=df2.columns).fillna(np.nan) |
|
df2=pd.concat([df2, add_df]) |
|
df2=df2.sort_index() |
|
df2=df2.ffill() |
|
df2=df2.fillna(0) |
|
|
|
rate=pd.concat([rate, add_df]) |
|
rate=rate.sort_index() |
|
rate=rate.ffill() |
|
rate=rate.fillna(-1) |
|
|
|
amount=pd.concat([amount, add_df]) |
|
amount=amount.sort_index() |
|
amount=amount.ffill() |
|
amount=amount.fillna(-1) |
|
|
|
df2.iloc[:,0:]=df2.iloc[:,0:].sub(df2.index,0) |
|
df2[df2>0]=1 |
|
df2[df2<0]=0 |
|
rate.iloc[:,0:]=df2.iloc[:,0:]*rate.iloc[:,0:] |
|
amount.iloc[:,0:]=df2.iloc[:,0:]*amount.iloc[:,0:] |
|
|
|
dataDic[hid]['Med']['signal']=df2.iloc[:,0:].to_dict(orient="list") |
|
dataDic[hid]['Med']['rate']=rate.iloc[:,0:].to_dict(orient="list") |
|
dataDic[hid]['Med']['amount']=amount.iloc[:,0:].to_dict(orient="list") |
|
|
|
|
|
if(self.feat_proc): |
|
feat=proc['itemid'].unique() |
|
df2=proc[proc['stay_id']==hid] |
|
if df2.shape[0]==0: |
|
df2=pd.DataFrame(np.zeros([los,len(feat)]),columns=feat) |
|
df2=df2.fillna(0) |
|
df2.columns=pd.MultiIndex.from_product([["PROC"], df2.columns]) |
|
else: |
|
df2['val']=1 |
|
|
|
df2=df2.pivot_table(index='start_time',columns='itemid',values='val') |
|
|
|
add_indices = pd.Index(range(los)).difference(df2.index) |
|
add_df = pd.DataFrame(index=add_indices, columns=df2.columns).fillna(np.nan) |
|
df2=pd.concat([df2, add_df]) |
|
df2=df2.sort_index() |
|
df2=df2.fillna(0) |
|
df2[df2>0]=1 |
|
|
|
dataDic[hid]['Proc']=df2.to_dict(orient="list") |
|
|
|
|
|
if(self.feat_out): |
|
feat=out['itemid'].unique() |
|
df2=out[out['stay_id']==hid] |
|
if df2.shape[0]==0: |
|
df2=pd.DataFrame(np.zeros([los,len(feat)]),columns=feat) |
|
df2=df2.fillna(0) |
|
df2.columns=pd.MultiIndex.from_product([["OUT"], df2.columns]) |
|
else: |
|
df2['val']=1 |
|
df2=df2.pivot_table(index='start_time',columns='itemid',values='val') |
|
|
|
add_indices = pd.Index(range(los)).difference(df2.index) |
|
add_df = pd.DataFrame(index=add_indices, columns=df2.columns).fillna(np.nan) |
|
df2=pd.concat([df2, add_df]) |
|
df2=df2.sort_index() |
|
df2=df2.fillna(0) |
|
df2[df2>0]=1 |
|
|
|
dataDic[hid]['Out']=df2.to_dict(orient="list") |
|
|
|
|
|
if(self.feat_chart): |
|
feat=chart['itemid'].unique() |
|
df2=chart[chart['stay_id']==hid] |
|
if df2.shape[0]==0: |
|
val=pd.DataFrame(np.zeros([los,len(feat)]),columns=feat) |
|
val=val.fillna(0) |
|
val.columns=pd.MultiIndex.from_product([["CHART"], val.columns]) |
|
else: |
|
val=df2.pivot_table(index='start_time',columns='itemid',values='valuenum') |
|
df2['val']=1 |
|
df2=df2.pivot_table(index='start_time',columns='itemid',values='val') |
|
|
|
add_indices = pd.Index(range(los)).difference(df2.index) |
|
add_df = pd.DataFrame(index=add_indices, columns=df2.columns).fillna(np.nan) |
|
df2=pd.concat([df2, add_df]) |
|
df2=df2.sort_index() |
|
df2=df2.fillna(0) |
|
|
|
val=pd.concat([val, add_df]) |
|
val=val.sort_index() |
|
if self.impute=='Mean': |
|
val=val.ffill() |
|
val=val.bfill() |
|
val=val.fillna(val.mean()) |
|
elif self.impute=='Median': |
|
val=val.ffill() |
|
val=val.bfill() |
|
val=val.fillna(val.median()) |
|
val=val.fillna(0) |
|
|
|
|
|
df2[df2>0]=1 |
|
df2[df2<0]=0 |
|
|
|
dataDic[hid]['Chart']['signal']=df2.iloc[:,0:].to_dict(orient="list") |
|
dataDic[hid]['Chart']['val']=val.iloc[:,0:].to_dict(orient="list") |
|
|
|
|
|
if(self.feat_cond): |
|
feat=self.cond['new_icd_code'].unique() |
|
grp=self.cond[self.cond['stay_id']==hid] |
|
if(grp.shape[0]==0): |
|
dataDic[hid]['Cond']={'fids':list(['<PAD>'])} |
|
else: |
|
dataDic[hid]['Cond']={'fids':list(grp['new_icd_code'])} |
|
|
|
|
|
|
|
metaDic={'Cond':{},'Proc':{},'Med':{},'Out':{},'Chart':{},'LOS':{}} |
|
metaDic['LOS']=los |
|
with open("./data/dict/"+self.task+"/dataDic", 'wb') as fp: |
|
pickle.dump(dataDic, fp) |
|
|
|
with open("./data/dict/"+self.task+"/hadmDic", 'wb') as fp: |
|
pickle.dump(self.hids, fp) |
|
|
|
with open("./data/dict/"+self.task+"/ethVocab", 'wb') as fp: |
|
pickle.dump(list(self.data['ethnicity'].unique()), fp) |
|
self.eth_vocab = self.data['ethnicity'].nunique() |
|
|
|
with open("./data/dict/"+self.task+"/ageVocab", 'wb') as fp: |
|
pickle.dump(list(self.data['Age'].unique()), fp) |
|
self.age_vocab = self.data['Age'].nunique() |
|
|
|
with open("./data/dict/"+self.task+"/insVocab", 'wb') as fp: |
|
pickle.dump(list(self.data['insurance'].unique()), fp) |
|
self.ins_vocab = self.data['insurance'].nunique() |
|
|
|
if(self.feat_med): |
|
with open("./data/dict/"+self.task+"/medVocab", 'wb') as fp: |
|
pickle.dump(list(meds['itemid'].unique()), fp) |
|
self.med_vocab = meds['itemid'].nunique() |
|
metaDic['Med']=self.med_per_adm |
|
|
|
if(self.feat_out): |
|
with open("./data/dict/"+self.task+"/outVocab", 'wb') as fp: |
|
pickle.dump(list(out['itemid'].unique()), fp) |
|
self.out_vocab = out['itemid'].nunique() |
|
metaDic['Out']=self.out_per_adm |
|
|
|
if(self.feat_chart): |
|
with open("./data/dict/"+self.task+"/chartVocab", 'wb') as fp: |
|
pickle.dump(list(chart['itemid'].unique()), fp) |
|
self.chart_vocab = chart['itemid'].nunique() |
|
metaDic['Chart']=self.chart_per_adm |
|
|
|
if(self.feat_cond): |
|
with open("./data/dict/"+self.task+"/condVocab", 'wb') as fp: |
|
pickle.dump(list(self.cond['new_icd_code'].unique()), fp) |
|
self.cond_vocab = self.cond['new_icd_code'].nunique() |
|
metaDic['Cond']=self.cond_per_adm |
|
|
|
if(self.feat_proc): |
|
with open("./data/dict/"+self.task+"/procVocab", 'wb') as fp: |
|
pickle.dump(list(proc['itemid'].unique()), fp) |
|
self.proc_vocab = proc['itemid'].nunique() |
|
metaDic['Proc']=self.proc_per_adm |
|
|
|
with open("./data/dict/"+self.task+"/metaDic", 'wb') as fp: |
|
pickle.dump(metaDic, fp) |