anonymous-submission-acl2025's picture
add 17
8a79f2e
import pandas as pd
import os
import shutil
import re
from functools import reduce
from datetime import datetime, timedelta
from lib.experiment_specs import study_config
from lib.data_helpers import data_utils
"""loads the phone data config from the provided config path"""
class BuilderUtils():
def get_config(self, config_path):
if os.path.isfile(config_path):
pd_config_df = pd.read_csv(config_path,index_col= "index")
pd_config_dict = pd_config_df.to_dict(orient = 'index')
return pd_config_dict
else:
return {}
"""
- Purpose: transports zipped files from PhoneDashboardPort and PCPort to the PhoneAddictionDropbox to the specified directory
- Inputs:
- port: specifies location of the port
- keyword: specifies the kind of inport from the source (e.g. budget, use, etc). the keyword must be in the file name for the function to work
- new_directory: the directory where the files will be transported
- """
def transport_new_zip_files(self,port,keyword,new_directory):
new_adds = []
added_files = os.listdir(new_directory)
empty_files_dir = os.listdir(os.path.join("data","external","input","PhoneDashboard","BuggyFiles","Empty"))
for zipfile in os.listdir(port):
if ".zip" not in zipfile:
continue
# if "UseIndiv" nearly exactly do process as "Use"
if keyword == "UseIndiv":
keyword = "Use"
# change zipfile name for pd use data
if ("full" in zipfile) & (keyword == "Use"):
new_zipfile = zipfile.replace("full","use")
os.rename(os.path.join(port, zipfile), os.path.join(port, new_zipfile))
zipfile = new_zipfile
# change zipfile name for pd custom delay data, as soon as possible
if ("snooze_delays" in zipfile):
new_zipfile = zipfile.replace("snooze_","")
os.rename(os.path.join(port, zipfile), os.path.join(port, new_zipfile))
zipfile = new_zipfile
if (keyword.lower() not in zipfile) and (keyword.upper() not in zipfile):
continue
#if it already exists, skip
if zipfile in added_files:
continue
#if in the empty or corrupt directory in PA dropbox, also place it in empty or corrupt dir in port
if zipfile in empty_files_dir:
try:
old_file = os.path.join(port, zipfile)
new_file = os.path.join(port, "Empty", zipfile)
os.rename(old_file, new_file)
except:
print(f"{zipfile}couldn't move zipfile to PDPort/Empty")
continue
#if out of date range, skip
match = re.search(r'\d{4}-\d{2}-\d{2}', zipfile)
zip_date = datetime.strptime(match.group(), '%Y-%m-%d')
if zip_date <= study_config.first_pull or zip_date >= study_config.last_pull:
continue
#else, copy and transfer it
else:
old_file_path = os.path.join(port,zipfile)
new_file_path = os.path.join(new_directory,zipfile)
new_adds.append(zipfile)
shutil.copy(old_file_path,new_file_path)
print(new_adds)
return new_adds
""" updates the existing config by adding the new config entries, and saves the updated config"""
def update_config(self,existing,new,config_path):
existing.update(new)
pd_config_df = pd.DataFrame.from_dict(existing, orient='index').reset_index()
pd_config_df.to_csv(config_path, index=False)
"""Default raw data processor invoked by event_puller.py"""
@staticmethod
def default_puller_process(df: pd.DataFrame, zip_file: str, event_puller):
for time_col in event_puller.time_cols:
df = data_utils.clean_iso_dates(df, time_col, keep_nan=False, orig_tz=event_puller.raw_timezone)
df = df.drop(columns=[time_col + "Date", time_col + "DatetimeHour", time_col + "EasternDatetimeHour"])
df = df.rename(columns={time_col + "Datetime": time_col})
if "TimeZone" in df.columns:
df = df.drop(columns=["TimeZone"])
match = re.search(r'\d{4}-\d{2}-\d{2}', zip_file)
df["AsOf"] = datetime.strptime(match.group(), '%Y-%m-%d')
df["AsOf"] = df["AsOf"].apply(lambda x: x.date())
return df
# add phase column to each obs based study_config survey start times
# start_buffer =1 means that days will be counted the day after the survey start
# end_buffer = -1 means that the days will be counted the day before the survey start
@staticmethod
def add_phase_label(raw_df, raw_df_date, start_buffer=1, end_buffer=-1):
df = raw_df.copy()
if "Phase" in df.columns.values:
df = df.drop(columns="Phase")
for phase, specs in study_config.phases.items():
# label use with phases if we're a day into a phase
if datetime.now() > specs["StartSurvey"]["Start"] + timedelta(1):
start_date = (study_config.phases[phase]["StartSurvey"]["Start"] + timedelta(start_buffer)).date()
end_date = (study_config.phases[phase]["EndSurvey"]["Start"] + timedelta(end_buffer)).date()
df.loc[(df[raw_df_date] >= start_date) & (df[raw_df_date] <= end_date), "Phase"] = phase
df["Phase"] = df["Phase"].astype('category')
return df
"""
Purpose: Iterates through a subsets dict and creates new avg daily use columns
One key-value pair of a subset dict:
"PCSC" : {
"Filters": {"SCBool":[True]},
"DenomCol": "DaysWithUse"},
"""
@staticmethod
def get_subsets_avg_use(df_p, subsets: dict):
subset_dfs = []
for label, specs in subsets.items():
filters = specs["Filters"]
denom_col = specs["DenomCol"]
num_cols = specs["NumCols"]
subset_df = BuilderUtils.subset_avg_use(df_p, label, filters, denom_col,num_cols)
subset_dfs.append(subset_df)
df_merged = reduce(lambda x, y: pd.merge(x, y, on='AppCode', how = 'outer'), subset_dfs)
# If they are in this df, then they recorded some use in the phase, so we convert all of their nan's
# (i.e. for a specfic subset) in the df to 0
df_merged = df_merged.fillna(0)
return df_merged
"""
Input:
- df: the event level df in the given phase
- label: the variable label
- specs: {variables to subset on: values of variables to keep}
- denom_col: the column name of the variable in the df which contains the denomenator value
- if == "NAN", the function will create it's own denomenator equal to days for which there is non-zero use for
the given subset
- num_cols: list of columns to sum over (often it's just [Use], but it can be [Checks,Pickups,Use]
"""
@staticmethod
def subset_avg_use(df: pd.DataFrame, label: str, filters: dict, denom_col: str, num_cols: list):
# if we don't want to subset the phase data at all
if len(filters) == 0:
pass
# go through each filter (note that at all filters for each variable must be met)
else:
for var, keep_vals in filters.items():
df = df.loc[df[var].isin(keep_vals),:]
for col in [denom_col]+[num_cols]:
df[col] = df[col].fillna(0)
sum_df = df.groupby(by=['AppCode',denom_col], as_index=False)[num_cols].sum()
sum_dfs = []
for num_col in num_cols:
sum_df = sum_df.rename(columns={num_col: f"{label}{num_col}Total"})
sum_df[f"{label}{num_col}Total"] = sum_df[f"{label}{num_col}Total"].round(0)
sum_df[f"{label}{num_col}"] = (sum_df[f"{label}{num_col}Total"] / (sum_df[denom_col])).round(0)
sum_dfs.append(sum_df[["AppCode", f"{label}{num_col}", f"{label}{num_col}Total"]])
final = reduce(lambda df1, df2: pd.merge(df1, df2, on='AppCode', how = 'outer'), sum_dfs)
return final
# add phase column to each obs based on time they completed the survey, indicating what phase they are in at the timestamp
# start_buffer =1 means that days will be counted the day after the survey start
# end_buffer = -1 means that the days will be counted the day before the survey start
@staticmethod
def add_personal_phase_label(raw_df, raw_master, raw_df_date, start_buffer=1, end_buffer=-1, drop_bool=True):
df = raw_df.copy()
if "Phase" in df.columns.values:
df = df.drop(columns="Phase")
for phase, specs in study_config.phases.items():
# label use with phases if we're a day into a phase
if datetime.now() > specs["StartSurvey"]["Start"] + timedelta(1):
raw_master = data_utils.inpute_missing_survey_datetimes(raw_master, phase)
old_code = study_config.phases[phase]["StartSurvey"]["Code"]
new_code = study_config.phases[phase]["EndSurvey"]["Code"]
start_col = f"{old_code}_SurveyEndDatetime"
end_col = f"{new_code}_SurveyStartDatetime"
df = df.merge(raw_master[["AppCode", start_col, end_col]], on="AppCode", how="inner")
for col in [start_col, end_col]:
df[col] = pd.to_datetime(df[col], infer_datetime_format=True).apply(lambda x: x.date())
df.loc[(df[raw_df_date] >= df[start_col].apply(lambda x: x + timedelta(start_buffer)))
& (df[raw_df_date] <= df[end_col].apply(lambda x: x + timedelta(end_buffer))), "Phase"] = phase
if drop_bool:
df = df.drop(columns=[start_col, end_col])
df["Phase"] = df["Phase"].astype('category')
return df