|
|
import os |
|
|
import sys |
|
|
import shutil |
|
|
import zipfile |
|
|
import multiprocessing |
|
|
import pandas as pd |
|
|
from datetime import datetime,timedelta |
|
|
|
|
|
from lib.data_helpers import data_utils |
|
|
from lib.experiment_specs import study_config |
|
|
from lib.data_helpers import test |
|
|
from lib.data_helpers.builder_utils import BuilderUtils |
|
|
from lib.data_helpers.gaming import Gaming |
|
|
|
|
|
from lib.utilities import serialize |
|
|
|
|
|
class PullEvents(): |
|
|
|
|
|
port_dir = {"PCDashboard": os.path.join("data","external","dropbox","..","PCPort"), |
|
|
"PhoneDashboard":os.path.join("data","external","dropbox","..","PhoneDashboardPort")} |
|
|
|
|
|
|
|
|
def __init__(self,source: str, keyword: str, scratch: bool, test: bool, |
|
|
time_cols: list, raw_timezone: str, appcode_col: str, identifying_cols: list, sort_cols: list, |
|
|
drop_cols: list, cat_cols: list, |
|
|
compress_type: str, processing_func = BuilderUtils.default_puller_process, file_reader = None): |
|
|
""" |
|
|
|
|
|
Purpose |
|
|
------- |
|
|
|
|
|
This class appends the raw data files housed in self.zipped_directory, cleans them a bit with the self.processing func, |
|
|
and saves the merged data file in self.raw_file. Unless self.scratch == True, the puller will only process new raw data files |
|
|
and appends them to the dataframe saved in self.raw_file. The puller documents which files have been processed in the config located in self.config_file_path |
|
|
|
|
|
WARNING: if self.scratch==TRUE, it may take hours or days to reprocess all the data. Call Michael to discuss if this needs to happen. |
|
|
|
|
|
Parameters |
|
|
---------- |
|
|
source - source the app data source (either PC or PD) |
|
|
keyword - the kind of data coming in (e.g. Snooze, Use etc) |
|
|
scratch - whether or not to start data processing from scratch |
|
|
test - whether or not to test the pulling pipeline (it will reprocess that latest three zipfiles) (I wouldn't play with this) |
|
|
time_cols - list of columns that contain datetimes |
|
|
raw_timezone - the timezone of the raw data |
|
|
appcode_col - the raw column name containing the appcodes |
|
|
identifying_cols - list of columns that should identify a unique row |
|
|
sort_cols - list of columns that describes how the df should be ordered. The df will be sorted from smallest to largest value of the sort_cols |
|
|
drop_cols - list of columns that appear in the raw zipfile or after zipfile processing that should be dropped |
|
|
cat_cols - list of columns that can be converted into the categorical data type |
|
|
compress_type - how the zipfiles are compressed (either "folder", "csv", or "txt") |
|
|
processing_func - the function used to process the raw data (e.g. clean_master dates) |
|
|
file_reader - function that reads in the raw file into memory. only required for file stored in zip folders. |
|
|
""" |
|
|
|
|
|
self.source = source |
|
|
self.keyword = keyword |
|
|
self.scratch = scratch |
|
|
self.pull_test = test |
|
|
self.time_cols = time_cols |
|
|
self.raw_timezone = raw_timezone |
|
|
self.appcode_col = appcode_col |
|
|
self.identifying_cols = identifying_cols |
|
|
self.sort_cols = sort_cols |
|
|
self.drop_cols = drop_cols |
|
|
self.cat_cols = cat_cols |
|
|
self.compress_type = compress_type |
|
|
self.processing_func = processing_func |
|
|
self.file_reader = file_reader |
|
|
|
|
|
|
|
|
self.zipped_directory = os.path.join("data","external", "input", source, f"Raw{self.keyword}Zipped") |
|
|
self.buggy_dir = os.path.join("data","external", "input", source, "BuggyFiles") |
|
|
self.config_file_path = os.path.join("data","external", "input", source, f"{self.source}{self.keyword}Config.csv") |
|
|
self.raw_file = os.path.join("data","external", "input", source, f"Raw{self.keyword}") |
|
|
self.test_file = os.path.join("data","external", "input_test", source, f"Raw{self.keyword}") |
|
|
|
|
|
self.config_user_dict = serialize.open_yaml("config_user.yaml") |
|
|
|
|
|
self.builder_utils = BuilderUtils() |
|
|
self.added_config = self.builder_utils.get_config(self.config_file_path) |
|
|
self.new_config = {} |
|
|
|
|
|
|
|
|
def update_data(self): |
|
|
print(f"\n Updating {self.keyword} data! {datetime.now()}") |
|
|
if self.config_user_dict["local"]["test"]==True: |
|
|
df = serialize.soft_df_open(self.test_file) |
|
|
df = self._memory_redux(df) |
|
|
print(df.memory_usage(deep=True) / 1024 ** 2) |
|
|
return df |
|
|
|
|
|
else: |
|
|
self.builder_utils.transport_new_zip_files(PullEvents.port_dir[self.source], self.keyword, self.zipped_directory) |
|
|
new_zip_files = [x for x in os.listdir(self.zipped_directory) if (x.replace(".zip", "") not in self.added_config) & (".zip" in x)] |
|
|
print(f"new zip files {len(new_zip_files)}") |
|
|
|
|
|
|
|
|
|
|
|
if ((self.scratch == False) & |
|
|
(len(new_zip_files) == 0) & |
|
|
(self.pull_test == False)): |
|
|
print("\t No New Data! Loading pickled raw data into memory") |
|
|
|
|
|
if(os.path.getsize(self.raw_file+".pickle") > 0) : |
|
|
df = serialize.soft_df_open(self.raw_file) |
|
|
else: |
|
|
print("Raw file is empty. Reading Backup Compression Zip") |
|
|
df = serialize.read_gz(self.raw_file+".gz") |
|
|
df = self._memory_redux(df) |
|
|
|
|
|
|
|
|
if (self.keyword == "Alternative") or (self.keyword == "Use"): |
|
|
|
|
|
df = df.loc[df["CreatedDatetimeHour"] >= study_config.surveys["Baseline"]["Start"]-timedelta(1)] |
|
|
|
|
|
return df |
|
|
|
|
|
else: |
|
|
print(f"Length of new zip files {len(new_zip_files)}") |
|
|
df = self.aggregate_data(new_zip_files) |
|
|
self.builder_utils.update_config(self.added_config, |
|
|
self.new_config, |
|
|
self.config_file_path) |
|
|
return df |
|
|
|
|
|
def aggregate_data(self,new_zip_files): |
|
|
old_df, new_zip_files = self._configure_data(new_zip_files) |
|
|
|
|
|
print(f"\t Going to add {len(new_zip_files)} files:") |
|
|
if self.config_user_dict['local']["parallel"] == False: |
|
|
df_dic = self._process_zips(new_zip_files) |
|
|
|
|
|
|
|
|
elif multiprocessing.cpu_count() < len(new_zip_files): |
|
|
df_dic = self._process_zips_mp(new_zip_files) |
|
|
|
|
|
else: |
|
|
print(f"not multiprocessing b/c number of cores {multiprocessing.cpu_count()} < {len(new_zip_files)}") |
|
|
df_dic = self._process_zips(new_zip_files) |
|
|
|
|
|
|
|
|
print("\t get data survey") |
|
|
cl_file = os.path.join(os.path.join("data","external","dropbox_confidential","ContactLists","Used"), |
|
|
study_config.kept_appcode_cl) |
|
|
a_df = pd.read_csv(cl_file) |
|
|
a_df = data_utils.add_A_to_appcode(a_df,"AppCode") |
|
|
appcodes = list(a_df["AppCode"]) |
|
|
print(len(appcodes)) |
|
|
assert len(appcodes) == study_config.number_of_kept_appcodes |
|
|
|
|
|
|
|
|
if len(df_dic)>0: |
|
|
print("\t adding new data to master") |
|
|
new_df = pd.concat(list(df_dic.values()), sort=True) |
|
|
new_df = new_df.loc[new_df["AppCode"].isin(appcodes)] |
|
|
parent_df = pd.concat([old_df,new_df], sort=True) |
|
|
else: |
|
|
print("\t no new data!") |
|
|
parent_df = old_df |
|
|
|
|
|
|
|
|
print(f"\t Len before duplicate drop {len(parent_df)}") |
|
|
parent_df = parent_df.sort_values(by = self.sort_cols) |
|
|
parent_df = parent_df.drop_duplicates(subset = self.identifying_cols, keep = 'last') |
|
|
print(f"\t Len after duplicate drop {len(parent_df)}") |
|
|
|
|
|
for col in parent_df.columns: |
|
|
if "Unnamed" in col: |
|
|
parent_df = parent_df.drop(columns= [col]) |
|
|
parent_df = parent_df.reset_index(drop=True) |
|
|
print(parent_df.memory_usage(deep=True) / 1024 ** 2) |
|
|
|
|
|
try: |
|
|
os.rename(self.raw_file+".pickle",self.raw_file+"PrevRun.pickle") |
|
|
except: |
|
|
print('could not find old pickle file') |
|
|
|
|
|
try: |
|
|
serialize.save_pickle(parent_df, self.raw_file) |
|
|
except: |
|
|
print("Failed to save pickle!") |
|
|
|
|
|
try: |
|
|
test.save_test_df(parent_df, self.test_file) |
|
|
except: |
|
|
print("Failed to save testfile") |
|
|
|
|
|
|
|
|
try: |
|
|
os.rename(self.raw_file + ".gz", self.raw_file + "PrevRun.gz") |
|
|
except: |
|
|
print("no old gz file") |
|
|
parent_df.to_csv(self.raw_file + '.gz', index=False, compression='gzip') |
|
|
|
|
|
print(f"\t Done Saving {datetime.now()}") |
|
|
|
|
|
|
|
|
for zip_file, df in df_dic.items(): |
|
|
try: |
|
|
latest_hour = str(df[self.time_cols[0]].max()) |
|
|
earliest_hour = str(df[self.time_cols[0]].min()) |
|
|
except: |
|
|
latest_hour = "NAN" |
|
|
earliest_hour = "NAN" |
|
|
self.new_config[zip_file.replace(".zip", "")] = {"EarliestHour": earliest_hour, |
|
|
"LatestHour": latest_hour, |
|
|
"ZipFile": zip_file} |
|
|
return parent_df |
|
|
|
|
|
""" modifes the reaggregation in the following way: |
|
|
- if self.scatch is True: all zipfiles in data dire are new, and old_df is empty |
|
|
- if self.test is True: reprocess the last three data files, and old_df is reloaded |
|
|
- else: process the new zip files with old df""" |
|
|
def _configure_data(self,new_zip_files): |
|
|
if self.scratch: |
|
|
all_zip_files = [x for x in os.listdir(self.zipped_directory) if ".zip" in x] |
|
|
new_zip_files = sorted(all_zip_files) |
|
|
|
|
|
|
|
|
continue_ans = "Y" |
|
|
if continue_ans == "N": |
|
|
sys.exit() |
|
|
print(f"DELETING ADDED EVENT CONFIG, CAUSING REAGGREGATION OF ALL {self.keyword} ZIPFILES") |
|
|
self.added_config = {} |
|
|
if self.keyword == "Use": |
|
|
print("\t Also Deleting all Granular Gaming Files") |
|
|
for gran_folder in [os.path.join(Gaming.gaming_dir, "FirstLast"), |
|
|
os.path.join(Gaming.gaming_dir, "Granular")]: |
|
|
shutil.rmtree(gran_folder) |
|
|
os.mkdir(gran_folder) |
|
|
old_df = pd.DataFrame() |
|
|
|
|
|
else: |
|
|
if self.pull_test: |
|
|
try: |
|
|
assert len(new_zip_files) == 0 |
|
|
except: |
|
|
print("There are actual new zip files to process! First process these files, then test") |
|
|
sys.exit() |
|
|
|
|
|
print("TEST: Reprocessing 3 files from zipped directory") |
|
|
test_files = os.listdir(self.zipped_directory)[-3:] |
|
|
for zip_file in test_files: |
|
|
del self.added_config[zip_file.replace(".zip", "")] |
|
|
new_zip_files = test_files |
|
|
|
|
|
if not os.path.exists((self.raw_file+".pickle")): |
|
|
old_df = pd.DataFrame() |
|
|
|
|
|
elif os.path.getsize(self.raw_file + ".pickle") > 0: |
|
|
old_df = serialize.soft_df_open(self.raw_file) |
|
|
else: |
|
|
print("Raw file is empty. Reading Backup Compression Zip") |
|
|
old_df = serialize.read_gz(self.raw_file + ".gz") |
|
|
|
|
|
if len(old_df)>1: |
|
|
old_df = self._memory_redux(old_df) |
|
|
|
|
|
|
|
|
if (self.keyword == "Alternative") or (self.keyword == "Use"): |
|
|
|
|
|
old_df = old_df.loc[old_df["CreatedDatetimeHour"] >= study_config.surveys["Baseline"]["Start"]-timedelta(1)] |
|
|
|
|
|
return old_df, new_zip_files |
|
|
|
|
|
def _process_zips_mp(self, new_zip_files: list): |
|
|
|
|
|
|
|
|
pool = multiprocessing.Pool(processes=study_config.cores) |
|
|
|
|
|
|
|
|
chunks = [new_zip_files[i::study_config.cores] for i in range(study_config.cores)] |
|
|
|
|
|
print(f"Multiprocessing with {study_config.cores} cpus") |
|
|
df_list_of_dics = pool.map(func = self._process_zips, iterable = chunks) |
|
|
pool.close() |
|
|
print("Done Multiprocessing") |
|
|
|
|
|
|
|
|
df_dic = {} |
|
|
for d in df_list_of_dics: |
|
|
for k, v in d.items(): |
|
|
df_dic[k] = v |
|
|
return df_dic |
|
|
|
|
|
def _process_zips(self,new_zip_files): |
|
|
df_dic = {} |
|
|
i = 1 |
|
|
for zip_file in new_zip_files: |
|
|
print(f"processing {zip_file} ({round((i * 100) / len(new_zip_files))}%)", end = "\r") |
|
|
df = self._process_zip(zip_file) |
|
|
if len(df) == 0: |
|
|
continue |
|
|
df_dic[zip_file] = df |
|
|
i = i + 1 |
|
|
return df_dic |
|
|
|
|
|
def _process_zip(self, zip_file): |
|
|
if ".zip" not in zip_file: |
|
|
return pd.DataFrame() |
|
|
|
|
|
|
|
|
if self.compress_type != "folder": |
|
|
df, problem = self._open_zipfile(zip_file) |
|
|
|
|
|
|
|
|
else: |
|
|
df = self._open_zipfolder(zip_file) |
|
|
problem = "" |
|
|
|
|
|
if len(df) == 0: |
|
|
if problem != "Corrupt": |
|
|
problem = "Empty" |
|
|
print(f"{zip_file} is empty ...archiving in {self.buggy_dir}/{problem}") |
|
|
|
|
|
old_file = os.path.join(self.zipped_directory, zip_file) |
|
|
new_file = os.path.join(self.buggy_dir, problem, zip_file) |
|
|
os.rename(old_file, new_file) |
|
|
return df |
|
|
|
|
|
df.columns = df.columns.str.replace(' ', '') |
|
|
df = df.rename(columns={self.appcode_col: "AppCode"}) |
|
|
df["AppCode"] = df["AppCode"].astype(str).apply(lambda x: "A" + x if x != "nan" and x.isnumeric() else x) |
|
|
df = self.processing_func(df, zip_file, self) |
|
|
|
|
|
df["Zipfile"] = zip_file |
|
|
df = self._memory_redux(df) |
|
|
|
|
|
return df |
|
|
|
|
|
def _open_zipfile(self,zip_file): |
|
|
if self.compress_type == "txt": |
|
|
separator = '\t' |
|
|
elif self.compress_type == "csv": |
|
|
separator = ',' |
|
|
else: |
|
|
separator = "" |
|
|
print("Illegal seperator") |
|
|
sys.exit() |
|
|
try: |
|
|
df = pd.read_csv(os.path.join(self.zipped_directory, zip_file), compression='zip', header=0, |
|
|
sep=separator, quotechar='"') |
|
|
problem = "None" |
|
|
except: |
|
|
print(f"{zip_file} is corrupt. Investigate in archive") |
|
|
df = pd.DataFrame() |
|
|
problem = "Corrupt" |
|
|
return df, problem |
|
|
|
|
|
def _open_zipfolder(self,zip_folder): |
|
|
|
|
|
|
|
|
zip_ref = zipfile.ZipFile(os.path.join(self.zipped_directory, zip_folder)) |
|
|
temp_folder = os.path.join(self.zipped_directory, zip_folder.replace(".zip", "FOLDER")) |
|
|
zip_ref.extractall(path=temp_folder) |
|
|
df_list = [] |
|
|
for file in os.listdir(temp_folder): |
|
|
file_path = os.path.join(temp_folder, file) |
|
|
clean_data = self.file_reader(file_path) |
|
|
df_list.append(clean_data) |
|
|
shutil.rmtree(temp_folder) |
|
|
df = pd.concat(df_list) |
|
|
return df |
|
|
|
|
|
def _open_zipfolders_install(self,zip_folders): |
|
|
"""for recovering install data""" |
|
|
for zip_folder in zip_folders: |
|
|
df = self._open_zipfolder(zip_folder) |
|
|
|
|
|
def _memory_redux(self,df): |
|
|
|
|
|
for drop_col in self.drop_cols + ["Server","AsOf"]: |
|
|
if drop_col in df.columns.values: |
|
|
df = df.drop(columns = [drop_col]) |
|
|
|
|
|
|
|
|
for cat_col in self.cat_cols + ["Zipfile"]: |
|
|
if cat_col in df.columns.values: |
|
|
try: |
|
|
df[cat_col] = df[cat_col].astype('category') |
|
|
except: |
|
|
print(f"\t categorizing {cat_col} failed!") |
|
|
|
|
|
return df |
|
|
|
|
|
def _archive_recruitment(self,df): |
|
|
if os.path.exists(self.raw_file + "RecruitmentPhase.pickle"): |
|
|
pass |
|
|
else: |
|
|
r_df_raw = df.loc[df["CreatedDatetimeHour"] < study_config.surveys["Baseline"]["Start"]-timedelta(1)] |
|
|
try: |
|
|
serialize.save_pickle(r_df_raw, self.raw_file + "RecruitmentPhase.pickle") |
|
|
except: |
|
|
r_df_raw.to_csv(self.raw_file + 'RecruitmentPhase.gz', index=False, compression='gzip') |
|
|
|