James McCool commited on
Commit
b4a377f
·
1 Parent(s): ebc0082

Update diversity score calculation to use chunked processing for improved performance in predict_dupes.py

Browse files
Files changed (1) hide show
  1. global_func/predict_dupes.py +1 -1
global_func/predict_dupes.py CHANGED
@@ -429,7 +429,7 @@ def predict_dupes(portfolio, maps_dict, site_var, type_var, Contest_Size, streng
429
  portfolio['Geomean'] = np.power((portfolio[own_columns] * 100).product(axis=1), 1 / len(own_columns))
430
 
431
  # Calculate similarity score based on actual player selection
432
- portfolio['Diversity'] = calculate_player_similarity_score_vectorized(portfolio, player_columns)
433
  # check_portfolio = portfolio.copy()
434
  portfolio = portfolio.drop(columns=dup_count_columns)
435
  portfolio = portfolio.drop(columns=own_columns)
 
429
  portfolio['Geomean'] = np.power((portfolio[own_columns] * 100).product(axis=1), 1 / len(own_columns))
430
 
431
  # Calculate similarity score based on actual player selection
432
+ portfolio['Diversity'] = calculate_player_similarity_score_chunked(portfolio, player_columns)
433
  # check_portfolio = portfolio.copy()
434
  portfolio = portfolio.drop(columns=dup_count_columns)
435
  portfolio = portfolio.drop(columns=own_columns)