From dba0faf125c3a4ded15056129cdd80cfa14f238f Mon Sep 17 00:00:00 2001 From: Nathan TeBlunthuis Date: Sat, 11 Jan 2025 20:02:36 -0800 Subject: [PATCH] bugfix --- similarities/weekly_cosine_similarities.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/similarities/weekly_cosine_similarities.py b/similarities/weekly_cosine_similarities.py index 0963697..862ceb9 100755 --- a/similarities/weekly_cosine_similarities.py +++ b/similarities/weekly_cosine_similarities.py @@ -79,7 +79,7 @@ def cosine_similarities_weekly_lsi(*args, n_components=100, lsi_model=None, **kw return cosine_similarities_weekly(*args, simfunc=simfunc, **kwargs) #tfidf = spark.read.parquet('/gscratch/comdata/users/nathante/competitive_exclusion_reddit/data/tfidf_weekly/comment_submission_terms_tfidf.parquet') -def cosine_similarities_weekly(tfidf_path, outfile, term_colname, included_subreddits = None, topN = None, simfunc=column_similarities, min_df=0, max_df=None, static_tfidf_path=None, clusters_path=None): +def cosine_similarities_weekly(tfidf_path, outfile, term_colname, included_subreddits = None, topN = None, simfunc=column_similarities, min_df=0, max_df=None, static_tfidf_path=None, clusters=None): print(outfile) # do this step in parallel if we have the memory for it. # should be doable with pool.map