From 39c0fa7a295ed0e79fe053b7e0401e92dff87ee0 Mon Sep 17 00:00:00 2001 From: Nathan TeBlunthuis Date: Tue, 3 Dec 2024 19:18:38 -0800 Subject: [PATCH] bugfix. --- ngrams/term_frequencies.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/ngrams/term_frequencies.py b/ngrams/term_frequencies.py index 28529dc..1253ded 100755 --- a/ngrams/term_frequencies.py +++ b/ngrams/term_frequencies.py @@ -62,7 +62,7 @@ def my_tokenizer(text, mwe_pass, mwe_tokenize, stopWords, ngram_output): for sentence in sentences: if random() <= 0.1: grams = list(chain(*map(lambda i : ngrams(sentence,i),range(4)))) - Path(ngram_output).mkdir(parents=True, exist_ok=True) + Path(ngram_output).parent.mkdir(parents=True, exist_ok=True) with open(ngram_output,'a') as gram_file: for ng in grams: gram_file.write(' '.join(ng) + '\n')