diff --git a/ngrams/term_frequencies.py b/ngrams/term_frequencies.py index f941d1a..28529dc 100755 --- a/ngrams/term_frequencies.py +++ b/ngrams/term_frequencies.py @@ -63,7 +63,7 @@ def my_tokenizer(text, mwe_pass, mwe_tokenize, stopWords, ngram_output): if random() <= 0.1: grams = list(chain(*map(lambda i : ngrams(sentence,i),range(4)))) Path(ngram_output).mkdir(parents=True, exist_ok=True) - with open(ngram_output','a') as gram_file: + with open(ngram_output,'a') as gram_file: for ng in grams: gram_file.write(' '.join(ng) + '\n') for token in sentence: