Every line of 'spacy stopwords' code snippets is scanned for vulnerabilities by our powerful machine learning engine that combs millions of open source libraries, ensuring your Python code is secure.
433 def stop_words_stem(self, stop_words=None): 434 if stop_words is not None: 435 stop_words_ = stop_words 436 else: 437 stop_words_ = self.stop_words 438 return list(set([stem(word) for word in stop_words_]))
Secure your code as it's written. Use Snyk Code to scan source code in minutes – no build needed – and fix issues immediately. Enable Snyk Code
73 def remove_stopwords(tokens, language): 74 """ 75 去除中文结束词 76 """ 77 from .stopwords import stopwords 78 79 # 删除结束词 80 tokens = set(tokens) - set(stopwords) 81 82 return tokens
231 def _remove_stop_words(phrase): 232 while len(phrase) > 0 and (phrase[0].is_stop 233 or str(phrase[0]).strip().lower() in Stopwords.get_words()): 234 phrase = phrase[1:] 235 while len(phrase) > 0 and (phrase[-1].is_stop 236 or str(phrase[-1]).strip().lower() in Stopwords.get_words()): 237 phrase = phrase[:-1] 238 return phrase
53 def remove_stopwords(tokens, stopwords=STOPWORDS): 54 """Remove stopwords using list from `gensim.parsing.preprocessing.STOPWORDS`. 55 56 Parameters 57 ---------- 58 tokens : iterable of str 59 Sequence of tokens. 60 stopwords : iterable of str, optional 61 Sequence of stopwords 62 63 Returns 64 ------- 65 list of str 66 List of tokens without `stopwords`. 67 68 """ 69 return [token for token in tokens if token not in stopwords]
18 def spacy_tok(text): 19 return [tok.text for tok in spacy_en.tokenizer(text)]