Top "Tokenize" questions

Tokenizing is the act of splitting a string into discrete elements called tokens.

Using boost::tokenizer with string delimiters

I've been looking boost::tokenizer, and I've found that the documentation is very thin. Is it possible to make it …

c++ string boost tokenize
Pass tokens to CountVectorizer

I have a text classification problem where i have two types of features: features which are n-grams (extracted by CountVectorizer) …

scikit-learn tokenize
Can a line of Python code know its indentation nesting level?

From something like this: print(get_indentation_level()) print(get_indentation_level()) print(get_indentation_level()) I would like to …

python reflection metaprogramming indentation tokenize