Overview - Token filters (lowercase, stemmer, synonym)
What is it?
Token filters are steps in text processing that change or refine words (tokens) after they are split from text. Common filters include lowercase, which makes all letters small; stemmer, which cuts words to their root form; and synonym, which replaces words with their equivalent meanings. These filters help search engines understand and match words better.
Why it matters
Without token filters, search engines would treat words like 'Running' and 'run' as completely different, missing relevant results. This would make searching frustrating and less useful. Token filters solve this by normalizing words, so searches find more helpful matches, improving user experience and accuracy.
Where it fits
Before learning token filters, you should understand how text is broken into tokens (tokenization). After mastering token filters, you can explore more advanced text analysis like custom analyzers and query tuning in Elasticsearch.