BERT tokenization breaks words into smaller pieces called tokens. The key metric to check is tokenization coverage, which shows how well the tokenizer splits words into known pieces. Good coverage means fewer unknown tokens, helping the model understand text better.
Another important metric is tokenization consistency, ensuring the same word is split the same way every time. This helps the model learn stable word meanings.