Tokenization
IntermediateConverting text into discrete units (tokens) for modeling; subword tokenizers balance vocabulary size and coverage.
AdvertisementAd space — term-top
Definition
Full Definition
Converting text into discrete units (tokens) for modeling; subword tokenizers balance vocabulary size and coverage.