Jump to content

Tokenization (data security)

From Wikipedia, the free encyclopedia
This is an old revision of this page, as edited by 216.94.11.2 (talk) at 10:34, 11 July 2002. The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

Tokenizing is the operation of replacing one set of symbols with another, typically to make the resulting set of symbols smaller.

The term is most commonly used in computers, where a programming language source, a set of symbols in an english-like format, is converted into another format that is much smaller. Most BASIC interpreters used this to save room, a command such as print would be replaced by a single number which uses much less room in memory. In fact most lossless compression systems use a form of tokenizing, although it's typically not referred to as such.