F# 3.0 Language Specification - F Sharp

Tokenization. The stream of Unicode characters is broken into a token stream by the lexical analysis described in §3. Lexical Filtering. The token stream is filtered by a state machine that implements the rules described in §15. Those rules describe how additional (artificial) tokens are inserted into the token stream and how some existing ... ................
................