A RetroSearch Logo

Home - News ( United States | United Kingdom | Italy | Germany ) - Football scores

Search Query:

Showing content from http://docs.python.org/lib/../c-api/../tutorial/../tutorial/../library/token.html below:

token — Constants used with Python parse trees — Python 3.13.5 documentation

token — Constants used with Python parse trees¶

Source code: Lib/token.py

This module provides constants which represent the numeric values of leaf nodes of the parse tree (terminal tokens). Refer to the file Grammar/Tokens in the Python distribution for the definitions of the names in the context of the language grammar. The specific numeric values which the names map to may change between Python versions.

The module also provides a mapping from numeric codes to names and some functions. The functions mirror definitions in the Python C header files.

Note that a token’s value may depend on tokenizer options. For example, a "+" token may be reported as either PLUS or OP, or a "match" token may be either NAME or SOFT_KEYWORD.

token.tok_name¶

Dictionary mapping the numeric values of the constants defined in this module back to name strings, allowing more human-readable representation of parse trees to be generated.

token.ISTERMINAL(x)¶

Return True for terminal token values.

token.ISNONTERMINAL(x)¶

Return True for non-terminal token values.

token.ISEOF(x)¶

Return True if x is the marker indicating the end of input.

The token constants are:

token.NAME¶

Token value that indicates an identifier. Note that keywords are also initially tokenized as NAME tokens.

token.NUMBER¶

Token value that indicates a numeric literal

token.STRING¶

Token value that indicates a string or byte literal, excluding formatted string literals. The token string is not interpreted: it includes the surrounding quotation marks and the prefix (if given); backslashes are included literally, without processing escape sequences.

token.OP¶

A generic token value that indicates an operator or delimiter.

CPython implementation detail: This value is only reported by the tokenize module. Internally, the tokenizer uses exact token types instead.

Token value used to indicate a comment. The parser ignores COMMENT tokens.

token.NEWLINE¶

Token value that indicates the end of a logical line.

token.NL¶

Token value used to indicate a non-terminating newline. NL tokens are generated when a logical line of code is continued over multiple physical lines. The parser ignores NL tokens.

token.INDENT¶

Token value used at the beginning of a logical line to indicate the start of an indented block.

token.DEDENT¶

Token value used at the beginning of a logical line to indicate the end of an indented block.

token.FSTRING_START¶

Token value used to indicate the beginning of an f-string literal.

CPython implementation detail: The token string includes the prefix and the opening quote(s), but none of the contents of the literal.

token.FSTRING_MIDDLE¶

Token value used for literal text inside an f-string literal, including format specifications.

CPython implementation detail: Replacement fields (that is, the non-literal parts of f-strings) use the same tokens as other expressions, and are delimited by LBRACE, RBRACE, EXCLAMATION and COLON tokens.

token.FSTRING_END¶

Token value used to indicate the end of a f-string.

CPython implementation detail: The token string contains the closing quote(s).

token.ENDMARKER¶

Token value that indicates the end of input.

token.ENCODING¶

Token value that indicates the encoding used to decode the source bytes into text. The first token returned by tokenize.tokenize() will always be an ENCODING token.

CPython implementation detail: This token type isn’t used by the C tokenizer but is needed for the tokenize module.

The following token types are not produced by the tokenize module, and are defined for special uses in the tokenizer or parser:

token.TYPE_IGNORE¶

Token value indicating that a type: ignore comment was recognized. Such tokens are produced instead of regular COMMENT tokens only with the PyCF_TYPE_COMMENTS flag.

Token value indicating that a type comment was recognized. Such tokens are produced instead of regular COMMENT tokens only with the PyCF_TYPE_COMMENTS flag.

token.SOFT_KEYWORD¶

Token value indicating a soft keyword.

The tokenizer never produces this value. To check for a soft keyword, pass a NAME token’s string to keyword.issoftkeyword().

token.ERRORTOKEN¶

Token value used to indicate wrong input.

The tokenize module generally indicates errors by raising exceptions instead of emitting this token. It can also emit tokens such as OP or NAME with strings that are later rejected by the parser.

The remaining tokens represent specific operators and delimiters. (The tokenize module reports these as OP; see exact_type in the tokenize documentation for details.)

Token

Value

token.LPAR¶

"("

token.RPAR¶

")"

token.LSQB¶

"["

token.RSQB¶

"]"

token.COLON¶

":"

token.COMMA¶

","

token.SEMI¶

";"

token.PLUS¶

"+"

token.MINUS¶

"-"

token.STAR¶

"*"

token.SLASH¶

"/"

token.VBAR¶

"|"

token.AMPER¶

"&"

token.LESS¶

"<"

token.GREATER¶

">"

token.EQUAL¶

"="

token.DOT¶

"."

token.PERCENT¶

"%"

token.LBRACE¶

"{"

token.RBRACE¶

"}"

token.EQEQUAL¶

"=="

token.NOTEQUAL¶

"!="

token.LESSEQUAL¶

"<="

token.GREATEREQUAL¶

">="

token.TILDE¶

"~"

token.CIRCUMFLEX¶

"^"

token.LEFTSHIFT¶

"<<"

token.RIGHTSHIFT¶

">>"

token.DOUBLESTAR¶

"**"

token.PLUSEQUAL¶

"+="

token.MINEQUAL¶

"-="

token.STAREQUAL¶

"*="

token.SLASHEQUAL¶

"/="

token.PERCENTEQUAL¶

"%="

token.AMPEREQUAL¶

"&="

token.VBAREQUAL¶

"|="

token.CIRCUMFLEXEQUAL¶

"^="

token.LEFTSHIFTEQUAL¶

"<<="

token.RIGHTSHIFTEQUAL¶

">>="

token.DOUBLESTAREQUAL¶

"**="

token.DOUBLESLASH¶

"//"

token.DOUBLESLASHEQUAL¶

"//="

token.AT¶

"@"

token.ATEQUAL¶

"@="

token.RARROW¶

"->"

token.ELLIPSIS¶

"..."

token.COLONEQUAL¶

":="

token.EXCLAMATION¶

"!"

The following non-token constants are provided:

token.N_TOKENS¶

The number of token types defined in this module.

token.EXACT_TOKEN_TYPES¶

A dictionary mapping the string representation of a token to its numeric code.

Added in version 3.8.

Changed in version 3.5: Added AWAIT and ASYNC tokens.

Changed in version 3.7: Added COMMENT, NL and ENCODING tokens.

Changed in version 3.7: Removed AWAIT and ASYNC tokens. “async” and “await” are now tokenized as NAME tokens.

Changed in version 3.8: Added TYPE_COMMENT, TYPE_IGNORE, COLONEQUAL. Added AWAIT and ASYNC tokens back (they’re needed to support parsing older Python versions for ast.parse() with feature_version set to 6 or lower).

Changed in version 3.12: Added EXCLAMATION.

Changed in version 3.13: Removed AWAIT and ASYNC tokens again.


RetroSearch is an open source project built by @garambo | Open a GitHub Issue

Search and Browse the WWW like it's 1997 | Search results from DuckDuckGo

HTML: 3.2 | Encoding: UTF-8 | Version: 0.7.4