Remove unused import, fix typo and rewrap docstrings.
This commit is contained in:
parent
7dde792e62
commit
43e4ea1b17
@ -1,12 +1,12 @@
|
|||||||
"""Tokenization help for Python programs.
|
"""Tokenization help for Python programs.
|
||||||
|
|
||||||
tokenize(readline) is a generator that breaks a stream of
|
tokenize(readline) is a generator that breaks a stream of bytes into
|
||||||
bytes into Python tokens. It decodes the bytes according to
|
Python tokens. It decodes the bytes according to PEP-0263 for
|
||||||
PEP-0263 for determining source file encoding.
|
determining source file encoding.
|
||||||
|
|
||||||
It accepts a readline-like method which is called
|
It accepts a readline-like method which is called repeatedly to get the
|
||||||
repeatedly to get the next line of input (or b"" for EOF). It generates
|
next line of input (or b"" for EOF). It generates 5-tuples with these
|
||||||
5-tuples with these members:
|
members:
|
||||||
|
|
||||||
the token type (see token.py)
|
the token type (see token.py)
|
||||||
the token (a string)
|
the token (a string)
|
||||||
@ -16,14 +16,16 @@ repeatedly to get the next line of input (or b"" for EOF). It generates
|
|||||||
|
|
||||||
It is designed to match the working of the Python tokenizer exactly, except
|
It is designed to match the working of the Python tokenizer exactly, except
|
||||||
that it produces COMMENT tokens for comments and gives type OP for all
|
that it produces COMMENT tokens for comments and gives type OP for all
|
||||||
operators. Aditionally, all token lists start with an ENCODING token
|
operators. Additionally, all token lists start with an ENCODING token
|
||||||
which tells you which encoding was used to decode the bytes stream."""
|
which tells you which encoding was used to decode the bytes stream.
|
||||||
|
"""
|
||||||
|
|
||||||
__author__ = 'Ka-Ping Yee <ping@lfw.org>'
|
__author__ = 'Ka-Ping Yee <ping@lfw.org>'
|
||||||
__credits__ = ('GvR, ESR, Tim Peters, Thomas Wouters, Fred Drake, '
|
__credits__ = ('GvR, ESR, Tim Peters, Thomas Wouters, Fred Drake, '
|
||||||
'Skip Montanaro, Raymond Hettinger, Trent Nelson, '
|
'Skip Montanaro, Raymond Hettinger, Trent Nelson, '
|
||||||
'Michael Foord')
|
'Michael Foord')
|
||||||
import re, string, sys
|
import re
|
||||||
|
import sys
|
||||||
from token import *
|
from token import *
|
||||||
from codecs import lookup, BOM_UTF8
|
from codecs import lookup, BOM_UTF8
|
||||||
cookie_re = re.compile("coding[:=]\s*([-\w.]+)")
|
cookie_re = re.compile("coding[:=]\s*([-\w.]+)")
|
||||||
@ -298,17 +300,16 @@ def _get_normal_name(orig_enc):
|
|||||||
def detect_encoding(readline):
|
def detect_encoding(readline):
|
||||||
"""
|
"""
|
||||||
The detect_encoding() function is used to detect the encoding that should
|
The detect_encoding() function is used to detect the encoding that should
|
||||||
be used to decode a Python source file. It requires one argment, readline,
|
be used to decode a Python source file. It requires one argment, readline,
|
||||||
in the same way as the tokenize() generator.
|
in the same way as the tokenize() generator.
|
||||||
|
|
||||||
It will call readline a maximum of twice, and return the encoding used
|
It will call readline a maximum of twice, and return the encoding used
|
||||||
(as a string) and a list of any lines (left as bytes) it has read
|
(as a string) and a list of any lines (left as bytes) it has read in.
|
||||||
in.
|
|
||||||
|
|
||||||
It detects the encoding from the presence of a utf-8 bom or an encoding
|
It detects the encoding from the presence of a utf-8 bom or an encoding
|
||||||
cookie as specified in pep-0263. If both a bom and a cookie are present, but
|
cookie as specified in pep-0263. If both a bom and a cookie are present,
|
||||||
disagree, a SyntaxError will be raised. If the encoding cookie is an invalid
|
but disagree, a SyntaxError will be raised. If the encoding cookie is an
|
||||||
charset, raise a SyntaxError. Note that if a utf-8 bom is found,
|
invalid charset, raise a SyntaxError. Note that if a utf-8 bom is found,
|
||||||
'utf-8-sig' is returned.
|
'utf-8-sig' is returned.
|
||||||
|
|
||||||
If no encoding is specified, then the default of 'utf-8' will be returned.
|
If no encoding is specified, then the default of 'utf-8' will be returned.
|
||||||
@ -372,7 +373,7 @@ def tokenize(readline):
|
|||||||
"""
|
"""
|
||||||
The tokenize() generator requires one argment, readline, which
|
The tokenize() generator requires one argment, readline, which
|
||||||
must be a callable object which provides the same interface as the
|
must be a callable object which provides the same interface as the
|
||||||
readline() method of built-in file objects. Each call to the function
|
readline() method of built-in file objects. Each call to the function
|
||||||
should return one line of input as bytes. Alternately, readline
|
should return one line of input as bytes. Alternately, readline
|
||||||
can be a callable function terminating with StopIteration:
|
can be a callable function terminating with StopIteration:
|
||||||
readline = open(myfile, 'rb').__next__ # Example of alternate readline
|
readline = open(myfile, 'rb').__next__ # Example of alternate readline
|
||||||
@ -381,7 +382,7 @@ def tokenize(readline):
|
|||||||
token string; a 2-tuple (srow, scol) of ints specifying the row and
|
token string; a 2-tuple (srow, scol) of ints specifying the row and
|
||||||
column where the token begins in the source; a 2-tuple (erow, ecol) of
|
column where the token begins in the source; a 2-tuple (erow, ecol) of
|
||||||
ints specifying the row and column where the token ends in the source;
|
ints specifying the row and column where the token ends in the source;
|
||||||
and the line on which the token was found. The line passed is the
|
and the line on which the token was found. The line passed is the
|
||||||
logical line; continuation lines are included.
|
logical line; continuation lines are included.
|
||||||
|
|
||||||
The first token sequence will always be an ENCODING token
|
The first token sequence will always be an ENCODING token
|
||||||
|
Loading…
x
Reference in New Issue
Block a user