removed the now redundant token.Token version in favor of tokenize.TokenInfo

This commit is contained in:
Dave Halter
2014-02-25 16:38:19 +01:00
parent 2db26abf72
commit 6439d6c848
2 changed files with 5 additions and 118 deletions

View File

@@ -318,9 +318,7 @@ class Parser(object):
or tok.string in breaks and level <= 0):
try:
# print 'parse_stmt', tok, tokenize.tok_name[token_type]
tok_list.append(
token_pr.Token.from_tuple(self._gen.current)
)
tok_list.append(tok)
if tok.string == 'as':
tok = next(self._gen)
if tok.type == tokenize.NAME:

View File

@@ -8,124 +8,13 @@ found that a flat object with slots is the best.
from inspect import cleandoc
from ast import literal_eval
from jedi._compatibility import u, unicode
from jedi._compatibility import unicode
from jedi.parser.tokenize import TokenInfo
class Token(object):
"""The token object is an efficient representation of the structure
(type, token, (start_pos_line, start_pos_col)). It has indexer
methods that maintain compatibility to existing code that expects the above
structure.
>>> tuple(Token(1,2,3,4))
(1, 2, (3, 4))
>>> repr(Token(1, "test", 1, 1))
"<Token: (1, 'test', (1, 1))>"
>>> Token(1, 2, 3, 4).__getstate__()
(1, 2, 3, 4)
>>> a = Token(0, 0, 0, 0)
>>> a.__setstate__((1, 2, 3, 4))
>>> a
<Token: (1, 2, (3, 4))>
>>> a.start_pos
(3, 4)
>>> a.token
2
>>> a.start_pos_col
4
>>> Token.from_tuple((6, 5, (4, 3)))
<Token: (6, 5, (4, 3))>
>>> Token(1, u("😷"), 1 ,1).string + "p" == u("😷p")
True
"""
__slots__ = ("_type", "string", "_start_pos_line", "_start_pos_col")
@classmethod
def from_tuple(cls, tp):
return Token(tp[0], tp[1], tp[2][0], tp[2][1])
def __init__(
self, type, token, start_pos_line, start_pos_col
):
self._type = type
self.string = token
self._start_pos_line = start_pos_line
self._start_pos_col = start_pos_col
def __repr__(self):
return "<%s: %s>" % (type(self).__name__, tuple(self))
# Backward compatibility
def __getitem__(self, key):
# Builds the same structure as tuple used to have
if key == 0:
return self.type
elif key == 1:
return self.token
elif key == 2:
return (self.start_pos_line, self.start_pos_col)
else:
raise IndexError("list index out of range")
@property
def type(self):
return self._type
@property
def token(self):
return self.string
@property
def start_pos_line(self):
return self._start_pos_line
@property
def start_pos_col(self):
return self._start_pos_col
@property
def start_pos(self):
return (self.start_pos_line, self.start_pos_col)
@property
def end_pos(self):
"""Returns end position respecting multiline tokens."""
end_pos_line = self.start_pos_line
lines = self.string.split('\n')
end_pos_line += len(lines) - 1
end_pos_col = self.start_pos_col
# Check for multiline token
if self.start_pos_line == end_pos_line:
end_pos_col += len(lines[-1])
else:
end_pos_col = len(lines[-1])
return (end_pos_line, end_pos_col)
# Make cache footprint smaller for faster unpickling
def __getstate__(self):
return (
self.type,
self.token,
self.start_pos_line,
self.start_pos_col,
)
def __setstate__(self, state):
self._type = state[0]
self.string = state[1]
self._start_pos_line = state[2]
self._start_pos_col = state[3]
class TokenNoCompat(Token):
class TokenNoCompat(TokenInfo):
__slots__ = ()
def __unicode__(self):
raise NotImplementedError("Compatibility only for basic token.")
def __str__(self):
raise NotImplementedError("Compatibility only for basic token.")
def __getitem__(self, key):
raise NotImplementedError("Compatibility only for basic token.")
@@ -150,4 +39,4 @@ class TokenDocstring(TokenNoCompat):
def as_string(self):
"""Returns a literal cleaned version of the token"""
return unicode(cleandoc(literal_eval(self.token)))
return unicode(cleandoc(literal_eval(self.string)))