diff --git a/jedi/interpret.py b/jedi/interpret.py index 1e17c05f..ffcf1e53 100644 --- a/jedi/interpret.py +++ b/jedi/interpret.py @@ -5,7 +5,7 @@ Module to handle interpreted Python objects. import itertools from jedi.parser import representation as pr -from jedi.parser import tokenizer as tokenize +from jedi.parser import tokenize from jedi.parser import token diff --git a/jedi/modules.py b/jedi/modules.py index 989b684f..6200918c 100644 --- a/jedi/modules.py +++ b/jedi/modules.py @@ -13,15 +13,13 @@ Basically :mod:`modules` offers the classes: Apart from those classes there's a ``sys.path`` fetching function, as well as `Virtual Env` and `Django` detection. """ -from __future__ import with_statement - import re import sys import os from jedi import cache from jedi.common import source_to_unicode -from jedi.parser import tokenizer as tokenize +from jedi.parser import tokenize from jedi.parser import fast from jedi import debug diff --git a/jedi/parser/__init__.py b/jedi/parser/__init__.py index afb438ce..54706c86 100644 --- a/jedi/parser/__init__.py +++ b/jedi/parser/__init__.py @@ -15,8 +15,6 @@ within the statement. This lowers memory usage and cpu time and reduces the complexity of the ``Parser`` (there's another parser sitting inside ``Statement``, which produces ``Array`` and ``Call``). """ -from __future__ import with_statement - import keyword from jedi._compatibility import next, StringIO @@ -24,7 +22,7 @@ from jedi import debug from jedi import common from jedi.parser import representation as pr from jedi.parser import token as token_pr -from jedi.parser import tokenizer as tokenize +from jedi.parser import tokenize class Parser(object): diff --git a/jedi/parser/fast.py b/jedi/parser/fast.py index 9935acc3..42072cb9 100644 --- a/jedi/parser/fast.py +++ b/jedi/parser/fast.py @@ -9,7 +9,7 @@ from jedi._compatibility import use_metaclass from jedi import settings from jedi.parser import Parser from jedi.parser import representation as pr -from jedi.parser import tokenizer as tokenize +from jedi.parser import tokenize from jedi import cache diff --git a/jedi/parser/representation.py b/jedi/parser/representation.py index aba49c3d..eff00198 100644 --- a/jedi/parser/representation.py +++ b/jedi/parser/representation.py @@ -44,7 +44,7 @@ from jedi._compatibility import next, Python3Method, encoding, unicode, is_py3k from jedi import common from jedi import debug from jedi import cache -from jedi.parser import tokenizer as tokenize +from jedi.parser import tokenize class Base(object): diff --git a/jedi/parser/tokenizer.py b/jedi/parser/tokenize.py similarity index 100% rename from jedi/parser/tokenizer.py rename to jedi/parser/tokenize.py index 36e4da84..da955860 100644 --- a/jedi/parser/tokenizer.py +++ b/jedi/parser/tokenize.py @@ -7,8 +7,8 @@ if the indentation is not right. The fast parser of jedi however requires Basically this is a stripped down version of the standard library module, so you can read the documentation there. """ - from __future__ import absolute_import + import string import re from token import *