1 Commits
v0.18.2 ... tmp

Author SHA1 Message Date
Dave Halter
354dab9503 Debug 2021-01-02 13:48:54 +01:00
88 changed files with 424 additions and 1128 deletions

View File

@@ -1,5 +1,5 @@
name: ci name: ci
on: [push, pull_request] on: push
jobs: jobs:
tests: tests:
@@ -7,8 +7,8 @@ jobs:
strategy: strategy:
matrix: matrix:
os: [ubuntu-20.04, windows-2019] os: [ubuntu-20.04, windows-2019]
python-version: ["3.10", "3.9", "3.8", "3.7", "3.6"] python-version: [3.9, 3.8, 3.7, 3.6]
environment: ['3.8', '3.10', '3.9', '3.7', '3.6', 'interpreter'] environment: ['3.8', '3.9', '3.7', '3.6', 'interpreter']
steps: steps:
- name: Checkout code - name: Checkout code
uses: actions/checkout@v2 uses: actions/checkout@v2
@@ -27,6 +27,9 @@ jobs:
- name: Install dependencies - name: Install dependencies
run: 'pip install .[testing]' run: 'pip install .[testing]'
- name: Setup tmate session
uses: mxschmitt/action-tmate@v3
- name: Run tests - name: Run tests
run: python -m pytest run: python -m pytest
env: env:

1
.gitignore vendored
View File

@@ -14,4 +14,3 @@ record.json
/.pytest_cache /.pytest_cache
/.mypy_cache /.mypy_cache
/venv/ /venv/
.nvimrc

View File

@@ -1,11 +1,2 @@
version: 2
python: python:
install: pip_install: true
- method: pip
path: .
extra_requirements:
- docs
submodules:
include: all

View File

@@ -1,4 +1,4 @@
Main Authors Main Authors
------------ ------------
- David Halter (@davidhalter) <davidhalter88@gmail.com> - David Halter (@davidhalter) <davidhalter88@gmail.com>
@@ -61,8 +61,6 @@ Code Contributors
- Vladislav Serebrennikov (@endilll) - Vladislav Serebrennikov (@endilll)
- Andrii Kolomoiets (@muffinmad) - Andrii Kolomoiets (@muffinmad)
- Leo Ryu (@Leo-Ryu) - Leo Ryu (@Leo-Ryu)
- Joseph Birkner (@josephbirkner)
- Márcio Mazza (@marciomazza)
And a few more "anonymous" contributors. And a few more "anonymous" contributors.

View File

@@ -6,22 +6,7 @@ Changelog
Unreleased Unreleased
++++++++++ ++++++++++
0.18.2 (2022-11-21)
+++++++++++++++++++
- Added dataclass-equivalent for attrs.define
- Find fixtures from Pytest entrypoints; Examples of pytest plugins installed
like this are pytest-django, pytest-sugar and Faker.
- Fixed Project.search, when a venv was involved, which is why for example
`:Pyimport django.db` did not work in some cases in jedi-vim.
- And many smaller bugfixes
0.18.1 (2021-11-17)
+++++++++++++++++++
- Implict namespaces are now a separate types in ``Name().type`` - Implict namespaces are now a separate types in ``Name().type``
- Python 3.10 support
- Mostly bugfixes
0.18.0 (2020-12-25) 0.18.0 (2020-12-25)
+++++++++++++++++++ +++++++++++++++++++

View File

@@ -57,7 +57,7 @@ Supported Python Features
Limitations Limitations
----------- -----------
In general Jedi's limit is quite high, but for very big projects or very In general Jedi's limit are quite high, but for very big projects or very
complex code, sometimes Jedi intentionally stops type inference, to avoid complex code, sometimes Jedi intentionally stops type inference, to avoid
hanging for a long time. hanging for a long time.

View File

@@ -27,7 +27,7 @@ ad
load load
""" """
__version__ = '0.18.2' __version__ = '0.18.0'
from jedi.api import Script, Interpreter, set_debug_function, preload_module from jedi.api import Script, Interpreter, set_debug_function, preload_module
from jedi import settings from jedi import settings

View File

@@ -7,6 +7,22 @@ import sys
import pickle import pickle
def cast_path(string):
"""
Take a bytes or str path and cast it to unicode.
Apparently it is perfectly fine to pass both byte and unicode objects into
the sys.path. This probably means that byte paths are normal at other
places as well.
Since this just really complicates everything and Python 2.7 will be EOL
soon anyway, just go with always strings.
"""
if isinstance(string, bytes):
return str(string, encoding='UTF-8', errors='replace')
return str(string)
def pickle_load(file): def pickle_load(file):
try: try:
return pickle.load(file) return pickle.load(file)

View File

@@ -13,6 +13,7 @@ from pathlib import Path
import parso import parso
from parso.python import tree from parso.python import tree
from jedi._compatibility import cast_path
from jedi.parser_utils import get_executable_nodes from jedi.parser_utils import get_executable_nodes
from jedi import debug from jedi import debug
from jedi import settings from jedi import settings
@@ -99,15 +100,13 @@ class Script:
""" """
def __init__(self, code=None, *, path=None, environment=None, project=None): def __init__(self, code=None, *, path=None, environment=None, project=None):
self._orig_path = path self._orig_path = path
# An empty path (also empty string) should always result in no path.
if isinstance(path, str): if isinstance(path, str):
path = Path(path) path = Path(path)
self.path = path.absolute() if path else None self.path = path.absolute() if path else None
if code is None: if code is None:
if path is None:
raise ValueError("Must provide at least one of code or path")
# TODO add a better warning than the traceback! # TODO add a better warning than the traceback!
with open(path, 'rb') as f: with open(path, 'rb') as f:
code = f.read() code = f.read()
@@ -153,7 +152,7 @@ class Script:
if self.path is None: if self.path is None:
file_io = None file_io = None
else: else:
file_io = KnownContentFileIO(self.path, self._code) file_io = KnownContentFileIO(cast_path(self.path), self._code)
if self.path is not None and self.path.suffix == '.pyi': if self.path is not None and self.path.suffix == '.pyi':
# We are in a stub file. Try to load the stub properly. # We are in a stub file. Try to load the stub properly.
stub_module = load_proper_stub_module( stub_module = load_proper_stub_module(
@@ -581,7 +580,7 @@ class Script:
@validate_line_column @validate_line_column
def extract_variable(self, line, column, *, new_name, until_line=None, until_column=None): def extract_variable(self, line, column, *, new_name, until_line=None, until_column=None):
""" """
Moves an expression to a new statement. Moves an expression to a new statemenet.
For example if you have the cursor on ``foo`` and provide a For example if you have the cursor on ``foo`` and provide a
``new_name`` called ``bar``:: ``new_name`` called ``bar``::
@@ -710,7 +709,7 @@ class Interpreter(Script):
""" """
_allow_descriptor_getattr_default = True _allow_descriptor_getattr_default = True
def __init__(self, code, namespaces, *, project=None, **kwds): def __init__(self, code, namespaces, **kwds):
try: try:
namespaces = [dict(n) for n in namespaces] namespaces = [dict(n) for n in namespaces]
except Exception: except Exception:
@@ -723,23 +722,16 @@ class Interpreter(Script):
if not isinstance(environment, InterpreterEnvironment): if not isinstance(environment, InterpreterEnvironment):
raise TypeError("The environment needs to be an InterpreterEnvironment subclass.") raise TypeError("The environment needs to be an InterpreterEnvironment subclass.")
if project is None: super().__init__(code, environment=environment,
project = Project(Path.cwd()) project=Project(Path.cwd()), **kwds)
super().__init__(code, environment=environment, project=project, **kwds)
self.namespaces = namespaces self.namespaces = namespaces
self._inference_state.allow_descriptor_getattr = self._allow_descriptor_getattr_default self._inference_state.allow_descriptor_getattr = self._allow_descriptor_getattr_default
@cache.memoize_method @cache.memoize_method
def _get_module_context(self): def _get_module_context(self):
if self.path is None:
file_io = None
else:
file_io = KnownContentFileIO(self.path, self._code)
tree_module_value = ModuleValue( tree_module_value = ModuleValue(
self._inference_state, self._module_node, self._inference_state, self._module_node,
file_io=file_io, file_io=KnownContentFileIO(str(self.path), self._code),
string_names=('__main__',), string_names=('__main__',),
code_lines=self._code_lines, code_lines=self._code_lines,
) )

View File

@@ -27,7 +27,7 @@ from jedi.inference.compiled.mixed import MixedName
from jedi.inference.names import ImportName, SubModuleName from jedi.inference.names import ImportName, SubModuleName
from jedi.inference.gradual.stub_value import StubModuleValue from jedi.inference.gradual.stub_value import StubModuleValue
from jedi.inference.gradual.conversion import convert_names, convert_values from jedi.inference.gradual.conversion import convert_names, convert_values
from jedi.inference.base_value import ValueSet, HasNoContext from jedi.inference.base_value import ValueSet
from jedi.api.keywords import KeywordName from jedi.api.keywords import KeywordName
from jedi.api import completion_cache from jedi.api import completion_cache
from jedi.api.helpers import filter_follow_imports from jedi.api.helpers import filter_follow_imports
@@ -37,17 +37,13 @@ def _sort_names_by_start_pos(names):
return sorted(names, key=lambda s: s.start_pos or (0, 0)) return sorted(names, key=lambda s: s.start_pos or (0, 0))
def defined_names(inference_state, value): def defined_names(inference_state, context):
""" """
List sub-definitions (e.g., methods in class). List sub-definitions (e.g., methods in class).
:type scope: Scope :type scope: Scope
:rtype: list of Name :rtype: list of Name
""" """
try:
context = value.as_context()
except HasNoContext:
return []
filter = next(context.get_filters()) filter = next(context.get_filters())
names = [name for name in filter.values()] names = [name for name in filter.values()]
return [Name(inference_state, n) for n in _sort_names_by_start_pos(names)] return [Name(inference_state, n) for n in _sort_names_by_start_pos(names)]
@@ -763,7 +759,7 @@ class Name(BaseName):
""" """
defs = self._name.infer() defs = self._name.infer()
return sorted( return sorted(
unite(defined_names(self._inference_state, d) for d in defs), unite(defined_names(self._inference_state, d.as_context()) for d in defs),
key=lambda s: s._name.start_pos or (0, 0) key=lambda s: s._name.start_pos or (0, 0)
) )

View File

@@ -18,8 +18,7 @@ from jedi.inference import imports
from jedi.inference.base_value import ValueSet from jedi.inference.base_value import ValueSet
from jedi.inference.helpers import infer_call_of_leaf, parse_dotted_names from jedi.inference.helpers import infer_call_of_leaf, parse_dotted_names
from jedi.inference.context import get_global_filters from jedi.inference.context import get_global_filters
from jedi.inference.value import TreeInstance from jedi.inference.value import TreeInstance, ModuleValue
from jedi.inference.docstring_utils import DocstringModule
from jedi.inference.names import ParamNameWrapper, SubModuleName from jedi.inference.names import ParamNameWrapper, SubModuleName
from jedi.inference.gradual.conversion import convert_values, convert_names from jedi.inference.gradual.conversion import convert_values, convert_names
from jedi.parser_utils import cut_value_at_position from jedi.parser_utils import cut_value_at_position
@@ -195,6 +194,7 @@ class Completion:
- In args: */**: no completion - In args: */**: no completion
- In params (also lambda): no completion before = - In params (also lambda): no completion before =
""" """
grammar = self._inference_state.grammar grammar = self._inference_state.grammar
self.stack = stack = None self.stack = stack = None
self._position = ( self._position = (
@@ -277,10 +277,6 @@ class Completion:
) )
elif nonterminals[-1] in ('trailer', 'dotted_name') and nodes[-1] == '.': elif nonterminals[-1] in ('trailer', 'dotted_name') and nodes[-1] == '.':
dot = self._module_node.get_leaf_for_position(self._position) dot = self._module_node.get_leaf_for_position(self._position)
if dot.type == "endmarker":
# This is a bit of a weird edge case, maybe we can somehow
# generalize this.
dot = leaf.get_previous_leaf()
cached_name, n = self._complete_trailer(dot.get_previous_leaf()) cached_name, n = self._complete_trailer(dot.get_previous_leaf())
completion_names += n completion_names += n
elif self._is_parameter_completion(): elif self._is_parameter_completion():
@@ -466,12 +462,12 @@ class Completion:
def _complete_code_lines(self, code_lines): def _complete_code_lines(self, code_lines):
module_node = self._inference_state.grammar.parse(''.join(code_lines)) module_node = self._inference_state.grammar.parse(''.join(code_lines))
module_value = DocstringModule( module_value = ModuleValue(
in_module_context=self._module_context, self._inference_state,
inference_state=self._inference_state, module_node,
module_node=module_node,
code_lines=code_lines, code_lines=code_lines,
) )
module_value.parent_context = self._module_context
return Completion( return Completion(
self._inference_state, self._inference_state,
module_value.as_context(), module_value.as_context(),

View File

@@ -17,7 +17,7 @@ import parso
_VersionInfo = namedtuple('VersionInfo', 'major minor micro') _VersionInfo = namedtuple('VersionInfo', 'major minor micro')
_SUPPORTED_PYTHONS = ['3.10', '3.9', '3.8', '3.7', '3.6'] _SUPPORTED_PYTHONS = ['3.9', '3.8', '3.7', '3.6']
_SAFE_PATHS = ['/usr/bin', '/usr/local/bin'] _SAFE_PATHS = ['/usr/bin', '/usr/local/bin']
_CONDA_VAR = 'CONDA_PREFIX' _CONDA_VAR = 'CONDA_PREFIX'
_CURRENT_VERSION = '%s.%s' % (sys.version_info.major, sys.version_info.minor) _CURRENT_VERSION = '%s.%s' % (sys.version_info.major, sys.version_info.minor)

View File

@@ -23,7 +23,7 @@ class RefactoringError(_JediError):
Refactorings can fail for various reasons. So if you work with refactorings Refactorings can fail for various reasons. So if you work with refactorings
like :meth:`.Script.rename`, :meth:`.Script.inline`, like :meth:`.Script.rename`, :meth:`.Script.inline`,
:meth:`.Script.extract_variable` and :meth:`.Script.extract_function`, make :meth:`.Script.extract_variable` and :meth:`.Script.extract_function`, make
sure to catch these. The descriptions in the errors are usually valuable sure to catch these. The descriptions in the errors are ususally valuable
for end users. for end users.
A typical ``RefactoringError`` would tell the user that inlining is not A typical ``RefactoringError`` would tell the user that inlining is not

View File

@@ -205,6 +205,7 @@ def filter_follow_imports(names, follow_builtin_imports=False):
class CallDetails: class CallDetails:
def __init__(self, bracket_leaf, children, position): def __init__(self, bracket_leaf, children, position):
['bracket_leaf', 'call_index', 'keyword_name_str']
self.bracket_leaf = bracket_leaf self.bracket_leaf = bracket_leaf
self._children = children self._children = children
self._position = position self._position = position
@@ -280,7 +281,7 @@ class CallDetails:
def count_positional_arguments(self): def count_positional_arguments(self):
count = 0 count = 0
for star_count, key_start, had_equal in self._list_arguments()[:-1]: for star_count, key_start, had_equal in self._list_arguments()[:-1]:
if star_count or key_start: if star_count:
break break
count += 1 count += 1
return count return count
@@ -306,7 +307,7 @@ def _iter_arguments(nodes, position):
first = node.children[0] first = node.children[0]
second = node.children[1] second = node.children[1]
if second == '=': if second == '=':
if second.start_pos < position and first.type == 'name': if second.start_pos < position:
yield 0, first.value, True yield 0, first.value, True
else: else:
yield 0, remove_after_pos(first), False yield 0, remove_after_pos(first), False

View File

@@ -106,16 +106,7 @@ class Project:
with open(self._get_json_path(self._path), 'w') as f: with open(self._get_json_path(self._path), 'w') as f:
return json.dump((_SERIALIZER_VERSION, data), f) return json.dump((_SERIALIZER_VERSION, data), f)
def __init__( def __init__(self, path, **kwargs):
self,
path,
*,
environment_path=None,
load_unsafe_extensions=False,
sys_path=None,
added_sys_path=(),
smart_sys_path=True,
) -> None:
""" """
:param path: The base path for this project. :param path: The base path for this project.
:param environment_path: The Python executable path, typically the path :param environment_path: The Python executable path, typically the path
@@ -134,22 +125,25 @@ class Project:
local directories. Otherwise you will have to rely on your packages local directories. Otherwise you will have to rely on your packages
being properly configured on the ``sys.path``. being properly configured on the ``sys.path``.
""" """
def py2_comp(path, environment_path=None, load_unsafe_extensions=False,
sys_path=None, added_sys_path=(), smart_sys_path=True):
if isinstance(path, str):
path = Path(path).absolute()
self._path = path
if isinstance(path, str): self._environment_path = environment_path
path = Path(path).absolute() if sys_path is not None:
self._path = path # Remap potential pathlib.Path entries
sys_path = list(map(str, sys_path))
self._environment_path = environment_path self._sys_path = sys_path
if sys_path is not None: self._smart_sys_path = smart_sys_path
self._load_unsafe_extensions = load_unsafe_extensions
self._django = False
# Remap potential pathlib.Path entries # Remap potential pathlib.Path entries
sys_path = list(map(str, sys_path)) self.added_sys_path = list(map(str, added_sys_path))
self._sys_path = sys_path """The sys path that is going to be added at the end of the """
self._smart_sys_path = smart_sys_path
self._load_unsafe_extensions = load_unsafe_extensions py2_comp(path, **kwargs)
self._django = False
# Remap potential pathlib.Path entries
self.added_sys_path = list(map(str, added_sys_path))
"""The sys path that is going to be added at the end of the """
@property @property
def path(self): def path(self):
@@ -334,8 +328,7 @@ class Project:
) )
# 2. Search for identifiers in the project. # 2. Search for identifiers in the project.
for module_context in search_in_file_ios(inference_state, file_ios, for module_context in search_in_file_ios(inference_state, file_ios, name):
name, complete=complete):
names = get_module_names(module_context.tree_node, all_scopes=all_scopes) names = get_module_names(module_context.tree_node, all_scopes=all_scopes)
names = [module_context.create_name(n) for n in names] names = [module_context.create_name(n) for n in names]
names = _remove_imports(names) names = _remove_imports(names)
@@ -352,8 +345,9 @@ class Project:
# 3. Search for modules on sys.path # 3. Search for modules on sys.path
sys_path = [ sys_path = [
p for p in self._get_sys_path(inference_state) p for p in self._get_sys_path(inference_state)
# Exclude the current folder which is handled by recursing the folders. # Exclude folders that are handled by recursing of the Python
if p != self._path # folders.
if not p.startswith(str(self._path))
] ]
names = list(iter_module_names(inference_state, empty_module_context, sys_path)) names = list(iter_module_names(inference_state, empty_module_context, sys_path))
yield from search_in_module( yield from search_in_module(
@@ -432,6 +426,7 @@ def get_default_project(path=None):
probable_path = dir probable_path = dir
if probable_path is not None: if probable_path is not None:
# TODO search for setup.py etc
return Project(probable_path) return Project(probable_path)
if first_no_init_file is not None: if first_no_init_file is not None:

View File

@@ -42,17 +42,11 @@ class ChangedFile:
if self._from_path is None: if self._from_path is None:
from_p = '' from_p = ''
else: else:
try: from_p = self._from_path.relative_to(project_path)
from_p = self._from_path.relative_to(project_path)
except ValueError: # Happens it the path is not on th project_path
from_p = self._from_path
if self._to_path is None: if self._to_path is None:
to_p = '' to_p = ''
else: else:
try: to_p = self._to_path.relative_to(project_path)
to_p = self._to_path.relative_to(project_path)
except ValueError:
to_p = self._to_path
diff = difflib.unified_diff( diff = difflib.unified_diff(
old_lines, new_lines, old_lines, new_lines,
fromfile=str(from_p), fromfile=str(from_p),

View File

@@ -106,7 +106,10 @@ def dbg(message, *args, color='GREEN'):
debug_function(color, i + 'dbg: ' + message % tuple(repr(a) for a in args)) debug_function(color, i + 'dbg: ' + message % tuple(repr(a) for a in args))
def warning(message, *args, format=True): def warning(message, *args, **kwargs):
format = kwargs.pop('format', True)
assert not kwargs
if debug_function and enable_warning: if debug_function and enable_warning:
i = ' ' * _debug_indent i = ' ' * _debug_indent
if format: if format:

View File

@@ -90,7 +90,7 @@ class InferenceState:
self.compiled_subprocess = environment.get_inference_state_subprocess(self) self.compiled_subprocess = environment.get_inference_state_subprocess(self)
self.grammar = environment.get_grammar() self.grammar = environment.get_grammar()
self.latest_grammar = parso.load_grammar(version='3.10') self.latest_grammar = parso.load_grammar(version='3.7')
self.memoize_cache = {} # for memoize decorators self.memoize_cache = {} # for memoize decorators
self.module_cache = imports.ModuleCache() # does the job of `sys.modules`. self.module_cache = imports.ModuleCache() # does the job of `sys.modules`.
self.stub_module_cache = {} # Dict[Tuple[str, ...], Optional[ModuleValue]] self.stub_module_cache = {} # Dict[Tuple[str, ...], Optional[ModuleValue]]
@@ -181,6 +181,8 @@ class InferenceState:
def parse_and_get_code(self, code=None, path=None, def parse_and_get_code(self, code=None, path=None,
use_latest_grammar=False, file_io=None, **kwargs): use_latest_grammar=False, file_io=None, **kwargs):
if path is not None:
path = str(path)
if code is None: if code is None:
if file_io is None: if file_io is None:
file_io = FileIO(path) file_io = FileIO(path)

View File

@@ -22,10 +22,6 @@ from jedi.cache import memoize_method
sentinel = object() sentinel = object()
class HasNoContext(Exception):
pass
class HelperValueMixin: class HelperValueMixin:
def get_root_context(self): def get_root_context(self):
value = self value = self
@@ -265,7 +261,7 @@ class Value(HelperValueMixin):
return self.parent_context.is_stub() return self.parent_context.is_stub()
def _as_context(self): def _as_context(self):
raise HasNoContext raise NotImplementedError('Not all values need to be converted to contexts: %s', self)
@property @property
def name(self): def name(self):
@@ -297,7 +293,7 @@ class Value(HelperValueMixin):
just the `_T` generic parameter. just the `_T` generic parameter.
`value_set`: represents the actual argument passed to the parameter `value_set`: represents the actual argument passed to the parameter
we're inferred for, or (for recursive calls) their types. In the we're inferrined for, or (for recursive calls) their types. In the
above example this would first be the representation of the list above example this would first be the representation of the list
`[1]` and then, when recursing, just of `1`. `[1]` and then, when recursing, just of `1`.
""" """

View File

@@ -8,8 +8,6 @@ import warnings
import re import re
import builtins import builtins
import typing import typing
from pathlib import Path
from typing import Optional
from jedi.inference.compiled.getattr_static import getattr_static from jedi.inference.compiled.getattr_static import getattr_static
@@ -181,9 +179,9 @@ class DirectObjectAccess:
def py__bool__(self): def py__bool__(self):
return bool(self._obj) return bool(self._obj)
def py__file__(self) -> Optional[Path]: def py__file__(self):
try: try:
return Path(self._obj.__file__) return self._obj.__file__
except AttributeError: except AttributeError:
return None return None
@@ -213,22 +211,7 @@ class DirectObjectAccess:
def py__getitem__all_values(self): def py__getitem__all_values(self):
if isinstance(self._obj, dict): if isinstance(self._obj, dict):
return [self._create_access_path(v) for v in self._obj.values()] return [self._create_access_path(v) for v in self._obj.values()]
if isinstance(self._obj, (list, tuple)): return self.py__iter__list()
return [self._create_access_path(v) for v in self._obj]
if self.is_instance():
cls = DirectObjectAccess(self._inference_state, self._obj.__class__)
return cls.py__getitem__all_values()
try:
getitem = self._obj.__getitem__
except AttributeError:
pass
else:
annotation = DirectObjectAccess(self._inference_state, getitem).get_return_annotation()
if annotation is not None:
return [annotation]
return None
def py__simple_getitem__(self, index): def py__simple_getitem__(self, index):
if type(self._obj) not in ALLOWED_GETITEM_TYPES: if type(self._obj) not in ALLOWED_GETITEM_TYPES:
@@ -238,14 +221,8 @@ class DirectObjectAccess:
return self._create_access_path(self._obj[index]) return self._create_access_path(self._obj[index])
def py__iter__list(self): def py__iter__list(self):
try: if not hasattr(self._obj, '__getitem__'):
iter_method = self._obj.__iter__
except AttributeError:
return None return None
else:
p = DirectObjectAccess(self._inference_state, iter_method).get_return_annotation()
if p is not None:
return [p]
if type(self._obj) not in ALLOWED_GETITEM_TYPES: if type(self._obj) not in ALLOWED_GETITEM_TYPES:
# Get rid of side effects, we won't call custom `__getitem__`s. # Get rid of side effects, we won't call custom `__getitem__`s.
@@ -329,9 +306,9 @@ class DirectObjectAccess:
except TypeError: except TypeError:
return False return False
def is_allowed_getattr(self, name, safe=True): def is_allowed_getattr(self, name, unsafe=False):
# TODO this API is ugly. # TODO this API is ugly.
if not safe: if unsafe:
# Unsafe is mostly used to check for __getattr__/__getattribute__. # Unsafe is mostly used to check for __getattr__/__getattribute__.
# getattr_static works for properties, but the underscore methods # getattr_static works for properties, but the underscore methods
# are just ignored (because it's safer and avoids more code # are just ignored (because it's safer and avoids more code
@@ -384,7 +361,7 @@ class DirectObjectAccess:
except AttributeError: except AttributeError:
pass pass
else: else:
if module is not None and isinstance(module, str): if module is not None:
try: try:
__import__(module) __import__(module)
# For some modules like _sqlite3, the __module__ for classes is # For some modules like _sqlite3, the __module__ for classes is

View File

@@ -34,7 +34,7 @@ class MixedObject(ValueWrapper):
This combined logic makes it possible to provide more powerful REPL This combined logic makes it possible to provide more powerful REPL
completion. It allows side effects that are not noticable with the default completion. It allows side effects that are not noticable with the default
parser structure to still be completable. parser structure to still be completeable.
The biggest difference from CompiledValue to MixedObject is that we are The biggest difference from CompiledValue to MixedObject is that we are
generally dealing with Python code and not with C code. This will generate generally dealing with Python code and not with C code. This will generate
@@ -187,7 +187,7 @@ def _find_syntax_node_name(inference_state, python_object):
try: try:
python_object = _get_object_to_check(python_object) python_object = _get_object_to_check(python_object)
path = inspect.getsourcefile(python_object) path = inspect.getsourcefile(python_object)
except (OSError, TypeError): except TypeError:
# The type might not be known (e.g. class_with_dict.__weakref__) # The type might not be known (e.g. class_with_dict.__weakref__)
return None return None
path = None if path is None else Path(path) path = None if path is None else Path(path)
@@ -267,7 +267,7 @@ def _find_syntax_node_name(inference_state, python_object):
@inference_state_function_cache() @inference_state_function_cache()
def _create(inference_state, compiled_value, module_context): def _create(inference_state, compiled_value, module_context):
# TODO accessing this is bad, but it probably doesn't matter that much, # TODO accessing this is bad, but it probably doesn't matter that much,
# because we're working with interpreters only here. # because we're working with interpreteters only here.
python_object = compiled_value.access_handle.access._obj python_object = compiled_value.access_handle.access._obj
result = _find_syntax_node_name(inference_state, python_object) result = _find_syntax_node_name(inference_state, python_object)
if result is None: if result is None:

View File

@@ -7,7 +7,6 @@ goals:
2. Make it possible to handle different Python versions as well as virtualenvs. 2. Make it possible to handle different Python versions as well as virtualenvs.
""" """
import collections
import os import os
import sys import sys
import queue import queue
@@ -169,7 +168,7 @@ class CompiledSubprocess:
def __init__(self, executable, env_vars=None): def __init__(self, executable, env_vars=None):
self._executable = executable self._executable = executable
self._env_vars = env_vars self._env_vars = env_vars
self._inference_state_deletion_queue = collections.deque() self._inference_state_deletion_queue = queue.deque()
self._cleanup_callable = lambda: None self._cleanup_callable = lambda: None
def __repr__(self): def __repr__(self):

View File

@@ -4,10 +4,10 @@ import inspect
import importlib import importlib
import warnings import warnings
from pathlib import Path from pathlib import Path
from zipfile import ZipFile from zipimport import zipimporter
from zipimport import zipimporter, ZipImportError
from importlib.machinery import all_suffixes from importlib.machinery import all_suffixes
from jedi._compatibility import cast_path
from jedi.inference.compiled import access from jedi.inference.compiled import access
from jedi import debug from jedi import debug
from jedi import parser_utils from jedi import parser_utils
@@ -15,7 +15,7 @@ from jedi.file_io import KnownContentFileIO, ZipFileIO
def get_sys_path(): def get_sys_path():
return sys.path return list(map(cast_path, sys.path))
def load_module(inference_state, **kwargs): def load_module(inference_state, **kwargs):
@@ -93,22 +93,15 @@ def _iter_module_names(inference_state, paths):
# Python modules/packages # Python modules/packages
for path in paths: for path in paths:
try: try:
dir_entries = ((entry.name, entry.is_dir()) for entry in os.scandir(path)) dirs = os.scandir(path)
except OSError: except OSError:
try: # The file might not exist or reading it might lead to an error.
zip_import_info = zipimporter(path) debug.warning("Not possible to list directory: %s", path)
# Unfortunately, there is no public way to access zipimporter's continue
# private _files member. We therefore have to use a for dir_entry in dirs:
# custom function to iterate over the files. name = dir_entry.name
dir_entries = _zip_list_subdirectory(
zip_import_info.archive, zip_import_info.prefix)
except ZipImportError:
# The file might not exist or reading it might lead to an error.
debug.warning("Not possible to list directory: %s", path)
continue
for name, is_dir in dir_entries:
# First Namespaces then modules/stubs # First Namespaces then modules/stubs
if is_dir: if dir_entry.is_dir():
# pycache is obviously not an interesting namespace. Also the # pycache is obviously not an interesting namespace. Also the
# name must be a valid identifier. # name must be a valid identifier.
if name != '__pycache__' and name.isidentifier(): if name != '__pycache__' and name.isidentifier():
@@ -151,11 +144,7 @@ def _find_module(string, path=None, full_name=None, is_global_search=True):
spec = find_spec(string, p) spec = find_spec(string, p)
if spec is not None: if spec is not None:
if spec.origin == "frozen":
continue
loader = spec.loader loader = spec.loader
if loader is None and not spec.has_location: if loader is None and not spec.has_location:
# This is a namespace package. # This is a namespace package.
full_name = string if not path else full_name full_name = string if not path else full_name
@@ -201,7 +190,7 @@ def _from_loader(loader, string):
except AttributeError: except AttributeError:
return None, is_package return None, is_package
else: else:
module_path = get_filename(string) module_path = cast_path(get_filename(string))
# To avoid unicode and read bytes, "overwrite" loader.get_source if # To avoid unicode and read bytes, "overwrite" loader.get_source if
# possible. # possible.
@@ -223,7 +212,7 @@ def _from_loader(loader, string):
if code is None: if code is None:
return None, is_package return None, is_package
if isinstance(loader, zipimporter): if isinstance(loader, zipimporter):
return ZipFileIO(module_path, code, Path(loader.archive)), is_package return ZipFileIO(module_path, code, Path(cast_path(loader.archive))), is_package
return KnownContentFileIO(module_path, code), is_package return KnownContentFileIO(module_path, code), is_package
@@ -241,17 +230,6 @@ def _get_source(loader, fullname):
name=fullname) name=fullname)
def _zip_list_subdirectory(zip_path, zip_subdir_path):
zip_file = ZipFile(zip_path)
zip_subdir_path = Path(zip_subdir_path)
zip_content_file_paths = zip_file.namelist()
for raw_file_name in zip_content_file_paths:
file_path = Path(raw_file_name)
if file_path.parent == zip_subdir_path:
file_path = file_path.relative_to(zip_subdir_path)
yield file_path.name, raw_file_name.endswith("/")
class ImplicitNSInfo: class ImplicitNSInfo:
"""Stores information returned from an implicit namespace spec""" """Stores information returned from an implicit namespace spec"""
def __init__(self, name, paths): def __init__(self, name, paths):

View File

@@ -5,10 +5,10 @@ import re
from functools import partial from functools import partial
from inspect import Parameter from inspect import Parameter
from pathlib import Path from pathlib import Path
from typing import Optional
from jedi import debug from jedi import debug
from jedi.inference.utils import to_list from jedi.inference.utils import to_list
from jedi._compatibility import cast_path
from jedi.cache import memoize_method from jedi.cache import memoize_method
from jedi.inference.filters import AbstractFilter from jedi.inference.filters import AbstractFilter
from jedi.inference.names import AbstractNameDefinition, ValueNameMixin, \ from jedi.inference.names import AbstractNameDefinition, ValueNameMixin, \
@@ -167,7 +167,7 @@ class CompiledValue(Value):
except AttributeError: except AttributeError:
return super().py__simple_getitem__(index) return super().py__simple_getitem__(index)
if access is None: if access is None:
return super().py__simple_getitem__(index) return NO_VALUES
return ValueSet([create_from_access_path(self.inference_state, access)]) return ValueSet([create_from_access_path(self.inference_state, access)])
@@ -293,7 +293,10 @@ class CompiledModule(CompiledValue):
return CompiledModuleContext(self) return CompiledModuleContext(self)
def py__path__(self): def py__path__(self):
return self.access_handle.py__path__() paths = self.access_handle.py__path__()
if paths is None:
return None
return map(cast_path, paths)
def is_package(self): def is_package(self):
return self.py__path__() is not None return self.py__path__() is not None
@@ -306,8 +309,11 @@ class CompiledModule(CompiledValue):
return () return ()
return tuple(name.split('.')) return tuple(name.split('.'))
def py__file__(self) -> Optional[Path]: def py__file__(self):
return self.access_handle.py__file__() # type: ignore[no-any-return] path = cast_path(self.access_handle.py__file__())
if path is None:
return None
return Path(path)
class CompiledName(AbstractNameDefinition): class CompiledName(AbstractNameDefinition):
@@ -434,7 +440,7 @@ class CompiledValueFilter(AbstractFilter):
access_handle = self.compiled_value.access_handle access_handle = self.compiled_value.access_handle
return self._get( return self._get(
name, name,
lambda name, safe: access_handle.is_allowed_getattr(name, safe=safe), lambda name, unsafe: access_handle.is_allowed_getattr(name, unsafe),
lambda name: name in access_handle.dir(), lambda name: name in access_handle.dir(),
check_has_attribute=True check_has_attribute=True
) )
@@ -448,7 +454,7 @@ class CompiledValueFilter(AbstractFilter):
has_attribute, is_descriptor = allowed_getattr_callback( has_attribute, is_descriptor = allowed_getattr_callback(
name, name,
safe=not self._inference_state.allow_descriptor_getattr unsafe=self._inference_state.allow_descriptor_getattr
) )
if check_has_attribute and not has_attribute: if check_has_attribute and not has_attribute:
return [] return []
@@ -472,7 +478,7 @@ class CompiledValueFilter(AbstractFilter):
from jedi.inference.compiled import builtin_from_name from jedi.inference.compiled import builtin_from_name
names = [] names = []
needs_type_completions, dir_infos = self.compiled_value.access_handle.get_dir_infos() needs_type_completions, dir_infos = self.compiled_value.access_handle.get_dir_infos()
# We could use `safe=False` here as well, especially as a parameter to # We could use `unsafe` here as well, especially as a parameter to
# get_dir_infos. But this would lead to a lot of property executions # get_dir_infos. But this would lead to a lot of property executions
# that are probably not wanted. The drawback for this is that we # that are probably not wanted. The drawback for this is that we
# have a different name for `get` and `values`. For `get` we always # have a different name for `get` and `values`. For `get` we always
@@ -480,7 +486,7 @@ class CompiledValueFilter(AbstractFilter):
for name in dir_infos: for name in dir_infos:
names += self._get( names += self._get(
name, name,
lambda name, safe: dir_infos[name], lambda name, unsafe: dir_infos[name],
lambda name: name in dir_infos, lambda name: name in dir_infos,
) )

View File

@@ -1,7 +1,5 @@
from abc import abstractmethod from abc import abstractmethod
from contextlib import contextmanager from contextlib import contextmanager
from pathlib import Path
from typing import Optional
from parso.tree import search_ancestor from parso.tree import search_ancestor
from parso.python.tree import Name from parso.python.tree import Name
@@ -309,8 +307,8 @@ class FunctionContext(TreeContextMixin, ValueContext):
class ModuleContext(TreeContextMixin, ValueContext): class ModuleContext(TreeContextMixin, ValueContext):
def py__file__(self) -> Optional[Path]: def py__file__(self):
return self._value.py__file__() # type: ignore[no-any-return] return self._value.py__file__()
def get_filters(self, until_position=None, origin_scope=None): def get_filters(self, until_position=None, origin_scope=None):
filters = self._value.get_filters(origin_scope) filters = self._value.get_filters(origin_scope)
@@ -327,7 +325,7 @@ class ModuleContext(TreeContextMixin, ValueContext):
yield from filters yield from filters
def get_global_filter(self): def get_global_filter(self):
return GlobalNameFilter(self) return GlobalNameFilter(self, self.tree_node)
@property @property
def string_names(self): def string_names(self):
@@ -357,8 +355,8 @@ class NamespaceContext(TreeContextMixin, ValueContext):
def string_names(self): def string_names(self):
return self._value.string_names return self._value.string_names
def py__file__(self) -> Optional[Path]: def py__file__(self):
return self._value.py__file__() # type: ignore[no-any-return] return self._value.py__file__()
class ClassContext(TreeContextMixin, ValueContext): class ClassContext(TreeContextMixin, ValueContext):
@@ -407,8 +405,8 @@ class CompiledModuleContext(CompiledContext):
def string_names(self): def string_names(self):
return self._value.string_names return self._value.string_names
def py__file__(self) -> Optional[Path]: def py__file__(self):
return self._value.py__file__() # type: ignore[no-any-return] return self._value.py__file__()
def _get_global_filters_for_name(context, name_or_none, position): def _get_global_filters_for_name(context, name_or_none, position):

View File

@@ -1,21 +0,0 @@
from jedi.inference.value import ModuleValue
from jedi.inference.context import ModuleContext
class DocstringModule(ModuleValue):
def __init__(self, in_module_context, **kwargs):
super().__init__(**kwargs)
self._in_module_context = in_module_context
def _as_context(self):
return DocstringModuleContext(self, self._in_module_context)
class DocstringModuleContext(ModuleContext):
def __init__(self, module_value, in_module_context):
super().__init__(module_value)
self._in_module_context = in_module_context
def get_filters(self, origin_scope=None, until_position=None):
yield from super().get_filters(until_position=until_position)
yield from self._in_module_context.get_filters()

View File

@@ -17,10 +17,12 @@ annotations.
import re import re
import warnings import warnings
from textwrap import dedent
from parso import parse, ParserSyntaxError from parso import parse, ParserSyntaxError
from jedi import debug from jedi import debug
from jedi.common import indent_block
from jedi.inference.cache import inference_state_method_cache from jedi.inference.cache import inference_state_method_cache
from jedi.inference.base_value import iterator_to_value_set, ValueSet, \ from jedi.inference.base_value import iterator_to_value_set, ValueSet, \
NO_VALUES NO_VALUES
@@ -180,40 +182,52 @@ def _strip_rst_role(type_str):
def _infer_for_statement_string(module_context, string): def _infer_for_statement_string(module_context, string):
code = dedent("""
def pseudo_docstring_stuff():
'''
Create a pseudo function for docstring statements.
Need this docstring so that if the below part is not valid Python this
is still a function.
'''
{}
""")
if string is None: if string is None:
return [] return []
potential_imports = re.findall(r'((?:\w+\.)*\w+)\.', string) for element in re.findall(r'((?:\w+\.)*\w+)\.', string):
# Try to import module part in dotted name. # Try to import module part in dotted name.
# (e.g., 'threading' in 'threading.Thread'). # (e.g., 'threading' in 'threading.Thread').
imports = "\n".join(f"import {p}" for p in potential_imports) string = 'import %s\n' % element + string
string = f'{imports}\n{string}'
debug.dbg('Parse docstring code %s', string, color='BLUE') debug.dbg('Parse docstring code %s', string, color='BLUE')
grammar = module_context.inference_state.grammar grammar = module_context.inference_state.grammar
try: try:
module = grammar.parse(string, error_recovery=False) module = grammar.parse(code.format(indent_block(string)), error_recovery=False)
except ParserSyntaxError: except ParserSyntaxError:
return [] return []
try: try:
# It's not the last item, because that's an end marker. funcdef = next(module.iter_funcdefs())
stmt = module.children[-2] # First pick suite, then simple_stmt and then the node,
# which is also not the last item, because there's a newline.
stmt = funcdef.children[-1].children[-1].children[-2]
except (AttributeError, IndexError): except (AttributeError, IndexError):
return [] return []
if stmt.type not in ('name', 'atom', 'atom_expr'): if stmt.type not in ('name', 'atom', 'atom_expr'):
return [] return []
# Here we basically use a fake module that also uses the filters in from jedi.inference.value import FunctionValue
# the actual module. function_value = FunctionValue(
from jedi.inference.docstring_utils import DocstringModule module_context.inference_state,
m = DocstringModule( module_context,
in_module_context=module_context, funcdef
inference_state=module_context.inference_state,
module_node=module,
code_lines=[],
) )
return list(_execute_types_in_stmt(m.as_context(), stmt)) func_execution_context = function_value.as_context()
# Use the module of the param.
# TODO this module is not the module of the param in case of a function
# call. In that case it's the module of the function call.
# stuffed with content from a function call.
return list(_execute_types_in_stmt(func_execution_context, stmt))
def _execute_types_in_stmt(module_context, stmt): def _execute_types_in_stmt(module_context, stmt):

View File

@@ -12,7 +12,7 @@ from parso.python.tree import Name, UsedNamesMapping
from jedi.inference import flow_analysis from jedi.inference import flow_analysis
from jedi.inference.base_value import ValueSet, ValueWrapper, \ from jedi.inference.base_value import ValueSet, ValueWrapper, \
LazyValueWrapper LazyValueWrapper
from jedi.parser_utils import get_cached_parent_scope, get_parso_cache_node from jedi.parser_utils import get_cached_parent_scope
from jedi.inference.utils import to_list from jedi.inference.utils import to_list
from jedi.inference.names import TreeNameDefinition, ParamName, \ from jedi.inference.names import TreeNameDefinition, ParamName, \
AnonymousParamName, AbstractNameDefinition, NameWrapper AnonymousParamName, AbstractNameDefinition, NameWrapper
@@ -54,15 +54,11 @@ class FilterWrapper:
return self.wrap_names(self._wrapped_filter.values()) return self.wrap_names(self._wrapped_filter.values())
def _get_definition_names(parso_cache_node, used_names, name_key): def _get_definition_names(used_names, name_key):
if parso_cache_node is None:
names = used_names.get(name_key, ())
return tuple(name for name in names if name.is_definition(include_setitem=True))
try: try:
for_module = _definition_name_cache[parso_cache_node] for_module = _definition_name_cache[used_names]
except KeyError: except KeyError:
for_module = _definition_name_cache[parso_cache_node] = {} for_module = _definition_name_cache[used_names] = {}
try: try:
return for_module[name_key] return for_module[name_key]
@@ -74,40 +70,18 @@ def _get_definition_names(parso_cache_node, used_names, name_key):
return result return result
class _AbstractUsedNamesFilter(AbstractFilter): class AbstractUsedNamesFilter(AbstractFilter):
name_class = TreeNameDefinition name_class = TreeNameDefinition
def __init__(self, parent_context, node_context=None): def __init__(self, parent_context, parser_scope):
if node_context is None: self._parser_scope = parser_scope
node_context = parent_context self._module_node = self._parser_scope.get_root_node()
self._node_context = node_context self._used_names = self._module_node.get_used_names()
self._parser_scope = node_context.tree_node
module_context = node_context.get_root_context()
# It is quite hacky that we have to use that. This is for caching
# certain things with a WeakKeyDictionary. However, parso intentionally
# uses slots (to save memory) and therefore we end up with having to
# have a weak reference to the object that caches the tree.
#
# Previously we have tried to solve this by using a weak reference onto
# used_names. However that also does not work, because it has a
# reference from the module, which itself is referenced by any node
# through parents.
path = module_context.py__file__()
if path is None:
# If the path is None, there is no guarantee that parso caches it.
self._parso_cache_node = None
else:
self._parso_cache_node = get_parso_cache_node(
module_context.inference_state.latest_grammar
if module_context.is_stub() else module_context.inference_state.grammar,
path
)
self._used_names = module_context.tree_node.get_used_names()
self.parent_context = parent_context self.parent_context = parent_context
def get(self, name): def get(self, name):
return self._convert_names(self._filter( return self._convert_names(self._filter(
_get_definition_names(self._parso_cache_node, self._used_names, name), _get_definition_names(self._used_names, name),
)) ))
def _convert_names(self, names): def _convert_names(self, names):
@@ -118,7 +92,7 @@ class _AbstractUsedNamesFilter(AbstractFilter):
name name
for name_key in self._used_names for name_key in self._used_names
for name in self._filter( for name in self._filter(
_get_definition_names(self._parso_cache_node, self._used_names, name_key), _get_definition_names(self._used_names, name_key),
) )
) )
@@ -126,7 +100,7 @@ class _AbstractUsedNamesFilter(AbstractFilter):
return '<%s: %s>' % (self.__class__.__name__, self.parent_context) return '<%s: %s>' % (self.__class__.__name__, self.parent_context)
class ParserTreeFilter(_AbstractUsedNamesFilter): class ParserTreeFilter(AbstractUsedNamesFilter):
def __init__(self, parent_context, node_context=None, until_position=None, def __init__(self, parent_context, node_context=None, until_position=None,
origin_scope=None): origin_scope=None):
""" """
@@ -135,7 +109,10 @@ class ParserTreeFilter(_AbstractUsedNamesFilter):
value, but for some type inference it's important to have a local value, but for some type inference it's important to have a local
value of the other classes. value of the other classes.
""" """
super().__init__(parent_context, node_context) if node_context is None:
node_context = parent_context
super().__init__(parent_context, node_context.tree_node)
self._node_context = node_context
self._origin_scope = origin_scope self._origin_scope = origin_scope
self._until_position = until_position self._until_position = until_position
@@ -149,7 +126,7 @@ class ParserTreeFilter(_AbstractUsedNamesFilter):
if parent.type == 'trailer': if parent.type == 'trailer':
return False return False
base_node = parent if parent.type in ('classdef', 'funcdef') else name base_node = parent if parent.type in ('classdef', 'funcdef') else name
return get_cached_parent_scope(self._parso_cache_node, base_node) == self._parser_scope return get_cached_parent_scope(self._used_names, base_node) == self._parser_scope
def _check_flows(self, names): def _check_flows(self, names):
for name in sorted(names, key=lambda name: name.start_pos, reverse=True): for name in sorted(names, key=lambda name: name.start_pos, reverse=True):
@@ -205,7 +182,7 @@ class AnonymousFunctionExecutionFilter(_FunctionExecutionFilter):
return AnonymousParamName(self._function_value, name) return AnonymousParamName(self._function_value, name)
class GlobalNameFilter(_AbstractUsedNamesFilter): class GlobalNameFilter(AbstractUsedNamesFilter):
def get(self, name): def get(self, name):
try: try:
names = self._used_names[name] names = self._used_names[name]

View File

@@ -196,43 +196,13 @@ def py__annotations__(funcdef):
return dct return dct
def resolve_forward_references(context, all_annotations):
def resolve(node):
if node is None or node.type != 'string':
return node
node = _get_forward_reference_node(
context,
context.inference_state.compiled_subprocess.safe_literal_eval(
node.value,
),
)
if node is None:
# There was a string, but it's not a valid annotation
return None
# The forward reference tree has an additional root node ('eval_input')
# that we don't want. Extract the node we do want, that is equivalent to
# the nodes returned by `py__annotations__` for a non-quoted node.
node = node.children[0]
return node
return {name: resolve(node) for name, node in all_annotations.items()}
@inference_state_method_cache() @inference_state_method_cache()
def infer_return_types(function, arguments): def infer_return_types(function, arguments):
""" """
Infers the type of a function's return value, Infers the type of a function's return value,
according to type annotations. according to type annotations.
""" """
context = function.get_default_param_context() all_annotations = py__annotations__(function.tree_node)
all_annotations = resolve_forward_references(
context,
py__annotations__(function.tree_node),
)
annotation = all_annotations.get("return", None) annotation = all_annotations.get("return", None)
if annotation is None: if annotation is None:
# If there is no Python 3-type annotation, look for an annotation # If there is no Python 3-type annotation, look for an annotation
@@ -247,10 +217,11 @@ def infer_return_types(function, arguments):
return NO_VALUES return NO_VALUES
return _infer_annotation_string( return _infer_annotation_string(
context, function.get_default_param_context(),
match.group(1).strip() match.group(1).strip()
).execute_annotation() ).execute_annotation()
context = function.get_default_param_context()
unknown_type_vars = find_unknown_type_vars(context, annotation) unknown_type_vars = find_unknown_type_vars(context, annotation)
annotation_values = infer_annotation(context, annotation) annotation_values = infer_annotation(context, annotation)
if not unknown_type_vars: if not unknown_type_vars:

View File

@@ -86,8 +86,6 @@ class StubFilter(ParserTreeFilter):
# Imports in stub files are only public if they have an "as" # Imports in stub files are only public if they have an "as"
# export. # export.
definition = name.get_definition() definition = name.get_definition()
if definition is None:
return False
if definition.type in ('import_from', 'import_name'): if definition.type in ('import_from', 'import_name'):
if name.parent.type not in ('import_as_name', 'dotted_as_name'): if name.parent.type not in ('import_as_name', 'dotted_as_name'):
return False return False

View File

@@ -7,6 +7,7 @@ from pathlib import Path
from jedi import settings from jedi import settings
from jedi.file_io import FileIO from jedi.file_io import FileIO
from jedi._compatibility import cast_path
from jedi.parser_utils import get_cached_code_lines from jedi.parser_utils import get_cached_code_lines
from jedi.inference.base_value import ValueSet, NO_VALUES from jedi.inference.base_value import ValueSet, NO_VALUES
from jedi.inference.gradual.stub_value import TypingModuleWrapper, StubModuleValue from jedi.inference.gradual.stub_value import TypingModuleWrapper, StubModuleValue
@@ -43,6 +44,7 @@ def _create_stub_map(directory_path_info):
return return
for entry in listed: for entry in listed:
entry = cast_path(entry)
path = os.path.join(directory_path_info.path, entry) path = os.path.join(directory_path_info.path, entry)
if os.path.isdir(path): if os.path.isdir(path):
init = os.path.join(path, '__init__.pyi') init = os.path.join(path, '__init__.pyi')
@@ -167,6 +169,7 @@ def _try_to_load_stub(inference_state, import_names, python_value_set,
if len(import_names) == 1: if len(import_names) == 1:
# foo-stubs # foo-stubs
for p in sys_path: for p in sys_path:
p = cast_path(p)
init = os.path.join(p, *import_names) + '-stubs' + os.path.sep + '__init__.pyi' init = os.path.join(p, *import_names) + '-stubs' + os.path.sep + '__init__.pyi'
m = _try_to_load_stub_from_file( m = _try_to_load_stub_from_file(
inference_state, inference_state,

View File

@@ -294,9 +294,6 @@ class Callable(BaseTypingInstance):
from jedi.inference.gradual.annotation import infer_return_for_callable from jedi.inference.gradual.annotation import infer_return_for_callable
return infer_return_for_callable(arguments, param_values, result_values) return infer_return_for_callable(arguments, param_values, result_values)
def py__get__(self, instance, class_value):
return ValueSet([self])
class Tuple(BaseTypingInstance): class Tuple(BaseTypingInstance):
def _is_homogenous(self): def _is_homogenous(self):
@@ -434,9 +431,6 @@ class NewType(Value):
from jedi.inference.compiled.value import CompiledValueName from jedi.inference.compiled.value import CompiledValueName
return CompiledValueName(self, 'NewType') return CompiledValueName(self, 'NewType')
def __repr__(self) -> str:
return '<NewType: %s>%s' % (self.tree_node, self._type_value_set)
class CastFunction(ValueWrapper): class CastFunction(ValueWrapper):
@repack_with_argument_clinic('type, object, /') @repack_with_argument_clinic('type, object, /')

View File

@@ -422,13 +422,20 @@ def import_module(inference_state, import_names, parent_module_value, sys_path):
# The module might not be a package. # The module might not be a package.
return NO_VALUES return NO_VALUES
file_io_or_ns, is_pkg = inference_state.compiled_subprocess.get_module_info( for path in paths:
string=import_names[-1], # At the moment we are only using one path. So this is
path=paths, # not important to be correct.
full_name=module_name, if not isinstance(path, list):
is_global_search=False, path = [path]
) file_io_or_ns, is_pkg = inference_state.compiled_subprocess.get_module_info(
if is_pkg is None: string=import_names[-1],
path=path,
full_name=module_name,
is_global_search=False,
)
if is_pkg is not None:
break
else:
return NO_VALUES return NO_VALUES
if isinstance(file_io_or_ns, ImplicitNSInfo): if isinstance(file_io_or_ns, ImplicitNSInfo):

View File

@@ -248,7 +248,7 @@ class ValueNameMixin:
def get_defining_qualified_value(self): def get_defining_qualified_value(self):
context = self.parent_context context = self.parent_context
if context is not None and (context.is_module() or context.is_class()): if context.is_module() or context.is_class():
return self.parent_context.get_value() # Might be None return self.parent_context.get_value() # Might be None
return None return None
@@ -341,12 +341,6 @@ class TreeNameDefinition(AbstractTreeName):
def py__doc__(self): def py__doc__(self):
api_type = self.api_type api_type = self.api_type
if api_type in ('function', 'class', 'property'): if api_type in ('function', 'class', 'property'):
if self.parent_context.get_root_context().is_stub():
from jedi.inference.gradual.conversion import convert_names
names = convert_names([self], prefer_stub_to_compiled=False)
if self not in names:
return _merge_name_docs(names)
# Make sure the names are not TreeNameDefinitions anymore. # Make sure the names are not TreeNameDefinitions anymore.
return clean_scope_docstring(self.tree_name.get_definition()) return clean_scope_docstring(self.tree_name.get_definition())
@@ -414,9 +408,6 @@ class ParamNameInterface(_ParamMixin):
return 2 return 2
return 0 return 0
def infer_default(self):
return NO_VALUES
class BaseTreeParamName(ParamNameInterface, AbstractTreeName): class BaseTreeParamName(ParamNameInterface, AbstractTreeName):
annotation_node = None annotation_node = None

View File

@@ -12,7 +12,7 @@ count the function calls.
Settings Settings
~~~~~~~~~~ ~~~~~~~~~~
Recursion settings are important if you don't want extremely Recursion settings are important if you don't want extremly
recursive python code to go absolutely crazy. recursive python code to go absolutely crazy.
The default values are based on experiments while completing the |jedi| library The default values are based on experiments while completing the |jedi| library

View File

@@ -180,34 +180,26 @@ def _check_fs(inference_state, file_io, regex):
return m.as_context() return m.as_context()
def gitignored_paths(folder_io, file_io): def gitignored_lines(folder_io, file_io):
ignored_paths_abs = set() ignored_paths = set()
ignored_paths_rel = set() ignored_names = set()
for l in file_io.read().splitlines(): for l in file_io.read().splitlines():
if not l or l.startswith(b'#') or l.startswith(b'!') or b'*' in l: if not l or l.startswith(b'#'):
continue continue
p = l.decode('utf-8', 'ignore').rstrip('/') p = l.decode('utf-8', 'ignore')
if '/' in p: if p.startswith('/'):
name = p.lstrip('/') name = p[1:]
ignored_paths_abs.add(os.path.join(folder_io.path, name)) if name.endswith(os.path.sep):
name = name[:-1]
ignored_paths.add(os.path.join(folder_io.path, name))
else: else:
name = p ignored_names.add(p)
ignored_paths_rel.add((folder_io.path, name)) return ignored_paths, ignored_names
return ignored_paths_abs, ignored_paths_rel
def expand_relative_ignore_paths(folder_io, relative_paths):
curr_path = folder_io.path
return {os.path.join(curr_path, p[1]) for p in relative_paths if curr_path.startswith(p[0])}
def recurse_find_python_folders_and_files(folder_io, except_paths=()): def recurse_find_python_folders_and_files(folder_io, except_paths=()):
except_paths = set(except_paths) except_paths = set(except_paths)
except_paths_relative = set()
for root_folder_io, folder_ios, file_ios in folder_io.walk(): for root_folder_io, folder_ios, file_ios in folder_io.walk():
# Delete folders that we don't want to iterate over. # Delete folders that we don't want to iterate over.
for file_io in file_ios: for file_io in file_ios:
@@ -217,21 +209,14 @@ def recurse_find_python_folders_and_files(folder_io, except_paths=()):
yield None, file_io yield None, file_io
if path.name == '.gitignore': if path.name == '.gitignore':
ignored_paths_abs, ignored_paths_rel = gitignored_paths( ignored_paths, ignored_names = \
root_folder_io, file_io gitignored_lines(root_folder_io, file_io)
) except_paths |= ignored_paths
except_paths |= ignored_paths_abs
except_paths_relative |= ignored_paths_rel
except_paths_relative_expanded = expand_relative_ignore_paths(
root_folder_io, except_paths_relative
)
folder_ios[:] = [ folder_ios[:] = [
folder_io folder_io
for folder_io in folder_ios for folder_io in folder_ios
if folder_io.path not in except_paths if folder_io.path not in except_paths
and folder_io.path not in except_paths_relative_expanded
and folder_io.get_base_name() not in _IGNORE_FOLDERS and folder_io.get_base_name() not in _IGNORE_FOLDERS
] ]
for folder_io in folder_ios: for folder_io in folder_ios:
@@ -297,13 +282,12 @@ def get_module_contexts_containing_name(inference_state, module_contexts, name,
limit_reduction=limit_reduction) limit_reduction=limit_reduction)
def search_in_file_ios(inference_state, file_io_iterator, name, def search_in_file_ios(inference_state, file_io_iterator, name, limit_reduction=1):
limit_reduction=1, complete=False):
parse_limit = _PARSED_FILE_LIMIT / limit_reduction parse_limit = _PARSED_FILE_LIMIT / limit_reduction
open_limit = _OPENED_FILE_LIMIT / limit_reduction open_limit = _OPENED_FILE_LIMIT / limit_reduction
file_io_count = 0 file_io_count = 0
parsed_file_count = 0 parsed_file_count = 0
regex = re.compile(r'\b' + re.escape(name) + (r'' if complete else r'\b')) regex = re.compile(r'\b' + re.escape(name) + r'\b')
for file_io in file_io_iterator: for file_io in file_io_iterator:
file_io_count += 1 file_io_count += 1
m = _check_fs(inference_state, file_io, regex) m = _check_fs(inference_state, file_io, regex)

View File

@@ -12,8 +12,6 @@ The signature here for bar should be `bar(b, c)` instead of bar(*args).
""" """
from inspect import Parameter from inspect import Parameter
from parso import tree
from jedi.inference.utils import to_list from jedi.inference.utils import to_list
from jedi.inference.names import ParamNameWrapper from jedi.inference.names import ParamNameWrapper
from jedi.inference.helpers import is_big_annoying_library from jedi.inference.helpers import is_big_annoying_library
@@ -24,11 +22,7 @@ def _iter_nodes_for_param(param_name):
from jedi.inference.arguments import TreeArguments from jedi.inference.arguments import TreeArguments
execution_context = param_name.parent_context execution_context = param_name.parent_context
# Walk up the parso tree to get the FunctionNode we want. We use the parso function_node = execution_context.tree_node
# tree rather than going via the execution context so that we're agnostic of
# the specific scope we're evaluating within (i.e: module or function,
# etc.).
function_node = tree.search_ancestor(param_name.tree_name, 'funcdef', 'lambdef')
module_node = function_node.get_root_node() module_node = function_node.get_root_node()
start = function_node.children[-1].start_pos start = function_node.children[-1].start_pos
end = function_node.children[-1].end_pos end = function_node.children[-1].end_pos

View File

@@ -2,7 +2,6 @@
Functions inferring the syntax tree. Functions inferring the syntax tree.
""" """
import copy import copy
import itertools
from parso.python import tree from parso.python import tree
@@ -516,20 +515,10 @@ def _literals_to_types(inference_state, result):
def _infer_comparison(context, left_values, operator, right_values): def _infer_comparison(context, left_values, operator, right_values):
state = context.inference_state state = context.inference_state
if isinstance(operator, str):
operator_str = operator
else:
operator_str = str(operator.value)
if not left_values or not right_values: if not left_values or not right_values:
# illegal slices e.g. cause left/right_result to be None # illegal slices e.g. cause left/right_result to be None
result = (left_values or NO_VALUES) | (right_values or NO_VALUES) result = (left_values or NO_VALUES) | (right_values or NO_VALUES)
return _literals_to_types(state, result) return _literals_to_types(state, result)
elif operator_str == "|" and all(
value.is_class() or value.is_compiled()
for value in itertools.chain(left_values, right_values)
):
# ^^^ A naive hack for PEP 604
return ValueSet.from_sets((left_values, right_values))
else: else:
# I don't think there's a reasonable chance that a string # I don't think there's a reasonable chance that a string
# operation is still correct, once we pass something like six # operation is still correct, once we pass something like six
@@ -749,13 +738,6 @@ def tree_name_to_values(inference_state, context, tree_name):
types = infer_expr_stmt(context, node, tree_name) types = infer_expr_stmt(context, node, tree_name)
elif typ == 'with_stmt': elif typ == 'with_stmt':
value_managers = context.infer_node(node.get_test_node_from_name(tree_name)) value_managers = context.infer_node(node.get_test_node_from_name(tree_name))
if node.parent.type == 'async_stmt':
# In the case of `async with` statements, we need to
# first get the coroutine from the `__aenter__` method,
# then "unwrap" via the `__await__` method
enter_methods = value_managers.py__getattribute__('__aenter__')
coro = enter_methods.execute_with_values()
return coro.py__await__().py__stop_iteration_returns()
enter_methods = value_managers.py__getattribute__('__enter__') enter_methods = value_managers.py__getattribute__('__enter__')
return enter_methods.execute_with_values() return enter_methods.execute_with_values()
elif typ in ('import_from', 'import_name'): elif typ in ('import_from', 'import_name'):

View File

@@ -186,6 +186,7 @@ def _get_buildout_script_paths(search_path: Path):
directory that look like python files. directory that look like python files.
:param search_path: absolute path to the module. :param search_path: absolute path to the module.
:type search_path: str
""" """
project_root = _get_parent_dir_with_file(search_path, 'buildout.cfg') project_root = _get_parent_dir_with_file(search_path, 'buildout.cfg')
if not project_root: if not project_root:
@@ -204,7 +205,7 @@ def _get_buildout_script_paths(search_path: Path):
except (UnicodeDecodeError, IOError) as e: except (UnicodeDecodeError, IOError) as e:
# Probably a binary file; permission error or race cond. because # Probably a binary file; permission error or race cond. because
# file got deleted. Ignore it. # file got deleted. Ignore it.
debug.warning(str(e)) debug.warning(e)
continue continue

View File

@@ -16,7 +16,7 @@ settings will stop this process.
It is important to note that: It is important to note that:
1. Array modifications work only in the current module. 1. Array modfications work only in the current module.
2. Jedi only checks Array additions; ``list.pop``, etc are ignored. 2. Jedi only checks Array additions; ``list.pop``, etc are ignored.
""" """
from jedi import debug from jedi import debug

View File

@@ -344,8 +344,7 @@ class BaseFunctionExecutionContext(ValueContext, TreeContextMixin):
GenericClass(c, TupleGenericManager(generics)) for c in async_classes GenericClass(c, TupleGenericManager(generics)) for c in async_classes
).execute_annotation() ).execute_annotation()
else: else:
# If there are annotations, prefer them over anything else. if self.is_generator():
if self.is_generator() and not self.infer_annotations():
return ValueSet([iterable.Generator(inference_state, self)]) return ValueSet([iterable.Generator(inference_state, self)])
else: else:
return self.get_return_values() return self.get_return_values()

View File

@@ -342,8 +342,6 @@ class SequenceLiteralValue(Sequence):
else: else:
with reraise_getitem_errors(TypeError, KeyError, IndexError): with reraise_getitem_errors(TypeError, KeyError, IndexError):
node = self.get_tree_entries()[index] node = self.get_tree_entries()[index]
if node == ':' or node.type == 'subscript':
return NO_VALUES
return self._defining_context.infer_node(node) return self._defining_context.infer_node(node)
def py__iter__(self, contextualized_node=None): def py__iter__(self, contextualized_node=None):
@@ -409,6 +407,16 @@ class SequenceLiteralValue(Sequence):
else: else:
return [array_node] return [array_node]
def exact_key_items(self):
"""
Returns a generator of tuples like dict.items(), where the key is
resolved (as a string) and the values are still lazy values.
"""
for key_node, value in self.get_tree_entries():
for key in self._defining_context.infer_node(key_node):
if is_string(key):
yield key.get_safe_value(), LazyTreeValue(self._defining_context, value)
def __repr__(self): def __repr__(self):
return "<%s of %s>" % (self.__class__.__name__, self.atom) return "<%s of %s>" % (self.__class__.__name__, self.atom)
@@ -464,16 +472,6 @@ class DictLiteralValue(_DictMixin, SequenceLiteralValue, _DictKeyMixin):
return ValueSet([FakeList(self.inference_state, lazy_values)]) return ValueSet([FakeList(self.inference_state, lazy_values)])
def exact_key_items(self):
"""
Returns a generator of tuples like dict.items(), where the key is
resolved (as a string) and the values are still lazy values.
"""
for key_node, value in self.get_tree_entries():
for key in self._defining_context.infer_node(key_node):
if is_string(key):
yield key.get_safe_value(), LazyTreeValue(self._defining_context, value)
def _dict_values(self): def _dict_values(self):
return ValueSet.from_sets( return ValueSet.from_sets(
self._defining_context.infer_node(v) self._defining_context.infer_node(v)

View File

@@ -78,8 +78,6 @@ class ClassName(TreeNameDefinition):
type_ = super().api_type type_ = super().api_type
if type_ == 'function': if type_ == 'function':
definition = self.tree_name.get_definition() definition = self.tree_name.get_definition()
if definition is None:
return type_
if function_is_property(definition): if function_is_property(definition):
# This essentially checks if there is an @property before # This essentially checks if there is an @property before
# the function. @property could be something different, but # the function. @property could be something different, but
@@ -116,10 +114,25 @@ class ClassFilter(ParserTreeFilter):
while node is not None: while node is not None:
if node == self._parser_scope or node == self.parent_context: if node == self._parser_scope or node == self.parent_context:
return True return True
node = get_cached_parent_scope(self._parso_cache_node, node) node = get_cached_parent_scope(self._used_names, node)
return False return False
def _access_possible(self, name): def _access_possible(self, name):
# Filter for ClassVar variables
# TODO this is not properly done, yet. It just checks for the string
# ClassVar in the annotation, which can be quite imprecise. If we
# wanted to do this correct, we would have to infer the ClassVar.
if not self._is_instance:
expr_stmt = name.get_definition()
if expr_stmt is not None and expr_stmt.type == 'expr_stmt':
annassign = expr_stmt.children[1]
if annassign.type == 'annassign':
# If there is an =, the variable is obviously also
# defined on the class.
if 'ClassVar' not in annassign.children[1].get_code() \
and '=' not in annassign.children:
return False
# Filter for name mangling of private variables like __foo # Filter for name mangling of private variables like __foo
return not name.value.startswith('__') or name.value.endswith('__') \ return not name.value.startswith('__') or name.value.endswith('__') \
or self._equals_origin_scope() or self._equals_origin_scope()

View File

@@ -64,7 +64,7 @@ class ModuleMixin(SubModuleDictMixin):
parent_context=self.as_context(), parent_context=self.as_context(),
origin_scope=origin_scope origin_scope=origin_scope
), ),
GlobalNameFilter(self.as_context()), GlobalNameFilter(self.as_context(), self.tree_node),
) )
yield DictFilter(self.sub_modules_dict()) yield DictFilter(self.sub_modules_dict())
yield DictFilter(self._module_attributes_dict()) yield DictFilter(self._module_attributes_dict())
@@ -148,7 +148,7 @@ class ModuleValue(ModuleMixin, TreeValue):
if file_io is None: if file_io is None:
self._path: Optional[Path] = None self._path: Optional[Path] = None
else: else:
self._path = file_io.path self._path = Path(file_io.path)
self.string_names = string_names # Optional[Tuple[str, ...]] self.string_names = string_names # Optional[Tuple[str, ...]]
self.code_lines = code_lines self.code_lines = code_lines
self._is_package = is_package self._is_package = is_package

View File

@@ -1,6 +1,3 @@
from pathlib import Path
from typing import Optional
from jedi.inference.cache import inference_state_method_cache from jedi.inference.cache import inference_state_method_cache
from jedi.inference.filters import DictFilter from jedi.inference.filters import DictFilter
from jedi.inference.names import ValueNameMixin, AbstractNameDefinition from jedi.inference.names import ValueNameMixin, AbstractNameDefinition
@@ -44,7 +41,7 @@ class ImplicitNamespaceValue(Value, SubModuleDictMixin):
string_name = self.py__package__()[-1] string_name = self.py__package__()[-1]
return ImplicitNSName(self, string_name) return ImplicitNSName(self, string_name)
def py__file__(self) -> Optional[Path]: def py__file__(self):
return None return None
def py__package__(self): def py__package__(self):

View File

@@ -216,14 +216,11 @@ def is_scope(node):
def _get_parent_scope_cache(func): def _get_parent_scope_cache(func):
cache = WeakKeyDictionary() cache = WeakKeyDictionary()
def wrapper(parso_cache_node, node, include_flows=False): def wrapper(used_names, node, include_flows=False):
if parso_cache_node is None:
return func(node, include_flows)
try: try:
for_module = cache[parso_cache_node] for_module = cache[used_names]
except KeyError: except KeyError:
for_module = cache[parso_cache_node] = {} for_module = cache[used_names] = {}
try: try:
return for_module[node] return for_module[node]
@@ -273,18 +270,7 @@ def get_cached_code_lines(grammar, path):
Basically access the cached code lines in parso. This is not the nicest way Basically access the cached code lines in parso. This is not the nicest way
to do this, but we avoid splitting all the lines again. to do this, but we avoid splitting all the lines again.
""" """
return get_parso_cache_node(grammar, path).lines return parser_cache[grammar._hashed][path].lines
def get_parso_cache_node(grammar, path):
"""
This is of course not public. But as long as I control parso, this
shouldn't be a problem. ~ Dave
The reason for this is mostly caching. This is obviously also a sign of a
broken caching architecture.
"""
return parser_cache[grammar._hashed][path]
def cut_value_at_position(leaf, position): def cut_value_at_position(leaf, position):

View File

@@ -2,7 +2,7 @@ from pathlib import Path
from parso.tree import search_ancestor from parso.tree import search_ancestor
from jedi.inference.cache import inference_state_method_cache from jedi.inference.cache import inference_state_method_cache
from jedi.inference.imports import goto_import, load_module_from_path from jedi.inference.imports import load_module_from_path
from jedi.inference.filters import ParserTreeFilter from jedi.inference.filters import ParserTreeFilter
from jedi.inference.base_value import NO_VALUES, ValueSet from jedi.inference.base_value import NO_VALUES, ValueSet
from jedi.inference.helpers import infer_call_of_leaf from jedi.inference.helpers import infer_call_of_leaf
@@ -31,15 +31,7 @@ def execute(callback):
def infer_anonymous_param(func): def infer_anonymous_param(func):
def get_returns(value): def get_returns(value):
if value.tree_node.annotation is not None: if value.tree_node.annotation is not None:
result = value.execute_with_values() return value.execute_with_values()
if any(v.name.get_qualified_names(include_module_names=True)
== ('typing', 'Generator')
for v in result):
return ValueSet.from_sets(
v.py__getattribute__('__next__').execute_annotation()
for v in result
)
return result
# In pytest we need to differentiate between generators and normal # In pytest we need to differentiate between generators and normal
# returns. # returns.
@@ -51,9 +43,6 @@ def infer_anonymous_param(func):
return function_context.get_return_values() return function_context.get_return_values()
def wrapper(param_name): def wrapper(param_name):
# parameters with an annotation do not need special handling
if param_name.annotation_node:
return func(param_name)
is_pytest_param, param_name_is_function_name = \ is_pytest_param, param_name_is_function_name = \
_is_a_pytest_param_and_inherited(param_name) _is_a_pytest_param_and_inherited(param_name)
if is_pytest_param: if is_pytest_param:
@@ -131,17 +120,6 @@ def _is_pytest_func(func_name, decorator_nodes):
or any('fixture' in n.get_code() for n in decorator_nodes) or any('fixture' in n.get_code() for n in decorator_nodes)
def _find_pytest_plugin_modules():
"""
Finds pytest plugin modules hooked by setuptools entry points
See https://docs.pytest.org/en/stable/how-to/writing_plugins.html#setuptools-entry-points
"""
from pkg_resources import iter_entry_points
return [ep.module_name.split(".") for ep in iter_entry_points(group="pytest11")]
@inference_state_method_cache() @inference_state_method_cache()
def _iter_pytest_modules(module_context, skip_own_module=False): def _iter_pytest_modules(module_context, skip_own_module=False):
if not skip_own_module: if not skip_own_module:
@@ -151,10 +129,6 @@ def _iter_pytest_modules(module_context, skip_own_module=False):
if file_io is not None: if file_io is not None:
folder = file_io.get_parent_folder() folder = file_io.get_parent_folder()
sys_path = module_context.inference_state.get_sys_path() sys_path = module_context.inference_state.get_sys_path()
# prevent an infinite loop when reaching the root of the current drive
last_folder = None
while any(folder.path.startswith(p) for p in sys_path): while any(folder.path.startswith(p) for p in sys_path):
file_io = folder.get_file_io('conftest.py') file_io = folder.get_file_io('conftest.py')
if Path(file_io.path) != module_context.py__file__(): if Path(file_io.path) != module_context.py__file__():
@@ -165,12 +139,7 @@ def _iter_pytest_modules(module_context, skip_own_module=False):
pass pass
folder = folder.get_parent_folder() folder = folder.get_parent_folder()
# prevent an infinite for loop if the same parent folder is return twice for names in _PYTEST_FIXTURE_MODULES:
if last_folder is not None and folder.path == last_folder.path:
break
last_folder = folder # keep track of the last found parent name
for names in _PYTEST_FIXTURE_MODULES + _find_pytest_plugin_modules():
for module_value in module_context.inference_state.import_module(names): for module_value in module_context.inference_state.import_module(names):
yield module_value.as_context() yield module_value.as_context()
@@ -178,28 +147,14 @@ def _iter_pytest_modules(module_context, skip_own_module=False):
class FixtureFilter(ParserTreeFilter): class FixtureFilter(ParserTreeFilter):
def _filter(self, names): def _filter(self, names):
for name in super()._filter(names): for name in super()._filter(names):
# look for fixture definitions of imported names funcdef = name.parent
if name.parent.type == "import_from": # Class fixtures are not supported
imported_names = goto_import(self.parent_context, name) if funcdef.type == 'funcdef':
if any( decorated = funcdef.parent
self._is_fixture(iname.parent_context, iname.tree_name) if decorated.type == 'decorated' and self._is_fixture(decorated):
for iname in imported_names
# discard imports of whole modules, that have no tree_name
if iname.tree_name
):
yield name yield name
elif self._is_fixture(self.parent_context, name): def _is_fixture(self, decorated):
yield name
def _is_fixture(self, context, name):
funcdef = name.parent
# Class fixtures are not supported
if funcdef.type != "funcdef":
return False
decorated = funcdef.parent
if decorated.type != "decorated":
return False
decorators = decorated.children[0] decorators = decorated.children[0]
if decorators.type == 'decorators': if decorators.type == 'decorators':
decorators = decorators.children decorators = decorators.children
@@ -216,12 +171,11 @@ class FixtureFilter(ParserTreeFilter):
last_leaf = last_trailer.get_last_leaf() last_leaf = last_trailer.get_last_leaf()
if last_leaf == ')': if last_leaf == ')':
values = infer_call_of_leaf( values = infer_call_of_leaf(
context, last_leaf, cut_own_trailer=True self.parent_context, last_leaf, cut_own_trailer=True)
)
else: else:
values = context.infer_node(dotted_name) values = self.parent_context.infer_node(dotted_name)
else: else:
values = context.infer_node(dotted_name) values = self.parent_context.infer_node(dotted_name)
for value in values: for value in values:
if value.name.get_qualified_names(include_module_names=True) \ if value.name.get_qualified_names(include_module_names=True) \
== ('_pytest', 'fixtures', 'fixture'): == ('_pytest', 'fixtures', 'fixture'):

View File

@@ -803,15 +803,6 @@ _implemented = {
# For now this works at least better than Jedi trying to understand it. # For now this works at least better than Jedi trying to understand it.
'dataclass': _dataclass 'dataclass': _dataclass
}, },
# attrs exposes declaration interface roughly compatible with dataclasses
# via attrs.define, attrs.frozen and attrs.mutable
# https://www.attrs.org/en/stable/names.html
'attr': {
'define': _dataclass,
},
'attrs': {
'define': _dataclass,
},
'os.path': { 'os.path': {
'dirname': _create_string_input_function(os.path.dirname), 'dirname': _create_string_input_function(os.path.dirname),
'abspath': _create_string_input_function(os.path.abspath), 'abspath': _create_string_input_function(os.path.abspath),

View File

@@ -35,46 +35,17 @@ setup(name='jedi',
install_requires=['parso>=0.8.0,<0.9.0'], install_requires=['parso>=0.8.0,<0.9.0'],
extras_require={ extras_require={
'testing': [ 'testing': [
'pytest<7.0.0', 'pytest<6.0.0',
# docopt for sith doctests # docopt for sith doctests
'docopt', 'docopt',
# coloroma for colored debug output # coloroma for colored debug output
'colorama', 'colorama',
'Django<3.1', # For now pin this. 'Django<3.1', # For now pin this.
'attrs',
], ],
'qa': [ 'qa': [
'flake8==3.8.3', 'flake8==3.8.3',
'mypy==0.782', 'mypy==0.782',
], ],
'docs': [
# Just pin all of these.
'Jinja2==2.11.3',
'MarkupSafe==1.1.1',
'Pygments==2.8.1',
'alabaster==0.7.12',
'babel==2.9.1',
'chardet==4.0.0',
'commonmark==0.8.1',
'docutils==0.17.1',
'future==0.18.2',
'idna==2.10',
'imagesize==1.2.0',
'mock==1.0.1',
'packaging==20.9',
'pyparsing==2.4.7',
'pytz==2021.1',
'readthedocs-sphinx-ext==2.1.4',
'recommonmark==0.5.0',
'requests==2.25.1',
'six==1.15.0',
'snowballstemmer==2.1.0',
'sphinx==1.8.5',
'sphinx-rtd-theme==0.4.3',
'sphinxcontrib-serializinghtml==1.1.4',
'sphinxcontrib-websupport==1.2.4',
'urllib3==1.26.4',
],
}, },
package_data={'jedi': ['*.pyi', 'third_party/typeshed/LICENSE', package_data={'jedi': ['*.pyi', 'third_party/typeshed/LICENSE',
'third_party/typeshed/README']}, 'third_party/typeshed/README']},
@@ -90,7 +61,6 @@ setup(name='jedi',
'Programming Language :: Python :: 3.7', 'Programming Language :: Python :: 3.7',
'Programming Language :: Python :: 3.8', 'Programming Language :: Python :: 3.8',
'Programming Language :: Python :: 3.9', 'Programming Language :: Python :: 3.9',
'Programming Language :: Python :: 3.10',
'Topic :: Software Development :: Libraries :: Python Modules', 'Topic :: Software Development :: Libraries :: Python Modules',
'Topic :: Text Editors :: Integrated Development Environments (IDE)', 'Topic :: Text Editors :: Integrated Development Environments (IDE)',
'Topic :: Utilities', 'Topic :: Utilities',

View File

@@ -44,8 +44,6 @@ b[int():]
#? list() #? list()
b[:] b[:]
#? int()
b[:, :-1]
#? 3 #? 3
b[:] b[:]
@@ -69,20 +67,6 @@ class _StrangeSlice():
#? slice() #? slice()
_StrangeSlice()[1:2] _StrangeSlice()[1:2]
for x in b[:]:
#? int()
x
for x in b[:, :-1]:
#?
x
class Foo:
def __getitem__(self, item):
return item
#?
Foo()[:, :-1][0]
# ----------------- # -----------------
# iterable multiplication # iterable multiplication

View File

@@ -26,6 +26,11 @@ async def y():
x().__await__().__next x().__await__().__next
return 2 return 2
async def x2():
async with open('asdf') as f:
#? ['readlines']
f.readlines
class A(): class A():
@staticmethod @staticmethod
async def b(c=1, d=2): async def b(c=1, d=2):
@@ -47,6 +52,8 @@ async def awaitable_test():
#? str() #? str()
foo foo
# python >= 3.6
async def asgen(): async def asgen():
yield 1 yield 1
await asyncio.sleep(0) await asyncio.sleep(0)
@@ -98,22 +105,3 @@ async def f():
f = await C().async_for_classmethod() f = await C().async_for_classmethod()
#? C() #? C()
f f
class AsyncCtxMgr:
def some_method():
pass
async def __aenter__(self):
return self
async def __aexit__(self, *args):
pass
async def asyncctxmgr():
async with AsyncCtxMgr() as acm:
#? AsyncCtxMgr()
acm
#? ['some_method']
acm.som

View File

@@ -23,7 +23,7 @@ def inheritance_fixture():
@pytest.fixture @pytest.fixture
def capsysbinary(capsysbinary): def testdir(testdir):
#? ['close'] #? ['chdir']
capsysbinary.clos testdir.chdir
return capsysbinary return testdir

View File

@@ -5,7 +5,6 @@ import uuid
from django.db import models from django.db import models
from django.contrib.auth.models import User from django.contrib.auth.models import User
from django.db.models.query_utils import DeferredAttribute from django.db.models.query_utils import DeferredAttribute
from django.db.models.manager import BaseManager
class TagManager(models.Manager): class TagManager(models.Manager):

View File

@@ -284,13 +284,6 @@ def doctest_with_space():
import_issu import_issu
""" """
def doctest_issue_github_1748():
"""From GitHub #1748
#? 10 []
This. Al
"""
pass
def docstring_rst_identifiers(): def docstring_rst_identifiers():
""" """

View File

@@ -1,3 +1,5 @@
# python >= 3.6
class Foo: class Foo:
bar = 1 bar = 1

View File

@@ -309,8 +309,3 @@ def annotation2() -> Iterator[float]:
next(annotation1()) next(annotation1())
#? float() #? float()
next(annotation2()) next(annotation2())
# annotations should override generator inference
#? float()
annotation1()

View File

@@ -110,4 +110,4 @@ class Test(object):
# nocond lambdas make no sense at all. # nocond lambdas make no sense at all.
#? int() #? int()
[a for a in [1,2] if (lambda: 3)][0] [a for a in [1,2] if lambda: 3][0]

View File

@@ -1 +0,0 @@
mod1_name = 'mod1'

View File

@@ -1 +0,0 @@
mod2_name = 'mod2'

View File

@@ -1,18 +0,0 @@
import sys
import os
from os.path import dirname
sys.path.insert(0, os.path.join(dirname(__file__), 'namespace2'))
sys.path.insert(0, os.path.join(dirname(__file__), 'namespace1'))
#? ['mod1']
import pkg1.pkg2.mod1
#? ['mod2']
import pkg1.pkg2.mod2
#? ['mod1_name']
pkg1.pkg2.mod1.mod1_name
#? ['mod2_name']
pkg1.pkg2.mod2.mod2_name

View File

@@ -23,9 +23,11 @@ def builtin_test():
import sqlite3 import sqlite3
# classes is a local module that has an __init__.py and can therefore not be # classes is a local module that has an __init__.py and can therefore not be
# found. # found. test can be found.
#? [] #? []
import classes import classes
#? ['test']
import test
#? ['timedelta'] #? ['timedelta']
from datetime import timedel from datetime import timedel

View File

@@ -1,50 +0,0 @@
""" Pep-0484 type hinted decorators """
from typing import Callable
def decorator(func):
def wrapper(*a, **k):
return str(func(*a, **k))
return wrapper
def typed_decorator(func: Callable[..., int]) -> Callable[..., str]:
...
# Functions
@decorator
def plain_func() -> int:
return 4
#? str()
plain_func()
@typed_decorator
def typed_func() -> int:
return 4
#? str()
typed_func()
# Methods
class X:
@decorator
def plain_method(self) -> int:
return 4
@typed_decorator
def typed_method(self) -> int:
return 4
inst = X()
#? str()
inst.plain_method()
#? str()
inst.typed_method()

View File

@@ -27,13 +27,13 @@ class PlainClass(object):
tpl = ("1", 2) tpl = ("1", 2)
tpl_typed: Tuple[str, int] = ("2", 3) tpl_typed = ("2", 3) # type: Tuple[str, int]
collection = {"a": 1} collection = {"a": 1}
collection_typed: Dict[str, int] = {"a": 1} collection_typed = {"a": 1} # type: Dict[str, int]
list_of_ints: List[int] = [42] list_of_ints = [42] # type: List[int]
list_of_funcs: List[Callable[[T], T]] = [foo] list_of_funcs = [foo] # type: List[Callable[[T], T]]
custom_generic = CustomGeneric(123.45) custom_generic = CustomGeneric(123.45)

View File

@@ -19,12 +19,12 @@ T_co = TypeVar('T_co', covariant=True)
V = TypeVar('V') V = TypeVar('V')
just_float: float = 42. just_float = 42. # type: float
optional_float: Optional[float] = 42. optional_float = 42. # type: Optional[float]
list_of_ints: List[int] = [42] list_of_ints = [42] # type: List[int]
list_of_floats: List[float] = [42.] list_of_floats = [42.] # type: List[float]
list_of_optional_floats: List[Optional[float]] = [x or None for x in list_of_floats] list_of_optional_floats = [x or None for x in list_of_floats] # type: List[Optional[float]]
list_of_ints_and_strs: List[Union[int, str]] = [42, 'abc'] list_of_ints_and_strs = [42, 'abc'] # type: List[Union[int, str]]
# Test that simple parameters are handled # Test that simple parameters are handled
def list_t_to_list_t(the_list: List[T]) -> List[T]: def list_t_to_list_t(the_list: List[T]) -> List[T]:
@@ -48,7 +48,7 @@ for z in list_t_to_list_t(list_of_ints_and_strs):
z z
list_of_int_type: List[Type[int]] = [int] list_of_int_type = [int] # type: List[Type[int]]
# Test that nested parameters are handled # Test that nested parameters are handled
def list_optional_t_to_list_t(the_list: List[Optional[T]]) -> List[T]: def list_optional_t_to_list_t(the_list: List[Optional[T]]) -> List[T]:
@@ -85,7 +85,7 @@ def optional_list_t_to_list_t(x: Optional[List[T]]) -> List[T]:
return x if x is not None else [] return x if x is not None else []
optional_list_float: Optional[List[float]] = None optional_list_float = None # type: Optional[List[float]]
for xc in optional_list_t_to_list_t(optional_list_float): for xc in optional_list_t_to_list_t(optional_list_float):
#? float() #? float()
xc xc
@@ -134,7 +134,7 @@ def list_tuple_t_to_tuple_list_t(the_list: List[Tuple[T]]) -> Tuple[List[T], ...
return tuple(list(x) for x in the_list) return tuple(list(x) for x in the_list)
list_of_int_tuples: List[Tuple[int]] = [(x,) for x in list_of_ints] list_of_int_tuples = [(x,) for x in list_of_ints] # type: List[Tuple[int]]
for b in list_tuple_t_to_tuple_list_t(list_of_int_tuples): for b in list_tuple_t_to_tuple_list_t(list_of_int_tuples):
#? int() #? int()
@@ -145,7 +145,7 @@ def list_tuple_t_elipsis_to_tuple_list_t(the_list: List[Tuple[T, ...]]) -> Tuple
return tuple(list(x) for x in the_list) return tuple(list(x) for x in the_list)
list_of_int_tuple_elipsis: List[Tuple[int, ...]] = [tuple(list_of_ints)] list_of_int_tuple_elipsis = [tuple(list_of_ints)] # type: List[Tuple[int, ...]]
for b in list_tuple_t_elipsis_to_tuple_list_t(list_of_int_tuple_elipsis): for b in list_tuple_t_elipsis_to_tuple_list_t(list_of_int_tuple_elipsis):
#? int() #? int()
@@ -157,7 +157,7 @@ def foo(x: int) -> int:
return x return x
list_of_funcs: List[Callable[[int], int]] = [foo] list_of_funcs = [foo] # type: List[Callable[[int], int]]
def list_func_t_to_list_func_type_t(the_list: List[Callable[[T], T]]) -> List[Callable[[Type[T]], T]]: def list_func_t_to_list_func_type_t(the_list: List[Callable[[T], T]]) -> List[Callable[[Type[T]], T]]:
def adapt(func: Callable[[T], T]) -> Callable[[Type[T]], T]: def adapt(func: Callable[[T], T]) -> Callable[[Type[T]], T]:
@@ -176,7 +176,7 @@ def bar(*a, **k) -> int:
return len(a) + len(k) return len(a) + len(k)
list_of_funcs_2: List[Callable[..., int]] = [bar] list_of_funcs_2 = [bar] # type: List[Callable[..., int]]
def list_func_t_passthrough(the_list: List[Callable[..., T]]) -> List[Callable[..., T]]: def list_func_t_passthrough(the_list: List[Callable[..., T]]) -> List[Callable[..., T]]:
return the_list return the_list
@@ -187,7 +187,7 @@ for b in list_func_t_passthrough(list_of_funcs_2):
b(None, x="x") b(None, x="x")
mapping_int_str: Dict[int, str] = {42: 'a'} mapping_int_str = {42: 'a'} # type: Dict[int, str]
# Test that mappings (that have more than one parameter) are handled # Test that mappings (that have more than one parameter) are handled
def invert_mapping(mapping: Mapping[K, V]) -> Mapping[V, K]: def invert_mapping(mapping: Mapping[K, V]) -> Mapping[V, K]:
@@ -210,11 +210,11 @@ first(mapping_int_str)
#? str() #? str()
first("abc") first("abc")
some_str: str = NotImplemented some_str = NotImplemented # type: str
#? str() #? str()
first(some_str) first(some_str)
annotated: List[ Callable[[Sequence[float]], int] ] = [len] annotated = [len] # type: List[ Callable[[Sequence[float]], int] ]
#? int() #? int()
first(annotated)() first(annotated)()
@@ -237,7 +237,7 @@ for b in values(mapping_int_str):
# #
# Tests that user-defined generic types are handled # Tests that user-defined generic types are handled
# #
list_ints: List[int] = [42] list_ints = [42] # type: List[int]
class CustomGeneric(Generic[T_co]): class CustomGeneric(Generic[T_co]):
def __init__(self, val: T_co) -> None: def __init__(self, val: T_co) -> None:
@@ -248,7 +248,7 @@ class CustomGeneric(Generic[T_co]):
def custom(x: CustomGeneric[T]) -> T: def custom(x: CustomGeneric[T]) -> T:
return x.val return x.val
custom_instance: CustomGeneric[int] = CustomGeneric(42) custom_instance = CustomGeneric(42) # type: CustomGeneric[int]
#? int() #? int()
custom(custom_instance) custom(custom_instance)
@@ -275,7 +275,7 @@ for x5 in wrap_custom(list_ints):
# Test extraction of type from a nested custom generic type # Test extraction of type from a nested custom generic type
list_custom_instances: List[CustomGeneric[int]] = [CustomGeneric(42)] list_custom_instances = [CustomGeneric(42)] # type: List[CustomGeneric[int]]
def unwrap_custom(iterable: Iterable[CustomGeneric[T]]) -> List[T]: def unwrap_custom(iterable: Iterable[CustomGeneric[T]]) -> List[T]:
return [x.val for x in iterable] return [x.val for x in iterable]
@@ -303,7 +303,7 @@ for xg in unwrap_custom(CustomGeneric(s) for s in 'abc'):
# Test extraction of type from type parameer nested within a custom generic type # Test extraction of type from type parameer nested within a custom generic type
custom_instance_list_int: CustomGeneric[List[int]] = CustomGeneric([42]) custom_instance_list_int = CustomGeneric([42]) # type: CustomGeneric[List[int]]
def unwrap_custom2(instance: CustomGeneric[Iterable[T]]) -> List[T]: def unwrap_custom2(instance: CustomGeneric[Iterable[T]]) -> List[T]:
return list(instance.val) return list(instance.val)
@@ -326,7 +326,7 @@ class Specialised(Mapping[int, str]):
pass pass
specialised_instance: Specialised = NotImplemented specialised_instance = NotImplemented # type: Specialised
#? int() #? int()
first(specialised_instance) first(specialised_instance)
@@ -341,7 +341,7 @@ class ChildOfSpecialised(Specialised):
pass pass
child_of_specialised_instance: ChildOfSpecialised = NotImplemented child_of_specialised_instance = NotImplemented # type: ChildOfSpecialised
#? int() #? int()
first(child_of_specialised_instance) first(child_of_specialised_instance)
@@ -355,13 +355,13 @@ class CustomPartialGeneric1(Mapping[str, T]):
pass pass
custom_partial1_instance: CustomPartialGeneric1[int] = NotImplemented custom_partial1_instance = NotImplemented # type: CustomPartialGeneric1[int]
#? str() #? str()
first(custom_partial1_instance) first(custom_partial1_instance)
custom_partial1_unbound_instance: CustomPartialGeneric1 = NotImplemented custom_partial1_unbound_instance = NotImplemented # type: CustomPartialGeneric1
#? str() #? str()
first(custom_partial1_unbound_instance) first(custom_partial1_unbound_instance)
@@ -371,7 +371,7 @@ class CustomPartialGeneric2(Mapping[T, str]):
pass pass
custom_partial2_instance: CustomPartialGeneric2[int] = NotImplemented custom_partial2_instance = NotImplemented # type: CustomPartialGeneric2[int]
#? int() #? int()
first(custom_partial2_instance) first(custom_partial2_instance)
@@ -380,7 +380,7 @@ first(custom_partial2_instance)
values(custom_partial2_instance)[0] values(custom_partial2_instance)[0]
custom_partial2_unbound_instance: CustomPartialGeneric2 = NotImplemented custom_partial2_unbound_instance = NotImplemented # type: CustomPartialGeneric2
#? [] #? []
first(custom_partial2_unbound_instance) first(custom_partial2_unbound_instance)

View File

@@ -19,16 +19,16 @@ TTypeAny = TypeVar('TTypeAny', bound=Type[Any])
TCallable = TypeVar('TCallable', bound=Callable[..., Any]) TCallable = TypeVar('TCallable', bound=Callable[..., Any])
untyped_list_str = ['abc', 'def'] untyped_list_str = ['abc', 'def']
typed_list_str: List[str] = ['abc', 'def'] typed_list_str = ['abc', 'def'] # type: List[str]
untyped_tuple_str = ('abc',) untyped_tuple_str = ('abc',)
typed_tuple_str: Tuple[str] = ('abc',) typed_tuple_str = ('abc',) # type: Tuple[str]
untyped_tuple_str_int = ('abc', 4) untyped_tuple_str_int = ('abc', 4)
typed_tuple_str_int: Tuple[str, int] = ('abc', 4) typed_tuple_str_int = ('abc', 4) # type: Tuple[str, int]
variadic_tuple_str: Tuple[str, ...] = ('abc',) variadic_tuple_str = ('abc',) # type: Tuple[str, ...]
variadic_tuple_str_int: Tuple[Union[str, int], ...] = ('abc', 4) variadic_tuple_str_int = ('abc', 4) # type: Tuple[Union[str, int], ...]
def untyped_passthrough(x): def untyped_passthrough(x):
@@ -58,16 +58,6 @@ def typed_bound_generic_passthrough(x: TList) -> TList:
return x return x
# Forward references are more likely with custom types, however this aims to
# test just the handling of the quoted type rather than any other part of the
# machinery.
def typed_quoted_return_generic_passthrough(x: T) -> 'List[T]':
return [x]
def typed_quoted_input_generic_passthrough(x: 'Tuple[T]') -> T:
x
return x[0]
for a in untyped_passthrough(untyped_list_str): for a in untyped_passthrough(untyped_list_str):
#? str() #? str()
@@ -156,23 +146,6 @@ for q in typed_bound_generic_passthrough(typed_list_str):
q q
for r in typed_quoted_return_generic_passthrough("something"):
#? str()
r
for s in typed_quoted_return_generic_passthrough(42):
#? int()
s
#? str()
typed_quoted_input_generic_passthrough(("something",))
#? int()
typed_quoted_input_generic_passthrough((42,))
class CustomList(List): class CustomList(List):
def get_first(self): def get_first(self):
return self[0] return self[0]

View File

@@ -1,9 +1,10 @@
# python >= 3.6
from typing import List, Dict, overload, Tuple, TypeVar from typing import List, Dict, overload, Tuple, TypeVar
lst: list lst: list
list_alias: List list_alias: List
list_str: List[str] list_str: List[str]
list_int: List[int] list_str: List[int]
# ------------------------- # -------------------------
# With base classes # With base classes

View File

@@ -2,14 +2,18 @@
Test the typing library, with docstrings and annotations Test the typing library, with docstrings and annotations
""" """
import typing import typing
from typing import Sequence, MutableSequence, List, Iterable, Iterator, \
AbstractSet, Tuple, Mapping, Dict, Union, Optional
class B: class B:
pass pass
def we_can_has_sequence(p: Sequence[int], q: Sequence[B], r: Sequence[int], def we_can_has_sequence(p, q, r, s, t, u):
s: Sequence["int"], t: MutableSequence[dict], u: List[float]): """
:type p: typing.Sequence[int]
:type q: typing.Sequence[B]
:type r: typing.Sequence[int]
:type s: typing.Sequence["int"]
:type t: typing.MutableSequence[dict]
:type u: typing.List[float]
"""
#? ["count"] #? ["count"]
p.c p.c
#? int() #? int()
@@ -39,8 +43,13 @@ def we_can_has_sequence(p: Sequence[int], q: Sequence[B], r: Sequence[int],
#? float() #? float()
u[1] u[1]
def iterators(ps: Iterable[int], qs: Iterator[str], rs: def iterators(ps, qs, rs, ts):
Sequence["ForwardReference"], ts: AbstractSet["float"]): """
:type ps: typing.Iterable[int]
:type qs: typing.Iterator[str]
:type rs: typing.Sequence["ForwardReference"]
:type ts: typing.AbstractSet["float"]
"""
for p in ps: for p in ps:
#? int() #? int()
p p
@@ -70,13 +79,22 @@ def iterators(ps: Iterable[int], qs: Iterator[str], rs:
#? float() #? float()
t t
def sets(p: AbstractSet[int], q: typing.MutableSet[float]): def sets(p, q):
"""
:type p: typing.AbstractSet[int]
:type q: typing.MutableSet[float]
"""
#? [] #? []
p.a p.a
#? ["add"] #? ["add"]
q.a q.a
def tuple(p: Tuple[int], q: Tuple[int, str, float], r: Tuple[B, ...]): def tuple(p, q, r):
"""
:type p: typing.Tuple[int]
:type q: typing.Tuple[int, str, float]
:type r: typing.Tuple[B, ...]
"""
#? int() #? int()
p[0] p[0]
#? ['index'] #? ['index']
@@ -109,14 +127,16 @@ class Key:
class Value: class Value:
pass pass
def mapping( def mapping(p, q, d, dd, r, s, t):
p: Mapping[Key, Value], """
q: typing.MutableMapping[Key, Value], :type p: typing.Mapping[Key, Value]
d: Dict[Key, Value], :type q: typing.MutableMapping[Key, Value]
dd: typing.DefaultDict[Key, Value], :type d: typing.Dict[Key, Value]
r: typing.KeysView[Key], :type dd: typing.DefaultDict[Key, Value]
s: typing.ValuesView[Value], :type r: typing.KeysView[Key]
t: typing.ItemsView[Key, Value]): :type s: typing.ValuesView[Value]
:type t: typing.ItemsView[Key, Value]
"""
#? [] #? []
p.setd p.setd
#? ["setdefault"] #? ["setdefault"]
@@ -178,12 +198,14 @@ def mapping(
#? Value() #? Value()
value value
def union( def union(p, q, r, s, t):
p: Union[int], """
q: Union[int, int], :type p: typing.Union[int]
r: Union[int, str, "int"], :type q: typing.Union[int, int]
s: Union[int, typing.Union[str, "typing.Union['float', 'dict']"]], :type r: typing.Union[int, str, "int"]
t: Union[int, None]): :type s: typing.Union[int, typing.Union[str, "typing.Union['float', 'dict']"]]
:type t: typing.Union[int, None]
"""
#? int() #? int()
p p
#? int() #? int()
@@ -195,8 +217,9 @@ def union(
#? int() None #? int() None
t t
def optional(p: Optional[int]): def optional(p):
""" """
:type p: typing.Optional[int]
Optional does not do anything special. However it should be recognised Optional does not do anything special. However it should be recognised
as being of that type. Jedi doesn't do anything with the extra into that as being of that type. Jedi doesn't do anything with the extra into that
it can be None as well it can be None as well
@@ -211,7 +234,10 @@ class TestDict(typing.Dict[str, int]):
def setdud(self): def setdud(self):
pass pass
def testdict(x: TestDict): def testdict(x):
"""
:type x: TestDict
"""
#? ["setdud", "setdefault"] #? ["setdud", "setdefault"]
x.setd x.setd
for key in x.keys(): for key in x.keys():
@@ -236,7 +262,10 @@ y = WrappingType(0) # Per https://github.com/davidhalter/jedi/issues/1015#issuec
#? str() #? str()
y y
def testnewtype(y: WrappingType): def testnewtype(y):
"""
:type y: WrappingType
"""
#? str() #? str()
y y
#? ["upper"] #? ["upper"]
@@ -244,7 +273,10 @@ def testnewtype(y: WrappingType):
WrappingType2 = typing.NewType() WrappingType2 = typing.NewType()
def testnewtype2(y: WrappingType2): def testnewtype2(y):
"""
:type y: WrappingType2
"""
#? #?
y y
#? [] #? []
@@ -265,7 +297,10 @@ class TestDefaultDict(typing.DefaultDict[str, int]):
def setdud(self): def setdud(self):
pass pass
def testdict(x: TestDefaultDict): def testdict(x):
"""
:type x: TestDefaultDict
"""
#? ["setdud", "setdefault"] #? ["setdud", "setdefault"]
x.setd x.setd
for key in x.keys(): for key in x.keys():

View File

@@ -1,6 +1,9 @@
""" """
PEP 526 introduced a way of using type annotations on variables. PEP 526 introduced a new way of using type annotations on variables. It was
introduced in Python 3.6.
""" """
# python >= 3.6
import typing import typing
asdf = '' asdf = ''
@@ -44,7 +47,7 @@ class Foo():
baz: typing.ClassVar[str] baz: typing.ClassVar[str]
#? int() #?
Foo.bar Foo.bar
#? int() #? int()
Foo().bar Foo().bar
@@ -58,7 +61,6 @@ class VarClass:
var_instance2: float var_instance2: float
var_class1: typing.ClassVar[str] = 1 var_class1: typing.ClassVar[str] = 1
var_class2: typing.ClassVar[bytes] var_class2: typing.ClassVar[bytes]
var_class3 = None
def __init__(self): def __init__(self):
#? int() #? int()
@@ -71,21 +73,15 @@ class VarClass:
d.var_class2 d.var_class2
#? [] #? []
d.int d.int
#? ['var_class1', 'var_class2', 'var_instance1', 'var_instance2', 'var_class3'] #? ['var_class1', 'var_class2', 'var_instance1', 'var_instance2']
self.var_ self.var_
class VarClass2(VarClass):
var_class3: typing.ClassVar[int]
def __init__(self): #? ['var_class1', 'var_class2', 'var_instance1']
#? int()
self.var_class3
#? ['var_class1', 'var_class2', 'var_instance1', 'var_class3', 'var_instance2']
VarClass.var_ VarClass.var_
#? int() #? int()
VarClass.var_instance1 VarClass.var_instance1
#? float() #?
VarClass.var_instance2 VarClass.var_instance2
#? str() #? str()
VarClass.var_class1 VarClass.var_class1
@@ -95,7 +91,7 @@ VarClass.var_class2
VarClass.int VarClass.int
d = VarClass() d = VarClass()
#? ['var_class1', 'var_class2', 'var_class3', 'var_instance1', 'var_instance2'] #? ['var_class1', 'var_class2', 'var_instance1', 'var_instance2']
d.var_ d.var_
#? int() #? int()
d.var_instance1 d.var_instance1

View File

@@ -1,46 +0,0 @@
from pep0484_generic_parameters import list_t_to_list_t
list_of_ints_and_strs: list[int | str]
# Test that unions are handled
x2 = list_t_to_list_t(list_of_ints_and_strs)[0]
#? int() str()
x2
for z in list_t_to_list_t(list_of_ints_and_strs):
#? int() str()
z
from pep0484_generic_passthroughs import (
typed_variadic_tuple_generic_passthrough,
)
variadic_tuple_str_int: tuple[int | str, ...]
for m in typed_variadic_tuple_generic_passthrough(variadic_tuple_str_int):
#? str() int()
m
def func_returns_byteslike() -> bytes | bytearray:
pass
#? bytes() bytearray()
func_returns_byteslike()
pep604_optional_1: int | str | None
pep604_optional_2: None | bytes
#? int() str() None
pep604_optional_1
#? None bytes()
pep604_optional_2
pep604_in_str: "int | bytes"
#? int() bytes()
pep604_in_str

View File

@@ -1,5 +1,3 @@
from typing import Generator
import pytest import pytest
from pytest import fixture from pytest import fixture
@@ -66,11 +64,6 @@ def lala(my_fixture):
def lala(my_fixture): def lala(my_fixture):
pass pass
# overriding types of a fixture should be possible
def test_x(my_yield_fixture: str):
#? str()
my_yield_fixture
# ----------------- # -----------------
# completion # completion
# ----------------- # -----------------
@@ -139,6 +132,9 @@ def test_p(monkeypatch):
#? ['capsysbinary'] #? ['capsysbinary']
def test_p(capsysbin def test_p(capsysbin
#? ['tmpdir', 'tmpdir_factory']
def test_p(tmpdi
def close_parens(): def close_parens():
pass pass
@@ -168,40 +164,3 @@ def test_inheritance_fixture(inheritance_fixture, caplog):
@pytest.fixture @pytest.fixture
def caplog(caplog): def caplog(caplog):
yield caplog yield caplog
# -----------------
# Generator with annotation
# -----------------
@pytest.fixture
def with_annot() -> Generator[float, None, None]:
pass
def test_with_annot(inheritance_fixture, with_annot):
#? float()
with_annot
# -----------------
# pytest external plugins
# -----------------
#? ['admin_user', 'admin_client']
def test_z(admin
#! 15 ['def admin_client']
def test_p(admin_client):
#? ['login', 'logout']
admin_client.log
@pytest.fixture
@some_decorator
#? ['admin_user']
def bla(admin_u
return
@pytest.fixture
@some_decorator
#! 12 ['def admin_user']
def bla(admin_user):
pass

View File

@@ -1,6 +1,7 @@
import os import os
import sys import sys
import subprocess import subprocess
from itertools import count
import pytest import pytest
@@ -9,6 +10,9 @@ from . import run
from . import refactor from . import refactor
from jedi import InterpreterEnvironment, get_system_environment from jedi import InterpreterEnvironment, get_system_environment
from jedi.inference.compiled.value import create_from_access_path from jedi.inference.compiled.value import create_from_access_path
from jedi.inference.imports import _load_python_module
from jedi.file_io import KnownContentFileIO
from jedi.inference.base_value import ValueSet
from jedi.api.interpreter import MixedModuleContext from jedi.api.interpreter import MixedModuleContext
# For interpreter tests sometimes the path of this directory is in the sys # For interpreter tests sometimes the path of this directory is in the sys
@@ -159,6 +163,19 @@ def create_compiled_object(inference_state):
) )
@pytest.fixture
def module_injector():
counter = count()
def module_injector(inference_state, names, code):
assert isinstance(names, tuple)
file_io = KnownContentFileIO('/foo/bar/module-injector-%s.py' % next(counter), code)
v = _load_python_module(inference_state, file_io, names)
inference_state.module_cache.add(names, ValueSet([v]))
return module_injector
@pytest.fixture(params=[False, True]) @pytest.fixture(params=[False, True])
def class_findable(monkeypatch, request): def class_findable(monkeypatch, request):
if not request.param: if not request.param:

View File

@@ -1,6 +0,0 @@
from pytest import fixture
@fixture()
def admin_user():
pass

View File

@@ -1,16 +0,0 @@
import pytest
from .fixtures import admin_user # noqa
@pytest.fixture()
def admin_client():
return Client()
class Client:
def login(self, **credentials):
...
def logout(self):
...

View File

@@ -104,14 +104,10 @@ import os
import re import re
import sys import sys
import operator import operator
if sys.version_info < (3, 8): from ast import literal_eval
literal_eval = eval
else:
from ast import literal_eval
from io import StringIO from io import StringIO
from functools import reduce from functools import reduce
from unittest.mock import ANY from unittest.mock import ANY
from pathlib import Path
import parso import parso
from _pytest.outcomes import Skipped from _pytest.outcomes import Skipped
@@ -126,7 +122,6 @@ from jedi.api.environment import get_default_environment, get_system_environment
from jedi.inference.gradual.conversion import convert_values from jedi.inference.gradual.conversion import convert_values
from jedi.inference.analysis import Warning from jedi.inference.analysis import Warning
test_dir = Path(__file__).absolute().parent
TEST_COMPLETIONS = 0 TEST_COMPLETIONS = 0
TEST_INFERENCE = 1 TEST_INFERENCE = 1
@@ -178,7 +173,6 @@ class IntegrationTestCase(BaseTestCase):
self.start = start self.start = start
self.line = line self.line = line
self.path = path self.path = path
self._project = jedi.Project(test_dir)
@property @property
def module_name(self): def module_name(self):
@@ -194,12 +188,7 @@ class IntegrationTestCase(BaseTestCase):
self.line_nr_test, self.line.rstrip()) self.line_nr_test, self.line.rstrip())
def script(self, environment): def script(self, environment):
return jedi.Script( return jedi.Script(self.source, path=self.path, environment=environment)
self.source,
path=self.path,
environment=environment,
project=self._project
)
def run(self, compare_cb, environment=None): def run(self, compare_cb, environment=None):
testers = { testers = {
@@ -209,7 +198,7 @@ class IntegrationTestCase(BaseTestCase):
TEST_REFERENCES: self.run_get_references, TEST_REFERENCES: self.run_get_references,
} }
if (self.path.endswith('pytest.py') or self.path.endswith('conftest.py')) \ if (self.path.endswith('pytest.py') or self.path.endswith('conftest.py')) \
and os.path.realpath(environment.executable) != os.path.realpath(sys.executable): and environment.executable != os.path.realpath(sys.executable):
# It's not guarantueed that pytest is installed in test # It's not guarantueed that pytest is installed in test
# environments, if we're not running in the same environment that # environments, if we're not running in the same environment that
# we're already in, so just skip that case. # we're already in, so just skip that case.
@@ -274,7 +263,7 @@ class IntegrationTestCase(BaseTestCase):
self.correct = self.correct.strip() self.correct = self.correct.strip()
compare = sorted( compare = sorted(
(('stub:' if r.is_stub() else '') (('stub:' if r.is_stub() else '')
+ re.sub(r'^completion\.', '', r.module_name), + re.sub(r'^test\.completion\.', '', r.module_name),
r.line, r.line,
r.column) r.column)
for r in result for r in result

View File

@@ -650,7 +650,6 @@ def test_cursor_after_signature(Script, column):
('abs(chr ( \nclass y: pass', 1, 8, 'abs', 0), ('abs(chr ( \nclass y: pass', 1, 8, 'abs', 0),
('abs(chr ( \nclass y: pass', 1, 9, 'abs', 0), ('abs(chr ( \nclass y: pass', 1, 9, 'abs', 0),
('abs(chr ( \nclass y: pass', 1, 10, 'chr', 0), ('abs(chr ( \nclass y: pass', 1, 10, 'chr', 0),
('abs(foo.bar=3)', 1, 13, 'abs', 0),
] ]
) )
def test_base_signatures(Script, code, line, column, name, index): def test_base_signatures(Script, code, line, column, name, index):

View File

@@ -1,16 +1,11 @@
from os.path import join, sep as s, dirname, expanduser from os.path import join, sep as s, dirname, expanduser
import os import os
from textwrap import dedent from textwrap import dedent
from itertools import count
from pathlib import Path
import pytest import pytest
from ..helpers import root_dir from ..helpers import root_dir
from jedi.api.helpers import _start_match, _fuzzy_match from jedi.api.helpers import _start_match, _fuzzy_match
from jedi.inference.imports import _load_python_module
from jedi.file_io import KnownContentFileIO
from jedi.inference.base_value import ValueSet
def test_in_whitespace(Script): def test_in_whitespace(Script):
@@ -405,22 +400,6 @@ def test_ellipsis_completion(Script):
assert Script('...').complete() == [] assert Script('...').complete() == []
@pytest.fixture
def module_injector():
counter = count()
def module_injector(inference_state, names, code):
assert isinstance(names, tuple)
file_io = KnownContentFileIO(
Path('foo/bar/module-injector-%s.py' % next(counter)).absolute(),
code
)
v = _load_python_module(inference_state, file_io, names)
inference_state.module_cache.add(names, ValueSet([v]))
return module_injector
def test_completion_cache(Script, module_injector): def test_completion_cache(Script, module_injector):
""" """
For some modules like numpy, tensorflow or pandas we cache docstrings and For some modules like numpy, tensorflow or pandas we cache docstrings and
@@ -457,7 +436,3 @@ def test_module_completions(Script, module):
# Just make sure that there are no errors # Just make sure that there are no errors
c.type c.type
c.docstring() c.docstring()
def test_whitespace_at_end_after_dot(Script):
assert 'strip' in [c.name for c in Script('str. ').complete()]

View File

@@ -37,17 +37,6 @@ def test_operator_doc(Script):
assert len(d.docstring()) > 100 assert len(d.docstring()) > 100
@pytest.mark.parametrize(
'code, help_part', [
('str', 'Create a new string object'),
('str.strip', 'Return a copy of the string'),
]
)
def test_stdlib_doc(Script, code, help_part):
h, = Script(code).help()
assert help_part in h.docstring(raw=True)
def test_lambda(Script): def test_lambda(Script):
d, = Script('lambda x: x').help(column=0) d, = Script('lambda x: x').help(column=0)
assert d.type == 'keyword' assert d.type == 'keyword'

View File

@@ -603,11 +603,8 @@ def test_dict_getitem(code, types):
@pytest.mark.parametrize( @pytest.mark.parametrize(
'code, expected', [ 'code, expected', [
('DunderCls()[0]', 'int'), ('DunderCls()[0]', 'int'),
('dunder[0]', 'int'),
('next(DunderCls())', 'float'), ('next(DunderCls())', 'float'),
('next(dunder)', 'float'),
('for x in DunderCls(): x', 'str'), ('for x in DunderCls(): x', 'str'),
#('for x in dunder: x', 'str'),
] ]
) )
def test_dunders(class_is_findable, code, expected): def test_dunders(class_is_findable, code, expected):
@@ -626,8 +623,6 @@ def test_dunders(class_is_findable, code, expected):
if not class_is_findable: if not class_is_findable:
DunderCls.__name__ = 'asdf' DunderCls.__name__ = 'asdf'
dunder = DunderCls()
n, = jedi.Interpreter(code, [locals()]).infer() n, = jedi.Interpreter(code, [locals()]).infer()
assert n.name == expected assert n.name == expected
@@ -711,46 +706,3 @@ def test_negate():
assert x.name == 'int' assert x.name == 'int'
value, = x._name.infer() value, = x._name.infer()
assert value.get_safe_value() == -3 assert value.get_safe_value() == -3
def test_complete_not_findable_class_source():
class TestClass():
ta=1
ta1=2
# Simulate the environment where the class is defined in
# an interactive session and therefore inspect module
# cannot find its source code and raises OSError (Py 3.10+) or TypeError.
TestClass.__module__ = "__main__"
# There is a pytest __main__ module we have to remove temporarily.
module = sys.modules.pop("__main__")
try:
interpreter = jedi.Interpreter("TestClass.", [locals()])
completions = interpreter.complete(column=10, line=1)
finally:
sys.modules["__main__"] = module
assert "ta" in [c.name for c in completions]
assert "ta1" in [c.name for c in completions]
def test_param_infer_default():
abs_sig, = jedi.Interpreter('abs(', [{'abs': abs}]).get_signatures()
param, = abs_sig.params
assert param.name == 'x'
assert param.infer_default() == []
@pytest.mark.parametrize(
'code, expected', [
("random.triangular(", ['high=', 'low=', 'mode=']),
("random.triangular(low=1, ", ['high=', 'mode=']),
("random.triangular(high=1, ", ['low=', 'mode=']),
("random.triangular(low=1, high=2, ", ['mode=']),
("random.triangular(low=1, mode=2, ", ['high=']),
],
)
def test_keyword_param_completion(code, expected):
import random
completions = jedi.Interpreter(code, [locals()]).complete()
assert expected == [c.name for c in completions if c.name.endswith('=')]

View File

@@ -189,9 +189,3 @@ def test_no_error(get_names):
def test_is_side_effect(get_names, code, index, is_side_effect): def test_is_side_effect(get_names, code, index, is_side_effect):
names = get_names(code, references=True, all_scopes=True) names = get_names(code, references=True, all_scopes=True)
assert names[index].is_side_effect() == is_side_effect assert names[index].is_side_effect() == is_side_effect
def test_no_defined_names(get_names):
definition, = get_names("x = (1, 2)")
assert not definition.defined_names()

View File

@@ -68,10 +68,6 @@ def test_load_save_project(tmpdir):
dict(all_scopes=True)), dict(all_scopes=True)),
('some_search_test_var', ['test_api.test_project.test_search.some_search_test_var'], ('some_search_test_var', ['test_api.test_project.test_search.some_search_test_var'],
dict(complete=True, all_scopes=True)), dict(complete=True, all_scopes=True)),
# Make sure that the searched name is not part of the file, by
# splitting it up.
('some_search_test_v' + 'a', ['test_api.test_project.test_search.some_search_test_var'],
dict(complete=True, all_scopes=True)),
('sample_int', ['helpers.sample_int'], {}), ('sample_int', ['helpers.sample_int'], {}),
('sample_int', ['helpers.sample_int'], dict(all_scopes=True)), ('sample_int', ['helpers.sample_int'], dict(all_scopes=True)),
@@ -150,7 +146,7 @@ def test_search(string, full_names, kwargs):
defs = project.complete_search(string, **kwargs) defs = project.complete_search(string, **kwargs)
else: else:
defs = project.search(string, **kwargs) defs = project.search(string, **kwargs)
assert sorted([('stub:' if d.is_stub() else '') + (d.full_name or d.name) for d in defs]) == full_names assert sorted([('stub:' if d.is_stub() else '') + d.full_name for d in defs]) == full_names
@pytest.mark.parametrize( @pytest.mark.parametrize(
@@ -178,7 +174,7 @@ def test_is_potential_project(path, expected):
if expected is None: if expected is None:
try: try:
expected = bool(set(_CONTAINS_POTENTIAL_PROJECT) & set(os.listdir(path))) expected = _CONTAINS_POTENTIAL_PROJECT in os.listdir(path)
except OSError: except OSError:
expected = False expected = False

View File

@@ -1,7 +1,6 @@
import os import os
from textwrap import dedent from textwrap import dedent
from pathlib import Path from pathlib import Path
import platform
import pytest import pytest
@@ -71,23 +70,3 @@ def test_diff_without_ending_newline(Script):
-a -a
+c +c
''') ''')
def test_diff_path_outside_of_project(Script):
if platform.system().lower() == 'windows':
abs_path = r'D:\unknown_dir\file.py'
else:
abs_path = '/unknown_dir/file.py'
script = Script(
code='foo = 1',
path=abs_path,
project=jedi.get_default_project()
)
diff = script.rename(line=1, column=0, new_name='bar').get_diff()
assert diff == dedent(f'''\
--- {abs_path}
+++ {abs_path}
@@ -1 +1 @@
-foo = 1
+bar = 1
''')

View File

@@ -64,6 +64,6 @@ def test_wrong_encoding(Script, tmpdir):
# Use both latin-1 and utf-8 (a really broken file). # Use both latin-1 and utf-8 (a really broken file).
x.write_binary('foobar = 1\nä'.encode('latin-1') + 'ä'.encode('utf-8')) x.write_binary('foobar = 1\nä'.encode('latin-1') + 'ä'.encode('utf-8'))
project = Project(tmpdir.strpath) project = Project('.', sys_path=[tmpdir.strpath])
c, = Script('import x; x.foo', project=project).complete() c, = Script('import x; x.foo', project=project).complete()
assert c.name == 'foobar' assert c.name == 'foobar'

View File

@@ -43,9 +43,6 @@ def test_implicit_namespace_package(Script):
solution = "foo = '%s'" % solution solution = "foo = '%s'" % solution
assert completion.description == solution assert completion.description == solution
c, = script_with_path('import pkg').complete()
assert c.docstring() == ""
def test_implicit_nested_namespace_package(Script): def test_implicit_nested_namespace_package(Script):
code = 'from implicit_nested_namespaces.namespace.pkg.module import CONST' code = 'from implicit_nested_namespaces.namespace.pkg.module import CONST'

View File

@@ -101,16 +101,6 @@ def test_correct_zip_package_behavior(Script, inference_state, environment, code
assert value.py__package__() == [] assert value.py__package__() == []
@pytest.mark.parametrize("code,names", [
("from pkg.", {"module", "nested", "namespace"}),
("from pkg.nested.", {"nested_module"})
])
def test_zip_package_import_complete(Script, environment, code, names):
sys_path = environment.get_sys_path() + [str(pkg_zip_path)]
completions = Script(code, project=Project('.', sys_path=sys_path)).complete()
assert names == {c.name for c in completions}
def test_find_module_not_package_zipped(Script, inference_state, environment): def test_find_module_not_package_zipped(Script, inference_state, environment):
path = get_example_dir('zipped_imports', 'not_pkg.zip') path = get_example_dir('zipped_imports', 'not_pkg.zip')
sys_path = environment.get_sys_path() + [path] sys_path = environment.get_sys_path() + [path]
@@ -297,6 +287,7 @@ def test_os_issues(Script):
# Github issue #759 # Github issue #759
s = 'import os, s' s = 'import os, s'
assert 'sys' in import_names(s) assert 'sys' in import_names(s)
assert 'path' not in import_names(s, column=len(s) - 1)
assert 'os' in import_names(s, column=len(s) - 3) assert 'os' in import_names(s, column=len(s) - 3)
# Some more checks # Some more checks
@@ -333,13 +324,12 @@ def test_compiled_import_none(monkeypatch, Script):
# context that was initially given, but now we just work with the file # context that was initially given, but now we just work with the file
# system. # system.
(os.path.join(THIS_DIR, 'test_docstring.py'), False, (os.path.join(THIS_DIR, 'test_docstring.py'), False,
('test_inference', 'test_imports')), ('test', 'test_inference', 'test_imports')),
(os.path.join(THIS_DIR, '__init__.py'), True, (os.path.join(THIS_DIR, '__init__.py'), True,
('test_inference', 'test_imports')), ('test', 'test_inference', 'test_imports')),
] ]
) )
def test_get_modules_containing_name(inference_state, path, goal, is_package): def test_get_modules_containing_name(inference_state, path, goal, is_package):
inference_state.project = Project(test_dir)
module = imports._load_python_module( module = imports._load_python_module(
inference_state, inference_state,
FileIO(path), FileIO(path),

View File

@@ -267,19 +267,19 @@ def test_pow_signature(Script, environment):
@pytest.mark.parametrize( @pytest.mark.parametrize(
'code, signature', [ 'code, signature', [
[dedent(''' [dedent('''
# identifier:A
import functools import functools
def f(x): def f(x):
pass pass
def x(f): def x(f):
@functools.wraps(f) @functools.wraps(f)
def wrapper(*args): def wrapper(*args):
# Have no arguments here, but because of wraps, the signature
# should still be f's.
return f(*args) return f(*args)
return wrapper return wrapper
x(f)('''), 'f(x, /)'], x(f)('''), 'f(x, /)'],
[dedent(''' [dedent('''
# identifier:B
import functools import functools
def f(x): def f(x):
pass pass
@@ -292,26 +292,6 @@ def test_pow_signature(Script, environment):
return wrapper return wrapper
x(f)('''), 'f()'], x(f)('''), 'f()'],
[dedent('''
# identifier:C
import functools
def f(x: int, y: float):
pass
@functools.wraps(f)
def wrapper(*args, **kwargs):
return f(*args, **kwargs)
wrapper('''), 'f(x: int, y: float)'],
[dedent('''
# identifier:D
def f(x: int, y: float):
pass
def wrapper(*args, **kwargs):
return f(*args, **kwargs)
wrapper('''), 'wrapper(x: int, y: float)'],
] ]
) )
def test_wraps_signature(Script, code, signature): def test_wraps_signature(Script, code, signature):
@@ -355,48 +335,6 @@ def test_dataclass_signature(Script, skip_pre_python37, start, start_params):
price, = sig.params[-2].infer() price, = sig.params[-2].infer()
assert price.name == 'float' assert price.name == 'float'
@pytest.mark.parametrize(
'start, start_params', [
['@define\nclass X:', []],
['@frozen\nclass X:', []],
['@define(eq=True)\nclass X:', []],
[dedent('''
class Y():
y: int
@define
class X(Y):'''), []],
[dedent('''
@define
class Y():
y: int
z = 5
@define
class X(Y):'''), ['y']],
]
)
def test_attrs_signature(Script, skip_pre_python37, start, start_params):
has_attrs = bool(Script('import attrs').infer())
if not has_attrs:
raise pytest.skip("attrs needed in target environment to run this test")
code = dedent('''
name: str
foo = 3
price: float
quantity: int = 0.0
X(''')
# attrs exposes two namespaces
code = 'from attrs import define, frozen\n' + start + code
sig, = Script(code).get_signatures()
assert [p.name for p in sig.params] == start_params + ['name', 'price', 'quantity']
quantity, = sig.params[-1].infer()
assert quantity.name == 'int'
price, = sig.params[-2].infer()
assert price.name == 'float'
@pytest.mark.parametrize( @pytest.mark.parametrize(
'stmt, expected', [ 'stmt, expected', [

View File

@@ -1,5 +1,4 @@
import os import os
from collections import namedtuple
import pytest import pytest
@@ -43,22 +42,6 @@ def test_completion(case, monkeypatch, environment, has_django):
if (not has_django) and case.path.endswith('django.py'): if (not has_django) and case.path.endswith('django.py'):
pytest.skip('Needs django to be installed to run this test.') pytest.skip('Needs django to be installed to run this test.')
if case.path.endswith("pytest.py"):
# to test finding pytest fixtures from external plugins
# add a stub pytest plugin to the project sys_path...
pytest_plugin_dir = str(helpers.get_example_dir("pytest_plugin_package"))
case._project.added_sys_path = [pytest_plugin_dir]
# ... and mock setuptools entry points to include it
# see https://docs.pytest.org/en/stable/how-to/writing_plugins.html#setuptools-entry-points
def mock_iter_entry_points(group):
assert group == "pytest11"
EntryPoint = namedtuple("EntryPoint", ["module_name"])
return [EntryPoint("pytest_plugin.plugin")]
monkeypatch.setattr("pkg_resources.iter_entry_points", mock_iter_entry_points)
repo_root = helpers.root_dir repo_root = helpers.root_dir
monkeypatch.chdir(os.path.join(repo_root, 'jedi')) monkeypatch.chdir(os.path.join(repo_root, 'jedi'))
case.run(assert_case_equal, environment) case.run(assert_case_equal, environment)

View File

@@ -1,9 +1,5 @@
import gc
from pathlib import Path
from jedi import parser_utils from jedi import parser_utils
from parso import parse from parso import parse
from parso.cache import parser_cache
from parso.python import tree from parso.python import tree
import pytest import pytest
@@ -71,18 +67,3 @@ def test_get_signature(code, signature):
if node.type == 'simple_stmt': if node.type == 'simple_stmt':
node = node.children[0] node = node.children[0]
assert parser_utils.get_signature(node) == signature assert parser_utils.get_signature(node) == signature
def test_parser_cache_clear(Script):
"""
If parso clears its cache, Jedi should not keep those resources, they
should be freed.
"""
script = Script("a = abs\na", path=Path(__file__).parent / 'parser_cache_test_foo.py')
script.complete()
module_id = id(script._module_node)
del parser_cache[script._inference_state.grammar._hashed][script.path]
del script
gc.collect()
assert module_id not in [id(m) for m in gc.get_referrers(tree.Module)]

View File

@@ -85,7 +85,7 @@ class TestSetupReadline(unittest.TestCase):
} }
# There are quite a few differences, because both Windows and Linux # There are quite a few differences, because both Windows and Linux
# (posix and nt) librariesare included. # (posix and nt) librariesare included.
assert len(difference) < 30 assert len(difference) < 15
def test_local_import(self): def test_local_import(self):
s = 'import test.test_utils' s = 'import test.test_utils'