mirror of
https://github.com/davidhalter/jedi.git
synced 2025-12-06 14:04:26 +08:00
Compare commits
1 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
354dab9503 |
9
.github/workflows/ci.yml
vendored
9
.github/workflows/ci.yml
vendored
@@ -1,5 +1,5 @@
|
||||
name: ci
|
||||
on: [push, pull_request]
|
||||
on: push
|
||||
|
||||
jobs:
|
||||
tests:
|
||||
@@ -7,8 +7,8 @@ jobs:
|
||||
strategy:
|
||||
matrix:
|
||||
os: [ubuntu-20.04, windows-2019]
|
||||
python-version: ["3.10", "3.9", "3.8", "3.7", "3.6"]
|
||||
environment: ['3.8', '3.10', '3.9', '3.7', '3.6', 'interpreter']
|
||||
python-version: [3.9, 3.8, 3.7, 3.6]
|
||||
environment: ['3.8', '3.9', '3.7', '3.6', 'interpreter']
|
||||
steps:
|
||||
- name: Checkout code
|
||||
uses: actions/checkout@v2
|
||||
@@ -27,6 +27,9 @@ jobs:
|
||||
- name: Install dependencies
|
||||
run: 'pip install .[testing]'
|
||||
|
||||
- name: Setup tmate session
|
||||
uses: mxschmitt/action-tmate@v3
|
||||
|
||||
- name: Run tests
|
||||
run: python -m pytest
|
||||
env:
|
||||
|
||||
1
.gitignore
vendored
1
.gitignore
vendored
@@ -14,4 +14,3 @@ record.json
|
||||
/.pytest_cache
|
||||
/.mypy_cache
|
||||
/venv/
|
||||
.nvimrc
|
||||
|
||||
@@ -1,11 +1,2 @@
|
||||
version: 2
|
||||
|
||||
python:
|
||||
install:
|
||||
- method: pip
|
||||
path: .
|
||||
extra_requirements:
|
||||
- docs
|
||||
|
||||
submodules:
|
||||
include: all
|
||||
pip_install: true
|
||||
|
||||
@@ -1,4 +1,4 @@
|
||||
Main Authors
|
||||
Main Authors
|
||||
------------
|
||||
|
||||
- David Halter (@davidhalter) <davidhalter88@gmail.com>
|
||||
@@ -61,8 +61,6 @@ Code Contributors
|
||||
- Vladislav Serebrennikov (@endilll)
|
||||
- Andrii Kolomoiets (@muffinmad)
|
||||
- Leo Ryu (@Leo-Ryu)
|
||||
- Joseph Birkner (@josephbirkner)
|
||||
- Márcio Mazza (@marciomazza)
|
||||
|
||||
And a few more "anonymous" contributors.
|
||||
|
||||
|
||||
@@ -6,22 +6,7 @@ Changelog
|
||||
Unreleased
|
||||
++++++++++
|
||||
|
||||
0.18.2 (2022-11-21)
|
||||
+++++++++++++++++++
|
||||
|
||||
- Added dataclass-equivalent for attrs.define
|
||||
- Find fixtures from Pytest entrypoints; Examples of pytest plugins installed
|
||||
like this are pytest-django, pytest-sugar and Faker.
|
||||
- Fixed Project.search, when a venv was involved, which is why for example
|
||||
`:Pyimport django.db` did not work in some cases in jedi-vim.
|
||||
- And many smaller bugfixes
|
||||
|
||||
0.18.1 (2021-11-17)
|
||||
+++++++++++++++++++
|
||||
|
||||
- Implict namespaces are now a separate types in ``Name().type``
|
||||
- Python 3.10 support
|
||||
- Mostly bugfixes
|
||||
|
||||
0.18.0 (2020-12-25)
|
||||
+++++++++++++++++++
|
||||
|
||||
@@ -57,7 +57,7 @@ Supported Python Features
|
||||
Limitations
|
||||
-----------
|
||||
|
||||
In general Jedi's limit is quite high, but for very big projects or very
|
||||
In general Jedi's limit are quite high, but for very big projects or very
|
||||
complex code, sometimes Jedi intentionally stops type inference, to avoid
|
||||
hanging for a long time.
|
||||
|
||||
|
||||
@@ -27,7 +27,7 @@ ad
|
||||
load
|
||||
"""
|
||||
|
||||
__version__ = '0.18.2'
|
||||
__version__ = '0.18.0'
|
||||
|
||||
from jedi.api import Script, Interpreter, set_debug_function, preload_module
|
||||
from jedi import settings
|
||||
|
||||
@@ -7,6 +7,22 @@ import sys
|
||||
import pickle
|
||||
|
||||
|
||||
def cast_path(string):
|
||||
"""
|
||||
Take a bytes or str path and cast it to unicode.
|
||||
|
||||
Apparently it is perfectly fine to pass both byte and unicode objects into
|
||||
the sys.path. This probably means that byte paths are normal at other
|
||||
places as well.
|
||||
|
||||
Since this just really complicates everything and Python 2.7 will be EOL
|
||||
soon anyway, just go with always strings.
|
||||
"""
|
||||
if isinstance(string, bytes):
|
||||
return str(string, encoding='UTF-8', errors='replace')
|
||||
return str(string)
|
||||
|
||||
|
||||
def pickle_load(file):
|
||||
try:
|
||||
return pickle.load(file)
|
||||
|
||||
@@ -13,6 +13,7 @@ from pathlib import Path
|
||||
import parso
|
||||
from parso.python import tree
|
||||
|
||||
from jedi._compatibility import cast_path
|
||||
from jedi.parser_utils import get_executable_nodes
|
||||
from jedi import debug
|
||||
from jedi import settings
|
||||
@@ -99,15 +100,13 @@ class Script:
|
||||
"""
|
||||
def __init__(self, code=None, *, path=None, environment=None, project=None):
|
||||
self._orig_path = path
|
||||
# An empty path (also empty string) should always result in no path.
|
||||
if isinstance(path, str):
|
||||
path = Path(path)
|
||||
|
||||
self.path = path.absolute() if path else None
|
||||
|
||||
if code is None:
|
||||
if path is None:
|
||||
raise ValueError("Must provide at least one of code or path")
|
||||
|
||||
# TODO add a better warning than the traceback!
|
||||
with open(path, 'rb') as f:
|
||||
code = f.read()
|
||||
@@ -153,7 +152,7 @@ class Script:
|
||||
if self.path is None:
|
||||
file_io = None
|
||||
else:
|
||||
file_io = KnownContentFileIO(self.path, self._code)
|
||||
file_io = KnownContentFileIO(cast_path(self.path), self._code)
|
||||
if self.path is not None and self.path.suffix == '.pyi':
|
||||
# We are in a stub file. Try to load the stub properly.
|
||||
stub_module = load_proper_stub_module(
|
||||
@@ -581,7 +580,7 @@ class Script:
|
||||
@validate_line_column
|
||||
def extract_variable(self, line, column, *, new_name, until_line=None, until_column=None):
|
||||
"""
|
||||
Moves an expression to a new statement.
|
||||
Moves an expression to a new statemenet.
|
||||
|
||||
For example if you have the cursor on ``foo`` and provide a
|
||||
``new_name`` called ``bar``::
|
||||
@@ -710,7 +709,7 @@ class Interpreter(Script):
|
||||
"""
|
||||
_allow_descriptor_getattr_default = True
|
||||
|
||||
def __init__(self, code, namespaces, *, project=None, **kwds):
|
||||
def __init__(self, code, namespaces, **kwds):
|
||||
try:
|
||||
namespaces = [dict(n) for n in namespaces]
|
||||
except Exception:
|
||||
@@ -723,23 +722,16 @@ class Interpreter(Script):
|
||||
if not isinstance(environment, InterpreterEnvironment):
|
||||
raise TypeError("The environment needs to be an InterpreterEnvironment subclass.")
|
||||
|
||||
if project is None:
|
||||
project = Project(Path.cwd())
|
||||
|
||||
super().__init__(code, environment=environment, project=project, **kwds)
|
||||
|
||||
super().__init__(code, environment=environment,
|
||||
project=Project(Path.cwd()), **kwds)
|
||||
self.namespaces = namespaces
|
||||
self._inference_state.allow_descriptor_getattr = self._allow_descriptor_getattr_default
|
||||
|
||||
@cache.memoize_method
|
||||
def _get_module_context(self):
|
||||
if self.path is None:
|
||||
file_io = None
|
||||
else:
|
||||
file_io = KnownContentFileIO(self.path, self._code)
|
||||
tree_module_value = ModuleValue(
|
||||
self._inference_state, self._module_node,
|
||||
file_io=file_io,
|
||||
file_io=KnownContentFileIO(str(self.path), self._code),
|
||||
string_names=('__main__',),
|
||||
code_lines=self._code_lines,
|
||||
)
|
||||
|
||||
@@ -27,7 +27,7 @@ from jedi.inference.compiled.mixed import MixedName
|
||||
from jedi.inference.names import ImportName, SubModuleName
|
||||
from jedi.inference.gradual.stub_value import StubModuleValue
|
||||
from jedi.inference.gradual.conversion import convert_names, convert_values
|
||||
from jedi.inference.base_value import ValueSet, HasNoContext
|
||||
from jedi.inference.base_value import ValueSet
|
||||
from jedi.api.keywords import KeywordName
|
||||
from jedi.api import completion_cache
|
||||
from jedi.api.helpers import filter_follow_imports
|
||||
@@ -37,17 +37,13 @@ def _sort_names_by_start_pos(names):
|
||||
return sorted(names, key=lambda s: s.start_pos or (0, 0))
|
||||
|
||||
|
||||
def defined_names(inference_state, value):
|
||||
def defined_names(inference_state, context):
|
||||
"""
|
||||
List sub-definitions (e.g., methods in class).
|
||||
|
||||
:type scope: Scope
|
||||
:rtype: list of Name
|
||||
"""
|
||||
try:
|
||||
context = value.as_context()
|
||||
except HasNoContext:
|
||||
return []
|
||||
filter = next(context.get_filters())
|
||||
names = [name for name in filter.values()]
|
||||
return [Name(inference_state, n) for n in _sort_names_by_start_pos(names)]
|
||||
@@ -763,7 +759,7 @@ class Name(BaseName):
|
||||
"""
|
||||
defs = self._name.infer()
|
||||
return sorted(
|
||||
unite(defined_names(self._inference_state, d) for d in defs),
|
||||
unite(defined_names(self._inference_state, d.as_context()) for d in defs),
|
||||
key=lambda s: s._name.start_pos or (0, 0)
|
||||
)
|
||||
|
||||
|
||||
@@ -18,8 +18,7 @@ from jedi.inference import imports
|
||||
from jedi.inference.base_value import ValueSet
|
||||
from jedi.inference.helpers import infer_call_of_leaf, parse_dotted_names
|
||||
from jedi.inference.context import get_global_filters
|
||||
from jedi.inference.value import TreeInstance
|
||||
from jedi.inference.docstring_utils import DocstringModule
|
||||
from jedi.inference.value import TreeInstance, ModuleValue
|
||||
from jedi.inference.names import ParamNameWrapper, SubModuleName
|
||||
from jedi.inference.gradual.conversion import convert_values, convert_names
|
||||
from jedi.parser_utils import cut_value_at_position
|
||||
@@ -195,6 +194,7 @@ class Completion:
|
||||
- In args: */**: no completion
|
||||
- In params (also lambda): no completion before =
|
||||
"""
|
||||
|
||||
grammar = self._inference_state.grammar
|
||||
self.stack = stack = None
|
||||
self._position = (
|
||||
@@ -277,10 +277,6 @@ class Completion:
|
||||
)
|
||||
elif nonterminals[-1] in ('trailer', 'dotted_name') and nodes[-1] == '.':
|
||||
dot = self._module_node.get_leaf_for_position(self._position)
|
||||
if dot.type == "endmarker":
|
||||
# This is a bit of a weird edge case, maybe we can somehow
|
||||
# generalize this.
|
||||
dot = leaf.get_previous_leaf()
|
||||
cached_name, n = self._complete_trailer(dot.get_previous_leaf())
|
||||
completion_names += n
|
||||
elif self._is_parameter_completion():
|
||||
@@ -466,12 +462,12 @@ class Completion:
|
||||
|
||||
def _complete_code_lines(self, code_lines):
|
||||
module_node = self._inference_state.grammar.parse(''.join(code_lines))
|
||||
module_value = DocstringModule(
|
||||
in_module_context=self._module_context,
|
||||
inference_state=self._inference_state,
|
||||
module_node=module_node,
|
||||
module_value = ModuleValue(
|
||||
self._inference_state,
|
||||
module_node,
|
||||
code_lines=code_lines,
|
||||
)
|
||||
module_value.parent_context = self._module_context
|
||||
return Completion(
|
||||
self._inference_state,
|
||||
module_value.as_context(),
|
||||
|
||||
@@ -17,7 +17,7 @@ import parso
|
||||
|
||||
_VersionInfo = namedtuple('VersionInfo', 'major minor micro')
|
||||
|
||||
_SUPPORTED_PYTHONS = ['3.10', '3.9', '3.8', '3.7', '3.6']
|
||||
_SUPPORTED_PYTHONS = ['3.9', '3.8', '3.7', '3.6']
|
||||
_SAFE_PATHS = ['/usr/bin', '/usr/local/bin']
|
||||
_CONDA_VAR = 'CONDA_PREFIX'
|
||||
_CURRENT_VERSION = '%s.%s' % (sys.version_info.major, sys.version_info.minor)
|
||||
|
||||
@@ -23,7 +23,7 @@ class RefactoringError(_JediError):
|
||||
Refactorings can fail for various reasons. So if you work with refactorings
|
||||
like :meth:`.Script.rename`, :meth:`.Script.inline`,
|
||||
:meth:`.Script.extract_variable` and :meth:`.Script.extract_function`, make
|
||||
sure to catch these. The descriptions in the errors are usually valuable
|
||||
sure to catch these. The descriptions in the errors are ususally valuable
|
||||
for end users.
|
||||
|
||||
A typical ``RefactoringError`` would tell the user that inlining is not
|
||||
|
||||
@@ -205,6 +205,7 @@ def filter_follow_imports(names, follow_builtin_imports=False):
|
||||
|
||||
class CallDetails:
|
||||
def __init__(self, bracket_leaf, children, position):
|
||||
['bracket_leaf', 'call_index', 'keyword_name_str']
|
||||
self.bracket_leaf = bracket_leaf
|
||||
self._children = children
|
||||
self._position = position
|
||||
@@ -280,7 +281,7 @@ class CallDetails:
|
||||
def count_positional_arguments(self):
|
||||
count = 0
|
||||
for star_count, key_start, had_equal in self._list_arguments()[:-1]:
|
||||
if star_count or key_start:
|
||||
if star_count:
|
||||
break
|
||||
count += 1
|
||||
return count
|
||||
@@ -306,7 +307,7 @@ def _iter_arguments(nodes, position):
|
||||
first = node.children[0]
|
||||
second = node.children[1]
|
||||
if second == '=':
|
||||
if second.start_pos < position and first.type == 'name':
|
||||
if second.start_pos < position:
|
||||
yield 0, first.value, True
|
||||
else:
|
||||
yield 0, remove_after_pos(first), False
|
||||
|
||||
@@ -106,16 +106,7 @@ class Project:
|
||||
with open(self._get_json_path(self._path), 'w') as f:
|
||||
return json.dump((_SERIALIZER_VERSION, data), f)
|
||||
|
||||
def __init__(
|
||||
self,
|
||||
path,
|
||||
*,
|
||||
environment_path=None,
|
||||
load_unsafe_extensions=False,
|
||||
sys_path=None,
|
||||
added_sys_path=(),
|
||||
smart_sys_path=True,
|
||||
) -> None:
|
||||
def __init__(self, path, **kwargs):
|
||||
"""
|
||||
:param path: The base path for this project.
|
||||
:param environment_path: The Python executable path, typically the path
|
||||
@@ -134,22 +125,25 @@ class Project:
|
||||
local directories. Otherwise you will have to rely on your packages
|
||||
being properly configured on the ``sys.path``.
|
||||
"""
|
||||
def py2_comp(path, environment_path=None, load_unsafe_extensions=False,
|
||||
sys_path=None, added_sys_path=(), smart_sys_path=True):
|
||||
if isinstance(path, str):
|
||||
path = Path(path).absolute()
|
||||
self._path = path
|
||||
|
||||
if isinstance(path, str):
|
||||
path = Path(path).absolute()
|
||||
self._path = path
|
||||
|
||||
self._environment_path = environment_path
|
||||
if sys_path is not None:
|
||||
self._environment_path = environment_path
|
||||
if sys_path is not None:
|
||||
# Remap potential pathlib.Path entries
|
||||
sys_path = list(map(str, sys_path))
|
||||
self._sys_path = sys_path
|
||||
self._smart_sys_path = smart_sys_path
|
||||
self._load_unsafe_extensions = load_unsafe_extensions
|
||||
self._django = False
|
||||
# Remap potential pathlib.Path entries
|
||||
sys_path = list(map(str, sys_path))
|
||||
self._sys_path = sys_path
|
||||
self._smart_sys_path = smart_sys_path
|
||||
self._load_unsafe_extensions = load_unsafe_extensions
|
||||
self._django = False
|
||||
# Remap potential pathlib.Path entries
|
||||
self.added_sys_path = list(map(str, added_sys_path))
|
||||
"""The sys path that is going to be added at the end of the """
|
||||
self.added_sys_path = list(map(str, added_sys_path))
|
||||
"""The sys path that is going to be added at the end of the """
|
||||
|
||||
py2_comp(path, **kwargs)
|
||||
|
||||
@property
|
||||
def path(self):
|
||||
@@ -334,8 +328,7 @@ class Project:
|
||||
)
|
||||
|
||||
# 2. Search for identifiers in the project.
|
||||
for module_context in search_in_file_ios(inference_state, file_ios,
|
||||
name, complete=complete):
|
||||
for module_context in search_in_file_ios(inference_state, file_ios, name):
|
||||
names = get_module_names(module_context.tree_node, all_scopes=all_scopes)
|
||||
names = [module_context.create_name(n) for n in names]
|
||||
names = _remove_imports(names)
|
||||
@@ -352,8 +345,9 @@ class Project:
|
||||
# 3. Search for modules on sys.path
|
||||
sys_path = [
|
||||
p for p in self._get_sys_path(inference_state)
|
||||
# Exclude the current folder which is handled by recursing the folders.
|
||||
if p != self._path
|
||||
# Exclude folders that are handled by recursing of the Python
|
||||
# folders.
|
||||
if not p.startswith(str(self._path))
|
||||
]
|
||||
names = list(iter_module_names(inference_state, empty_module_context, sys_path))
|
||||
yield from search_in_module(
|
||||
@@ -432,6 +426,7 @@ def get_default_project(path=None):
|
||||
probable_path = dir
|
||||
|
||||
if probable_path is not None:
|
||||
# TODO search for setup.py etc
|
||||
return Project(probable_path)
|
||||
|
||||
if first_no_init_file is not None:
|
||||
|
||||
@@ -42,17 +42,11 @@ class ChangedFile:
|
||||
if self._from_path is None:
|
||||
from_p = ''
|
||||
else:
|
||||
try:
|
||||
from_p = self._from_path.relative_to(project_path)
|
||||
except ValueError: # Happens it the path is not on th project_path
|
||||
from_p = self._from_path
|
||||
from_p = self._from_path.relative_to(project_path)
|
||||
if self._to_path is None:
|
||||
to_p = ''
|
||||
else:
|
||||
try:
|
||||
to_p = self._to_path.relative_to(project_path)
|
||||
except ValueError:
|
||||
to_p = self._to_path
|
||||
to_p = self._to_path.relative_to(project_path)
|
||||
diff = difflib.unified_diff(
|
||||
old_lines, new_lines,
|
||||
fromfile=str(from_p),
|
||||
|
||||
@@ -106,7 +106,10 @@ def dbg(message, *args, color='GREEN'):
|
||||
debug_function(color, i + 'dbg: ' + message % tuple(repr(a) for a in args))
|
||||
|
||||
|
||||
def warning(message, *args, format=True):
|
||||
def warning(message, *args, **kwargs):
|
||||
format = kwargs.pop('format', True)
|
||||
assert not kwargs
|
||||
|
||||
if debug_function and enable_warning:
|
||||
i = ' ' * _debug_indent
|
||||
if format:
|
||||
|
||||
@@ -90,7 +90,7 @@ class InferenceState:
|
||||
self.compiled_subprocess = environment.get_inference_state_subprocess(self)
|
||||
self.grammar = environment.get_grammar()
|
||||
|
||||
self.latest_grammar = parso.load_grammar(version='3.10')
|
||||
self.latest_grammar = parso.load_grammar(version='3.7')
|
||||
self.memoize_cache = {} # for memoize decorators
|
||||
self.module_cache = imports.ModuleCache() # does the job of `sys.modules`.
|
||||
self.stub_module_cache = {} # Dict[Tuple[str, ...], Optional[ModuleValue]]
|
||||
@@ -181,6 +181,8 @@ class InferenceState:
|
||||
|
||||
def parse_and_get_code(self, code=None, path=None,
|
||||
use_latest_grammar=False, file_io=None, **kwargs):
|
||||
if path is not None:
|
||||
path = str(path)
|
||||
if code is None:
|
||||
if file_io is None:
|
||||
file_io = FileIO(path)
|
||||
|
||||
@@ -22,10 +22,6 @@ from jedi.cache import memoize_method
|
||||
sentinel = object()
|
||||
|
||||
|
||||
class HasNoContext(Exception):
|
||||
pass
|
||||
|
||||
|
||||
class HelperValueMixin:
|
||||
def get_root_context(self):
|
||||
value = self
|
||||
@@ -265,7 +261,7 @@ class Value(HelperValueMixin):
|
||||
return self.parent_context.is_stub()
|
||||
|
||||
def _as_context(self):
|
||||
raise HasNoContext
|
||||
raise NotImplementedError('Not all values need to be converted to contexts: %s', self)
|
||||
|
||||
@property
|
||||
def name(self):
|
||||
@@ -297,7 +293,7 @@ class Value(HelperValueMixin):
|
||||
just the `_T` generic parameter.
|
||||
|
||||
`value_set`: represents the actual argument passed to the parameter
|
||||
we're inferred for, or (for recursive calls) their types. In the
|
||||
we're inferrined for, or (for recursive calls) their types. In the
|
||||
above example this would first be the representation of the list
|
||||
`[1]` and then, when recursing, just of `1`.
|
||||
"""
|
||||
|
||||
@@ -8,8 +8,6 @@ import warnings
|
||||
import re
|
||||
import builtins
|
||||
import typing
|
||||
from pathlib import Path
|
||||
from typing import Optional
|
||||
|
||||
from jedi.inference.compiled.getattr_static import getattr_static
|
||||
|
||||
@@ -181,9 +179,9 @@ class DirectObjectAccess:
|
||||
def py__bool__(self):
|
||||
return bool(self._obj)
|
||||
|
||||
def py__file__(self) -> Optional[Path]:
|
||||
def py__file__(self):
|
||||
try:
|
||||
return Path(self._obj.__file__)
|
||||
return self._obj.__file__
|
||||
except AttributeError:
|
||||
return None
|
||||
|
||||
@@ -213,22 +211,7 @@ class DirectObjectAccess:
|
||||
def py__getitem__all_values(self):
|
||||
if isinstance(self._obj, dict):
|
||||
return [self._create_access_path(v) for v in self._obj.values()]
|
||||
if isinstance(self._obj, (list, tuple)):
|
||||
return [self._create_access_path(v) for v in self._obj]
|
||||
|
||||
if self.is_instance():
|
||||
cls = DirectObjectAccess(self._inference_state, self._obj.__class__)
|
||||
return cls.py__getitem__all_values()
|
||||
|
||||
try:
|
||||
getitem = self._obj.__getitem__
|
||||
except AttributeError:
|
||||
pass
|
||||
else:
|
||||
annotation = DirectObjectAccess(self._inference_state, getitem).get_return_annotation()
|
||||
if annotation is not None:
|
||||
return [annotation]
|
||||
return None
|
||||
return self.py__iter__list()
|
||||
|
||||
def py__simple_getitem__(self, index):
|
||||
if type(self._obj) not in ALLOWED_GETITEM_TYPES:
|
||||
@@ -238,14 +221,8 @@ class DirectObjectAccess:
|
||||
return self._create_access_path(self._obj[index])
|
||||
|
||||
def py__iter__list(self):
|
||||
try:
|
||||
iter_method = self._obj.__iter__
|
||||
except AttributeError:
|
||||
if not hasattr(self._obj, '__getitem__'):
|
||||
return None
|
||||
else:
|
||||
p = DirectObjectAccess(self._inference_state, iter_method).get_return_annotation()
|
||||
if p is not None:
|
||||
return [p]
|
||||
|
||||
if type(self._obj) not in ALLOWED_GETITEM_TYPES:
|
||||
# Get rid of side effects, we won't call custom `__getitem__`s.
|
||||
@@ -329,9 +306,9 @@ class DirectObjectAccess:
|
||||
except TypeError:
|
||||
return False
|
||||
|
||||
def is_allowed_getattr(self, name, safe=True):
|
||||
def is_allowed_getattr(self, name, unsafe=False):
|
||||
# TODO this API is ugly.
|
||||
if not safe:
|
||||
if unsafe:
|
||||
# Unsafe is mostly used to check for __getattr__/__getattribute__.
|
||||
# getattr_static works for properties, but the underscore methods
|
||||
# are just ignored (because it's safer and avoids more code
|
||||
@@ -384,7 +361,7 @@ class DirectObjectAccess:
|
||||
except AttributeError:
|
||||
pass
|
||||
else:
|
||||
if module is not None and isinstance(module, str):
|
||||
if module is not None:
|
||||
try:
|
||||
__import__(module)
|
||||
# For some modules like _sqlite3, the __module__ for classes is
|
||||
|
||||
@@ -34,7 +34,7 @@ class MixedObject(ValueWrapper):
|
||||
|
||||
This combined logic makes it possible to provide more powerful REPL
|
||||
completion. It allows side effects that are not noticable with the default
|
||||
parser structure to still be completable.
|
||||
parser structure to still be completeable.
|
||||
|
||||
The biggest difference from CompiledValue to MixedObject is that we are
|
||||
generally dealing with Python code and not with C code. This will generate
|
||||
@@ -187,7 +187,7 @@ def _find_syntax_node_name(inference_state, python_object):
|
||||
try:
|
||||
python_object = _get_object_to_check(python_object)
|
||||
path = inspect.getsourcefile(python_object)
|
||||
except (OSError, TypeError):
|
||||
except TypeError:
|
||||
# The type might not be known (e.g. class_with_dict.__weakref__)
|
||||
return None
|
||||
path = None if path is None else Path(path)
|
||||
@@ -267,7 +267,7 @@ def _find_syntax_node_name(inference_state, python_object):
|
||||
@inference_state_function_cache()
|
||||
def _create(inference_state, compiled_value, module_context):
|
||||
# TODO accessing this is bad, but it probably doesn't matter that much,
|
||||
# because we're working with interpreters only here.
|
||||
# because we're working with interpreteters only here.
|
||||
python_object = compiled_value.access_handle.access._obj
|
||||
result = _find_syntax_node_name(inference_state, python_object)
|
||||
if result is None:
|
||||
|
||||
@@ -7,7 +7,6 @@ goals:
|
||||
2. Make it possible to handle different Python versions as well as virtualenvs.
|
||||
"""
|
||||
|
||||
import collections
|
||||
import os
|
||||
import sys
|
||||
import queue
|
||||
@@ -169,7 +168,7 @@ class CompiledSubprocess:
|
||||
def __init__(self, executable, env_vars=None):
|
||||
self._executable = executable
|
||||
self._env_vars = env_vars
|
||||
self._inference_state_deletion_queue = collections.deque()
|
||||
self._inference_state_deletion_queue = queue.deque()
|
||||
self._cleanup_callable = lambda: None
|
||||
|
||||
def __repr__(self):
|
||||
|
||||
@@ -4,10 +4,10 @@ import inspect
|
||||
import importlib
|
||||
import warnings
|
||||
from pathlib import Path
|
||||
from zipfile import ZipFile
|
||||
from zipimport import zipimporter, ZipImportError
|
||||
from zipimport import zipimporter
|
||||
from importlib.machinery import all_suffixes
|
||||
|
||||
from jedi._compatibility import cast_path
|
||||
from jedi.inference.compiled import access
|
||||
from jedi import debug
|
||||
from jedi import parser_utils
|
||||
@@ -15,7 +15,7 @@ from jedi.file_io import KnownContentFileIO, ZipFileIO
|
||||
|
||||
|
||||
def get_sys_path():
|
||||
return sys.path
|
||||
return list(map(cast_path, sys.path))
|
||||
|
||||
|
||||
def load_module(inference_state, **kwargs):
|
||||
@@ -93,22 +93,15 @@ def _iter_module_names(inference_state, paths):
|
||||
# Python modules/packages
|
||||
for path in paths:
|
||||
try:
|
||||
dir_entries = ((entry.name, entry.is_dir()) for entry in os.scandir(path))
|
||||
dirs = os.scandir(path)
|
||||
except OSError:
|
||||
try:
|
||||
zip_import_info = zipimporter(path)
|
||||
# Unfortunately, there is no public way to access zipimporter's
|
||||
# private _files member. We therefore have to use a
|
||||
# custom function to iterate over the files.
|
||||
dir_entries = _zip_list_subdirectory(
|
||||
zip_import_info.archive, zip_import_info.prefix)
|
||||
except ZipImportError:
|
||||
# The file might not exist or reading it might lead to an error.
|
||||
debug.warning("Not possible to list directory: %s", path)
|
||||
continue
|
||||
for name, is_dir in dir_entries:
|
||||
# The file might not exist or reading it might lead to an error.
|
||||
debug.warning("Not possible to list directory: %s", path)
|
||||
continue
|
||||
for dir_entry in dirs:
|
||||
name = dir_entry.name
|
||||
# First Namespaces then modules/stubs
|
||||
if is_dir:
|
||||
if dir_entry.is_dir():
|
||||
# pycache is obviously not an interesting namespace. Also the
|
||||
# name must be a valid identifier.
|
||||
if name != '__pycache__' and name.isidentifier():
|
||||
@@ -151,11 +144,7 @@ def _find_module(string, path=None, full_name=None, is_global_search=True):
|
||||
|
||||
spec = find_spec(string, p)
|
||||
if spec is not None:
|
||||
if spec.origin == "frozen":
|
||||
continue
|
||||
|
||||
loader = spec.loader
|
||||
|
||||
if loader is None and not spec.has_location:
|
||||
# This is a namespace package.
|
||||
full_name = string if not path else full_name
|
||||
@@ -201,7 +190,7 @@ def _from_loader(loader, string):
|
||||
except AttributeError:
|
||||
return None, is_package
|
||||
else:
|
||||
module_path = get_filename(string)
|
||||
module_path = cast_path(get_filename(string))
|
||||
|
||||
# To avoid unicode and read bytes, "overwrite" loader.get_source if
|
||||
# possible.
|
||||
@@ -223,7 +212,7 @@ def _from_loader(loader, string):
|
||||
if code is None:
|
||||
return None, is_package
|
||||
if isinstance(loader, zipimporter):
|
||||
return ZipFileIO(module_path, code, Path(loader.archive)), is_package
|
||||
return ZipFileIO(module_path, code, Path(cast_path(loader.archive))), is_package
|
||||
|
||||
return KnownContentFileIO(module_path, code), is_package
|
||||
|
||||
@@ -241,17 +230,6 @@ def _get_source(loader, fullname):
|
||||
name=fullname)
|
||||
|
||||
|
||||
def _zip_list_subdirectory(zip_path, zip_subdir_path):
|
||||
zip_file = ZipFile(zip_path)
|
||||
zip_subdir_path = Path(zip_subdir_path)
|
||||
zip_content_file_paths = zip_file.namelist()
|
||||
for raw_file_name in zip_content_file_paths:
|
||||
file_path = Path(raw_file_name)
|
||||
if file_path.parent == zip_subdir_path:
|
||||
file_path = file_path.relative_to(zip_subdir_path)
|
||||
yield file_path.name, raw_file_name.endswith("/")
|
||||
|
||||
|
||||
class ImplicitNSInfo:
|
||||
"""Stores information returned from an implicit namespace spec"""
|
||||
def __init__(self, name, paths):
|
||||
|
||||
@@ -5,10 +5,10 @@ import re
|
||||
from functools import partial
|
||||
from inspect import Parameter
|
||||
from pathlib import Path
|
||||
from typing import Optional
|
||||
|
||||
from jedi import debug
|
||||
from jedi.inference.utils import to_list
|
||||
from jedi._compatibility import cast_path
|
||||
from jedi.cache import memoize_method
|
||||
from jedi.inference.filters import AbstractFilter
|
||||
from jedi.inference.names import AbstractNameDefinition, ValueNameMixin, \
|
||||
@@ -167,7 +167,7 @@ class CompiledValue(Value):
|
||||
except AttributeError:
|
||||
return super().py__simple_getitem__(index)
|
||||
if access is None:
|
||||
return super().py__simple_getitem__(index)
|
||||
return NO_VALUES
|
||||
|
||||
return ValueSet([create_from_access_path(self.inference_state, access)])
|
||||
|
||||
@@ -293,7 +293,10 @@ class CompiledModule(CompiledValue):
|
||||
return CompiledModuleContext(self)
|
||||
|
||||
def py__path__(self):
|
||||
return self.access_handle.py__path__()
|
||||
paths = self.access_handle.py__path__()
|
||||
if paths is None:
|
||||
return None
|
||||
return map(cast_path, paths)
|
||||
|
||||
def is_package(self):
|
||||
return self.py__path__() is not None
|
||||
@@ -306,8 +309,11 @@ class CompiledModule(CompiledValue):
|
||||
return ()
|
||||
return tuple(name.split('.'))
|
||||
|
||||
def py__file__(self) -> Optional[Path]:
|
||||
return self.access_handle.py__file__() # type: ignore[no-any-return]
|
||||
def py__file__(self):
|
||||
path = cast_path(self.access_handle.py__file__())
|
||||
if path is None:
|
||||
return None
|
||||
return Path(path)
|
||||
|
||||
|
||||
class CompiledName(AbstractNameDefinition):
|
||||
@@ -434,7 +440,7 @@ class CompiledValueFilter(AbstractFilter):
|
||||
access_handle = self.compiled_value.access_handle
|
||||
return self._get(
|
||||
name,
|
||||
lambda name, safe: access_handle.is_allowed_getattr(name, safe=safe),
|
||||
lambda name, unsafe: access_handle.is_allowed_getattr(name, unsafe),
|
||||
lambda name: name in access_handle.dir(),
|
||||
check_has_attribute=True
|
||||
)
|
||||
@@ -448,7 +454,7 @@ class CompiledValueFilter(AbstractFilter):
|
||||
|
||||
has_attribute, is_descriptor = allowed_getattr_callback(
|
||||
name,
|
||||
safe=not self._inference_state.allow_descriptor_getattr
|
||||
unsafe=self._inference_state.allow_descriptor_getattr
|
||||
)
|
||||
if check_has_attribute and not has_attribute:
|
||||
return []
|
||||
@@ -472,7 +478,7 @@ class CompiledValueFilter(AbstractFilter):
|
||||
from jedi.inference.compiled import builtin_from_name
|
||||
names = []
|
||||
needs_type_completions, dir_infos = self.compiled_value.access_handle.get_dir_infos()
|
||||
# We could use `safe=False` here as well, especially as a parameter to
|
||||
# We could use `unsafe` here as well, especially as a parameter to
|
||||
# get_dir_infos. But this would lead to a lot of property executions
|
||||
# that are probably not wanted. The drawback for this is that we
|
||||
# have a different name for `get` and `values`. For `get` we always
|
||||
@@ -480,7 +486,7 @@ class CompiledValueFilter(AbstractFilter):
|
||||
for name in dir_infos:
|
||||
names += self._get(
|
||||
name,
|
||||
lambda name, safe: dir_infos[name],
|
||||
lambda name, unsafe: dir_infos[name],
|
||||
lambda name: name in dir_infos,
|
||||
)
|
||||
|
||||
|
||||
@@ -1,7 +1,5 @@
|
||||
from abc import abstractmethod
|
||||
from contextlib import contextmanager
|
||||
from pathlib import Path
|
||||
from typing import Optional
|
||||
|
||||
from parso.tree import search_ancestor
|
||||
from parso.python.tree import Name
|
||||
@@ -309,8 +307,8 @@ class FunctionContext(TreeContextMixin, ValueContext):
|
||||
|
||||
|
||||
class ModuleContext(TreeContextMixin, ValueContext):
|
||||
def py__file__(self) -> Optional[Path]:
|
||||
return self._value.py__file__() # type: ignore[no-any-return]
|
||||
def py__file__(self):
|
||||
return self._value.py__file__()
|
||||
|
||||
def get_filters(self, until_position=None, origin_scope=None):
|
||||
filters = self._value.get_filters(origin_scope)
|
||||
@@ -327,7 +325,7 @@ class ModuleContext(TreeContextMixin, ValueContext):
|
||||
yield from filters
|
||||
|
||||
def get_global_filter(self):
|
||||
return GlobalNameFilter(self)
|
||||
return GlobalNameFilter(self, self.tree_node)
|
||||
|
||||
@property
|
||||
def string_names(self):
|
||||
@@ -357,8 +355,8 @@ class NamespaceContext(TreeContextMixin, ValueContext):
|
||||
def string_names(self):
|
||||
return self._value.string_names
|
||||
|
||||
def py__file__(self) -> Optional[Path]:
|
||||
return self._value.py__file__() # type: ignore[no-any-return]
|
||||
def py__file__(self):
|
||||
return self._value.py__file__()
|
||||
|
||||
|
||||
class ClassContext(TreeContextMixin, ValueContext):
|
||||
@@ -407,8 +405,8 @@ class CompiledModuleContext(CompiledContext):
|
||||
def string_names(self):
|
||||
return self._value.string_names
|
||||
|
||||
def py__file__(self) -> Optional[Path]:
|
||||
return self._value.py__file__() # type: ignore[no-any-return]
|
||||
def py__file__(self):
|
||||
return self._value.py__file__()
|
||||
|
||||
|
||||
def _get_global_filters_for_name(context, name_or_none, position):
|
||||
|
||||
@@ -1,21 +0,0 @@
|
||||
from jedi.inference.value import ModuleValue
|
||||
from jedi.inference.context import ModuleContext
|
||||
|
||||
|
||||
class DocstringModule(ModuleValue):
|
||||
def __init__(self, in_module_context, **kwargs):
|
||||
super().__init__(**kwargs)
|
||||
self._in_module_context = in_module_context
|
||||
|
||||
def _as_context(self):
|
||||
return DocstringModuleContext(self, self._in_module_context)
|
||||
|
||||
|
||||
class DocstringModuleContext(ModuleContext):
|
||||
def __init__(self, module_value, in_module_context):
|
||||
super().__init__(module_value)
|
||||
self._in_module_context = in_module_context
|
||||
|
||||
def get_filters(self, origin_scope=None, until_position=None):
|
||||
yield from super().get_filters(until_position=until_position)
|
||||
yield from self._in_module_context.get_filters()
|
||||
@@ -17,10 +17,12 @@ annotations.
|
||||
|
||||
import re
|
||||
import warnings
|
||||
from textwrap import dedent
|
||||
|
||||
from parso import parse, ParserSyntaxError
|
||||
|
||||
from jedi import debug
|
||||
from jedi.common import indent_block
|
||||
from jedi.inference.cache import inference_state_method_cache
|
||||
from jedi.inference.base_value import iterator_to_value_set, ValueSet, \
|
||||
NO_VALUES
|
||||
@@ -180,40 +182,52 @@ def _strip_rst_role(type_str):
|
||||
|
||||
|
||||
def _infer_for_statement_string(module_context, string):
|
||||
code = dedent("""
|
||||
def pseudo_docstring_stuff():
|
||||
'''
|
||||
Create a pseudo function for docstring statements.
|
||||
Need this docstring so that if the below part is not valid Python this
|
||||
is still a function.
|
||||
'''
|
||||
{}
|
||||
""")
|
||||
if string is None:
|
||||
return []
|
||||
|
||||
potential_imports = re.findall(r'((?:\w+\.)*\w+)\.', string)
|
||||
# Try to import module part in dotted name.
|
||||
# (e.g., 'threading' in 'threading.Thread').
|
||||
imports = "\n".join(f"import {p}" for p in potential_imports)
|
||||
string = f'{imports}\n{string}'
|
||||
for element in re.findall(r'((?:\w+\.)*\w+)\.', string):
|
||||
# Try to import module part in dotted name.
|
||||
# (e.g., 'threading' in 'threading.Thread').
|
||||
string = 'import %s\n' % element + string
|
||||
|
||||
debug.dbg('Parse docstring code %s', string, color='BLUE')
|
||||
grammar = module_context.inference_state.grammar
|
||||
try:
|
||||
module = grammar.parse(string, error_recovery=False)
|
||||
module = grammar.parse(code.format(indent_block(string)), error_recovery=False)
|
||||
except ParserSyntaxError:
|
||||
return []
|
||||
try:
|
||||
# It's not the last item, because that's an end marker.
|
||||
stmt = module.children[-2]
|
||||
funcdef = next(module.iter_funcdefs())
|
||||
# First pick suite, then simple_stmt and then the node,
|
||||
# which is also not the last item, because there's a newline.
|
||||
stmt = funcdef.children[-1].children[-1].children[-2]
|
||||
except (AttributeError, IndexError):
|
||||
return []
|
||||
|
||||
if stmt.type not in ('name', 'atom', 'atom_expr'):
|
||||
return []
|
||||
|
||||
# Here we basically use a fake module that also uses the filters in
|
||||
# the actual module.
|
||||
from jedi.inference.docstring_utils import DocstringModule
|
||||
m = DocstringModule(
|
||||
in_module_context=module_context,
|
||||
inference_state=module_context.inference_state,
|
||||
module_node=module,
|
||||
code_lines=[],
|
||||
from jedi.inference.value import FunctionValue
|
||||
function_value = FunctionValue(
|
||||
module_context.inference_state,
|
||||
module_context,
|
||||
funcdef
|
||||
)
|
||||
return list(_execute_types_in_stmt(m.as_context(), stmt))
|
||||
func_execution_context = function_value.as_context()
|
||||
# Use the module of the param.
|
||||
# TODO this module is not the module of the param in case of a function
|
||||
# call. In that case it's the module of the function call.
|
||||
# stuffed with content from a function call.
|
||||
return list(_execute_types_in_stmt(func_execution_context, stmt))
|
||||
|
||||
|
||||
def _execute_types_in_stmt(module_context, stmt):
|
||||
|
||||
@@ -12,7 +12,7 @@ from parso.python.tree import Name, UsedNamesMapping
|
||||
from jedi.inference import flow_analysis
|
||||
from jedi.inference.base_value import ValueSet, ValueWrapper, \
|
||||
LazyValueWrapper
|
||||
from jedi.parser_utils import get_cached_parent_scope, get_parso_cache_node
|
||||
from jedi.parser_utils import get_cached_parent_scope
|
||||
from jedi.inference.utils import to_list
|
||||
from jedi.inference.names import TreeNameDefinition, ParamName, \
|
||||
AnonymousParamName, AbstractNameDefinition, NameWrapper
|
||||
@@ -54,15 +54,11 @@ class FilterWrapper:
|
||||
return self.wrap_names(self._wrapped_filter.values())
|
||||
|
||||
|
||||
def _get_definition_names(parso_cache_node, used_names, name_key):
|
||||
if parso_cache_node is None:
|
||||
names = used_names.get(name_key, ())
|
||||
return tuple(name for name in names if name.is_definition(include_setitem=True))
|
||||
|
||||
def _get_definition_names(used_names, name_key):
|
||||
try:
|
||||
for_module = _definition_name_cache[parso_cache_node]
|
||||
for_module = _definition_name_cache[used_names]
|
||||
except KeyError:
|
||||
for_module = _definition_name_cache[parso_cache_node] = {}
|
||||
for_module = _definition_name_cache[used_names] = {}
|
||||
|
||||
try:
|
||||
return for_module[name_key]
|
||||
@@ -74,40 +70,18 @@ def _get_definition_names(parso_cache_node, used_names, name_key):
|
||||
return result
|
||||
|
||||
|
||||
class _AbstractUsedNamesFilter(AbstractFilter):
|
||||
class AbstractUsedNamesFilter(AbstractFilter):
|
||||
name_class = TreeNameDefinition
|
||||
|
||||
def __init__(self, parent_context, node_context=None):
|
||||
if node_context is None:
|
||||
node_context = parent_context
|
||||
self._node_context = node_context
|
||||
self._parser_scope = node_context.tree_node
|
||||
module_context = node_context.get_root_context()
|
||||
# It is quite hacky that we have to use that. This is for caching
|
||||
# certain things with a WeakKeyDictionary. However, parso intentionally
|
||||
# uses slots (to save memory) and therefore we end up with having to
|
||||
# have a weak reference to the object that caches the tree.
|
||||
#
|
||||
# Previously we have tried to solve this by using a weak reference onto
|
||||
# used_names. However that also does not work, because it has a
|
||||
# reference from the module, which itself is referenced by any node
|
||||
# through parents.
|
||||
path = module_context.py__file__()
|
||||
if path is None:
|
||||
# If the path is None, there is no guarantee that parso caches it.
|
||||
self._parso_cache_node = None
|
||||
else:
|
||||
self._parso_cache_node = get_parso_cache_node(
|
||||
module_context.inference_state.latest_grammar
|
||||
if module_context.is_stub() else module_context.inference_state.grammar,
|
||||
path
|
||||
)
|
||||
self._used_names = module_context.tree_node.get_used_names()
|
||||
def __init__(self, parent_context, parser_scope):
|
||||
self._parser_scope = parser_scope
|
||||
self._module_node = self._parser_scope.get_root_node()
|
||||
self._used_names = self._module_node.get_used_names()
|
||||
self.parent_context = parent_context
|
||||
|
||||
def get(self, name):
|
||||
return self._convert_names(self._filter(
|
||||
_get_definition_names(self._parso_cache_node, self._used_names, name),
|
||||
_get_definition_names(self._used_names, name),
|
||||
))
|
||||
|
||||
def _convert_names(self, names):
|
||||
@@ -118,7 +92,7 @@ class _AbstractUsedNamesFilter(AbstractFilter):
|
||||
name
|
||||
for name_key in self._used_names
|
||||
for name in self._filter(
|
||||
_get_definition_names(self._parso_cache_node, self._used_names, name_key),
|
||||
_get_definition_names(self._used_names, name_key),
|
||||
)
|
||||
)
|
||||
|
||||
@@ -126,7 +100,7 @@ class _AbstractUsedNamesFilter(AbstractFilter):
|
||||
return '<%s: %s>' % (self.__class__.__name__, self.parent_context)
|
||||
|
||||
|
||||
class ParserTreeFilter(_AbstractUsedNamesFilter):
|
||||
class ParserTreeFilter(AbstractUsedNamesFilter):
|
||||
def __init__(self, parent_context, node_context=None, until_position=None,
|
||||
origin_scope=None):
|
||||
"""
|
||||
@@ -135,7 +109,10 @@ class ParserTreeFilter(_AbstractUsedNamesFilter):
|
||||
value, but for some type inference it's important to have a local
|
||||
value of the other classes.
|
||||
"""
|
||||
super().__init__(parent_context, node_context)
|
||||
if node_context is None:
|
||||
node_context = parent_context
|
||||
super().__init__(parent_context, node_context.tree_node)
|
||||
self._node_context = node_context
|
||||
self._origin_scope = origin_scope
|
||||
self._until_position = until_position
|
||||
|
||||
@@ -149,7 +126,7 @@ class ParserTreeFilter(_AbstractUsedNamesFilter):
|
||||
if parent.type == 'trailer':
|
||||
return False
|
||||
base_node = parent if parent.type in ('classdef', 'funcdef') else name
|
||||
return get_cached_parent_scope(self._parso_cache_node, base_node) == self._parser_scope
|
||||
return get_cached_parent_scope(self._used_names, base_node) == self._parser_scope
|
||||
|
||||
def _check_flows(self, names):
|
||||
for name in sorted(names, key=lambda name: name.start_pos, reverse=True):
|
||||
@@ -205,7 +182,7 @@ class AnonymousFunctionExecutionFilter(_FunctionExecutionFilter):
|
||||
return AnonymousParamName(self._function_value, name)
|
||||
|
||||
|
||||
class GlobalNameFilter(_AbstractUsedNamesFilter):
|
||||
class GlobalNameFilter(AbstractUsedNamesFilter):
|
||||
def get(self, name):
|
||||
try:
|
||||
names = self._used_names[name]
|
||||
|
||||
@@ -196,43 +196,13 @@ def py__annotations__(funcdef):
|
||||
return dct
|
||||
|
||||
|
||||
def resolve_forward_references(context, all_annotations):
|
||||
def resolve(node):
|
||||
if node is None or node.type != 'string':
|
||||
return node
|
||||
|
||||
node = _get_forward_reference_node(
|
||||
context,
|
||||
context.inference_state.compiled_subprocess.safe_literal_eval(
|
||||
node.value,
|
||||
),
|
||||
)
|
||||
|
||||
if node is None:
|
||||
# There was a string, but it's not a valid annotation
|
||||
return None
|
||||
|
||||
# The forward reference tree has an additional root node ('eval_input')
|
||||
# that we don't want. Extract the node we do want, that is equivalent to
|
||||
# the nodes returned by `py__annotations__` for a non-quoted node.
|
||||
node = node.children[0]
|
||||
|
||||
return node
|
||||
|
||||
return {name: resolve(node) for name, node in all_annotations.items()}
|
||||
|
||||
|
||||
@inference_state_method_cache()
|
||||
def infer_return_types(function, arguments):
|
||||
"""
|
||||
Infers the type of a function's return value,
|
||||
according to type annotations.
|
||||
"""
|
||||
context = function.get_default_param_context()
|
||||
all_annotations = resolve_forward_references(
|
||||
context,
|
||||
py__annotations__(function.tree_node),
|
||||
)
|
||||
all_annotations = py__annotations__(function.tree_node)
|
||||
annotation = all_annotations.get("return", None)
|
||||
if annotation is None:
|
||||
# If there is no Python 3-type annotation, look for an annotation
|
||||
@@ -247,10 +217,11 @@ def infer_return_types(function, arguments):
|
||||
return NO_VALUES
|
||||
|
||||
return _infer_annotation_string(
|
||||
context,
|
||||
function.get_default_param_context(),
|
||||
match.group(1).strip()
|
||||
).execute_annotation()
|
||||
|
||||
context = function.get_default_param_context()
|
||||
unknown_type_vars = find_unknown_type_vars(context, annotation)
|
||||
annotation_values = infer_annotation(context, annotation)
|
||||
if not unknown_type_vars:
|
||||
|
||||
@@ -86,8 +86,6 @@ class StubFilter(ParserTreeFilter):
|
||||
# Imports in stub files are only public if they have an "as"
|
||||
# export.
|
||||
definition = name.get_definition()
|
||||
if definition is None:
|
||||
return False
|
||||
if definition.type in ('import_from', 'import_name'):
|
||||
if name.parent.type not in ('import_as_name', 'dotted_as_name'):
|
||||
return False
|
||||
|
||||
@@ -7,6 +7,7 @@ from pathlib import Path
|
||||
|
||||
from jedi import settings
|
||||
from jedi.file_io import FileIO
|
||||
from jedi._compatibility import cast_path
|
||||
from jedi.parser_utils import get_cached_code_lines
|
||||
from jedi.inference.base_value import ValueSet, NO_VALUES
|
||||
from jedi.inference.gradual.stub_value import TypingModuleWrapper, StubModuleValue
|
||||
@@ -43,6 +44,7 @@ def _create_stub_map(directory_path_info):
|
||||
return
|
||||
|
||||
for entry in listed:
|
||||
entry = cast_path(entry)
|
||||
path = os.path.join(directory_path_info.path, entry)
|
||||
if os.path.isdir(path):
|
||||
init = os.path.join(path, '__init__.pyi')
|
||||
@@ -167,6 +169,7 @@ def _try_to_load_stub(inference_state, import_names, python_value_set,
|
||||
if len(import_names) == 1:
|
||||
# foo-stubs
|
||||
for p in sys_path:
|
||||
p = cast_path(p)
|
||||
init = os.path.join(p, *import_names) + '-stubs' + os.path.sep + '__init__.pyi'
|
||||
m = _try_to_load_stub_from_file(
|
||||
inference_state,
|
||||
|
||||
@@ -294,9 +294,6 @@ class Callable(BaseTypingInstance):
|
||||
from jedi.inference.gradual.annotation import infer_return_for_callable
|
||||
return infer_return_for_callable(arguments, param_values, result_values)
|
||||
|
||||
def py__get__(self, instance, class_value):
|
||||
return ValueSet([self])
|
||||
|
||||
|
||||
class Tuple(BaseTypingInstance):
|
||||
def _is_homogenous(self):
|
||||
@@ -434,9 +431,6 @@ class NewType(Value):
|
||||
from jedi.inference.compiled.value import CompiledValueName
|
||||
return CompiledValueName(self, 'NewType')
|
||||
|
||||
def __repr__(self) -> str:
|
||||
return '<NewType: %s>%s' % (self.tree_node, self._type_value_set)
|
||||
|
||||
|
||||
class CastFunction(ValueWrapper):
|
||||
@repack_with_argument_clinic('type, object, /')
|
||||
|
||||
@@ -422,13 +422,20 @@ def import_module(inference_state, import_names, parent_module_value, sys_path):
|
||||
# The module might not be a package.
|
||||
return NO_VALUES
|
||||
|
||||
file_io_or_ns, is_pkg = inference_state.compiled_subprocess.get_module_info(
|
||||
string=import_names[-1],
|
||||
path=paths,
|
||||
full_name=module_name,
|
||||
is_global_search=False,
|
||||
)
|
||||
if is_pkg is None:
|
||||
for path in paths:
|
||||
# At the moment we are only using one path. So this is
|
||||
# not important to be correct.
|
||||
if not isinstance(path, list):
|
||||
path = [path]
|
||||
file_io_or_ns, is_pkg = inference_state.compiled_subprocess.get_module_info(
|
||||
string=import_names[-1],
|
||||
path=path,
|
||||
full_name=module_name,
|
||||
is_global_search=False,
|
||||
)
|
||||
if is_pkg is not None:
|
||||
break
|
||||
else:
|
||||
return NO_VALUES
|
||||
|
||||
if isinstance(file_io_or_ns, ImplicitNSInfo):
|
||||
|
||||
@@ -248,7 +248,7 @@ class ValueNameMixin:
|
||||
|
||||
def get_defining_qualified_value(self):
|
||||
context = self.parent_context
|
||||
if context is not None and (context.is_module() or context.is_class()):
|
||||
if context.is_module() or context.is_class():
|
||||
return self.parent_context.get_value() # Might be None
|
||||
return None
|
||||
|
||||
@@ -341,12 +341,6 @@ class TreeNameDefinition(AbstractTreeName):
|
||||
def py__doc__(self):
|
||||
api_type = self.api_type
|
||||
if api_type in ('function', 'class', 'property'):
|
||||
if self.parent_context.get_root_context().is_stub():
|
||||
from jedi.inference.gradual.conversion import convert_names
|
||||
names = convert_names([self], prefer_stub_to_compiled=False)
|
||||
if self not in names:
|
||||
return _merge_name_docs(names)
|
||||
|
||||
# Make sure the names are not TreeNameDefinitions anymore.
|
||||
return clean_scope_docstring(self.tree_name.get_definition())
|
||||
|
||||
@@ -414,9 +408,6 @@ class ParamNameInterface(_ParamMixin):
|
||||
return 2
|
||||
return 0
|
||||
|
||||
def infer_default(self):
|
||||
return NO_VALUES
|
||||
|
||||
|
||||
class BaseTreeParamName(ParamNameInterface, AbstractTreeName):
|
||||
annotation_node = None
|
||||
|
||||
@@ -12,7 +12,7 @@ count the function calls.
|
||||
Settings
|
||||
~~~~~~~~~~
|
||||
|
||||
Recursion settings are important if you don't want extremely
|
||||
Recursion settings are important if you don't want extremly
|
||||
recursive python code to go absolutely crazy.
|
||||
|
||||
The default values are based on experiments while completing the |jedi| library
|
||||
|
||||
@@ -180,34 +180,26 @@ def _check_fs(inference_state, file_io, regex):
|
||||
return m.as_context()
|
||||
|
||||
|
||||
def gitignored_paths(folder_io, file_io):
|
||||
ignored_paths_abs = set()
|
||||
ignored_paths_rel = set()
|
||||
|
||||
def gitignored_lines(folder_io, file_io):
|
||||
ignored_paths = set()
|
||||
ignored_names = set()
|
||||
for l in file_io.read().splitlines():
|
||||
if not l or l.startswith(b'#') or l.startswith(b'!') or b'*' in l:
|
||||
if not l or l.startswith(b'#'):
|
||||
continue
|
||||
|
||||
p = l.decode('utf-8', 'ignore').rstrip('/')
|
||||
if '/' in p:
|
||||
name = p.lstrip('/')
|
||||
ignored_paths_abs.add(os.path.join(folder_io.path, name))
|
||||
p = l.decode('utf-8', 'ignore')
|
||||
if p.startswith('/'):
|
||||
name = p[1:]
|
||||
if name.endswith(os.path.sep):
|
||||
name = name[:-1]
|
||||
ignored_paths.add(os.path.join(folder_io.path, name))
|
||||
else:
|
||||
name = p
|
||||
ignored_paths_rel.add((folder_io.path, name))
|
||||
|
||||
return ignored_paths_abs, ignored_paths_rel
|
||||
|
||||
|
||||
def expand_relative_ignore_paths(folder_io, relative_paths):
|
||||
curr_path = folder_io.path
|
||||
return {os.path.join(curr_path, p[1]) for p in relative_paths if curr_path.startswith(p[0])}
|
||||
ignored_names.add(p)
|
||||
return ignored_paths, ignored_names
|
||||
|
||||
|
||||
def recurse_find_python_folders_and_files(folder_io, except_paths=()):
|
||||
except_paths = set(except_paths)
|
||||
except_paths_relative = set()
|
||||
|
||||
for root_folder_io, folder_ios, file_ios in folder_io.walk():
|
||||
# Delete folders that we don't want to iterate over.
|
||||
for file_io in file_ios:
|
||||
@@ -217,21 +209,14 @@ def recurse_find_python_folders_and_files(folder_io, except_paths=()):
|
||||
yield None, file_io
|
||||
|
||||
if path.name == '.gitignore':
|
||||
ignored_paths_abs, ignored_paths_rel = gitignored_paths(
|
||||
root_folder_io, file_io
|
||||
)
|
||||
except_paths |= ignored_paths_abs
|
||||
except_paths_relative |= ignored_paths_rel
|
||||
|
||||
except_paths_relative_expanded = expand_relative_ignore_paths(
|
||||
root_folder_io, except_paths_relative
|
||||
)
|
||||
ignored_paths, ignored_names = \
|
||||
gitignored_lines(root_folder_io, file_io)
|
||||
except_paths |= ignored_paths
|
||||
|
||||
folder_ios[:] = [
|
||||
folder_io
|
||||
for folder_io in folder_ios
|
||||
if folder_io.path not in except_paths
|
||||
and folder_io.path not in except_paths_relative_expanded
|
||||
and folder_io.get_base_name() not in _IGNORE_FOLDERS
|
||||
]
|
||||
for folder_io in folder_ios:
|
||||
@@ -297,13 +282,12 @@ def get_module_contexts_containing_name(inference_state, module_contexts, name,
|
||||
limit_reduction=limit_reduction)
|
||||
|
||||
|
||||
def search_in_file_ios(inference_state, file_io_iterator, name,
|
||||
limit_reduction=1, complete=False):
|
||||
def search_in_file_ios(inference_state, file_io_iterator, name, limit_reduction=1):
|
||||
parse_limit = _PARSED_FILE_LIMIT / limit_reduction
|
||||
open_limit = _OPENED_FILE_LIMIT / limit_reduction
|
||||
file_io_count = 0
|
||||
parsed_file_count = 0
|
||||
regex = re.compile(r'\b' + re.escape(name) + (r'' if complete else r'\b'))
|
||||
regex = re.compile(r'\b' + re.escape(name) + r'\b')
|
||||
for file_io in file_io_iterator:
|
||||
file_io_count += 1
|
||||
m = _check_fs(inference_state, file_io, regex)
|
||||
|
||||
@@ -12,8 +12,6 @@ The signature here for bar should be `bar(b, c)` instead of bar(*args).
|
||||
"""
|
||||
from inspect import Parameter
|
||||
|
||||
from parso import tree
|
||||
|
||||
from jedi.inference.utils import to_list
|
||||
from jedi.inference.names import ParamNameWrapper
|
||||
from jedi.inference.helpers import is_big_annoying_library
|
||||
@@ -24,11 +22,7 @@ def _iter_nodes_for_param(param_name):
|
||||
from jedi.inference.arguments import TreeArguments
|
||||
|
||||
execution_context = param_name.parent_context
|
||||
# Walk up the parso tree to get the FunctionNode we want. We use the parso
|
||||
# tree rather than going via the execution context so that we're agnostic of
|
||||
# the specific scope we're evaluating within (i.e: module or function,
|
||||
# etc.).
|
||||
function_node = tree.search_ancestor(param_name.tree_name, 'funcdef', 'lambdef')
|
||||
function_node = execution_context.tree_node
|
||||
module_node = function_node.get_root_node()
|
||||
start = function_node.children[-1].start_pos
|
||||
end = function_node.children[-1].end_pos
|
||||
|
||||
@@ -2,7 +2,6 @@
|
||||
Functions inferring the syntax tree.
|
||||
"""
|
||||
import copy
|
||||
import itertools
|
||||
|
||||
from parso.python import tree
|
||||
|
||||
@@ -516,20 +515,10 @@ def _literals_to_types(inference_state, result):
|
||||
|
||||
def _infer_comparison(context, left_values, operator, right_values):
|
||||
state = context.inference_state
|
||||
if isinstance(operator, str):
|
||||
operator_str = operator
|
||||
else:
|
||||
operator_str = str(operator.value)
|
||||
if not left_values or not right_values:
|
||||
# illegal slices e.g. cause left/right_result to be None
|
||||
result = (left_values or NO_VALUES) | (right_values or NO_VALUES)
|
||||
return _literals_to_types(state, result)
|
||||
elif operator_str == "|" and all(
|
||||
value.is_class() or value.is_compiled()
|
||||
for value in itertools.chain(left_values, right_values)
|
||||
):
|
||||
# ^^^ A naive hack for PEP 604
|
||||
return ValueSet.from_sets((left_values, right_values))
|
||||
else:
|
||||
# I don't think there's a reasonable chance that a string
|
||||
# operation is still correct, once we pass something like six
|
||||
@@ -749,13 +738,6 @@ def tree_name_to_values(inference_state, context, tree_name):
|
||||
types = infer_expr_stmt(context, node, tree_name)
|
||||
elif typ == 'with_stmt':
|
||||
value_managers = context.infer_node(node.get_test_node_from_name(tree_name))
|
||||
if node.parent.type == 'async_stmt':
|
||||
# In the case of `async with` statements, we need to
|
||||
# first get the coroutine from the `__aenter__` method,
|
||||
# then "unwrap" via the `__await__` method
|
||||
enter_methods = value_managers.py__getattribute__('__aenter__')
|
||||
coro = enter_methods.execute_with_values()
|
||||
return coro.py__await__().py__stop_iteration_returns()
|
||||
enter_methods = value_managers.py__getattribute__('__enter__')
|
||||
return enter_methods.execute_with_values()
|
||||
elif typ in ('import_from', 'import_name'):
|
||||
|
||||
@@ -186,6 +186,7 @@ def _get_buildout_script_paths(search_path: Path):
|
||||
directory that look like python files.
|
||||
|
||||
:param search_path: absolute path to the module.
|
||||
:type search_path: str
|
||||
"""
|
||||
project_root = _get_parent_dir_with_file(search_path, 'buildout.cfg')
|
||||
if not project_root:
|
||||
@@ -204,7 +205,7 @@ def _get_buildout_script_paths(search_path: Path):
|
||||
except (UnicodeDecodeError, IOError) as e:
|
||||
# Probably a binary file; permission error or race cond. because
|
||||
# file got deleted. Ignore it.
|
||||
debug.warning(str(e))
|
||||
debug.warning(e)
|
||||
continue
|
||||
|
||||
|
||||
|
||||
@@ -16,7 +16,7 @@ settings will stop this process.
|
||||
|
||||
It is important to note that:
|
||||
|
||||
1. Array modifications work only in the current module.
|
||||
1. Array modfications work only in the current module.
|
||||
2. Jedi only checks Array additions; ``list.pop``, etc are ignored.
|
||||
"""
|
||||
from jedi import debug
|
||||
|
||||
@@ -344,8 +344,7 @@ class BaseFunctionExecutionContext(ValueContext, TreeContextMixin):
|
||||
GenericClass(c, TupleGenericManager(generics)) for c in async_classes
|
||||
).execute_annotation()
|
||||
else:
|
||||
# If there are annotations, prefer them over anything else.
|
||||
if self.is_generator() and not self.infer_annotations():
|
||||
if self.is_generator():
|
||||
return ValueSet([iterable.Generator(inference_state, self)])
|
||||
else:
|
||||
return self.get_return_values()
|
||||
|
||||
@@ -342,8 +342,6 @@ class SequenceLiteralValue(Sequence):
|
||||
else:
|
||||
with reraise_getitem_errors(TypeError, KeyError, IndexError):
|
||||
node = self.get_tree_entries()[index]
|
||||
if node == ':' or node.type == 'subscript':
|
||||
return NO_VALUES
|
||||
return self._defining_context.infer_node(node)
|
||||
|
||||
def py__iter__(self, contextualized_node=None):
|
||||
@@ -409,6 +407,16 @@ class SequenceLiteralValue(Sequence):
|
||||
else:
|
||||
return [array_node]
|
||||
|
||||
def exact_key_items(self):
|
||||
"""
|
||||
Returns a generator of tuples like dict.items(), where the key is
|
||||
resolved (as a string) and the values are still lazy values.
|
||||
"""
|
||||
for key_node, value in self.get_tree_entries():
|
||||
for key in self._defining_context.infer_node(key_node):
|
||||
if is_string(key):
|
||||
yield key.get_safe_value(), LazyTreeValue(self._defining_context, value)
|
||||
|
||||
def __repr__(self):
|
||||
return "<%s of %s>" % (self.__class__.__name__, self.atom)
|
||||
|
||||
@@ -464,16 +472,6 @@ class DictLiteralValue(_DictMixin, SequenceLiteralValue, _DictKeyMixin):
|
||||
|
||||
return ValueSet([FakeList(self.inference_state, lazy_values)])
|
||||
|
||||
def exact_key_items(self):
|
||||
"""
|
||||
Returns a generator of tuples like dict.items(), where the key is
|
||||
resolved (as a string) and the values are still lazy values.
|
||||
"""
|
||||
for key_node, value in self.get_tree_entries():
|
||||
for key in self._defining_context.infer_node(key_node):
|
||||
if is_string(key):
|
||||
yield key.get_safe_value(), LazyTreeValue(self._defining_context, value)
|
||||
|
||||
def _dict_values(self):
|
||||
return ValueSet.from_sets(
|
||||
self._defining_context.infer_node(v)
|
||||
|
||||
@@ -78,8 +78,6 @@ class ClassName(TreeNameDefinition):
|
||||
type_ = super().api_type
|
||||
if type_ == 'function':
|
||||
definition = self.tree_name.get_definition()
|
||||
if definition is None:
|
||||
return type_
|
||||
if function_is_property(definition):
|
||||
# This essentially checks if there is an @property before
|
||||
# the function. @property could be something different, but
|
||||
@@ -116,10 +114,25 @@ class ClassFilter(ParserTreeFilter):
|
||||
while node is not None:
|
||||
if node == self._parser_scope or node == self.parent_context:
|
||||
return True
|
||||
node = get_cached_parent_scope(self._parso_cache_node, node)
|
||||
node = get_cached_parent_scope(self._used_names, node)
|
||||
return False
|
||||
|
||||
def _access_possible(self, name):
|
||||
# Filter for ClassVar variables
|
||||
# TODO this is not properly done, yet. It just checks for the string
|
||||
# ClassVar in the annotation, which can be quite imprecise. If we
|
||||
# wanted to do this correct, we would have to infer the ClassVar.
|
||||
if not self._is_instance:
|
||||
expr_stmt = name.get_definition()
|
||||
if expr_stmt is not None and expr_stmt.type == 'expr_stmt':
|
||||
annassign = expr_stmt.children[1]
|
||||
if annassign.type == 'annassign':
|
||||
# If there is an =, the variable is obviously also
|
||||
# defined on the class.
|
||||
if 'ClassVar' not in annassign.children[1].get_code() \
|
||||
and '=' not in annassign.children:
|
||||
return False
|
||||
|
||||
# Filter for name mangling of private variables like __foo
|
||||
return not name.value.startswith('__') or name.value.endswith('__') \
|
||||
or self._equals_origin_scope()
|
||||
|
||||
@@ -64,7 +64,7 @@ class ModuleMixin(SubModuleDictMixin):
|
||||
parent_context=self.as_context(),
|
||||
origin_scope=origin_scope
|
||||
),
|
||||
GlobalNameFilter(self.as_context()),
|
||||
GlobalNameFilter(self.as_context(), self.tree_node),
|
||||
)
|
||||
yield DictFilter(self.sub_modules_dict())
|
||||
yield DictFilter(self._module_attributes_dict())
|
||||
@@ -148,7 +148,7 @@ class ModuleValue(ModuleMixin, TreeValue):
|
||||
if file_io is None:
|
||||
self._path: Optional[Path] = None
|
||||
else:
|
||||
self._path = file_io.path
|
||||
self._path = Path(file_io.path)
|
||||
self.string_names = string_names # Optional[Tuple[str, ...]]
|
||||
self.code_lines = code_lines
|
||||
self._is_package = is_package
|
||||
|
||||
@@ -1,6 +1,3 @@
|
||||
from pathlib import Path
|
||||
from typing import Optional
|
||||
|
||||
from jedi.inference.cache import inference_state_method_cache
|
||||
from jedi.inference.filters import DictFilter
|
||||
from jedi.inference.names import ValueNameMixin, AbstractNameDefinition
|
||||
@@ -44,7 +41,7 @@ class ImplicitNamespaceValue(Value, SubModuleDictMixin):
|
||||
string_name = self.py__package__()[-1]
|
||||
return ImplicitNSName(self, string_name)
|
||||
|
||||
def py__file__(self) -> Optional[Path]:
|
||||
def py__file__(self):
|
||||
return None
|
||||
|
||||
def py__package__(self):
|
||||
|
||||
@@ -216,14 +216,11 @@ def is_scope(node):
|
||||
def _get_parent_scope_cache(func):
|
||||
cache = WeakKeyDictionary()
|
||||
|
||||
def wrapper(parso_cache_node, node, include_flows=False):
|
||||
if parso_cache_node is None:
|
||||
return func(node, include_flows)
|
||||
|
||||
def wrapper(used_names, node, include_flows=False):
|
||||
try:
|
||||
for_module = cache[parso_cache_node]
|
||||
for_module = cache[used_names]
|
||||
except KeyError:
|
||||
for_module = cache[parso_cache_node] = {}
|
||||
for_module = cache[used_names] = {}
|
||||
|
||||
try:
|
||||
return for_module[node]
|
||||
@@ -273,18 +270,7 @@ def get_cached_code_lines(grammar, path):
|
||||
Basically access the cached code lines in parso. This is not the nicest way
|
||||
to do this, but we avoid splitting all the lines again.
|
||||
"""
|
||||
return get_parso_cache_node(grammar, path).lines
|
||||
|
||||
|
||||
def get_parso_cache_node(grammar, path):
|
||||
"""
|
||||
This is of course not public. But as long as I control parso, this
|
||||
shouldn't be a problem. ~ Dave
|
||||
|
||||
The reason for this is mostly caching. This is obviously also a sign of a
|
||||
broken caching architecture.
|
||||
"""
|
||||
return parser_cache[grammar._hashed][path]
|
||||
return parser_cache[grammar._hashed][path].lines
|
||||
|
||||
|
||||
def cut_value_at_position(leaf, position):
|
||||
|
||||
@@ -2,7 +2,7 @@ from pathlib import Path
|
||||
|
||||
from parso.tree import search_ancestor
|
||||
from jedi.inference.cache import inference_state_method_cache
|
||||
from jedi.inference.imports import goto_import, load_module_from_path
|
||||
from jedi.inference.imports import load_module_from_path
|
||||
from jedi.inference.filters import ParserTreeFilter
|
||||
from jedi.inference.base_value import NO_VALUES, ValueSet
|
||||
from jedi.inference.helpers import infer_call_of_leaf
|
||||
@@ -31,15 +31,7 @@ def execute(callback):
|
||||
def infer_anonymous_param(func):
|
||||
def get_returns(value):
|
||||
if value.tree_node.annotation is not None:
|
||||
result = value.execute_with_values()
|
||||
if any(v.name.get_qualified_names(include_module_names=True)
|
||||
== ('typing', 'Generator')
|
||||
for v in result):
|
||||
return ValueSet.from_sets(
|
||||
v.py__getattribute__('__next__').execute_annotation()
|
||||
for v in result
|
||||
)
|
||||
return result
|
||||
return value.execute_with_values()
|
||||
|
||||
# In pytest we need to differentiate between generators and normal
|
||||
# returns.
|
||||
@@ -51,9 +43,6 @@ def infer_anonymous_param(func):
|
||||
return function_context.get_return_values()
|
||||
|
||||
def wrapper(param_name):
|
||||
# parameters with an annotation do not need special handling
|
||||
if param_name.annotation_node:
|
||||
return func(param_name)
|
||||
is_pytest_param, param_name_is_function_name = \
|
||||
_is_a_pytest_param_and_inherited(param_name)
|
||||
if is_pytest_param:
|
||||
@@ -131,17 +120,6 @@ def _is_pytest_func(func_name, decorator_nodes):
|
||||
or any('fixture' in n.get_code() for n in decorator_nodes)
|
||||
|
||||
|
||||
def _find_pytest_plugin_modules():
|
||||
"""
|
||||
Finds pytest plugin modules hooked by setuptools entry points
|
||||
|
||||
See https://docs.pytest.org/en/stable/how-to/writing_plugins.html#setuptools-entry-points
|
||||
"""
|
||||
from pkg_resources import iter_entry_points
|
||||
|
||||
return [ep.module_name.split(".") for ep in iter_entry_points(group="pytest11")]
|
||||
|
||||
|
||||
@inference_state_method_cache()
|
||||
def _iter_pytest_modules(module_context, skip_own_module=False):
|
||||
if not skip_own_module:
|
||||
@@ -151,10 +129,6 @@ def _iter_pytest_modules(module_context, skip_own_module=False):
|
||||
if file_io is not None:
|
||||
folder = file_io.get_parent_folder()
|
||||
sys_path = module_context.inference_state.get_sys_path()
|
||||
|
||||
# prevent an infinite loop when reaching the root of the current drive
|
||||
last_folder = None
|
||||
|
||||
while any(folder.path.startswith(p) for p in sys_path):
|
||||
file_io = folder.get_file_io('conftest.py')
|
||||
if Path(file_io.path) != module_context.py__file__():
|
||||
@@ -165,12 +139,7 @@ def _iter_pytest_modules(module_context, skip_own_module=False):
|
||||
pass
|
||||
folder = folder.get_parent_folder()
|
||||
|
||||
# prevent an infinite for loop if the same parent folder is return twice
|
||||
if last_folder is not None and folder.path == last_folder.path:
|
||||
break
|
||||
last_folder = folder # keep track of the last found parent name
|
||||
|
||||
for names in _PYTEST_FIXTURE_MODULES + _find_pytest_plugin_modules():
|
||||
for names in _PYTEST_FIXTURE_MODULES:
|
||||
for module_value in module_context.inference_state.import_module(names):
|
||||
yield module_value.as_context()
|
||||
|
||||
@@ -178,28 +147,14 @@ def _iter_pytest_modules(module_context, skip_own_module=False):
|
||||
class FixtureFilter(ParserTreeFilter):
|
||||
def _filter(self, names):
|
||||
for name in super()._filter(names):
|
||||
# look for fixture definitions of imported names
|
||||
if name.parent.type == "import_from":
|
||||
imported_names = goto_import(self.parent_context, name)
|
||||
if any(
|
||||
self._is_fixture(iname.parent_context, iname.tree_name)
|
||||
for iname in imported_names
|
||||
# discard imports of whole modules, that have no tree_name
|
||||
if iname.tree_name
|
||||
):
|
||||
funcdef = name.parent
|
||||
# Class fixtures are not supported
|
||||
if funcdef.type == 'funcdef':
|
||||
decorated = funcdef.parent
|
||||
if decorated.type == 'decorated' and self._is_fixture(decorated):
|
||||
yield name
|
||||
|
||||
elif self._is_fixture(self.parent_context, name):
|
||||
yield name
|
||||
|
||||
def _is_fixture(self, context, name):
|
||||
funcdef = name.parent
|
||||
# Class fixtures are not supported
|
||||
if funcdef.type != "funcdef":
|
||||
return False
|
||||
decorated = funcdef.parent
|
||||
if decorated.type != "decorated":
|
||||
return False
|
||||
def _is_fixture(self, decorated):
|
||||
decorators = decorated.children[0]
|
||||
if decorators.type == 'decorators':
|
||||
decorators = decorators.children
|
||||
@@ -216,12 +171,11 @@ class FixtureFilter(ParserTreeFilter):
|
||||
last_leaf = last_trailer.get_last_leaf()
|
||||
if last_leaf == ')':
|
||||
values = infer_call_of_leaf(
|
||||
context, last_leaf, cut_own_trailer=True
|
||||
)
|
||||
self.parent_context, last_leaf, cut_own_trailer=True)
|
||||
else:
|
||||
values = context.infer_node(dotted_name)
|
||||
values = self.parent_context.infer_node(dotted_name)
|
||||
else:
|
||||
values = context.infer_node(dotted_name)
|
||||
values = self.parent_context.infer_node(dotted_name)
|
||||
for value in values:
|
||||
if value.name.get_qualified_names(include_module_names=True) \
|
||||
== ('_pytest', 'fixtures', 'fixture'):
|
||||
|
||||
@@ -803,15 +803,6 @@ _implemented = {
|
||||
# For now this works at least better than Jedi trying to understand it.
|
||||
'dataclass': _dataclass
|
||||
},
|
||||
# attrs exposes declaration interface roughly compatible with dataclasses
|
||||
# via attrs.define, attrs.frozen and attrs.mutable
|
||||
# https://www.attrs.org/en/stable/names.html
|
||||
'attr': {
|
||||
'define': _dataclass,
|
||||
},
|
||||
'attrs': {
|
||||
'define': _dataclass,
|
||||
},
|
||||
'os.path': {
|
||||
'dirname': _create_string_input_function(os.path.dirname),
|
||||
'abspath': _create_string_input_function(os.path.abspath),
|
||||
|
||||
32
setup.py
32
setup.py
@@ -35,46 +35,17 @@ setup(name='jedi',
|
||||
install_requires=['parso>=0.8.0,<0.9.0'],
|
||||
extras_require={
|
||||
'testing': [
|
||||
'pytest<7.0.0',
|
||||
'pytest<6.0.0',
|
||||
# docopt for sith doctests
|
||||
'docopt',
|
||||
# coloroma for colored debug output
|
||||
'colorama',
|
||||
'Django<3.1', # For now pin this.
|
||||
'attrs',
|
||||
],
|
||||
'qa': [
|
||||
'flake8==3.8.3',
|
||||
'mypy==0.782',
|
||||
],
|
||||
'docs': [
|
||||
# Just pin all of these.
|
||||
'Jinja2==2.11.3',
|
||||
'MarkupSafe==1.1.1',
|
||||
'Pygments==2.8.1',
|
||||
'alabaster==0.7.12',
|
||||
'babel==2.9.1',
|
||||
'chardet==4.0.0',
|
||||
'commonmark==0.8.1',
|
||||
'docutils==0.17.1',
|
||||
'future==0.18.2',
|
||||
'idna==2.10',
|
||||
'imagesize==1.2.0',
|
||||
'mock==1.0.1',
|
||||
'packaging==20.9',
|
||||
'pyparsing==2.4.7',
|
||||
'pytz==2021.1',
|
||||
'readthedocs-sphinx-ext==2.1.4',
|
||||
'recommonmark==0.5.0',
|
||||
'requests==2.25.1',
|
||||
'six==1.15.0',
|
||||
'snowballstemmer==2.1.0',
|
||||
'sphinx==1.8.5',
|
||||
'sphinx-rtd-theme==0.4.3',
|
||||
'sphinxcontrib-serializinghtml==1.1.4',
|
||||
'sphinxcontrib-websupport==1.2.4',
|
||||
'urllib3==1.26.4',
|
||||
],
|
||||
},
|
||||
package_data={'jedi': ['*.pyi', 'third_party/typeshed/LICENSE',
|
||||
'third_party/typeshed/README']},
|
||||
@@ -90,7 +61,6 @@ setup(name='jedi',
|
||||
'Programming Language :: Python :: 3.7',
|
||||
'Programming Language :: Python :: 3.8',
|
||||
'Programming Language :: Python :: 3.9',
|
||||
'Programming Language :: Python :: 3.10',
|
||||
'Topic :: Software Development :: Libraries :: Python Modules',
|
||||
'Topic :: Text Editors :: Integrated Development Environments (IDE)',
|
||||
'Topic :: Utilities',
|
||||
|
||||
@@ -44,8 +44,6 @@ b[int():]
|
||||
|
||||
#? list()
|
||||
b[:]
|
||||
#? int()
|
||||
b[:, :-1]
|
||||
|
||||
#? 3
|
||||
b[:]
|
||||
@@ -69,20 +67,6 @@ class _StrangeSlice():
|
||||
#? slice()
|
||||
_StrangeSlice()[1:2]
|
||||
|
||||
for x in b[:]:
|
||||
#? int()
|
||||
x
|
||||
|
||||
for x in b[:, :-1]:
|
||||
#?
|
||||
x
|
||||
|
||||
class Foo:
|
||||
def __getitem__(self, item):
|
||||
return item
|
||||
|
||||
#?
|
||||
Foo()[:, :-1][0]
|
||||
|
||||
# -----------------
|
||||
# iterable multiplication
|
||||
|
||||
@@ -26,6 +26,11 @@ async def y():
|
||||
x().__await__().__next
|
||||
return 2
|
||||
|
||||
async def x2():
|
||||
async with open('asdf') as f:
|
||||
#? ['readlines']
|
||||
f.readlines
|
||||
|
||||
class A():
|
||||
@staticmethod
|
||||
async def b(c=1, d=2):
|
||||
@@ -47,6 +52,8 @@ async def awaitable_test():
|
||||
#? str()
|
||||
foo
|
||||
|
||||
# python >= 3.6
|
||||
|
||||
async def asgen():
|
||||
yield 1
|
||||
await asyncio.sleep(0)
|
||||
@@ -98,22 +105,3 @@ async def f():
|
||||
f = await C().async_for_classmethod()
|
||||
#? C()
|
||||
f
|
||||
|
||||
|
||||
class AsyncCtxMgr:
|
||||
def some_method():
|
||||
pass
|
||||
|
||||
async def __aenter__(self):
|
||||
return self
|
||||
|
||||
async def __aexit__(self, *args):
|
||||
pass
|
||||
|
||||
|
||||
async def asyncctxmgr():
|
||||
async with AsyncCtxMgr() as acm:
|
||||
#? AsyncCtxMgr()
|
||||
acm
|
||||
#? ['some_method']
|
||||
acm.som
|
||||
|
||||
@@ -23,7 +23,7 @@ def inheritance_fixture():
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
def capsysbinary(capsysbinary):
|
||||
#? ['close']
|
||||
capsysbinary.clos
|
||||
return capsysbinary
|
||||
def testdir(testdir):
|
||||
#? ['chdir']
|
||||
testdir.chdir
|
||||
return testdir
|
||||
|
||||
@@ -5,7 +5,6 @@ import uuid
|
||||
from django.db import models
|
||||
from django.contrib.auth.models import User
|
||||
from django.db.models.query_utils import DeferredAttribute
|
||||
from django.db.models.manager import BaseManager
|
||||
|
||||
|
||||
class TagManager(models.Manager):
|
||||
|
||||
@@ -284,13 +284,6 @@ def doctest_with_space():
|
||||
import_issu
|
||||
"""
|
||||
|
||||
def doctest_issue_github_1748():
|
||||
"""From GitHub #1748
|
||||
#? 10 []
|
||||
This. Al
|
||||
"""
|
||||
pass
|
||||
|
||||
|
||||
def docstring_rst_identifiers():
|
||||
"""
|
||||
|
||||
@@ -1,3 +1,5 @@
|
||||
# python >= 3.6
|
||||
|
||||
class Foo:
|
||||
bar = 1
|
||||
|
||||
|
||||
@@ -309,8 +309,3 @@ def annotation2() -> Iterator[float]:
|
||||
next(annotation1())
|
||||
#? float()
|
||||
next(annotation2())
|
||||
|
||||
|
||||
# annotations should override generator inference
|
||||
#? float()
|
||||
annotation1()
|
||||
|
||||
@@ -110,4 +110,4 @@ class Test(object):
|
||||
# nocond lambdas make no sense at all.
|
||||
|
||||
#? int()
|
||||
[a for a in [1,2] if (lambda: 3)][0]
|
||||
[a for a in [1,2] if lambda: 3][0]
|
||||
|
||||
@@ -1 +0,0 @@
|
||||
mod1_name = 'mod1'
|
||||
@@ -1 +0,0 @@
|
||||
mod2_name = 'mod2'
|
||||
@@ -1,18 +0,0 @@
|
||||
import sys
|
||||
import os
|
||||
from os.path import dirname
|
||||
|
||||
sys.path.insert(0, os.path.join(dirname(__file__), 'namespace2'))
|
||||
sys.path.insert(0, os.path.join(dirname(__file__), 'namespace1'))
|
||||
|
||||
#? ['mod1']
|
||||
import pkg1.pkg2.mod1
|
||||
|
||||
#? ['mod2']
|
||||
import pkg1.pkg2.mod2
|
||||
|
||||
#? ['mod1_name']
|
||||
pkg1.pkg2.mod1.mod1_name
|
||||
|
||||
#? ['mod2_name']
|
||||
pkg1.pkg2.mod2.mod2_name
|
||||
@@ -23,9 +23,11 @@ def builtin_test():
|
||||
import sqlite3
|
||||
|
||||
# classes is a local module that has an __init__.py and can therefore not be
|
||||
# found.
|
||||
# found. test can be found.
|
||||
#? []
|
||||
import classes
|
||||
#? ['test']
|
||||
import test
|
||||
|
||||
#? ['timedelta']
|
||||
from datetime import timedel
|
||||
|
||||
@@ -1,50 +0,0 @@
|
||||
""" Pep-0484 type hinted decorators """
|
||||
|
||||
from typing import Callable
|
||||
|
||||
|
||||
def decorator(func):
|
||||
def wrapper(*a, **k):
|
||||
return str(func(*a, **k))
|
||||
return wrapper
|
||||
|
||||
|
||||
def typed_decorator(func: Callable[..., int]) -> Callable[..., str]:
|
||||
...
|
||||
|
||||
# Functions
|
||||
|
||||
@decorator
|
||||
def plain_func() -> int:
|
||||
return 4
|
||||
|
||||
#? str()
|
||||
plain_func()
|
||||
|
||||
|
||||
@typed_decorator
|
||||
def typed_func() -> int:
|
||||
return 4
|
||||
|
||||
#? str()
|
||||
typed_func()
|
||||
|
||||
|
||||
# Methods
|
||||
|
||||
class X:
|
||||
@decorator
|
||||
def plain_method(self) -> int:
|
||||
return 4
|
||||
|
||||
@typed_decorator
|
||||
def typed_method(self) -> int:
|
||||
return 4
|
||||
|
||||
inst = X()
|
||||
|
||||
#? str()
|
||||
inst.plain_method()
|
||||
|
||||
#? str()
|
||||
inst.typed_method()
|
||||
@@ -27,13 +27,13 @@ class PlainClass(object):
|
||||
|
||||
|
||||
tpl = ("1", 2)
|
||||
tpl_typed: Tuple[str, int] = ("2", 3)
|
||||
tpl_typed = ("2", 3) # type: Tuple[str, int]
|
||||
|
||||
collection = {"a": 1}
|
||||
collection_typed: Dict[str, int] = {"a": 1}
|
||||
collection_typed = {"a": 1} # type: Dict[str, int]
|
||||
|
||||
list_of_ints: List[int] = [42]
|
||||
list_of_funcs: List[Callable[[T], T]] = [foo]
|
||||
list_of_ints = [42] # type: List[int]
|
||||
list_of_funcs = [foo] # type: List[Callable[[T], T]]
|
||||
|
||||
custom_generic = CustomGeneric(123.45)
|
||||
|
||||
|
||||
@@ -19,12 +19,12 @@ T_co = TypeVar('T_co', covariant=True)
|
||||
V = TypeVar('V')
|
||||
|
||||
|
||||
just_float: float = 42.
|
||||
optional_float: Optional[float] = 42.
|
||||
list_of_ints: List[int] = [42]
|
||||
list_of_floats: List[float] = [42.]
|
||||
list_of_optional_floats: List[Optional[float]] = [x or None for x in list_of_floats]
|
||||
list_of_ints_and_strs: List[Union[int, str]] = [42, 'abc']
|
||||
just_float = 42. # type: float
|
||||
optional_float = 42. # type: Optional[float]
|
||||
list_of_ints = [42] # type: List[int]
|
||||
list_of_floats = [42.] # type: List[float]
|
||||
list_of_optional_floats = [x or None for x in list_of_floats] # type: List[Optional[float]]
|
||||
list_of_ints_and_strs = [42, 'abc'] # type: List[Union[int, str]]
|
||||
|
||||
# Test that simple parameters are handled
|
||||
def list_t_to_list_t(the_list: List[T]) -> List[T]:
|
||||
@@ -48,7 +48,7 @@ for z in list_t_to_list_t(list_of_ints_and_strs):
|
||||
z
|
||||
|
||||
|
||||
list_of_int_type: List[Type[int]] = [int]
|
||||
list_of_int_type = [int] # type: List[Type[int]]
|
||||
|
||||
# Test that nested parameters are handled
|
||||
def list_optional_t_to_list_t(the_list: List[Optional[T]]) -> List[T]:
|
||||
@@ -85,7 +85,7 @@ def optional_list_t_to_list_t(x: Optional[List[T]]) -> List[T]:
|
||||
return x if x is not None else []
|
||||
|
||||
|
||||
optional_list_float: Optional[List[float]] = None
|
||||
optional_list_float = None # type: Optional[List[float]]
|
||||
for xc in optional_list_t_to_list_t(optional_list_float):
|
||||
#? float()
|
||||
xc
|
||||
@@ -134,7 +134,7 @@ def list_tuple_t_to_tuple_list_t(the_list: List[Tuple[T]]) -> Tuple[List[T], ...
|
||||
return tuple(list(x) for x in the_list)
|
||||
|
||||
|
||||
list_of_int_tuples: List[Tuple[int]] = [(x,) for x in list_of_ints]
|
||||
list_of_int_tuples = [(x,) for x in list_of_ints] # type: List[Tuple[int]]
|
||||
|
||||
for b in list_tuple_t_to_tuple_list_t(list_of_int_tuples):
|
||||
#? int()
|
||||
@@ -145,7 +145,7 @@ def list_tuple_t_elipsis_to_tuple_list_t(the_list: List[Tuple[T, ...]]) -> Tuple
|
||||
return tuple(list(x) for x in the_list)
|
||||
|
||||
|
||||
list_of_int_tuple_elipsis: List[Tuple[int, ...]] = [tuple(list_of_ints)]
|
||||
list_of_int_tuple_elipsis = [tuple(list_of_ints)] # type: List[Tuple[int, ...]]
|
||||
|
||||
for b in list_tuple_t_elipsis_to_tuple_list_t(list_of_int_tuple_elipsis):
|
||||
#? int()
|
||||
@@ -157,7 +157,7 @@ def foo(x: int) -> int:
|
||||
return x
|
||||
|
||||
|
||||
list_of_funcs: List[Callable[[int], int]] = [foo]
|
||||
list_of_funcs = [foo] # type: List[Callable[[int], int]]
|
||||
|
||||
def list_func_t_to_list_func_type_t(the_list: List[Callable[[T], T]]) -> List[Callable[[Type[T]], T]]:
|
||||
def adapt(func: Callable[[T], T]) -> Callable[[Type[T]], T]:
|
||||
@@ -176,7 +176,7 @@ def bar(*a, **k) -> int:
|
||||
return len(a) + len(k)
|
||||
|
||||
|
||||
list_of_funcs_2: List[Callable[..., int]] = [bar]
|
||||
list_of_funcs_2 = [bar] # type: List[Callable[..., int]]
|
||||
|
||||
def list_func_t_passthrough(the_list: List[Callable[..., T]]) -> List[Callable[..., T]]:
|
||||
return the_list
|
||||
@@ -187,7 +187,7 @@ for b in list_func_t_passthrough(list_of_funcs_2):
|
||||
b(None, x="x")
|
||||
|
||||
|
||||
mapping_int_str: Dict[int, str] = {42: 'a'}
|
||||
mapping_int_str = {42: 'a'} # type: Dict[int, str]
|
||||
|
||||
# Test that mappings (that have more than one parameter) are handled
|
||||
def invert_mapping(mapping: Mapping[K, V]) -> Mapping[V, K]:
|
||||
@@ -210,11 +210,11 @@ first(mapping_int_str)
|
||||
#? str()
|
||||
first("abc")
|
||||
|
||||
some_str: str = NotImplemented
|
||||
some_str = NotImplemented # type: str
|
||||
#? str()
|
||||
first(some_str)
|
||||
|
||||
annotated: List[ Callable[[Sequence[float]], int] ] = [len]
|
||||
annotated = [len] # type: List[ Callable[[Sequence[float]], int] ]
|
||||
#? int()
|
||||
first(annotated)()
|
||||
|
||||
@@ -237,7 +237,7 @@ for b in values(mapping_int_str):
|
||||
#
|
||||
# Tests that user-defined generic types are handled
|
||||
#
|
||||
list_ints: List[int] = [42]
|
||||
list_ints = [42] # type: List[int]
|
||||
|
||||
class CustomGeneric(Generic[T_co]):
|
||||
def __init__(self, val: T_co) -> None:
|
||||
@@ -248,7 +248,7 @@ class CustomGeneric(Generic[T_co]):
|
||||
def custom(x: CustomGeneric[T]) -> T:
|
||||
return x.val
|
||||
|
||||
custom_instance: CustomGeneric[int] = CustomGeneric(42)
|
||||
custom_instance = CustomGeneric(42) # type: CustomGeneric[int]
|
||||
|
||||
#? int()
|
||||
custom(custom_instance)
|
||||
@@ -275,7 +275,7 @@ for x5 in wrap_custom(list_ints):
|
||||
|
||||
|
||||
# Test extraction of type from a nested custom generic type
|
||||
list_custom_instances: List[CustomGeneric[int]] = [CustomGeneric(42)]
|
||||
list_custom_instances = [CustomGeneric(42)] # type: List[CustomGeneric[int]]
|
||||
|
||||
def unwrap_custom(iterable: Iterable[CustomGeneric[T]]) -> List[T]:
|
||||
return [x.val for x in iterable]
|
||||
@@ -303,7 +303,7 @@ for xg in unwrap_custom(CustomGeneric(s) for s in 'abc'):
|
||||
|
||||
|
||||
# Test extraction of type from type parameer nested within a custom generic type
|
||||
custom_instance_list_int: CustomGeneric[List[int]] = CustomGeneric([42])
|
||||
custom_instance_list_int = CustomGeneric([42]) # type: CustomGeneric[List[int]]
|
||||
|
||||
def unwrap_custom2(instance: CustomGeneric[Iterable[T]]) -> List[T]:
|
||||
return list(instance.val)
|
||||
@@ -326,7 +326,7 @@ class Specialised(Mapping[int, str]):
|
||||
pass
|
||||
|
||||
|
||||
specialised_instance: Specialised = NotImplemented
|
||||
specialised_instance = NotImplemented # type: Specialised
|
||||
|
||||
#? int()
|
||||
first(specialised_instance)
|
||||
@@ -341,7 +341,7 @@ class ChildOfSpecialised(Specialised):
|
||||
pass
|
||||
|
||||
|
||||
child_of_specialised_instance: ChildOfSpecialised = NotImplemented
|
||||
child_of_specialised_instance = NotImplemented # type: ChildOfSpecialised
|
||||
|
||||
#? int()
|
||||
first(child_of_specialised_instance)
|
||||
@@ -355,13 +355,13 @@ class CustomPartialGeneric1(Mapping[str, T]):
|
||||
pass
|
||||
|
||||
|
||||
custom_partial1_instance: CustomPartialGeneric1[int] = NotImplemented
|
||||
custom_partial1_instance = NotImplemented # type: CustomPartialGeneric1[int]
|
||||
|
||||
#? str()
|
||||
first(custom_partial1_instance)
|
||||
|
||||
|
||||
custom_partial1_unbound_instance: CustomPartialGeneric1 = NotImplemented
|
||||
custom_partial1_unbound_instance = NotImplemented # type: CustomPartialGeneric1
|
||||
|
||||
#? str()
|
||||
first(custom_partial1_unbound_instance)
|
||||
@@ -371,7 +371,7 @@ class CustomPartialGeneric2(Mapping[T, str]):
|
||||
pass
|
||||
|
||||
|
||||
custom_partial2_instance: CustomPartialGeneric2[int] = NotImplemented
|
||||
custom_partial2_instance = NotImplemented # type: CustomPartialGeneric2[int]
|
||||
|
||||
#? int()
|
||||
first(custom_partial2_instance)
|
||||
@@ -380,7 +380,7 @@ first(custom_partial2_instance)
|
||||
values(custom_partial2_instance)[0]
|
||||
|
||||
|
||||
custom_partial2_unbound_instance: CustomPartialGeneric2 = NotImplemented
|
||||
custom_partial2_unbound_instance = NotImplemented # type: CustomPartialGeneric2
|
||||
|
||||
#? []
|
||||
first(custom_partial2_unbound_instance)
|
||||
|
||||
@@ -19,16 +19,16 @@ TTypeAny = TypeVar('TTypeAny', bound=Type[Any])
|
||||
TCallable = TypeVar('TCallable', bound=Callable[..., Any])
|
||||
|
||||
untyped_list_str = ['abc', 'def']
|
||||
typed_list_str: List[str] = ['abc', 'def']
|
||||
typed_list_str = ['abc', 'def'] # type: List[str]
|
||||
|
||||
untyped_tuple_str = ('abc',)
|
||||
typed_tuple_str: Tuple[str] = ('abc',)
|
||||
typed_tuple_str = ('abc',) # type: Tuple[str]
|
||||
|
||||
untyped_tuple_str_int = ('abc', 4)
|
||||
typed_tuple_str_int: Tuple[str, int] = ('abc', 4)
|
||||
typed_tuple_str_int = ('abc', 4) # type: Tuple[str, int]
|
||||
|
||||
variadic_tuple_str: Tuple[str, ...] = ('abc',)
|
||||
variadic_tuple_str_int: Tuple[Union[str, int], ...] = ('abc', 4)
|
||||
variadic_tuple_str = ('abc',) # type: Tuple[str, ...]
|
||||
variadic_tuple_str_int = ('abc', 4) # type: Tuple[Union[str, int], ...]
|
||||
|
||||
|
||||
def untyped_passthrough(x):
|
||||
@@ -58,16 +58,6 @@ def typed_bound_generic_passthrough(x: TList) -> TList:
|
||||
|
||||
return x
|
||||
|
||||
# Forward references are more likely with custom types, however this aims to
|
||||
# test just the handling of the quoted type rather than any other part of the
|
||||
# machinery.
|
||||
def typed_quoted_return_generic_passthrough(x: T) -> 'List[T]':
|
||||
return [x]
|
||||
|
||||
def typed_quoted_input_generic_passthrough(x: 'Tuple[T]') -> T:
|
||||
x
|
||||
return x[0]
|
||||
|
||||
|
||||
for a in untyped_passthrough(untyped_list_str):
|
||||
#? str()
|
||||
@@ -156,23 +146,6 @@ for q in typed_bound_generic_passthrough(typed_list_str):
|
||||
q
|
||||
|
||||
|
||||
for r in typed_quoted_return_generic_passthrough("something"):
|
||||
#? str()
|
||||
r
|
||||
|
||||
for s in typed_quoted_return_generic_passthrough(42):
|
||||
#? int()
|
||||
s
|
||||
|
||||
|
||||
#? str()
|
||||
typed_quoted_input_generic_passthrough(("something",))
|
||||
|
||||
#? int()
|
||||
typed_quoted_input_generic_passthrough((42,))
|
||||
|
||||
|
||||
|
||||
class CustomList(List):
|
||||
def get_first(self):
|
||||
return self[0]
|
||||
|
||||
@@ -1,9 +1,10 @@
|
||||
# python >= 3.6
|
||||
from typing import List, Dict, overload, Tuple, TypeVar
|
||||
|
||||
lst: list
|
||||
list_alias: List
|
||||
list_str: List[str]
|
||||
list_int: List[int]
|
||||
list_str: List[int]
|
||||
|
||||
# -------------------------
|
||||
# With base classes
|
||||
|
||||
@@ -2,14 +2,18 @@
|
||||
Test the typing library, with docstrings and annotations
|
||||
"""
|
||||
import typing
|
||||
from typing import Sequence, MutableSequence, List, Iterable, Iterator, \
|
||||
AbstractSet, Tuple, Mapping, Dict, Union, Optional
|
||||
|
||||
class B:
|
||||
pass
|
||||
|
||||
def we_can_has_sequence(p: Sequence[int], q: Sequence[B], r: Sequence[int],
|
||||
s: Sequence["int"], t: MutableSequence[dict], u: List[float]):
|
||||
def we_can_has_sequence(p, q, r, s, t, u):
|
||||
"""
|
||||
:type p: typing.Sequence[int]
|
||||
:type q: typing.Sequence[B]
|
||||
:type r: typing.Sequence[int]
|
||||
:type s: typing.Sequence["int"]
|
||||
:type t: typing.MutableSequence[dict]
|
||||
:type u: typing.List[float]
|
||||
"""
|
||||
#? ["count"]
|
||||
p.c
|
||||
#? int()
|
||||
@@ -39,8 +43,13 @@ def we_can_has_sequence(p: Sequence[int], q: Sequence[B], r: Sequence[int],
|
||||
#? float()
|
||||
u[1]
|
||||
|
||||
def iterators(ps: Iterable[int], qs: Iterator[str], rs:
|
||||
Sequence["ForwardReference"], ts: AbstractSet["float"]):
|
||||
def iterators(ps, qs, rs, ts):
|
||||
"""
|
||||
:type ps: typing.Iterable[int]
|
||||
:type qs: typing.Iterator[str]
|
||||
:type rs: typing.Sequence["ForwardReference"]
|
||||
:type ts: typing.AbstractSet["float"]
|
||||
"""
|
||||
for p in ps:
|
||||
#? int()
|
||||
p
|
||||
@@ -70,13 +79,22 @@ def iterators(ps: Iterable[int], qs: Iterator[str], rs:
|
||||
#? float()
|
||||
t
|
||||
|
||||
def sets(p: AbstractSet[int], q: typing.MutableSet[float]):
|
||||
def sets(p, q):
|
||||
"""
|
||||
:type p: typing.AbstractSet[int]
|
||||
:type q: typing.MutableSet[float]
|
||||
"""
|
||||
#? []
|
||||
p.a
|
||||
#? ["add"]
|
||||
q.a
|
||||
|
||||
def tuple(p: Tuple[int], q: Tuple[int, str, float], r: Tuple[B, ...]):
|
||||
def tuple(p, q, r):
|
||||
"""
|
||||
:type p: typing.Tuple[int]
|
||||
:type q: typing.Tuple[int, str, float]
|
||||
:type r: typing.Tuple[B, ...]
|
||||
"""
|
||||
#? int()
|
||||
p[0]
|
||||
#? ['index']
|
||||
@@ -109,14 +127,16 @@ class Key:
|
||||
class Value:
|
||||
pass
|
||||
|
||||
def mapping(
|
||||
p: Mapping[Key, Value],
|
||||
q: typing.MutableMapping[Key, Value],
|
||||
d: Dict[Key, Value],
|
||||
dd: typing.DefaultDict[Key, Value],
|
||||
r: typing.KeysView[Key],
|
||||
s: typing.ValuesView[Value],
|
||||
t: typing.ItemsView[Key, Value]):
|
||||
def mapping(p, q, d, dd, r, s, t):
|
||||
"""
|
||||
:type p: typing.Mapping[Key, Value]
|
||||
:type q: typing.MutableMapping[Key, Value]
|
||||
:type d: typing.Dict[Key, Value]
|
||||
:type dd: typing.DefaultDict[Key, Value]
|
||||
:type r: typing.KeysView[Key]
|
||||
:type s: typing.ValuesView[Value]
|
||||
:type t: typing.ItemsView[Key, Value]
|
||||
"""
|
||||
#? []
|
||||
p.setd
|
||||
#? ["setdefault"]
|
||||
@@ -178,12 +198,14 @@ def mapping(
|
||||
#? Value()
|
||||
value
|
||||
|
||||
def union(
|
||||
p: Union[int],
|
||||
q: Union[int, int],
|
||||
r: Union[int, str, "int"],
|
||||
s: Union[int, typing.Union[str, "typing.Union['float', 'dict']"]],
|
||||
t: Union[int, None]):
|
||||
def union(p, q, r, s, t):
|
||||
"""
|
||||
:type p: typing.Union[int]
|
||||
:type q: typing.Union[int, int]
|
||||
:type r: typing.Union[int, str, "int"]
|
||||
:type s: typing.Union[int, typing.Union[str, "typing.Union['float', 'dict']"]]
|
||||
:type t: typing.Union[int, None]
|
||||
"""
|
||||
#? int()
|
||||
p
|
||||
#? int()
|
||||
@@ -195,8 +217,9 @@ def union(
|
||||
#? int() None
|
||||
t
|
||||
|
||||
def optional(p: Optional[int]):
|
||||
def optional(p):
|
||||
"""
|
||||
:type p: typing.Optional[int]
|
||||
Optional does not do anything special. However it should be recognised
|
||||
as being of that type. Jedi doesn't do anything with the extra into that
|
||||
it can be None as well
|
||||
@@ -211,7 +234,10 @@ class TestDict(typing.Dict[str, int]):
|
||||
def setdud(self):
|
||||
pass
|
||||
|
||||
def testdict(x: TestDict):
|
||||
def testdict(x):
|
||||
"""
|
||||
:type x: TestDict
|
||||
"""
|
||||
#? ["setdud", "setdefault"]
|
||||
x.setd
|
||||
for key in x.keys():
|
||||
@@ -236,7 +262,10 @@ y = WrappingType(0) # Per https://github.com/davidhalter/jedi/issues/1015#issuec
|
||||
#? str()
|
||||
y
|
||||
|
||||
def testnewtype(y: WrappingType):
|
||||
def testnewtype(y):
|
||||
"""
|
||||
:type y: WrappingType
|
||||
"""
|
||||
#? str()
|
||||
y
|
||||
#? ["upper"]
|
||||
@@ -244,7 +273,10 @@ def testnewtype(y: WrappingType):
|
||||
|
||||
WrappingType2 = typing.NewType()
|
||||
|
||||
def testnewtype2(y: WrappingType2):
|
||||
def testnewtype2(y):
|
||||
"""
|
||||
:type y: WrappingType2
|
||||
"""
|
||||
#?
|
||||
y
|
||||
#? []
|
||||
@@ -265,7 +297,10 @@ class TestDefaultDict(typing.DefaultDict[str, int]):
|
||||
def setdud(self):
|
||||
pass
|
||||
|
||||
def testdict(x: TestDefaultDict):
|
||||
def testdict(x):
|
||||
"""
|
||||
:type x: TestDefaultDict
|
||||
"""
|
||||
#? ["setdud", "setdefault"]
|
||||
x.setd
|
||||
for key in x.keys():
|
||||
|
||||
@@ -1,6 +1,9 @@
|
||||
"""
|
||||
PEP 526 introduced a way of using type annotations on variables.
|
||||
PEP 526 introduced a new way of using type annotations on variables. It was
|
||||
introduced in Python 3.6.
|
||||
"""
|
||||
# python >= 3.6
|
||||
|
||||
import typing
|
||||
|
||||
asdf = ''
|
||||
@@ -44,7 +47,7 @@ class Foo():
|
||||
baz: typing.ClassVar[str]
|
||||
|
||||
|
||||
#? int()
|
||||
#?
|
||||
Foo.bar
|
||||
#? int()
|
||||
Foo().bar
|
||||
@@ -58,7 +61,6 @@ class VarClass:
|
||||
var_instance2: float
|
||||
var_class1: typing.ClassVar[str] = 1
|
||||
var_class2: typing.ClassVar[bytes]
|
||||
var_class3 = None
|
||||
|
||||
def __init__(self):
|
||||
#? int()
|
||||
@@ -71,21 +73,15 @@ class VarClass:
|
||||
d.var_class2
|
||||
#? []
|
||||
d.int
|
||||
#? ['var_class1', 'var_class2', 'var_instance1', 'var_instance2', 'var_class3']
|
||||
#? ['var_class1', 'var_class2', 'var_instance1', 'var_instance2']
|
||||
self.var_
|
||||
|
||||
class VarClass2(VarClass):
|
||||
var_class3: typing.ClassVar[int]
|
||||
|
||||
def __init__(self):
|
||||
#? int()
|
||||
self.var_class3
|
||||
|
||||
#? ['var_class1', 'var_class2', 'var_instance1', 'var_class3', 'var_instance2']
|
||||
#? ['var_class1', 'var_class2', 'var_instance1']
|
||||
VarClass.var_
|
||||
#? int()
|
||||
VarClass.var_instance1
|
||||
#? float()
|
||||
#?
|
||||
VarClass.var_instance2
|
||||
#? str()
|
||||
VarClass.var_class1
|
||||
@@ -95,7 +91,7 @@ VarClass.var_class2
|
||||
VarClass.int
|
||||
|
||||
d = VarClass()
|
||||
#? ['var_class1', 'var_class2', 'var_class3', 'var_instance1', 'var_instance2']
|
||||
#? ['var_class1', 'var_class2', 'var_instance1', 'var_instance2']
|
||||
d.var_
|
||||
#? int()
|
||||
d.var_instance1
|
||||
|
||||
@@ -1,46 +0,0 @@
|
||||
from pep0484_generic_parameters import list_t_to_list_t
|
||||
|
||||
list_of_ints_and_strs: list[int | str]
|
||||
|
||||
# Test that unions are handled
|
||||
x2 = list_t_to_list_t(list_of_ints_and_strs)[0]
|
||||
#? int() str()
|
||||
x2
|
||||
|
||||
for z in list_t_to_list_t(list_of_ints_and_strs):
|
||||
#? int() str()
|
||||
z
|
||||
|
||||
|
||||
from pep0484_generic_passthroughs import (
|
||||
typed_variadic_tuple_generic_passthrough,
|
||||
)
|
||||
|
||||
variadic_tuple_str_int: tuple[int | str, ...]
|
||||
|
||||
for m in typed_variadic_tuple_generic_passthrough(variadic_tuple_str_int):
|
||||
#? str() int()
|
||||
m
|
||||
|
||||
|
||||
def func_returns_byteslike() -> bytes | bytearray:
|
||||
pass
|
||||
|
||||
#? bytes() bytearray()
|
||||
func_returns_byteslike()
|
||||
|
||||
|
||||
pep604_optional_1: int | str | None
|
||||
pep604_optional_2: None | bytes
|
||||
|
||||
#? int() str() None
|
||||
pep604_optional_1
|
||||
|
||||
#? None bytes()
|
||||
pep604_optional_2
|
||||
|
||||
|
||||
pep604_in_str: "int | bytes"
|
||||
|
||||
#? int() bytes()
|
||||
pep604_in_str
|
||||
@@ -1,5 +1,3 @@
|
||||
from typing import Generator
|
||||
|
||||
import pytest
|
||||
from pytest import fixture
|
||||
|
||||
@@ -66,11 +64,6 @@ def lala(my_fixture):
|
||||
def lala(my_fixture):
|
||||
pass
|
||||
|
||||
# overriding types of a fixture should be possible
|
||||
def test_x(my_yield_fixture: str):
|
||||
#? str()
|
||||
my_yield_fixture
|
||||
|
||||
# -----------------
|
||||
# completion
|
||||
# -----------------
|
||||
@@ -139,6 +132,9 @@ def test_p(monkeypatch):
|
||||
#? ['capsysbinary']
|
||||
def test_p(capsysbin
|
||||
|
||||
#? ['tmpdir', 'tmpdir_factory']
|
||||
def test_p(tmpdi
|
||||
|
||||
|
||||
def close_parens():
|
||||
pass
|
||||
@@ -168,40 +164,3 @@ def test_inheritance_fixture(inheritance_fixture, caplog):
|
||||
@pytest.fixture
|
||||
def caplog(caplog):
|
||||
yield caplog
|
||||
|
||||
# -----------------
|
||||
# Generator with annotation
|
||||
# -----------------
|
||||
|
||||
@pytest.fixture
|
||||
def with_annot() -> Generator[float, None, None]:
|
||||
pass
|
||||
|
||||
def test_with_annot(inheritance_fixture, with_annot):
|
||||
#? float()
|
||||
with_annot
|
||||
|
||||
# -----------------
|
||||
# pytest external plugins
|
||||
# -----------------
|
||||
|
||||
#? ['admin_user', 'admin_client']
|
||||
def test_z(admin
|
||||
|
||||
#! 15 ['def admin_client']
|
||||
def test_p(admin_client):
|
||||
#? ['login', 'logout']
|
||||
admin_client.log
|
||||
|
||||
@pytest.fixture
|
||||
@some_decorator
|
||||
#? ['admin_user']
|
||||
def bla(admin_u
|
||||
return
|
||||
|
||||
@pytest.fixture
|
||||
@some_decorator
|
||||
#! 12 ['def admin_user']
|
||||
def bla(admin_user):
|
||||
pass
|
||||
|
||||
|
||||
@@ -1,6 +1,7 @@
|
||||
import os
|
||||
import sys
|
||||
import subprocess
|
||||
from itertools import count
|
||||
|
||||
import pytest
|
||||
|
||||
@@ -9,6 +10,9 @@ from . import run
|
||||
from . import refactor
|
||||
from jedi import InterpreterEnvironment, get_system_environment
|
||||
from jedi.inference.compiled.value import create_from_access_path
|
||||
from jedi.inference.imports import _load_python_module
|
||||
from jedi.file_io import KnownContentFileIO
|
||||
from jedi.inference.base_value import ValueSet
|
||||
from jedi.api.interpreter import MixedModuleContext
|
||||
|
||||
# For interpreter tests sometimes the path of this directory is in the sys
|
||||
@@ -159,6 +163,19 @@ def create_compiled_object(inference_state):
|
||||
)
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
def module_injector():
|
||||
counter = count()
|
||||
|
||||
def module_injector(inference_state, names, code):
|
||||
assert isinstance(names, tuple)
|
||||
file_io = KnownContentFileIO('/foo/bar/module-injector-%s.py' % next(counter), code)
|
||||
v = _load_python_module(inference_state, file_io, names)
|
||||
inference_state.module_cache.add(names, ValueSet([v]))
|
||||
|
||||
return module_injector
|
||||
|
||||
|
||||
@pytest.fixture(params=[False, True])
|
||||
def class_findable(monkeypatch, request):
|
||||
if not request.param:
|
||||
|
||||
@@ -1,6 +0,0 @@
|
||||
from pytest import fixture
|
||||
|
||||
|
||||
@fixture()
|
||||
def admin_user():
|
||||
pass
|
||||
@@ -1,16 +0,0 @@
|
||||
import pytest
|
||||
|
||||
from .fixtures import admin_user # noqa
|
||||
|
||||
|
||||
@pytest.fixture()
|
||||
def admin_client():
|
||||
return Client()
|
||||
|
||||
|
||||
class Client:
|
||||
def login(self, **credentials):
|
||||
...
|
||||
|
||||
def logout(self):
|
||||
...
|
||||
19
test/run.py
19
test/run.py
@@ -104,14 +104,10 @@ import os
|
||||
import re
|
||||
import sys
|
||||
import operator
|
||||
if sys.version_info < (3, 8):
|
||||
literal_eval = eval
|
||||
else:
|
||||
from ast import literal_eval
|
||||
from ast import literal_eval
|
||||
from io import StringIO
|
||||
from functools import reduce
|
||||
from unittest.mock import ANY
|
||||
from pathlib import Path
|
||||
|
||||
import parso
|
||||
from _pytest.outcomes import Skipped
|
||||
@@ -126,7 +122,6 @@ from jedi.api.environment import get_default_environment, get_system_environment
|
||||
from jedi.inference.gradual.conversion import convert_values
|
||||
from jedi.inference.analysis import Warning
|
||||
|
||||
test_dir = Path(__file__).absolute().parent
|
||||
|
||||
TEST_COMPLETIONS = 0
|
||||
TEST_INFERENCE = 1
|
||||
@@ -178,7 +173,6 @@ class IntegrationTestCase(BaseTestCase):
|
||||
self.start = start
|
||||
self.line = line
|
||||
self.path = path
|
||||
self._project = jedi.Project(test_dir)
|
||||
|
||||
@property
|
||||
def module_name(self):
|
||||
@@ -194,12 +188,7 @@ class IntegrationTestCase(BaseTestCase):
|
||||
self.line_nr_test, self.line.rstrip())
|
||||
|
||||
def script(self, environment):
|
||||
return jedi.Script(
|
||||
self.source,
|
||||
path=self.path,
|
||||
environment=environment,
|
||||
project=self._project
|
||||
)
|
||||
return jedi.Script(self.source, path=self.path, environment=environment)
|
||||
|
||||
def run(self, compare_cb, environment=None):
|
||||
testers = {
|
||||
@@ -209,7 +198,7 @@ class IntegrationTestCase(BaseTestCase):
|
||||
TEST_REFERENCES: self.run_get_references,
|
||||
}
|
||||
if (self.path.endswith('pytest.py') or self.path.endswith('conftest.py')) \
|
||||
and os.path.realpath(environment.executable) != os.path.realpath(sys.executable):
|
||||
and environment.executable != os.path.realpath(sys.executable):
|
||||
# It's not guarantueed that pytest is installed in test
|
||||
# environments, if we're not running in the same environment that
|
||||
# we're already in, so just skip that case.
|
||||
@@ -274,7 +263,7 @@ class IntegrationTestCase(BaseTestCase):
|
||||
self.correct = self.correct.strip()
|
||||
compare = sorted(
|
||||
(('stub:' if r.is_stub() else '')
|
||||
+ re.sub(r'^completion\.', '', r.module_name),
|
||||
+ re.sub(r'^test\.completion\.', '', r.module_name),
|
||||
r.line,
|
||||
r.column)
|
||||
for r in result
|
||||
|
||||
@@ -650,7 +650,6 @@ def test_cursor_after_signature(Script, column):
|
||||
('abs(chr ( \nclass y: pass', 1, 8, 'abs', 0),
|
||||
('abs(chr ( \nclass y: pass', 1, 9, 'abs', 0),
|
||||
('abs(chr ( \nclass y: pass', 1, 10, 'chr', 0),
|
||||
('abs(foo.bar=3)', 1, 13, 'abs', 0),
|
||||
]
|
||||
)
|
||||
def test_base_signatures(Script, code, line, column, name, index):
|
||||
|
||||
@@ -1,16 +1,11 @@
|
||||
from os.path import join, sep as s, dirname, expanduser
|
||||
import os
|
||||
from textwrap import dedent
|
||||
from itertools import count
|
||||
from pathlib import Path
|
||||
|
||||
import pytest
|
||||
|
||||
from ..helpers import root_dir
|
||||
from jedi.api.helpers import _start_match, _fuzzy_match
|
||||
from jedi.inference.imports import _load_python_module
|
||||
from jedi.file_io import KnownContentFileIO
|
||||
from jedi.inference.base_value import ValueSet
|
||||
|
||||
|
||||
def test_in_whitespace(Script):
|
||||
@@ -405,22 +400,6 @@ def test_ellipsis_completion(Script):
|
||||
assert Script('...').complete() == []
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
def module_injector():
|
||||
counter = count()
|
||||
|
||||
def module_injector(inference_state, names, code):
|
||||
assert isinstance(names, tuple)
|
||||
file_io = KnownContentFileIO(
|
||||
Path('foo/bar/module-injector-%s.py' % next(counter)).absolute(),
|
||||
code
|
||||
)
|
||||
v = _load_python_module(inference_state, file_io, names)
|
||||
inference_state.module_cache.add(names, ValueSet([v]))
|
||||
|
||||
return module_injector
|
||||
|
||||
|
||||
def test_completion_cache(Script, module_injector):
|
||||
"""
|
||||
For some modules like numpy, tensorflow or pandas we cache docstrings and
|
||||
@@ -457,7 +436,3 @@ def test_module_completions(Script, module):
|
||||
# Just make sure that there are no errors
|
||||
c.type
|
||||
c.docstring()
|
||||
|
||||
|
||||
def test_whitespace_at_end_after_dot(Script):
|
||||
assert 'strip' in [c.name for c in Script('str. ').complete()]
|
||||
|
||||
@@ -37,17 +37,6 @@ def test_operator_doc(Script):
|
||||
assert len(d.docstring()) > 100
|
||||
|
||||
|
||||
@pytest.mark.parametrize(
|
||||
'code, help_part', [
|
||||
('str', 'Create a new string object'),
|
||||
('str.strip', 'Return a copy of the string'),
|
||||
]
|
||||
)
|
||||
def test_stdlib_doc(Script, code, help_part):
|
||||
h, = Script(code).help()
|
||||
assert help_part in h.docstring(raw=True)
|
||||
|
||||
|
||||
def test_lambda(Script):
|
||||
d, = Script('lambda x: x').help(column=0)
|
||||
assert d.type == 'keyword'
|
||||
|
||||
@@ -603,11 +603,8 @@ def test_dict_getitem(code, types):
|
||||
@pytest.mark.parametrize(
|
||||
'code, expected', [
|
||||
('DunderCls()[0]', 'int'),
|
||||
('dunder[0]', 'int'),
|
||||
('next(DunderCls())', 'float'),
|
||||
('next(dunder)', 'float'),
|
||||
('for x in DunderCls(): x', 'str'),
|
||||
#('for x in dunder: x', 'str'),
|
||||
]
|
||||
)
|
||||
def test_dunders(class_is_findable, code, expected):
|
||||
@@ -626,8 +623,6 @@ def test_dunders(class_is_findable, code, expected):
|
||||
if not class_is_findable:
|
||||
DunderCls.__name__ = 'asdf'
|
||||
|
||||
dunder = DunderCls()
|
||||
|
||||
n, = jedi.Interpreter(code, [locals()]).infer()
|
||||
assert n.name == expected
|
||||
|
||||
@@ -711,46 +706,3 @@ def test_negate():
|
||||
assert x.name == 'int'
|
||||
value, = x._name.infer()
|
||||
assert value.get_safe_value() == -3
|
||||
|
||||
|
||||
def test_complete_not_findable_class_source():
|
||||
class TestClass():
|
||||
ta=1
|
||||
ta1=2
|
||||
|
||||
# Simulate the environment where the class is defined in
|
||||
# an interactive session and therefore inspect module
|
||||
# cannot find its source code and raises OSError (Py 3.10+) or TypeError.
|
||||
TestClass.__module__ = "__main__"
|
||||
# There is a pytest __main__ module we have to remove temporarily.
|
||||
module = sys.modules.pop("__main__")
|
||||
try:
|
||||
interpreter = jedi.Interpreter("TestClass.", [locals()])
|
||||
completions = interpreter.complete(column=10, line=1)
|
||||
finally:
|
||||
sys.modules["__main__"] = module
|
||||
|
||||
assert "ta" in [c.name for c in completions]
|
||||
assert "ta1" in [c.name for c in completions]
|
||||
|
||||
|
||||
def test_param_infer_default():
|
||||
abs_sig, = jedi.Interpreter('abs(', [{'abs': abs}]).get_signatures()
|
||||
param, = abs_sig.params
|
||||
assert param.name == 'x'
|
||||
assert param.infer_default() == []
|
||||
|
||||
|
||||
@pytest.mark.parametrize(
|
||||
'code, expected', [
|
||||
("random.triangular(", ['high=', 'low=', 'mode=']),
|
||||
("random.triangular(low=1, ", ['high=', 'mode=']),
|
||||
("random.triangular(high=1, ", ['low=', 'mode=']),
|
||||
("random.triangular(low=1, high=2, ", ['mode=']),
|
||||
("random.triangular(low=1, mode=2, ", ['high=']),
|
||||
],
|
||||
)
|
||||
def test_keyword_param_completion(code, expected):
|
||||
import random
|
||||
completions = jedi.Interpreter(code, [locals()]).complete()
|
||||
assert expected == [c.name for c in completions if c.name.endswith('=')]
|
||||
|
||||
@@ -189,9 +189,3 @@ def test_no_error(get_names):
|
||||
def test_is_side_effect(get_names, code, index, is_side_effect):
|
||||
names = get_names(code, references=True, all_scopes=True)
|
||||
assert names[index].is_side_effect() == is_side_effect
|
||||
|
||||
|
||||
def test_no_defined_names(get_names):
|
||||
definition, = get_names("x = (1, 2)")
|
||||
|
||||
assert not definition.defined_names()
|
||||
|
||||
@@ -68,10 +68,6 @@ def test_load_save_project(tmpdir):
|
||||
dict(all_scopes=True)),
|
||||
('some_search_test_var', ['test_api.test_project.test_search.some_search_test_var'],
|
||||
dict(complete=True, all_scopes=True)),
|
||||
# Make sure that the searched name is not part of the file, by
|
||||
# splitting it up.
|
||||
('some_search_test_v' + 'a', ['test_api.test_project.test_search.some_search_test_var'],
|
||||
dict(complete=True, all_scopes=True)),
|
||||
|
||||
('sample_int', ['helpers.sample_int'], {}),
|
||||
('sample_int', ['helpers.sample_int'], dict(all_scopes=True)),
|
||||
@@ -150,7 +146,7 @@ def test_search(string, full_names, kwargs):
|
||||
defs = project.complete_search(string, **kwargs)
|
||||
else:
|
||||
defs = project.search(string, **kwargs)
|
||||
assert sorted([('stub:' if d.is_stub() else '') + (d.full_name or d.name) for d in defs]) == full_names
|
||||
assert sorted([('stub:' if d.is_stub() else '') + d.full_name for d in defs]) == full_names
|
||||
|
||||
|
||||
@pytest.mark.parametrize(
|
||||
@@ -178,7 +174,7 @@ def test_is_potential_project(path, expected):
|
||||
|
||||
if expected is None:
|
||||
try:
|
||||
expected = bool(set(_CONTAINS_POTENTIAL_PROJECT) & set(os.listdir(path)))
|
||||
expected = _CONTAINS_POTENTIAL_PROJECT in os.listdir(path)
|
||||
except OSError:
|
||||
expected = False
|
||||
|
||||
|
||||
@@ -1,7 +1,6 @@
|
||||
import os
|
||||
from textwrap import dedent
|
||||
from pathlib import Path
|
||||
import platform
|
||||
|
||||
import pytest
|
||||
|
||||
@@ -71,23 +70,3 @@ def test_diff_without_ending_newline(Script):
|
||||
-a
|
||||
+c
|
||||
''')
|
||||
|
||||
|
||||
def test_diff_path_outside_of_project(Script):
|
||||
if platform.system().lower() == 'windows':
|
||||
abs_path = r'D:\unknown_dir\file.py'
|
||||
else:
|
||||
abs_path = '/unknown_dir/file.py'
|
||||
script = Script(
|
||||
code='foo = 1',
|
||||
path=abs_path,
|
||||
project=jedi.get_default_project()
|
||||
)
|
||||
diff = script.rename(line=1, column=0, new_name='bar').get_diff()
|
||||
assert diff == dedent(f'''\
|
||||
--- {abs_path}
|
||||
+++ {abs_path}
|
||||
@@ -1 +1 @@
|
||||
-foo = 1
|
||||
+bar = 1
|
||||
''')
|
||||
|
||||
@@ -64,6 +64,6 @@ def test_wrong_encoding(Script, tmpdir):
|
||||
# Use both latin-1 and utf-8 (a really broken file).
|
||||
x.write_binary('foobar = 1\nä'.encode('latin-1') + 'ä'.encode('utf-8'))
|
||||
|
||||
project = Project(tmpdir.strpath)
|
||||
project = Project('.', sys_path=[tmpdir.strpath])
|
||||
c, = Script('import x; x.foo', project=project).complete()
|
||||
assert c.name == 'foobar'
|
||||
|
||||
@@ -43,9 +43,6 @@ def test_implicit_namespace_package(Script):
|
||||
solution = "foo = '%s'" % solution
|
||||
assert completion.description == solution
|
||||
|
||||
c, = script_with_path('import pkg').complete()
|
||||
assert c.docstring() == ""
|
||||
|
||||
|
||||
def test_implicit_nested_namespace_package(Script):
|
||||
code = 'from implicit_nested_namespaces.namespace.pkg.module import CONST'
|
||||
|
||||
@@ -101,16 +101,6 @@ def test_correct_zip_package_behavior(Script, inference_state, environment, code
|
||||
assert value.py__package__() == []
|
||||
|
||||
|
||||
@pytest.mark.parametrize("code,names", [
|
||||
("from pkg.", {"module", "nested", "namespace"}),
|
||||
("from pkg.nested.", {"nested_module"})
|
||||
])
|
||||
def test_zip_package_import_complete(Script, environment, code, names):
|
||||
sys_path = environment.get_sys_path() + [str(pkg_zip_path)]
|
||||
completions = Script(code, project=Project('.', sys_path=sys_path)).complete()
|
||||
assert names == {c.name for c in completions}
|
||||
|
||||
|
||||
def test_find_module_not_package_zipped(Script, inference_state, environment):
|
||||
path = get_example_dir('zipped_imports', 'not_pkg.zip')
|
||||
sys_path = environment.get_sys_path() + [path]
|
||||
@@ -297,6 +287,7 @@ def test_os_issues(Script):
|
||||
# Github issue #759
|
||||
s = 'import os, s'
|
||||
assert 'sys' in import_names(s)
|
||||
assert 'path' not in import_names(s, column=len(s) - 1)
|
||||
assert 'os' in import_names(s, column=len(s) - 3)
|
||||
|
||||
# Some more checks
|
||||
@@ -333,13 +324,12 @@ def test_compiled_import_none(monkeypatch, Script):
|
||||
# context that was initially given, but now we just work with the file
|
||||
# system.
|
||||
(os.path.join(THIS_DIR, 'test_docstring.py'), False,
|
||||
('test_inference', 'test_imports')),
|
||||
('test', 'test_inference', 'test_imports')),
|
||||
(os.path.join(THIS_DIR, '__init__.py'), True,
|
||||
('test_inference', 'test_imports')),
|
||||
('test', 'test_inference', 'test_imports')),
|
||||
]
|
||||
)
|
||||
def test_get_modules_containing_name(inference_state, path, goal, is_package):
|
||||
inference_state.project = Project(test_dir)
|
||||
module = imports._load_python_module(
|
||||
inference_state,
|
||||
FileIO(path),
|
||||
|
||||
@@ -267,19 +267,19 @@ def test_pow_signature(Script, environment):
|
||||
@pytest.mark.parametrize(
|
||||
'code, signature', [
|
||||
[dedent('''
|
||||
# identifier:A
|
||||
import functools
|
||||
def f(x):
|
||||
pass
|
||||
def x(f):
|
||||
@functools.wraps(f)
|
||||
def wrapper(*args):
|
||||
# Have no arguments here, but because of wraps, the signature
|
||||
# should still be f's.
|
||||
return f(*args)
|
||||
return wrapper
|
||||
|
||||
x(f)('''), 'f(x, /)'],
|
||||
[dedent('''
|
||||
# identifier:B
|
||||
import functools
|
||||
def f(x):
|
||||
pass
|
||||
@@ -292,26 +292,6 @@ def test_pow_signature(Script, environment):
|
||||
return wrapper
|
||||
|
||||
x(f)('''), 'f()'],
|
||||
[dedent('''
|
||||
# identifier:C
|
||||
import functools
|
||||
def f(x: int, y: float):
|
||||
pass
|
||||
|
||||
@functools.wraps(f)
|
||||
def wrapper(*args, **kwargs):
|
||||
return f(*args, **kwargs)
|
||||
|
||||
wrapper('''), 'f(x: int, y: float)'],
|
||||
[dedent('''
|
||||
# identifier:D
|
||||
def f(x: int, y: float):
|
||||
pass
|
||||
|
||||
def wrapper(*args, **kwargs):
|
||||
return f(*args, **kwargs)
|
||||
|
||||
wrapper('''), 'wrapper(x: int, y: float)'],
|
||||
]
|
||||
)
|
||||
def test_wraps_signature(Script, code, signature):
|
||||
@@ -355,48 +335,6 @@ def test_dataclass_signature(Script, skip_pre_python37, start, start_params):
|
||||
price, = sig.params[-2].infer()
|
||||
assert price.name == 'float'
|
||||
|
||||
@pytest.mark.parametrize(
|
||||
'start, start_params', [
|
||||
['@define\nclass X:', []],
|
||||
['@frozen\nclass X:', []],
|
||||
['@define(eq=True)\nclass X:', []],
|
||||
[dedent('''
|
||||
class Y():
|
||||
y: int
|
||||
@define
|
||||
class X(Y):'''), []],
|
||||
[dedent('''
|
||||
@define
|
||||
class Y():
|
||||
y: int
|
||||
z = 5
|
||||
@define
|
||||
class X(Y):'''), ['y']],
|
||||
]
|
||||
)
|
||||
def test_attrs_signature(Script, skip_pre_python37, start, start_params):
|
||||
has_attrs = bool(Script('import attrs').infer())
|
||||
if not has_attrs:
|
||||
raise pytest.skip("attrs needed in target environment to run this test")
|
||||
|
||||
code = dedent('''
|
||||
name: str
|
||||
foo = 3
|
||||
price: float
|
||||
quantity: int = 0.0
|
||||
|
||||
X(''')
|
||||
|
||||
# attrs exposes two namespaces
|
||||
code = 'from attrs import define, frozen\n' + start + code
|
||||
|
||||
sig, = Script(code).get_signatures()
|
||||
assert [p.name for p in sig.params] == start_params + ['name', 'price', 'quantity']
|
||||
quantity, = sig.params[-1].infer()
|
||||
assert quantity.name == 'int'
|
||||
price, = sig.params[-2].infer()
|
||||
assert price.name == 'float'
|
||||
|
||||
|
||||
@pytest.mark.parametrize(
|
||||
'stmt, expected', [
|
||||
|
||||
@@ -1,5 +1,4 @@
|
||||
import os
|
||||
from collections import namedtuple
|
||||
|
||||
import pytest
|
||||
|
||||
@@ -43,22 +42,6 @@ def test_completion(case, monkeypatch, environment, has_django):
|
||||
|
||||
if (not has_django) and case.path.endswith('django.py'):
|
||||
pytest.skip('Needs django to be installed to run this test.')
|
||||
|
||||
if case.path.endswith("pytest.py"):
|
||||
# to test finding pytest fixtures from external plugins
|
||||
# add a stub pytest plugin to the project sys_path...
|
||||
pytest_plugin_dir = str(helpers.get_example_dir("pytest_plugin_package"))
|
||||
case._project.added_sys_path = [pytest_plugin_dir]
|
||||
|
||||
# ... and mock setuptools entry points to include it
|
||||
# see https://docs.pytest.org/en/stable/how-to/writing_plugins.html#setuptools-entry-points
|
||||
def mock_iter_entry_points(group):
|
||||
assert group == "pytest11"
|
||||
EntryPoint = namedtuple("EntryPoint", ["module_name"])
|
||||
return [EntryPoint("pytest_plugin.plugin")]
|
||||
|
||||
monkeypatch.setattr("pkg_resources.iter_entry_points", mock_iter_entry_points)
|
||||
|
||||
repo_root = helpers.root_dir
|
||||
monkeypatch.chdir(os.path.join(repo_root, 'jedi'))
|
||||
case.run(assert_case_equal, environment)
|
||||
|
||||
@@ -1,9 +1,5 @@
|
||||
import gc
|
||||
from pathlib import Path
|
||||
|
||||
from jedi import parser_utils
|
||||
from parso import parse
|
||||
from parso.cache import parser_cache
|
||||
from parso.python import tree
|
||||
|
||||
import pytest
|
||||
@@ -71,18 +67,3 @@ def test_get_signature(code, signature):
|
||||
if node.type == 'simple_stmt':
|
||||
node = node.children[0]
|
||||
assert parser_utils.get_signature(node) == signature
|
||||
|
||||
|
||||
def test_parser_cache_clear(Script):
|
||||
"""
|
||||
If parso clears its cache, Jedi should not keep those resources, they
|
||||
should be freed.
|
||||
"""
|
||||
script = Script("a = abs\na", path=Path(__file__).parent / 'parser_cache_test_foo.py')
|
||||
script.complete()
|
||||
module_id = id(script._module_node)
|
||||
del parser_cache[script._inference_state.grammar._hashed][script.path]
|
||||
del script
|
||||
|
||||
gc.collect()
|
||||
assert module_id not in [id(m) for m in gc.get_referrers(tree.Module)]
|
||||
|
||||
@@ -85,7 +85,7 @@ class TestSetupReadline(unittest.TestCase):
|
||||
}
|
||||
# There are quite a few differences, because both Windows and Linux
|
||||
# (posix and nt) librariesare included.
|
||||
assert len(difference) < 30
|
||||
assert len(difference) < 15
|
||||
|
||||
def test_local_import(self):
|
||||
s = 'import test.test_utils'
|
||||
|
||||
Reference in New Issue
Block a user