forked from VimPlug/jedi
I don't really understand why this wasn't an issue before, but it looks like we have to catch both IOError and and socket.error in Python2.
335 lines
11 KiB
Python
335 lines
11 KiB
Python
"""
|
|
Makes it possible to do the compiled analysis in a subprocess. This has two
|
|
goals:
|
|
|
|
1. Making it safer - Segfaults and RuntimeErrors as well as stdout/stderr can
|
|
be ignored and dealt with.
|
|
2. Make it possible to handle different Python versions as well as virtualenvs.
|
|
"""
|
|
|
|
import os
|
|
import sys
|
|
import subprocess
|
|
import socket
|
|
import errno
|
|
import weakref
|
|
import traceback
|
|
from functools import partial
|
|
|
|
from jedi._compatibility import queue, is_py3, force_unicode, \
|
|
pickle_dump, pickle_load
|
|
from jedi.cache import memoize_method
|
|
from jedi.evaluate.compiled.subprocess import functions
|
|
from jedi.evaluate.compiled.access import DirectObjectAccess, AccessPath, \
|
|
SignatureParam
|
|
from jedi.api.exceptions import InternalError
|
|
|
|
_subprocesses = {}
|
|
|
|
_MAIN_PATH = os.path.join(os.path.dirname(__file__), '__main__.py')
|
|
|
|
|
|
def get_subprocess(executable):
|
|
try:
|
|
return _subprocesses[executable]
|
|
except KeyError:
|
|
sub = _subprocesses[executable] = _CompiledSubprocess(executable)
|
|
return sub
|
|
|
|
|
|
def _get_function(name):
|
|
return getattr(functions, name)
|
|
|
|
|
|
class _EvaluatorProcess(object):
|
|
def __init__(self, evaluator):
|
|
self._evaluator_weakref = weakref.ref(evaluator)
|
|
self._evaluator_id = id(evaluator)
|
|
self._handles = {}
|
|
|
|
def get_or_create_access_handle(self, obj):
|
|
id_ = id(obj)
|
|
try:
|
|
return self.get_access_handle(id_)
|
|
except KeyError:
|
|
access = DirectObjectAccess(self._evaluator_weakref(), obj)
|
|
handle = AccessHandle(self, access, id_)
|
|
self.set_access_handle(handle)
|
|
return handle
|
|
|
|
def get_access_handle(self, id_):
|
|
return self._handles[id_]
|
|
|
|
def set_access_handle(self, handle):
|
|
self._handles[handle.id] = handle
|
|
|
|
|
|
class EvaluatorSameProcess(_EvaluatorProcess):
|
|
"""
|
|
Basically just an easy access to functions.py. It has the same API
|
|
as EvaluatorSubprocess and does the same thing without using a subprocess.
|
|
This is necessary for the Interpreter process.
|
|
"""
|
|
def __getattr__(self, name):
|
|
return partial(_get_function(name), self._evaluator_weakref())
|
|
|
|
|
|
class EvaluatorSubprocess(_EvaluatorProcess):
|
|
def __init__(self, evaluator, compiled_subprocess):
|
|
super(EvaluatorSubprocess, self).__init__(evaluator)
|
|
self._used = False
|
|
self._compiled_subprocess = compiled_subprocess
|
|
|
|
def __getattr__(self, name):
|
|
func = _get_function(name)
|
|
|
|
def wrapper(*args, **kwargs):
|
|
self._used = True
|
|
|
|
result = self._compiled_subprocess.run(
|
|
self._evaluator_weakref(),
|
|
func,
|
|
args=args,
|
|
kwargs=kwargs,
|
|
)
|
|
# IMO it should be possible to create a hook in pickle.load to
|
|
# mess with the loaded objects. However it's extremely complicated
|
|
# to work around this so just do it with this call. ~ dave
|
|
return self._convert_access_handles(result)
|
|
|
|
return wrapper
|
|
|
|
def _convert_access_handles(self, obj):
|
|
if isinstance(obj, SignatureParam):
|
|
return SignatureParam(*self._convert_access_handles(tuple(obj)))
|
|
elif isinstance(obj, tuple):
|
|
return tuple(self._convert_access_handles(o) for o in obj)
|
|
elif isinstance(obj, list):
|
|
return [self._convert_access_handles(o) for o in obj]
|
|
elif isinstance(obj, AccessHandle):
|
|
try:
|
|
# Rewrite the access handle to one we're already having.
|
|
obj = self.get_access_handle(obj.id)
|
|
except KeyError:
|
|
obj.add_subprocess(self)
|
|
self.set_access_handle(obj)
|
|
elif isinstance(obj, AccessPath):
|
|
return AccessPath(self._convert_access_handles(obj.accesses))
|
|
return obj
|
|
|
|
def __del__(self):
|
|
if self._used:
|
|
self._compiled_subprocess.delete_evaluator(self._evaluator_id)
|
|
|
|
|
|
class _CompiledSubprocess(object):
|
|
_crashed = False
|
|
|
|
def __init__(self, executable):
|
|
self._executable = executable
|
|
self._evaluator_deletion_queue = queue.deque()
|
|
|
|
@property
|
|
@memoize_method
|
|
def _process(self):
|
|
parso_path = sys.modules['parso'].__file__
|
|
args = (
|
|
self._executable,
|
|
_MAIN_PATH,
|
|
os.path.dirname(os.path.dirname(parso_path))
|
|
)
|
|
return subprocess.Popen(
|
|
args,
|
|
stdin=subprocess.PIPE,
|
|
stdout=subprocess.PIPE,
|
|
# stderr=subprocess.PIPE
|
|
)
|
|
|
|
def run(self, evaluator, function, args=(), kwargs={}):
|
|
# Delete old evaluators.
|
|
while True:
|
|
try:
|
|
evaluator_id = self._evaluator_deletion_queue.pop()
|
|
except IndexError:
|
|
break
|
|
else:
|
|
self._send(evaluator_id, None)
|
|
|
|
assert callable(function)
|
|
return self._send(id(evaluator), function, args, kwargs)
|
|
|
|
def get_sys_path(self):
|
|
return self._send(None, functions.get_sys_path, (), {})
|
|
|
|
def kill(self):
|
|
self._crashed = True
|
|
try:
|
|
subprocess = _subprocesses[self._executable]
|
|
except KeyError:
|
|
# Fine it was already removed from the cache.
|
|
pass
|
|
else:
|
|
# In the `!=` case there is already a new subprocess in place
|
|
# and we don't need to do anything here anymore.
|
|
if subprocess == self:
|
|
del _subprocesses[self._executable]
|
|
|
|
self._process.kill()
|
|
|
|
def _send(self, evaluator_id, function, args=(), kwargs={}):
|
|
if self._crashed:
|
|
raise InternalError("The subprocess has crashed.")
|
|
|
|
if not is_py3:
|
|
# Python 2 compatibility
|
|
kwargs = {force_unicode(key): value for key, value in kwargs.items()}
|
|
|
|
data = evaluator_id, function, args, kwargs
|
|
try:
|
|
pickle_dump(data, self._process.stdin)
|
|
self._process.stdin.flush()
|
|
except (socket.error, IOError) as e:
|
|
# Once Python2 will be removed we can just use `BrokenPipeError`.
|
|
if e.errno != errno.EPIPE:
|
|
# Not a broken pipe
|
|
raise
|
|
self.kill()
|
|
raise InternalError("The subprocess was killed. Maybe out of memory?")
|
|
|
|
try:
|
|
is_exception, traceback, result = pickle_load(self._process.stdout)
|
|
except EOFError:
|
|
self.kill()
|
|
raise InternalError("The subprocess crashed.")
|
|
|
|
if is_exception:
|
|
# Replace the attribute error message with a the traceback. It's
|
|
# way more informative.
|
|
result.args = (traceback,)
|
|
raise result
|
|
return result
|
|
|
|
def delete_evaluator(self, evaluator_id):
|
|
"""
|
|
Currently we are not deleting evalutors instantly. They only get
|
|
deleted once the subprocess is used again. It would probably a better
|
|
solution to move all of this into a thread. However, the memory usage
|
|
of a single evaluator shouldn't be that high.
|
|
"""
|
|
# With an argument - the evaluator gets deleted.
|
|
self._evaluator_deletion_queue.append(evaluator_id)
|
|
|
|
|
|
class Listener(object):
|
|
def __init__(self):
|
|
self._evaluators = {}
|
|
# TODO refactor so we don't need to process anymore just handle
|
|
# controlling.
|
|
self._process = _EvaluatorProcess(Listener)
|
|
|
|
def _get_evaluator(self, function, evaluator_id):
|
|
from jedi.evaluate import Evaluator
|
|
|
|
try:
|
|
evaluator = self._evaluators[evaluator_id]
|
|
except KeyError:
|
|
from jedi.api.environment import InterpreterEnvironment
|
|
evaluator = Evaluator(
|
|
# The project is not actually needed. Nothing should need to
|
|
# access it.
|
|
project=None,
|
|
environment=InterpreterEnvironment()
|
|
)
|
|
self._evaluators[evaluator_id] = evaluator
|
|
return evaluator
|
|
|
|
def _run(self, evaluator_id, function, args, kwargs):
|
|
if evaluator_id is None:
|
|
return function(*args, **kwargs)
|
|
elif function is None:
|
|
del self._evaluators[evaluator_id]
|
|
else:
|
|
evaluator = self._get_evaluator(function, evaluator_id)
|
|
|
|
# Exchange all handles
|
|
args = list(args)
|
|
for i, arg in enumerate(args):
|
|
if isinstance(arg, AccessHandle):
|
|
args[i] = evaluator.compiled_subprocess.get_access_handle(arg.id)
|
|
for key, value in kwargs.items():
|
|
if isinstance(value, AccessHandle):
|
|
kwargs[key] = evaluator.compiled_subprocess.get_access_handle(value.id)
|
|
|
|
return function(evaluator, *args, **kwargs)
|
|
|
|
def listen(self):
|
|
stdout = sys.stdout
|
|
stdin = sys.stdin
|
|
if sys.version_info[0] > 2:
|
|
stdout = stdout.buffer
|
|
stdin = stdin.buffer
|
|
|
|
while True:
|
|
try:
|
|
payload = pickle_load(stdin)
|
|
except EOFError:
|
|
# It looks like the parent process closed. Don't make a big fuss
|
|
# here and just exit.
|
|
exit(1)
|
|
try:
|
|
result = False, None, self._run(*payload)
|
|
except Exception as e:
|
|
result = True, traceback.format_exc(), e
|
|
|
|
pickle_dump(result, file=stdout)
|
|
stdout.flush()
|
|
|
|
|
|
class AccessHandle(object):
|
|
def __init__(self, subprocess, access, id_):
|
|
self.access = access
|
|
self._subprocess = subprocess
|
|
self.id = id_
|
|
|
|
def add_subprocess(self, subprocess):
|
|
self._subprocess = subprocess
|
|
|
|
def __repr__(self):
|
|
try:
|
|
detail = self.access
|
|
except AttributeError:
|
|
detail = '#' + str(self.id)
|
|
return '<%s of %s>' % (self.__class__.__name__, detail)
|
|
|
|
def __getstate__(self):
|
|
return self.id
|
|
|
|
def __setstate__(self, state):
|
|
self.id = state
|
|
|
|
def __getattr__(self, name):
|
|
if name in ('id', 'access') or name.startswith('_'):
|
|
raise AttributeError("Something went wrong with unpickling")
|
|
|
|
#if not is_py3: print >> sys.stderr, name
|
|
#print('getattr', name, file=sys.stderr)
|
|
return partial(self._workaround, force_unicode(name))
|
|
|
|
def _workaround(self, name, *args, **kwargs):
|
|
"""
|
|
TODO Currently we're passing slice objects around. This should not
|
|
happen. They are also the only unhashable objects that we're passing
|
|
around.
|
|
"""
|
|
if args and isinstance(args[0], slice):
|
|
return self._subprocess.get_compiled_method_return(self.id, name, *args, **kwargs)
|
|
return self._cached_results(name, *args, **kwargs)
|
|
|
|
@memoize_method
|
|
def _cached_results(self, name, *args, **kwargs):
|
|
#if type(self._subprocess) == EvaluatorSubprocess:
|
|
#print(name, args, kwargs,
|
|
#self._subprocess.get_compiled_method_return(self.id, name, *args, **kwargs)
|
|
#)
|
|
return self._subprocess.get_compiled_method_return(self.id, name, *args, **kwargs)
|