forked from VimPlug/jedi
More test fixes.
This commit is contained in:
@@ -136,7 +136,7 @@ class Completion:
|
|||||||
try:
|
try:
|
||||||
stack = helpers.get_stack_at_position(grammar, self._source, self._module, pos)
|
stack = helpers.get_stack_at_position(grammar, self._source, self._module, pos)
|
||||||
except helpers.OnErrorLeaf:
|
except helpers.OnErrorLeaf:
|
||||||
return []
|
return self._simple_complete(completion_parts)
|
||||||
|
|
||||||
allowed_keywords, allowed_tokens = \
|
allowed_keywords, allowed_tokens = \
|
||||||
helpers.get_possible_completion_types(grammar, stack)
|
helpers.get_possible_completion_types(grammar, stack)
|
||||||
|
|||||||
@@ -98,6 +98,7 @@ def get_stack_at_position(grammar, source, module, pos):
|
|||||||
if pos <= user_stmt.start_pos:
|
if pos <= user_stmt.start_pos:
|
||||||
try:
|
try:
|
||||||
leaf = user_stmt.get_previous_leaf()
|
leaf = user_stmt.get_previous_leaf()
|
||||||
|
print(user_stmt, leaf)
|
||||||
except IndexError:
|
except IndexError:
|
||||||
pass
|
pass
|
||||||
else:
|
else:
|
||||||
|
|||||||
@@ -414,23 +414,24 @@ class ParserWithRecovery(Parser):
|
|||||||
|
|
||||||
#print('err', token.tok_name[typ], repr(value), start_pos, len(stack), index)
|
#print('err', token.tok_name[typ], repr(value), start_pos, len(stack), index)
|
||||||
if self._stack_removal(grammar, stack, arcs, index + 1, value, start_pos):
|
if self._stack_removal(grammar, stack, arcs, index + 1, value, start_pos):
|
||||||
#add_token_callback(typ, value, prefix, start_pos)
|
add_token_callback(typ, value, prefix, start_pos)
|
||||||
pass
|
if typ == INDENT:
|
||||||
|
# For every deleted INDENT we have to delete a DEDENT as well.
|
||||||
|
# Otherwise the parser will get into trouble and DEDENT too early.
|
||||||
|
self._omit_dedent_list.append(self._indent_counter)
|
||||||
|
|
||||||
else:
|
else:
|
||||||
#error_leaf = ErrorToken(self.position_modifier, value, start_pos, prefix)
|
#error_leaf = ErrorToken(self.position_modifier, value, start_pos, prefix)
|
||||||
#stack = [(None, [error_leaf])]
|
#stack = [(None, [error_leaf])]
|
||||||
# TODO document the shizzle!
|
# TODO document the shizzle!
|
||||||
#self._error_statements.append(ErrorStatement(stack, None, None,
|
#self._error_statements.append(ErrorStatement(stack, None, None,
|
||||||
# self.position_modifier, error_leaf.end_pos))
|
# self.position_modifier, error_leaf.end_pos))
|
||||||
error_leaf = pt.ErrorLeaf(self.position_modifier, value, start_pos, prefix)
|
error_leaf = pt.ErrorLeaf(self.position_modifier, typ, value, start_pos, prefix)
|
||||||
stack[-1][2][1].append(error_leaf)
|
stack[-1][2][1].append(error_leaf)
|
||||||
|
|
||||||
return
|
return
|
||||||
|
|
||||||
if typ == INDENT:
|
'''
|
||||||
# For every deleted INDENT we have to delete a DEDENT as well.
|
|
||||||
# Otherwise the parser will get into trouble and DEDENT too early.
|
|
||||||
self._omit_dedent_list.append(self._indent_counter)
|
|
||||||
|
|
||||||
if value in ('import', 'class', 'def', 'try', 'while', 'return', '\n'):
|
if value in ('import', 'class', 'def', 'try', 'while', 'return', '\n'):
|
||||||
# Those can always be new statements.
|
# Those can always be new statements.
|
||||||
add_token_callback(typ, value, prefix, start_pos)
|
add_token_callback(typ, value, prefix, start_pos)
|
||||||
@@ -445,6 +446,7 @@ class ParserWithRecovery(Parser):
|
|||||||
# anyway (compile.c does that for Python), because Python's grammar
|
# anyway (compile.c does that for Python), because Python's grammar
|
||||||
# doesn't stop you from defining `continue` in a module, etc.
|
# doesn't stop you from defining `continue` in a module, etc.
|
||||||
add_token_callback(typ, value, prefix, start_pos)
|
add_token_callback(typ, value, prefix, start_pos)
|
||||||
|
'''
|
||||||
|
|
||||||
def _stack_removal(self, grammar, stack, arcs, start_index, value, start_pos):
|
def _stack_removal(self, grammar, stack, arcs, start_index, value, start_pos):
|
||||||
def clear_names(children):
|
def clear_names(children):
|
||||||
|
|||||||
@@ -121,11 +121,9 @@ class PgenParser(object):
|
|||||||
break
|
break
|
||||||
else:
|
else:
|
||||||
# We never broke out -- EOF is too soon -- Unfinished statement.
|
# We never broke out -- EOF is too soon -- Unfinished statement.
|
||||||
# TODO the arcs argument [] is not correctly defined.
|
# However, the error recovery might have added the token again, if
|
||||||
self.error_recovery(self.grammar, self.stack, [], type_, value,
|
# the stack is empty, we're fine.
|
||||||
start_pos, prefix, self.addtoken)
|
if self.stack:
|
||||||
# Add the ENDMARKER again.
|
|
||||||
if not self.addtoken(type_, value, prefix, start_pos):
|
|
||||||
raise InternalParseError("incomplete input", type_, value, start_pos)
|
raise InternalParseError("incomplete input", type_, value, start_pos)
|
||||||
return self.rootnode
|
return self.rootnode
|
||||||
|
|
||||||
|
|||||||
@@ -43,6 +43,7 @@ from jedi import common
|
|||||||
from jedi._compatibility import (Python3Method, encoding, is_py3, utf8_repr,
|
from jedi._compatibility import (Python3Method, encoding, is_py3, utf8_repr,
|
||||||
literal_eval, use_metaclass, unicode)
|
literal_eval, use_metaclass, unicode)
|
||||||
from jedi import cache
|
from jedi import cache
|
||||||
|
from jedi.parser import token
|
||||||
|
|
||||||
|
|
||||||
def is_node(node, *symbol_names):
|
def is_node(node, *symbol_names):
|
||||||
@@ -653,9 +654,17 @@ class ErrorLeaf(Leaf):
|
|||||||
"""
|
"""
|
||||||
TODO doc
|
TODO doc
|
||||||
"""
|
"""
|
||||||
__slots__ = ()
|
__slots__ = ('original_type')
|
||||||
type = 'error_leaf'
|
type = 'error_leaf'
|
||||||
|
|
||||||
|
def __init__(self, position_modifier, original_type, value, start_pos, prefix=''):
|
||||||
|
super(ErrorLeaf, self).__init__(position_modifier, value, start_pos, prefix)
|
||||||
|
self.original_type = original_type
|
||||||
|
|
||||||
|
def __repr__(self):
|
||||||
|
token_type = token.tok_name[self.original_type]
|
||||||
|
return "<%s: %s, %s)>" % (type(self).__name__, token_type, repr(self.value))
|
||||||
|
|
||||||
|
|
||||||
class IsScopeMeta(type):
|
class IsScopeMeta(type):
|
||||||
def __instancecheck__(self, other):
|
def __instancecheck__(self, other):
|
||||||
|
|||||||
@@ -10,7 +10,7 @@ import os
|
|||||||
import sys
|
import sys
|
||||||
|
|
||||||
from jedi import Interpreter
|
from jedi import Interpreter
|
||||||
from jedi.api.helpers import completion_parts
|
from jedi.api.helpers import get_completion_parts
|
||||||
from jedi.parser.user_context import UserContext
|
from jedi.parser.user_context import UserContext
|
||||||
|
|
||||||
|
|
||||||
@@ -73,7 +73,7 @@ def setup_readline(namespace_module=__main__):
|
|||||||
interpreter = Interpreter(text, [namespace_module.__dict__])
|
interpreter = Interpreter(text, [namespace_module.__dict__])
|
||||||
|
|
||||||
path = UserContext(text, (1, len(text))).get_path_until_cursor()
|
path = UserContext(text, (1, len(text))).get_path_until_cursor()
|
||||||
path, dot, like = completion_parts(path)
|
path, dot, like = get_completion_parts(path)
|
||||||
before = text[:len(text) - len(like)]
|
before = text[:len(text) - len(like)]
|
||||||
completions = interpreter.completions()
|
completions = interpreter.completions()
|
||||||
finally:
|
finally:
|
||||||
|
|||||||
Reference in New Issue
Block a user