本文整理汇总了Python中pygments.lexer.RegexLexer类的典型用法代码示例。如果您正苦于以下问题:Python RegexLexer类的具体用法?Python RegexLexer怎么用?Python RegexLexer使用的例子?那么恭喜您, 这里精选的类代码示例或许可以为您提供帮助。
在下文中一共展示了RegexLexer类的20个代码示例,这些例子默认根据受欢迎程度排序。您可以为喜欢或者感觉有用的代码点赞,您的评价将有助于我们的系统推荐出更棒的Python代码示例。
示例1: __init__
def __init__(self, **options):
from pygments.lexers._vimbuiltins import command, option, auto
self._cmd = command
self._opt = option
self._aut = auto
RegexLexer.__init__(self, **options)
开发者ID:spencerlyon2,项目名称:pygments,代码行数:7,代码来源:textedit.py
示例2: __init__
def __init__(self, **options):
level = options.get('i6t', '+i6t-not-inline')
if level not in self._all_tokens:
self._tokens = self.__class__.process_tokendef(level)
else:
self._tokens = self._all_tokens[level]
RegexLexer.__init__(self, **options)
开发者ID:spencerlyon2,项目名称:pygments,代码行数:7,代码来源:int_fiction.py
示例3: __init__
def __init__(self, **options):
level = get_choice_opt(options, 'unicodelevel', list(self.tokens), 'basic')
if level not in self._all_tokens:
# compile the regexes now
self._tokens = self.__class__.process_tokendef(level)
else:
self._tokens = self._all_tokens[level]
RegexLexer.__init__(self, **options)
开发者ID:0038lana,项目名称:Test-Task,代码行数:9,代码来源:dotnet.py
示例4: __init__
def __init__(self, **options):
self.smhighlighting = get_bool_opt(options,
'sourcemod', True)
self._functions = set()
if self.smhighlighting:
from pygments.lexers._sourcemodbuiltins import FUNCTIONS
self._functions.update(FUNCTIONS)
RegexLexer.__init__(self, **options)
开发者ID:spencerlyon2,项目名称:pygments,代码行数:9,代码来源:pawn.py
示例5: __init__
def __init__(self, **options):
self.func_name_highlighting = get_bool_opt(
options, 'func_name_highlighting', True)
self.disabled_modules = get_list_opt(options, 'disabled_modules', [])
self._functions = set()
if self.func_name_highlighting:
from pygments.lexers._lua_builtins import MODULES
for mod, func in MODULES.iteritems():
if mod not in self.disabled_modules:
self._functions.update(func)
RegexLexer.__init__(self, **options)
开发者ID:marvin-h,项目名称:tarantool,代码行数:12,代码来源:LuaLexer.py
示例6: __init__
def __init__(self, **options):
from pygments.lexers._clbuiltins import BUILTIN_FUNCTIONS, \
SPECIAL_FORMS, MACROS, LAMBDA_LIST_KEYWORDS, DECLARATIONS, \
BUILTIN_TYPES, BUILTIN_CLASSES
self.builtin_function = BUILTIN_FUNCTIONS
self.special_forms = SPECIAL_FORMS
self.macros = MACROS
self.lambda_list_keywords = LAMBDA_LIST_KEYWORDS
self.declarations = DECLARATIONS
self.builtin_types = BUILTIN_TYPES
self.builtin_classes = BUILTIN_CLASSES
RegexLexer.__init__(self, **options)
开发者ID:Mekyi,项目名称:crunchy,代码行数:12,代码来源:functional.py
示例7: __init__
def __init__(self, **options):
self.builtinshighlighting = get_bool_opt(options, "builtinshighlighting", True)
self.requiredelimiters = get_bool_opt(options, "requiredelimiters", False)
self._builtins = set()
self._members = set()
if self.builtinshighlighting:
from pygments.lexers._lasso_builtins import BUILTINS, MEMBERS
for key, value in iteritems(BUILTINS):
self._builtins.update(value)
for key, value in iteritems(MEMBERS):
self._members.update(value)
RegexLexer.__init__(self, **options)
开发者ID:knossos-project,项目名称:PythonQt,代码行数:14,代码来源:javascript.py
示例8: get_tokens_unprocessed
def get_tokens_unprocessed(self, text):
# cut at the beginning of the interface and information tab stuff
substrings = text.partition('@#$#@#$#@')
text = substrings[0]
stack = ['root']
for item in RegexLexer.get_tokens_unprocessed(self, text, stack):
yield item
开发者ID:cdzhang,项目名称:netlogo_latex_minted,代码行数:7,代码来源:netlogo.py
示例9: get_tokens_unprocessed
def get_tokens_unprocessed(self, text):
stack = ['root']
for index, token, value in RegexLexer.get_tokens_unprocessed(self, text, stack):
if token is Name.Variable:
if value in self.builtin_function:
yield index, Name.Builtin, value
continue
if value in self.special_forms:
yield index, Keyword, value
continue
if value in self.macros:
yield index, Name.Builtin, value
continue
if value in self.lambda_list_keywords:
yield index, Keyword, value
continue
if value in self.declarations:
yield index, Keyword, value
continue
if value in self.builtin_types:
yield index, Keyword.Type, value
continue
if value in self.builtin_classes:
yield index, Name.Class, value
continue
yield index, token, value
开发者ID:Mekyi,项目名称:crunchy,代码行数:26,代码来源:functional.py
示例10: get_tokens_unprocessed
def get_tokens_unprocessed(self, text):
for index, token, value in (
RegexLexer.get_tokens_unprocessed(self, text)
):
if token is Name.Function and value in ('init', 'del'):
token = Keyword.Pseudo
yield index, token, value
开发者ID:DawidvC,项目名称:Cbang,代码行数:7,代码来源:pygments_cbang.py
示例11: get_tokens_unprocessed
def get_tokens_unprocessed(self, text):
# 'in' is either a keyword or an operator.
# If the token two tokens after 'in' is ')', 'in' is a keyword:
# objectloop(a in b)
# Otherwise, it is an operator:
# objectloop(a in b && true)
objectloop_queue = []
objectloop_token_count = -1
previous_token = None
for index, token, value in RegexLexer.get_tokens_unprocessed(self,
text):
if previous_token is Name.Variable and value == 'in':
objectloop_queue = [[index, token, value]]
objectloop_token_count = 2
elif objectloop_token_count > 0:
if token not in Comment and token not in Text:
objectloop_token_count -= 1
objectloop_queue.append((index, token, value))
else:
if objectloop_token_count == 0:
if objectloop_queue[-1][2] == ')':
objectloop_queue[0][1] = Keyword
while objectloop_queue:
yield objectloop_queue.pop(0)
objectloop_token_count = -1
yield index, token, value
if token not in Comment and token not in Text:
previous_token = token
while objectloop_queue:
yield objectloop_queue.pop(0)
开发者ID:spencerlyon2,项目名称:pygments,代码行数:30,代码来源:int_fiction.py
示例12: get_tokens_unprocessed
def get_tokens_unprocessed(self, text):
for index, token, value in RegexLexer.get_tokens_unprocessed(self, text):
# Convention: mark all upper case names as constants
if token is Name:
if value.isupper():
token = Name.Constant
yield index, token, value
开发者ID:testmana2,项目名称:eric,代码行数:7,代码来源:hdl.py
示例13: get_tokens_unprocessed
def get_tokens_unprocessed(self, text):
for index, token, value in RegexLexer.get_tokens_unprocessed(self, text):
if token is Name:
if value in builtins.BUILTIN_FUNCTIONS:
token = Name.Builtin
elif value in builtins.BUILTIN_OPERATIONS:
token = Name.Builtin.Pseudo
yield index, token, value
开发者ID:yamad,项目名称:igor-pygments-lexer,代码行数:8,代码来源:igorlexer.py
示例14: get_tokens_unprocessed
def get_tokens_unprocessed(self, text):
stack = ['root']
for index, token, value in \
RegexLexer.get_tokens_unprocessed(self, text, stack):
if token is Name and value in self.EXTRA_KEYWORDS:
yield index, Name.Builtin, value
else:
yield index, token, value
开发者ID:spencerlyon2,项目名称:pygments,代码行数:8,代码来源:haskell.py
示例15: get_tokens_unprocessed
def get_tokens_unprocessed(self, text, stack=('root', )):
ma = MathematicaAnnotations()
annotations = (ma.builtins, ma.unicode, ma.lexical_scope)
for index, token, value in RegexLexer.get_tokens_unprocessed(self, text):
result = (index, token, value)
for func in annotations:
result = func(*result)
yield result
开发者ID:andandandand,项目名称:pygments-mathematica,代码行数:9,代码来源:lexer.py
示例16: get_tokens_unprocessed
def get_tokens_unprocessed(self, text):
from pygments.lexers._asy_builtins import ASYFUNCNAME, ASYVARNAME
for index, token, value in \
RegexLexer.get_tokens_unprocessed(self, text):
if token is Name and value in ASYFUNCNAME:
token = Name.Function
elif token is Name and value in ASYVARNAME:
token = Name.Variable
yield index, token, value
开发者ID:axil,项目名称:blog,代码行数:9,代码来源:graphics.py
示例17: get_tokens_unprocessed
def get_tokens_unprocessed(self,text):
for index,token,value in RegexLexer.get_tokens_unprocessed(self,text):
if token == Generic:
value = "\\(" + value + "\\)"
yield (index,token,value)
elif token == XMath:
yield (index,Generic,value)
else:
yield (index,token,value)
开发者ID:felix-lang,项目名称:felix,代码行数:9,代码来源:xfelix.py
示例18: __init__
def __init__(self, **options):
self.funcnamehighlighting = get_bool_opt(options, "funcnamehighlighting", True)
self.disabledmodules = get_list_opt(options, "disabledmodules", ["unknown"])
self.startinline = get_bool_opt(options, "startinline", False)
# private option argument for the lexer itself
if "_startinline" in options:
self.startinline = options.pop("_startinline")
# collect activated functions in a set
self._functions = set()
if self.funcnamehighlighting:
from pygments.lexers._phpbuiltins import MODULES
for key, value in MODULES.iteritems():
if key not in self.disabledmodules:
self._functions.update(value)
RegexLexer.__init__(self, **options)
开发者ID:dlutcat,项目名称:pastebin,代码行数:18,代码来源:web.py
示例19: get_tokens_unprocessed
def get_tokens_unprocessed(self, text):
from pygments.lexers._cocoa_builtins import COCOA_INTERFACES, COCOA_PROTOCOLS, COCOA_PRIMITIVES
for index, token, value in RegexLexer.get_tokens_unprocessed(self, text):
if token is Name or token is Name.Class:
if value in COCOA_INTERFACES or value in COCOA_PROTOCOLS or value in COCOA_PRIMITIVES:
token = Name.Builtin.Pseudo
yield index, token, value
开发者ID:AlexStef,项目名称:stef-sublime-conf,代码行数:9,代码来源:objective.py
示例20: __init__
def __init__(self, **options):
self.funcnamehighlighting = get_bool_opt(
options, 'funcnamehighlighting', True)
self.disabledmodules = get_list_opt(
options, 'disabledmodules', ['unknown'])
self.startinline = get_bool_opt(options, 'startinline', False)
# private option argument for the lexer itself
if '_startinline' in options:
self.startinline = options.pop('_startinline')
# collect activated functions in a set
self._functions = set()
if self.funcnamehighlighting:
from pygments.lexers._php_builtins import MODULES
for key, value in iteritems(MODULES):
if key not in self.disabledmodules:
self._functions.update(value)
RegexLexer.__init__(self, **options)
开发者ID:axil,项目名称:blog,代码行数:19,代码来源:php.py
注:本文中的pygments.lexer.RegexLexer类示例由纯净天空整理自Github/MSDocs等源码及文档管理平台,相关代码片段筛选自各路编程大神贡献的开源项目,源码版权归原作者所有,传播和使用请参考对应项目的License;未经允许,请勿转载。 |
请发表评论