Mock Version: 2.6 Mock Version: 2.6 Mock Version: 2.6 ENTER ['do_with_status'](['bash', '--login', '-c', '/usr/bin/rpmbuild -bs --target noarch --nodeps /builddir/build/SPECS/python-jsonpath-ng.spec'], chrootPath='/var/lib/mock/f34-build-23952183-2401411/root'env={'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'}shell=Falselogger=timeout=201600uid=1000gid=425user='mockbuild'nspawn_args=[]unshare_net=TrueprintOutput=False) Executing command: ['bash', '--login', '-c', '/usr/bin/rpmbuild -bs --target noarch --nodeps /builddir/build/SPECS/python-jsonpath-ng.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'} and shell False Building target platforms: noarch Building for target noarch setting SOURCE_DATE_EPOCH=1599436800 Wrote: /builddir/build/SRPMS/python-jsonpath-ng-1.5.1-2.fc34.src.rpm Child return code was: 0 ENTER ['do_with_status'](['bash', '--login', '-c', '/usr/bin/rpmbuild -bb --target noarch --nodeps /builddir/build/SPECS/python-jsonpath-ng.spec'], chrootPath='/var/lib/mock/f34-build-23952183-2401411/root'env={'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'}shell=Falselogger=timeout=201600uid=1000gid=425user='mockbuild'nspawn_args=[]unshare_net=TrueprintOutput=False) Executing command: ['bash', '--login', '-c', '/usr/bin/rpmbuild -bb --target noarch --nodeps /builddir/build/SPECS/python-jsonpath-ng.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'} and shell False Building target platforms: noarch Building for target noarch setting SOURCE_DATE_EPOCH=1599436800 Executing(%prep): /bin/sh -e /var/tmp/rpm-tmp.D937q9 + umask 022 + cd /builddir/build/BUILD + cd /builddir/build/BUILD + rm -rf jsonpath-ng-1.5.1 + /usr/bin/gzip -dc /builddir/build/SOURCES/jsonpath-ng-1.5.1.tar.gz + /usr/bin/tar -xof - + STATUS=0 + '[' 0 -ne 0 ']' + cd jsonpath-ng-1.5.1 + /usr/bin/chmod -Rf a+rX,u+w,g-w,o-w . + sed -i -e '/^#!\//, 1d' jsonpath_ng/bin/jsonpath.py + rm -rf jsonpath-ng.egg-info + RPM_EC=0 ++ jobs -p + exit 0 Executing(%build): /bin/sh -e /var/tmp/rpm-tmp.Yndhh9 + umask 022 + cd /builddir/build/BUILD + cd jsonpath-ng-1.5.1 + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection' + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld ' + /usr/bin/python3 setup.py build '--executable=/usr/bin/python3 -s' running build running build_py creating build creating build/lib creating build/lib/jsonpath_ng copying jsonpath_ng/parser.py -> build/lib/jsonpath_ng copying jsonpath_ng/jsonpath.py -> build/lib/jsonpath_ng copying jsonpath_ng/lexer.py -> build/lib/jsonpath_ng copying jsonpath_ng/__init__.py -> build/lib/jsonpath_ng creating build/lib/jsonpath_ng/bin copying jsonpath_ng/bin/jsonpath.py -> build/lib/jsonpath_ng/bin copying jsonpath_ng/bin/__init__.py -> build/lib/jsonpath_ng/bin creating build/lib/jsonpath_ng/ext copying jsonpath_ng/ext/parser.py -> build/lib/jsonpath_ng/ext copying jsonpath_ng/ext/filter.py -> build/lib/jsonpath_ng/ext copying jsonpath_ng/ext/iterable.py -> build/lib/jsonpath_ng/ext copying jsonpath_ng/ext/string.py -> build/lib/jsonpath_ng/ext copying jsonpath_ng/ext/arithmetic.py -> build/lib/jsonpath_ng/ext copying jsonpath_ng/ext/__init__.py -> build/lib/jsonpath_ng/ext + RPM_EC=0 ++ jobs -p + exit 0 Executing(%install): /bin/sh -e /var/tmp/rpm-tmp.I18ux6 + umask 022 + cd /builddir/build/BUILD + '[' /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch '!=' / ']' + rm -rf /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch ++ dirname /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch + mkdir -p /builddir/build/BUILDROOT + mkdir /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch + cd jsonpath-ng-1.5.1 + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection' + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld ' + /usr/bin/python3 setup.py install -O1 --skip-build --root /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch running install running install_lib creating /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr creating /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/lib creating /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/lib/python3.9 creating /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/lib/python3.9/site-packages creating /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/lib/python3.9/site-packages/jsonpath_ng copying build/lib/jsonpath_ng/parser.py -> /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/lib/python3.9/site-packages/jsonpath_ng copying build/lib/jsonpath_ng/jsonpath.py -> /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/lib/python3.9/site-packages/jsonpath_ng creating /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/lib/python3.9/site-packages/jsonpath_ng/bin copying build/lib/jsonpath_ng/bin/jsonpath.py -> /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/lib/python3.9/site-packages/jsonpath_ng/bin copying build/lib/jsonpath_ng/bin/__init__.py -> /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/lib/python3.9/site-packages/jsonpath_ng/bin creating /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/lib/python3.9/site-packages/jsonpath_ng/ext copying build/lib/jsonpath_ng/ext/parser.py -> /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/lib/python3.9/site-packages/jsonpath_ng/ext copying build/lib/jsonpath_ng/ext/filter.py -> /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/lib/python3.9/site-packages/jsonpath_ng/ext copying build/lib/jsonpath_ng/ext/iterable.py -> /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/lib/python3.9/site-packages/jsonpath_ng/ext copying build/lib/jsonpath_ng/ext/string.py -> /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/lib/python3.9/site-packages/jsonpath_ng/ext copying build/lib/jsonpath_ng/ext/arithmetic.py -> /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/lib/python3.9/site-packages/jsonpath_ng/ext copying build/lib/jsonpath_ng/ext/__init__.py -> /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/lib/python3.9/site-packages/jsonpath_ng/ext copying build/lib/jsonpath_ng/lexer.py -> /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/lib/python3.9/site-packages/jsonpath_ng copying build/lib/jsonpath_ng/__init__.py -> /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/lib/python3.9/site-packages/jsonpath_ng byte-compiling /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/lib/python3.9/site-packages/jsonpath_ng/parser.py to parser.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/lib/python3.9/site-packages/jsonpath_ng/jsonpath.py to jsonpath.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/lib/python3.9/site-packages/jsonpath_ng/bin/jsonpath.py to jsonpath.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/lib/python3.9/site-packages/jsonpath_ng/bin/__init__.py to __init__.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/lib/python3.9/site-packages/jsonpath_ng/ext/parser.py to parser.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/lib/python3.9/site-packages/jsonpath_ng/ext/filter.py to filter.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/lib/python3.9/site-packages/jsonpath_ng/ext/iterable.py to iterable.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/lib/python3.9/site-packages/jsonpath_ng/ext/string.py to string.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/lib/python3.9/site-packages/jsonpath_ng/ext/arithmetic.py to arithmetic.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/lib/python3.9/site-packages/jsonpath_ng/ext/__init__.py to __init__.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/lib/python3.9/site-packages/jsonpath_ng/lexer.py to lexer.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/lib/python3.9/site-packages/jsonpath_ng/__init__.py to __init__.cpython-39.pyc writing byte-compilation script '/tmp/tmpncp_hhuo.py' /usr/bin/python3 /tmp/tmpncp_hhuo.py removing /tmp/tmpncp_hhuo.py running install_egg_info running egg_info writing jsonpath_ng.egg-info/PKG-INFO writing dependency_links to jsonpath_ng.egg-info/dependency_links.txt writing entry points to jsonpath_ng.egg-info/entry_points.txt writing requirements to jsonpath_ng.egg-info/requires.txt writing top-level names to jsonpath_ng.egg-info/top_level.txt reading manifest file 'jsonpath_ng.egg-info/SOURCES.txt' reading manifest template 'MANIFEST.in' writing manifest file 'jsonpath_ng.egg-info/SOURCES.txt' Copying jsonpath_ng.egg-info to /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/lib/python3.9/site-packages/jsonpath_ng-1.5.1-py3.9.egg-info running install_scripts Installing jsonpath_ng script to /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/bin + rm -rfv /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/bin/__pycache__ + /usr/lib/rpm/find-debuginfo.sh -j5 --strict-build-id -m -i --build-id-seed 1.5.1-2.fc34 --unique-debug-suffix -1.5.1-2.fc34.noarch --unique-debug-src-base python-jsonpath-ng-1.5.1-2.fc34.noarch --run-dwz --dwz-low-mem-die-limit 10000000 --dwz-max-die-limit 50000000 -S debugsourcefiles.list /builddir/build/BUILD/jsonpath-ng-1.5.1 find: 'debug': No such file or directory + /usr/lib/rpm/check-buildroot + /usr/lib/rpm/redhat/brp-ldconfig + /usr/lib/rpm/brp-compress + /usr/lib/rpm/redhat/brp-strip-lto /usr/bin/strip + /usr/lib/rpm/brp-strip-static-archive /usr/bin/strip + /usr/lib/rpm/redhat/brp-python-bytecompile '' 1 0 Bytecompiling .py files below /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/lib/python3.9 using /usr/bin/python3.9 + /usr/lib/rpm/brp-python-hardlink + /usr/lib/rpm/redhat/brp-mangle-shebangs Executing(%check): /bin/sh -e /var/tmp/rpm-tmp.yZXOD7 + umask 022 + cd /builddir/build/BUILD + cd jsonpath-ng-1.5.1 + /usr/bin/python3 setup.py test running test WARNING: Testing via this command is deprecated and will be removed in a future version. Users looking for a generic test entry point independent of test runner are encouraged to use tox. running egg_info writing jsonpath_ng.egg-info/PKG-INFO writing dependency_links to jsonpath_ng.egg-info/dependency_links.txt writing entry points to jsonpath_ng.egg-info/entry_points.txt writing requirements to jsonpath_ng.egg-info/requires.txt writing top-level names to jsonpath_ng.egg-info/top_level.txt reading manifest file 'jsonpath_ng.egg-info/SOURCES.txt' reading manifest template 'MANIFEST.in' writing manifest file 'jsonpath_ng.egg-info/SOURCES.txt' running build_ext test_basic_errors (tests.test_lexer.TestLexer) ... lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' ok test_simple_inputs (tests.test_lexer.TestLexer) ... lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' ok test_child_auto_id (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_child_auto_id ... ok test_child_paths (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_child_paths ... ok test_child_value (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_child_value ... ok test_descendants_auto_id (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_descendants_auto_id ... ok test_descendants_paths (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_descendants_paths ... ok test_descendants_value (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_descendants_value ... ok test_fields_auto_id (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_fields_auto_id ... ok test_fields_paths (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_fields_paths ... ok test_fields_value (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_fields_value ... ok test_hyphen_key (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_hyphen_key ... ok test_index_auto_id (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_index_auto_id ... ok test_index_paths (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_index_paths ... ok test_index_value (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_index_value ... ok test_parent_value (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_parent_value ... ok test_root_auto_id (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_root_auto_id ... ok test_root_paths (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_root_paths ... ok test_root_value (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_root_value ... ok test_slice_auto_id (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_slice_auto_id ... ok test_slice_paths (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_slice_paths ... ok test_slice_value (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_slice_value ... ok test_this_auto_id (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_this_auto_id ... ok test_this_paths (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_this_paths ... ok test_this_value (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_this_value ... ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sorted_list) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts /builddir/build/BUILD/jsonpath-ng-1.5.1/jsonpath_ng/parser.py:47: ResourceWarning: unclosed file <_io.TextIOWrapper name='/builddir/build/BUILD/jsonpath-ng-1.5.1/jsonpath_ng/parser.out' mode='w' encoding='UTF-8'> new_parser = ply.yacc.yacc(module=self, ResourceWarning: Enable tracemalloc to get the object allocation traceback ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sorted_list_indexed) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sorted_dict) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sorted_dict_indexed) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (len_list) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (len_dict) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (len_str) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (filter_exists_syntax1) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (filter_exists_syntax2) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (filter_exists_syntax3) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (filter_exists_syntax4) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (filter_eq1) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (filter_eq2) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (filter_eq3) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (filter_gt) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (filter_and) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (filter_float_gt) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sort1) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sort1_indexed) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sort2) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sort2_indexed) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sort3) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sort3_indexed) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sort4) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sort4_indexed) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sort5_twofields) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sort5_indexed) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_number_only) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_mul1) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_mul2) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_mul3) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_mul4) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_mul5) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_mul6) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_mul7) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_str0) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_str1) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_str2) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_str3) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_str4) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_list1) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_list2) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_list_err1) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_err1) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_err2) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (real_life_example1) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (real_life_example2) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (real_life_example3) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (real_life_example4) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sub1) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sub2) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (str1) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (split1) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (split2) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (bug-#2-correct) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (bug-#2-wrong) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (boolean-filter-true) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (boolean-filter-false) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (boolean-filter-other-datatypes-involved) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (boolean-filter-string-true-string-literal) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_filename_mode (tests.bin.test_jsonpath.TestJsonPathScript) ... ok test_stdin_mode (tests.bin.test_jsonpath.TestJsonPathScript) ... ok test_atomic (tests.test_parser.TestParser) ... ok test_nested (tests.test_parser.TestParser) ... ok test_DatumInContext_in_context (tests.test_jsonpath.TestDatumInContext) ... ok test_DatumInContext_init (tests.test_jsonpath.TestDatumInContext) ... ok test_child_auto_id (tests.test_jsonpath.TestJsonPath) ... ok test_child_paths (tests.test_jsonpath.TestJsonPath) ... ok test_child_value (tests.test_jsonpath.TestJsonPath) ... ok test_descendants_auto_id (tests.test_jsonpath.TestJsonPath) ... ok test_descendants_paths (tests.test_jsonpath.TestJsonPath) ... ok test_descendants_value (tests.test_jsonpath.TestJsonPath) ... ok test_fields_auto_id (tests.test_jsonpath.TestJsonPath) ... ok test_fields_paths (tests.test_jsonpath.TestJsonPath) ... ok test_fields_value (tests.test_jsonpath.TestJsonPath) ... ok test_hyphen_key (tests.test_jsonpath.TestJsonPath) ... ok test_index_auto_id (tests.test_jsonpath.TestJsonPath) ... ok test_index_paths (tests.test_jsonpath.TestJsonPath) ... ok test_index_value (tests.test_jsonpath.TestJsonPath) ... ok LexToken($,'$',1,0) LexToken($,'$',-1,-1) LexToken(ID,'hello',1,6) LexToken(ID,'hello',-1,-1) LexToken(ID,'goodbye',1,8) LexToken(ID,'goodbye',-1,-1) LexToken(ID,'doublequote"',1,13) LexToken(ID,'doublequote"',-1,-1) LexToken(ID,'doublequote"',1,14) LexToken(ID,'doublequote"',-1,-1) LexToken(ID,"singlequote'",1,14) LexToken(ID,"singlequote'",-1,-1) LexToken(ID,"singlequote'",1,13) LexToken(ID,"singlequote'",-1,-1) LexToken(ID,'fuzz',1,0) LexToken(ID,'fuzz',-1,-1) LexToken(NUMBER,1,1,0) LexToken(NUMBER,1,-1,-1) LexToken(NUMBER,45,1,0) LexToken(NUMBER,45,-1,-1) LexToken(NUMBER,-1,1,0) LexToken(NUMBER,-1,-1,-1) LexToken(NUMBER,-13,1,1) LexToken(NUMBER,-13,-1,-1) LexToken(ID,'fuzz.bang',1,10) LexToken(ID,'fuzz.bang',-1,-1) LexToken(ID,'fuzz',1,0) LexToken(ID,'fuzz',-1,-1) LexToken(.,'.',1,4) LexToken(.,'.',-1,-1) LexToken(ID,'bang',1,5) LexToken(ID,'bang',-1,-1) LexToken(ID,'fuzz',1,0) LexToken(ID,'fuzz',-1,-1) LexToken(.,'.',1,4) LexToken(.,'.',-1,-1) LexToken(*,'*',1,5) LexToken(*,'*',-1,-1) LexToken(ID,'fuzz',1,0) LexToken(ID,'fuzz',-1,-1) LexToken(DOUBLEDOT,'..',1,4) LexToken(DOUBLEDOT,'..',-1,-1) LexToken(ID,'bang',1,6) LexToken(ID,'bang',-1,-1) LexToken(&,'&',1,0) LexToken(&,'&',-1,-1) LexToken(ID,'@',1,0) LexToken(ID,'@',-1,-1) LexToken(NAMED_OPERATOR,'this',1,5) LexToken(NAMED_OPERATOR,'this',-1,-1) LexToken(|,'|',1,0) LexToken(|,'|',-1,-1) LexToken(WHERE,'where',1,0) LexToken(WHERE,'where',-1,-1) parse("foo.baz.id").find({'foo': {'baz': 3}}) =?= ['foo.baz'] parse("foo.baz.id").find({'foo': {'baz': [3]}}) =?= ['foo.baz'] parse("foo.baz.id").find({'foo': {'id': 'bizzle', 'baz': 3}}) =?= ['bizzle.baz'] parse("foo.baz.id").find({'foo': {'baz': {'id': 'hi'}}}) =?= ['foo.hi'] parse("foo.baz.bizzle.id").find({'foo': {'baz': {'bizzle': 5}}}) =?= ['foo.baz.bizzle'] parse("foo.baz").find({'foo': {'baz': 3}}).paths =?= ['foo.baz'] parse("foo.baz").find({'foo': {'baz': [3]}}).paths =?= ['foo.baz'] parse("foo.baz.bizzle").find({'foo': {'baz': {'bizzle': 5}}}).paths =?= ['foo.baz.bizzle'] parse("foo.baz").find({'foo': {'baz': 3}}) =?= [3] parse("foo.baz").find({'foo': {'baz': [3]}}) =?= [[3]] parse("foo.baz.bizzle").find({'foo': {'baz': {'bizzle': 5}}}) =?= [5] parse("foo..baz.id").find({'foo': {'baz': 1, 'bing': {'baz': 2}}}) =?= ['foo.baz', 'foo.bing.baz'] parse("foo..baz").find({'foo': {'baz': 1, 'bing': {'baz': 2}}}).paths =?= ['foo.baz', 'foo.bing.baz'] parse("foo..baz").find({'foo': {'baz': 1, 'bing': {'baz': 2}}}) =?= [1, 2] parse("foo..baz").find({'foo': [{'baz': 1}, {'baz': 2}]}) =?= [1, 2] parse("foo.id").find({'foo': 'baz'}) =?= ['foo'] parse("foo.id").find({'foo': {'id': 'baz'}}) =?= ['baz'] parse("foo,baz.id").find({'foo': 1, 'baz': 2}) =?= ['foo', 'baz'] parse("*.id").find({'foo': {'id': 1}, 'baz': 2}) =?= {'1', 'baz'} parse("foo").find({'foo': 'baz'}).paths =?= ['foo'] parse("foo,baz").find({'foo': 1, 'baz': 2}).paths =?= ['foo', 'baz'] parse("*").find({'foo': 1, 'baz': 2}).paths =?= {'foo', 'baz'} parse("*").find({'foo': 1, 'baz': 2}).paths =?= {'foo', 'id', 'baz'} parse("foo").find({'foo': 'baz'}) =?= ['baz'] parse("foo,baz").find({'foo': 1, 'baz': 2}) =?= [1, 2] parse("@foo").find({'@foo': 1}) =?= [1] parse("*").find({'foo': 1, 'baz': 2}) =?= {1, 2} parse("*").find({'foo': 1, 'baz': 2}) =?= {1, 2, '`this`'} parse("foo."bar-baz"").find({'foo': {'bar-baz': 3}}) =?= [3] parse("foo.["bar-baz","blah-blah"]").find({'foo': {'bar-baz': 3, 'blah-blah': 5}}) =?= [3, 5] parse("[0].id").find([42]) =?= ['[0]'] parse("[2].id").find([34, 65, 29, 59]) =?= ['[2]'] parse("[0]").find([42]).paths =?= ['[0]'] parse("[2]").find([34, 65, 29, 59]).paths =?= ['[2]'] parse("[0]").find([42]) =?= [42] parse("[5]").find([42]) =?= [] parse("[2]").find([34, 65, 29, 59]) =?= [29] parse("foo.baz.`parent`").find({'foo': {'baz': 3}}) =?= [{'baz': 3}] parse("foo.`parent`.foo.baz.`parent`.baz.bizzle").find({'foo': {'baz': {'bizzle': 5}}}) =?= [5] parse("$.id").find({'foo': 'baz'}) =?= ['$'] parse("foo.$.id").find({'foo': 'baz', 'id': 'bizzle'}) =?= ['bizzle'] parse("foo.$.baz.id").find({'foo': 4, 'baz': 3}) =?= ['baz'] parse("$").find({'foo': 'baz'}).paths =?= ['$'] parse("foo.$").find({'foo': 'baz'}).paths =?= ['$'] parse("foo.$.foo").find({'foo': 'baz'}).paths =?= ['foo'] parse("$").find({'foo': 'baz'}) =?= [{'foo': 'baz'}] parse("foo.$").find({'foo': 'baz'}) =?= [{'foo': 'baz'}] parse("foo.$.foo").find({'foo': 'baz'}) =?= ['baz'] parse("[*].id").find([1, 2, 3]) =?= ['[0]', '[1]', '[2]'] parse("[1:].id").find([1, 2, 3, 4]) =?= ['[1]', '[2]', '[3]'] parse("[*]").find([1, 2, 3]).paths =?= ['[0]', '[1]', '[2]'] parse("[1:]").find([1, 2, 3, 4]).paths =?= ['[1]', '[2]', '[3]'] parse("[*]").find([1, 2, 3]) =?= [1, 2, 3] parse("[*]").find(range(1, 4)) =?= [1, 2, 3] parse("[1:]").find([1, 2, 3, 4]) =?= [2, 3, 4] parse("[:2]").find([1, 2, 3, 4]) =?= [1, 2] parse("[*]").find(1) =?= [1] parse("[0:]").find(1) =?= [1] parse("[*]").find({'foo': 1}) =?= [{'foo': 1}] parse("[*].foo").find({'foo': 1}) =?= [1] parse("id").find({'foo': 'baz'}) =?= ['`this`'] parse("foo.`this`.id").find({'foo': 'baz'}) =?= ['foo'] parse("foo.`this`.baz.id").find({'foo': {'baz': 3}}) =?= ['foo.baz'] parse("`this`").find({'foo': 'baz'}).paths =?= ['`this`'] parse("foo.`this`").find({'foo': 'baz'}).paths =?= ['foo'] parse("foo.`this`.baz").find({'foo': {'baz': 3}}).paths =?= ['foo.baz'] parse("`this`").find({'foo': 'baz'}) =?= [{'foo': 'baz'}] parse("foo.`this`").find({'foo': 'baz'}) =?= ['baz'] parse("foo.`this`.baz").find({'foo': {'baz': 3}}) =?= [3] foo =?= foo * =?= * baz,bizzle =?= baz,bizzle [1] =?= [1] [1:] =?= [1] [:] =?= [*] [*] =?= [*] [:2] =?= [:2] [1:2] =?= [1:2] [5:-2] =?= [5:-2] foo.baz =?= foo.baz foo.baz,bizzle =?= foo.baz,bizzle foo where baz =?= foo where baz foo..baz =?= foo..baz foo..baz.bing =?= foo..baz.bing parse("foo.baz.id").find({'foo': {'baz': 3}}) =?= ['foo.baz'] parse("foo.baz.id").find({'foo': {'baz': [3]}}) =?= ['foo.baz'] parse("foo.baz.id").find({'foo': {'id': 'bizzle', 'baz': 3}}) =?= ['bizzle.baz'] parse("foo.baz.id").find({'foo': {'baz': {'id': 'hi'}}}) =?= ['foo.hi'] parse("foo.baz.bizzle.id").find({'foo': {'baz': {'bizzle': 5}}}) =?= ['foo.baz.bizzle'] parse("foo.baz").find({'foo': {'baz': 3}}).paths =?= ['foo.baz'] parse("foo.baz").find({'foo': {'baz': [3]}}).paths =?= ['foo.baz'] parse("foo.baz.bizzle").find({'foo': {'baz': {'bizzle': 5}}}).paths =?= ['foo.baz.bizzle'] parse("foo.baz").find({'foo': {'baz': 3}}) =?= [3] parse("foo.baz").find({'foo': {'baz': [3]}}) =?= [[3]] parse("foo.baz.bizzle").find({'foo': {'baz': {'bizzle': 5}}}) =?= [5] parse("foo..baz.id").find({'foo': {'baz': 1, 'bing': {'baz': 2}}}) =?= ['foo.baz', 'foo.bing.baz'] parse("foo..baz").find({'foo': {'baz': 1, 'bing': {'baz': 2}}}).paths =?= ['foo.baz', 'foo.bing.baz'] parse("foo..baz").find({'foo': {'baz': 1, 'bing': {'baz': 2}}}) =?= [1, 2] parse("foo..baz").find({'foo': [{'baz': 1}, {'baz': 2}]}) =?= [1, 2] parse("foo.id").find({'foo': 'baz'}) =?= ['foo'] parse("foo.id").find({'foo': {'id': 'baz'}}) =?= ['baz'] parse("foo,baz.id").find({'foo': 1, 'baz': 2}) =?= ['foo', 'baz'] parse("*.id").find({'foo': {'id': 1}, 'baz': 2}) =?= {'1', 'baz'} parse("foo").find({'foo': 'baz'}).paths =?= ['foo'] parse("foo,baz").find({'foo': 1, 'baz': 2}).paths =?= ['foo', 'baz'] parse("*").find({'foo': 1, 'baz': 2}).paths =?= {'foo', 'baz'} parse("*").find({'foo': 1, 'baz': 2}).paths =?= {'foo', 'id', 'baz'} parse("foo").find({'foo': 'baz'}) =?= ['baz'] parse("foo,baz").find({'foo': 1, 'baz': 2}) =?= [1, 2] parse("@foo").find({'@foo': 1}) =?= [1] parse("*").find({'foo': 1, 'baz': 2}) =?= {1, 2} parse("*").find({'foo': 1, 'baz': 2}) =?= {1, 2, '`this`'} parse("foo.bar-baz").find({'foo': {'bar-baz': 3}}) =?= [3] parse("foo.[bar-baz,blah-blah]").find({'foo': {'bar-baz': 3, 'blah-blah': 5}}) =?= [3, 5] parse("foo.-baz").find({'foo': {'-baz': 8}}) =?= [8] parse("[0].id").find([42]) =?= ['[0]'] parse("[2].id").find([34, 65, 29, 59]) =?= ['[2]'] parse("[0]").find([42]).paths =?= ['[0]'] parse("[2]").find([34, 65, 29, 59]).paths =?= ['[2]'] parse("[0]").find([42]) =?= [42] parse("[5]").find([42]) =?= [] parse("[2]").find([34, 65, 29, 59]) =?= [29] parse("[0]").find(None) =?= [] parse("foo.baz.`parent`").find({'foo': {'baz': 3}}) =?= [{'baz': 3}]test_parent_value (tests.test_jsonpath.TestJsonPath) ... ok test_root_auto_id (tests.test_jsonpath.TestJsonPath) ... ok test_root_paths (tests.test_jsonpath.TestJsonPath) ... ok test_root_value (tests.test_jsonpath.TestJsonPath) ... ok test_slice_auto_id (tests.test_jsonpath.TestJsonPath) ... ok test_slice_paths (tests.test_jsonpath.TestJsonPath) ... ok test_slice_value (tests.test_jsonpath.TestJsonPath) ... ok test_this_auto_id (tests.test_jsonpath.TestJsonPath) ... ok test_this_paths (tests.test_jsonpath.TestJsonPath) ... ok test_this_value (tests.test_jsonpath.TestJsonPath) ... ok test_update_child (tests.test_jsonpath.TestJsonPath) ... ok test_update_descendants (tests.test_jsonpath.TestJsonPath) ... ok test_update_descendants_where (tests.test_jsonpath.TestJsonPath) ... ok test_update_fields (tests.test_jsonpath.TestJsonPath) ... ok test_update_index (tests.test_jsonpath.TestJsonPath) ... ok test_update_root (tests.test_jsonpath.TestJsonPath) ... ok test_update_slice (tests.test_jsonpath.TestJsonPath) ... ok test_update_this (tests.test_jsonpath.TestJsonPath) ... ok test_update_where (tests.test_jsonpath.TestJsonPath) ... ok ---------------------------------------------------------------------- Ran 123 tests in 8.805s OK parse("foo.`parent`.foo.baz.`parent`.baz.bizzle").find({'foo': {'baz': {'bizzle': 5}}}) =?= [5] parse("$.id").find({'foo': 'baz'}) =?= ['$'] parse("foo.$.id").find({'foo': 'baz', 'id': 'bizzle'}) =?= ['bizzle'] parse("foo.$.baz.id").find({'foo': 4, 'baz': 3}) =?= ['baz'] parse("$").find({'foo': 'baz'}).paths =?= ['$'] parse("foo.$").find({'foo': 'baz'}).paths =?= ['$'] parse("foo.$.foo").find({'foo': 'baz'}).paths =?= ['foo'] parse("$").find({'foo': 'baz'}) =?= [{'foo': 'baz'}] parse("foo.$").find({'foo': 'baz'}) =?= [{'foo': 'baz'}] parse("foo.$.foo").find({'foo': 'baz'}) =?= ['baz'] parse("[*].id").find([1, 2, 3]) =?= ['[0]', '[1]', '[2]'] parse("[1:].id").find([1, 2, 3, 4]) =?= ['[1]', '[2]', '[3]'] parse("[*]").find([1, 2, 3]).paths =?= ['[0]', '[1]', '[2]'] parse("[1:]").find([1, 2, 3, 4]).paths =?= ['[1]', '[2]', '[3]'] parse("[*]").find([1, 2, 3]) =?= [1, 2, 3] parse("[*]").find(range(1, 4)) =?= [1, 2, 3] parse("[1:]").find([1, 2, 3, 4]) =?= [2, 3, 4] parse("[:2]").find([1, 2, 3, 4]) =?= [1, 2] parse("[*]").find(1) =?= [1] parse("[0:]").find(1) =?= [1] parse("[*]").find({'foo': 1}) =?= [{'foo': 1}] parse("[*].foo").find({'foo': 1}) =?= [1] parse("id").find({'foo': 'baz'}) =?= ['`this`'] parse("foo.`this`.id").find({'foo': 'baz'}) =?= ['foo'] parse("foo.`this`.baz.id").find({'foo': {'baz': 3}}) =?= ['foo.baz'] parse("`this`").find({'foo': 'baz'}).paths =?= ['`this`'] parse("foo.`this`").find({'foo': 'baz'}).paths =?= ['foo'] parse("foo.`this`.baz").find({'foo': {'baz': 3}}).paths =?= ['foo.baz'] parse("`this`").find({'foo': 'baz'}) =?= [{'foo': 'baz'}] parse("foo.`this`").find({'foo': 'baz'}) =?= ['baz'] parse("foo.`this`.baz").find({'foo': {'baz': 3}}) =?= [3] parse('$.foo').update({'foo': 'bar'}, 'baz') =?= {'foo': 'baz'} parse('foo.bar').update({'foo': {'bar': 1}}, 'baz') =?= {'foo': {'bar': 'baz'}} parse('$..somefield').update({'somefield': 1}, 42) =?= {'somefield': 42} parse('$..nestedfield').update({'outer': {'nestedfield': 1}}, 42) =?= {'outer': {'nestedfield': 42}} parse('$..bar').update({'outs': {'bar': 1, 'ins': {'bar': 9}}, 'outs2': {'bar': 2}}, 42) =?= {'outs': {'bar': 42, 'ins': {'bar': 42}}, 'outs2': {'bar': 42}} parse('(* where flag) .. bar').update({'foo': {'bar': 1, 'flag': 1}, 'baz': {'bar': 2}}, 3) =?= {'foo': {'bar': 3, 'flag': 1}, 'baz': {'bar': 2}} parse('foo').update({'foo': 1}, 5) =?= {'foo': 5} parse('$.*').update({'foo': 1, 'bar': 2}, 3) =?= {'foo': 3, 'bar': 3} parse('[0]').update(['foo', 'bar', 'baz'], 'test') =?= ['test', 'bar', 'baz'] parse('$').update('foo', 'bar') =?= 'bar' parse('[0:2]').update(['foo', 'bar', 'baz'], 'test') =?= ['test', 'test', 'baz'] parse('`this`').update('foo', 'bar') =?= 'bar' parse('*.bar where baz').update({'foo': {'bar': {'baz': 1}}, 'bar': {'baz': 2}}, 5) =?= {'foo': {'bar': 5}, 'bar': {'baz': 2}} + RPM_EC=0 ++ jobs -p + exit 0 Processing files: python3-jsonpath-ng-1.5.1-2.fc34.noarch Executing(%doc): /bin/sh -e /var/tmp/rpm-tmp.C2BQe6 + umask 022 + cd /builddir/build/BUILD + cd jsonpath-ng-1.5.1 + DOCDIR=/builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/share/doc/python3-jsonpath-ng + export LC_ALL=C + LC_ALL=C + export DOCDIR + /usr/bin/mkdir -p /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/share/doc/python3-jsonpath-ng + cp -pr README.rst /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch/usr/share/doc/python3-jsonpath-ng + RPM_EC=0 ++ jobs -p + exit 0 Provides: python-jsonpath-ng = 1.5.1-2.fc34 python3-jsonpath-ng = 1.5.1-2.fc34 python3.9-jsonpath-ng = 1.5.1-2.fc34 python3.9dist(jsonpath-ng) = 1.5.1 python3dist(jsonpath-ng) = 1.5.1 Requires(rpmlib): rpmlib(CompressedFileNames) <= 3.0.4-1 rpmlib(FileDigests) <= 4.6.0-1 rpmlib(PartialHardlinkSets) <= 4.0.4-1 rpmlib(PayloadFilesHavePrefix) <= 4.0-1 Requires: /usr/bin/python3 python(abi) = 3.9 python3.9dist(decorator) python3.9dist(ply) python3.9dist(setuptools) python3.9dist(six) Obsoletes: python-jsonpath-ng < 1.5.1-2.fc34 Checking for unpackaged file(s): /usr/lib/rpm/check-files /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch Wrote: /builddir/build/RPMS/python3-jsonpath-ng-1.5.1-2.fc34.noarch.rpm Executing(%clean): /bin/sh -e /var/tmp/rpm-tmp.VeleL5 + umask 022 + cd /builddir/build/BUILD + cd jsonpath-ng-1.5.1 + /usr/bin/rm -rf /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-2.fc34.noarch + RPM_EC=0 ++ jobs -p + exit 0 Child return code was: 0