Mock Version: 5.6 Mock Version: 5.6 Mock Version: 5.6 ENTER ['do_with_status'](['bash', '--login', '-c', '/usr/bin/rpmbuild -bs --noclean --target noarch --nodeps /builddir/build/SPECS/python-jsonpath-ng.spec'], chrootPath='/var/lib/mock/f41-build-51914954-6216538/root'env={'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'}shell=Falselogger=timeout=201600uid=1000gid=425user='mockbuild'nspawn_args=['--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.sv_9o15w:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11']unshare_net=TrueprintOutput=False) Using nspawn with args ['--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.sv_9o15w:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11'] Executing command: ['/usr/bin/systemd-nspawn', '-q', '-M', 'ecf0c3772dc44aa2ae0d6f8fda3f12a9', '-D', '/var/lib/mock/f41-build-51914954-6216538/root', '-a', '-u', 'mockbuild', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.sv_9o15w:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11', '--console=pipe', '--setenv=TERM=vt100', '--setenv=SHELL=/bin/bash', '--setenv=HOME=/builddir', '--setenv=HOSTNAME=mock', '--setenv=PATH=/usr/bin:/bin:/usr/sbin:/sbin', '--setenv=PROMPT_COMMAND=printf "\\033]0;\\007"', '--setenv=PS1= \\s-\\v\\$ ', '--setenv=LANG=C.UTF-8', '--resolv-conf=off', 'bash', '--login', '-c', '/usr/bin/rpmbuild -bs --noclean --target noarch --nodeps /builddir/build/SPECS/python-jsonpath-ng.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8', 'SYSTEMD_NSPAWN_TMPFS_TMP': '0', 'SYSTEMD_SECCOMP': '0'} and shell False Building target platforms: noarch Building for target noarch setting SOURCE_DATE_EPOCH=1717804800 Wrote: /builddir/build/SRPMS/python-jsonpath-ng-1.5.1-14.fc41.src.rpm Child return code was: 0 ENTER ['do_with_status'](['bash', '--login', '-c', '/usr/bin/rpmbuild -bb --noclean --target noarch --nodeps /builddir/build/SPECS/python-jsonpath-ng.spec'], chrootPath='/var/lib/mock/f41-build-51914954-6216538/root'env={'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'}shell=Falselogger=timeout=201600uid=1000gid=425user='mockbuild'nspawn_args=['--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.sv_9o15w:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11']unshare_net=TrueprintOutput=False) Using nspawn with args ['--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.sv_9o15w:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11'] Executing command: ['/usr/bin/systemd-nspawn', '-q', '-M', '83237efcbfbf4141802d98032c915005', '-D', '/var/lib/mock/f41-build-51914954-6216538/root', '-a', '-u', 'mockbuild', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.sv_9o15w:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11', '--console=pipe', '--setenv=TERM=vt100', '--setenv=SHELL=/bin/bash', '--setenv=HOME=/builddir', '--setenv=HOSTNAME=mock', '--setenv=PATH=/usr/bin:/bin:/usr/sbin:/sbin', '--setenv=PROMPT_COMMAND=printf "\\033]0;\\007"', '--setenv=PS1= \\s-\\v\\$ ', '--setenv=LANG=C.UTF-8', '--resolv-conf=off', 'bash', '--login', '-c', '/usr/bin/rpmbuild -bb --noclean --target noarch --nodeps /builddir/build/SPECS/python-jsonpath-ng.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8', 'SYSTEMD_NSPAWN_TMPFS_TMP': '0', 'SYSTEMD_SECCOMP': '0'} and shell False Building target platforms: noarch Building for target noarch setting SOURCE_DATE_EPOCH=1717804800 Executing(%mkbuilddir): /bin/sh -e /var/tmp/rpm-tmp.NPRwYN + umask 022 + cd /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build + test -d /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build + /usr/bin/chmod -Rf a+rX,u+w,g-w,o-w /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build + /usr/bin/rm -rf /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build + /usr/bin/mkdir -p /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build + /usr/bin/mkdir -p /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/SPECPARTS + RPM_EC=0 ++ jobs -p + exit 0 Executing(%prep): /bin/sh -e /var/tmp/rpm-tmp.TGsAH3 + umask 022 + cd /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build + cd /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build + rm -rf jsonpath-ng-1.5.1 + /usr/lib/rpm/rpmuncompress -x /builddir/build/SOURCES/jsonpath-ng-1.5.1.tar.gz + STATUS=0 + '[' 0 -ne 0 ']' + cd jsonpath-ng-1.5.1 + /usr/bin/chmod -Rf a+rX,u+w,g-w,o-w . + sed -i -e '/^#!\//, 1d' jsonpath_ng/bin/jsonpath.py + rm -rf jsonpath-ng.egg-info + RPM_EC=0 ++ jobs -p + exit 0 Executing(%build): /bin/sh -e /var/tmp/rpm-tmp.A4UdGX + umask 022 + cd /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -mtls-dialect=gnu2 -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer ' + export CFLAGS + CXXFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -mtls-dialect=gnu2 -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer ' + export CXXFLAGS + FFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -mtls-dialect=gnu2 -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer -I/usr/lib/gfortran/modules ' + export FFLAGS + FCFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -mtls-dialect=gnu2 -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer -I/usr/lib/gfortran/modules ' + export FCFLAGS + VALAFLAGS=-g + export VALAFLAGS + RUSTFLAGS='-Copt-level=3 -Cdebuginfo=2 -Ccodegen-units=1 -Cstrip=none -Cforce-frame-pointers=yes --cap-lints=warn' + export RUSTFLAGS + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -Wl,--build-id=sha1 ' + export LDFLAGS + LT_SYS_LIBRARY_PATH=/usr/lib: + export LT_SYS_LIBRARY_PATH + CC=gcc + export CC + CXX=g++ + export CXX + cd jsonpath-ng-1.5.1 + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -mtls-dialect=gnu2 -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer ' + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -Wl,--build-id=sha1 ' + /usr/bin/python3 setup.py build '--executable=/usr/bin/python3 -sP' running build running build_py creating build creating build/lib creating build/lib/jsonpath_ng copying jsonpath_ng/__init__.py -> build/lib/jsonpath_ng copying jsonpath_ng/parser.py -> build/lib/jsonpath_ng copying jsonpath_ng/lexer.py -> build/lib/jsonpath_ng copying jsonpath_ng/jsonpath.py -> build/lib/jsonpath_ng creating build/lib/jsonpath_ng/bin copying jsonpath_ng/bin/__init__.py -> build/lib/jsonpath_ng/bin copying jsonpath_ng/bin/jsonpath.py -> build/lib/jsonpath_ng/bin creating build/lib/jsonpath_ng/ext copying jsonpath_ng/ext/arithmetic.py -> build/lib/jsonpath_ng/ext copying jsonpath_ng/ext/iterable.py -> build/lib/jsonpath_ng/ext copying jsonpath_ng/ext/__init__.py -> build/lib/jsonpath_ng/ext copying jsonpath_ng/ext/parser.py -> build/lib/jsonpath_ng/ext copying jsonpath_ng/ext/filter.py -> build/lib/jsonpath_ng/ext copying jsonpath_ng/ext/string.py -> build/lib/jsonpath_ng/ext + RPM_EC=0 ++ jobs -p + exit 0 Executing(%install): /bin/sh -e /var/tmp/rpm-tmp.Mu7RPX + umask 022 + cd /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build + '[' /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT '!=' / ']' + rm -rf /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT ++ dirname /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT + mkdir -p /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build + mkdir /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -mtls-dialect=gnu2 -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer ' + export CFLAGS + CXXFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -mtls-dialect=gnu2 -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer ' + export CXXFLAGS + FFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -mtls-dialect=gnu2 -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer -I/usr/lib/gfortran/modules ' + export FFLAGS + FCFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -mtls-dialect=gnu2 -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer -I/usr/lib/gfortran/modules ' + export FCFLAGS + VALAFLAGS=-g + export VALAFLAGS + RUSTFLAGS='-Copt-level=3 -Cdebuginfo=2 -Ccodegen-units=1 -Cstrip=none -Cforce-frame-pointers=yes --cap-lints=warn' + export RUSTFLAGS + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -Wl,--build-id=sha1 ' + export LDFLAGS + LT_SYS_LIBRARY_PATH=/usr/lib: + export LT_SYS_LIBRARY_PATH + CC=gcc + export CC + CXX=g++ + export CXX + cd jsonpath-ng-1.5.1 + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -mtls-dialect=gnu2 -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer ' + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -Wl,--build-id=sha1 ' + /usr/bin/python3 setup.py install -O1 --skip-build --root /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT --prefix /usr running install /usr/lib/python3.13/site-packages/setuptools/_distutils/cmd.py:66: SetuptoolsDeprecationWarning: setup.py install is deprecated. !! ******************************************************************************** Please avoid running ``setup.py`` directly. Instead, use pypa/build, pypa/installer or other standards-based tools. Follow the current Python packaging guidelines when building Python RPM packages. See https://blog.ganssle.io/articles/2021/10/setup-py-deprecated.html and https://docs.fedoraproject.org/en-US/packaging-guidelines/Python/ for details. ******************************************************************************** !! self.initialize_options() running install_lib creating /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr creating /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib creating /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13 creating /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages creating /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng copying build/lib/jsonpath_ng/__init__.py -> /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng copying build/lib/jsonpath_ng/parser.py -> /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng copying build/lib/jsonpath_ng/lexer.py -> /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng copying build/lib/jsonpath_ng/jsonpath.py -> /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng creating /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/bin copying build/lib/jsonpath_ng/bin/__init__.py -> /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/bin copying build/lib/jsonpath_ng/bin/jsonpath.py -> /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/bin creating /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/ext copying build/lib/jsonpath_ng/ext/arithmetic.py -> /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/ext copying build/lib/jsonpath_ng/ext/iterable.py -> /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/ext copying build/lib/jsonpath_ng/ext/__init__.py -> /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/ext copying build/lib/jsonpath_ng/ext/parser.py -> /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/ext copying build/lib/jsonpath_ng/ext/filter.py -> /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/ext copying build/lib/jsonpath_ng/ext/string.py -> /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/ext byte-compiling /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/__init__.py to __init__.cpython-313.pyc byte-compiling /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/parser.py to parser.cpython-313.pyc byte-compiling /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/lexer.py to lexer.cpython-313.pyc byte-compiling /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/jsonpath.py to jsonpath.cpython-313.pyc byte-compiling /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/bin/__init__.py to __init__.cpython-313.pyc byte-compiling /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/bin/jsonpath.py to jsonpath.cpython-313.pyc byte-compiling /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/ext/arithmetic.py to arithmetic.cpython-313.pyc byte-compiling /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/ext/iterable.py to iterable.cpython-313.pyc byte-compiling /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/ext/__init__.py to __init__.cpython-313.pyc byte-compiling /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/ext/parser.py to parser.cpython-313.pyc byte-compiling /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/ext/filter.py to filter.cpython-313.pyc byte-compiling /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/ext/string.py to string.cpython-313.pyc /usr/lib/python3.13/site-packages/jsonpath_ng/ext/string.py:18: SyntaxWarning: invalid escape sequence '\(' /usr/lib/python3.13/site-packages/jsonpath_ng/ext/string.py:19: SyntaxWarning: invalid escape sequence '\(' /usr/lib/python3.13/site-packages/jsonpath_ng/ext/string.py:20: SyntaxWarning: invalid escape sequence '\(' writing byte-compilation script '/tmp/tmpzjkk54au.py' /usr/bin/python3 /tmp/tmpzjkk54au.py /usr/lib/python3.13/site-packages/jsonpath_ng/ext/string.py:18: SyntaxWarning: invalid escape sequence '\(' /usr/lib/python3.13/site-packages/jsonpath_ng/ext/string.py:19: SyntaxWarning: invalid escape sequence '\(' /usr/lib/python3.13/site-packages/jsonpath_ng/ext/string.py:20: SyntaxWarning: invalid escape sequence '\(' removing /tmp/tmpzjkk54au.py running install_egg_info running egg_info writing jsonpath_ng.egg-info/PKG-INFO writing dependency_links to jsonpath_ng.egg-info/dependency_links.txt writing entry points to jsonpath_ng.egg-info/entry_points.txt writing requirements to jsonpath_ng.egg-info/requires.txt writing top-level names to jsonpath_ng.egg-info/top_level.txt reading manifest file 'jsonpath_ng.egg-info/SOURCES.txt' reading manifest template 'MANIFEST.in' writing manifest file 'jsonpath_ng.egg-info/SOURCES.txt' Copying jsonpath_ng.egg-info to /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng-1.5.1-py3.13.egg-info running install_scripts Installing jsonpath_ng script to /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/bin + rm -rfv /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/bin/__pycache__ + /usr/lib/rpm/check-buildroot + /usr/lib/rpm/redhat/brp-ldconfig + /usr/lib/rpm/brp-compress + /usr/lib/rpm/brp-strip /usr/bin/strip + /usr/lib/rpm/brp-strip-comment-note /usr/bin/strip /usr/bin/objdump + /usr/lib/rpm/redhat/brp-strip-lto /usr/bin/strip + /usr/lib/rpm/brp-strip-static-archive /usr/bin/strip + /usr/lib/rpm/check-rpaths + /usr/lib/rpm/redhat/brp-mangle-shebangs + /usr/lib/rpm/brp-remove-la-files + env /usr/lib/rpm/redhat/brp-python-bytecompile '' 1 0 -j48 Bytecompiling .py files below /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13 using python3.13 /usr/lib/python3.13/site-packages/jsonpath_ng/ext/string.py:18: SyntaxWarning: invalid escape sequence '\(' /usr/lib/python3.13/site-packages/jsonpath_ng/ext/string.py:19: SyntaxWarning: invalid escape sequence '\(' /usr/lib/python3.13/site-packages/jsonpath_ng/ext/string.py:20: SyntaxWarning: invalid escape sequence '\(' /usr/lib/python3.13/site-packages/jsonpath_ng/ext/string.py:18: SyntaxWarning: invalid escape sequence '\(' /usr/lib/python3.13/site-packages/jsonpath_ng/ext/string.py:19: SyntaxWarning: invalid escape sequence '\(' /usr/lib/python3.13/site-packages/jsonpath_ng/ext/string.py:20: SyntaxWarning: invalid escape sequence '\(' + /usr/lib/rpm/redhat/brp-python-hardlink + /usr/bin/add-determinism --brp -j48 /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/bin/__pycache__/__init__.cpython-313.pyc: rewriting with normalized contents /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/bin/__pycache__/jsonpath.cpython-313.pyc: rewriting with normalized contents /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/ext/__pycache__/arithmetic.cpython-313.pyc: rewriting with normalized contents /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/ext/__pycache__/__init__.cpython-313.pyc: rewriting with normalized contents /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/ext/__pycache__/filter.cpython-313.pyc: rewriting with normalized contents /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/ext/__pycache__/iterable.cpython-313.pyc: rewriting with normalized contents /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/__pycache__/__init__.cpython-313.pyc: rewriting with normalized contents /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/__pycache__/jsonpath.cpython-313.pyc: rewriting with normalized contents /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/__pycache__/lexer.cpython-313.pyc: rewriting with normalized contents /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/__pycache__/parser.cpython-313.pyc: rewriting with normalized contents /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/ext/__pycache__/string.cpython-313.pyc: rewriting with normalized contents /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/lib/python3.13/site-packages/jsonpath_ng/ext/__pycache__/parser.cpython-313.pyc: rewriting with normalized contents Scanned 13 directories and 43 files, processed 12 inodes, 12 modified (0 replaced + 12 rewritten), 0 unsupported format, 0 errors Executing(%check): /bin/sh -e /var/tmp/rpm-tmp.6elFV3 + umask 022 + cd /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -mtls-dialect=gnu2 -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer ' + export CFLAGS + CXXFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -mtls-dialect=gnu2 -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer ' + export CXXFLAGS + FFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -mtls-dialect=gnu2 -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer -I/usr/lib/gfortran/modules ' + export FFLAGS + FCFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -mtls-dialect=gnu2 -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer -I/usr/lib/gfortran/modules ' + export FCFLAGS + VALAFLAGS=-g + export VALAFLAGS + RUSTFLAGS='-Copt-level=3 -Cdebuginfo=2 -Ccodegen-units=1 -Cstrip=none -Cforce-frame-pointers=yes --cap-lints=warn' + export RUSTFLAGS + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -Wl,--build-id=sha1 ' + export LDFLAGS + LT_SYS_LIBRARY_PATH=/usr/lib: + export LT_SYS_LIBRARY_PATH + CC=gcc + export CC + CXX=g++ + export CXX + cd jsonpath-ng-1.5.1 + /usr/bin/python3 setup.py test running test WARNING: Testing via this command is deprecated and will be removed in a future version. Users looking for a generic test entry point independent of test runner are encouraged to use tox. /usr/lib/python3.13/site-packages/setuptools/command/test.py:193: _DeprecatedInstaller: setuptools.installer and fetch_build_eggs are deprecated. !! ******************************************************************************** Requirements should be satisfied by a PEP 517 installer. If you are using pip, you can try `pip install --use-pep517`. ******************************************************************************** !! ir_d = dist.fetch_build_eggs(dist.install_requires) WARNING: The wheel package is not available. /usr/lib/python3.13/site-packages/setuptools/command/test.py:194: _DeprecatedInstaller: setuptools.installer and fetch_build_eggs are deprecated. !! ******************************************************************************** Requirements should be satisfied by a PEP 517 installer. If you are using pip, you can try `pip install --use-pep517`. ******************************************************************************** !! tr_d = dist.fetch_build_eggs(dist.tests_require or []) WARNING: The wheel package is not available. /usr/lib/python3.13/site-packages/setuptools/command/test.py:195: _DeprecatedInstaller: setuptools.installer and fetch_build_eggs are deprecated. !! ******************************************************************************** Requirements should be satisfied by a PEP 517 installer. If you are using pip, you can try `pip install --use-pep517`. ******************************************************************************** !! er_d = dist.fetch_build_eggs( WARNING: The wheel package is not available. running egg_info writing jsonpath_ng.egg-info/PKG-INFO writing dependency_links to jsonpath_ng.egg-info/dependency_links.txt writing entry points to jsonpath_ng.egg-info/entry_points.txt writing requirements to jsonpath_ng.egg-info/requires.txt writing top-level names to jsonpath_ng.egg-info/top_level.txt reading manifest file 'jsonpath_ng.egg-info/SOURCES.txt' reading manifest template 'MANIFEST.in' writing manifest file 'jsonpath_ng.egg-info/SOURCES.txt' running build_ext /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/jsonpath-ng-1.5.1/tests/test_jsonpath_rw_ext.py:129: SyntaxWarning: invalid escape sequence '\c' ('sort2', dict(string='objects[\cat]', /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/jsonpath-ng-1.5.1/tests/test_jsonpath_rw_ext.py:132: SyntaxWarning: invalid escape sequence '\c' ('sort2_indexed', dict(string='objects[\cat][-1].cat', /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/jsonpath-ng-1.5.1/tests/test_jsonpath_rw_ext.py:136: SyntaxWarning: invalid escape sequence '\c' ('sort3', dict(string='objects[/cow,\cat]', /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/jsonpath-ng-1.5.1/tests/test_jsonpath_rw_ext.py:145: SyntaxWarning: invalid escape sequence '\c' ('sort3_indexed', dict(string='objects[/cow,\cat][0].cat', /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/jsonpath-ng-1.5.1/jsonpath_ng/ext/string.py:18: SyntaxWarning: invalid escape sequence '\(' SUB = re.compile("sub\(/(.*)/,\s+(.*)\)") /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/jsonpath-ng-1.5.1/jsonpath_ng/ext/string.py:19: SyntaxWarning: invalid escape sequence '\(' SPLIT = re.compile("split\((.),\s+(\d+),\s+(\d+|-1)\)") /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/jsonpath-ng-1.5.1/jsonpath_ng/ext/string.py:20: SyntaxWarning: invalid escape sequence '\(' STR = re.compile("str\(\)") test_atomic (tests.test_parser.TestParser.test_atomic) ... /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/jsonpath-ng-1.5.1/jsonpath_ng/parser.py:47: ResourceWarning: unclosed file <_io.TextIOWrapper name='/builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/jsonpath-ng-1.5.1/jsonpath_ng/parser.out' mode='w' encoding='UTF-8'> new_parser = ply.yacc.yacc(module=self, ResourceWarning: Enable tracemalloc to get the object allocation traceback ok test_nested (tests.test_parser.TestParser.test_nested) ... ok test_child_auto_id (tests.test_jsonpath_rw_ext.TestJsonPath.test_child_auto_id) tests.test_jsonpath_rw_ext.TestJsonPath.test_child_auto_id ... ok test_child_paths (tests.test_jsonpath_rw_ext.TestJsonPath.test_child_paths) tests.test_jsonpath_rw_ext.TestJsonPath.test_child_paths ... ok test_child_value (tests.test_jsonpath_rw_ext.TestJsonPath.test_child_value) tests.test_jsonpath_rw_ext.TestJsonPath.test_child_value ... ok test_descendants_auto_id (tests.test_jsonpath_rw_ext.TestJsonPath.test_descendants_auto_id) tests.test_jsonpath_rw_ext.TestJsonPath.test_descendants_auto_id ... ok test_descendants_paths (tests.test_jsonpath_rw_ext.TestJsonPath.test_descendants_paths) tests.test_jsonpath_rw_ext.TestJsonPath.test_descendants_paths ... ok test_descendants_value (tests.test_jsonpath_rw_ext.TestJsonPath.test_descendants_value) tests.test_jsonpath_rw_ext.TestJsonPath.test_descendants_value ... ok test_fields_auto_id (tests.test_jsonpath_rw_ext.TestJsonPath.test_fields_auto_id) tests.test_jsonpath_rw_ext.TestJsonPath.test_fields_auto_id ... ok test_fields_paths (tests.test_jsonpath_rw_ext.TestJsonPath.test_fields_paths) tests.test_jsonpath_rw_ext.TestJsonPath.test_fields_paths ... ok test_fields_value (tests.test_jsonpath_rw_ext.TestJsonPath.test_fields_value) tests.test_jsonpath_rw_ext.TestJsonPath.test_fields_value ... ok test_hyphen_key (tests.test_jsonpath_rw_ext.TestJsonPath.test_hyphen_key) tests.test_jsonpath_rw_ext.TestJsonPath.test_hyphen_key ... ok test_index_auto_id (tests.test_jsonpath_rw_ext.TestJsonPath.test_index_auto_id) tests.test_jsonpath_rw_ext.TestJsonPath.test_index_auto_id ... ok test_index_paths (tests.test_jsonpath_rw_ext.TestJsonPath.test_index_paths) tests.test_jsonpath_rw_ext.TestJsonPath.test_index_paths ... ok test_index_value (tests.test_jsonpath_rw_ext.TestJsonPath.test_index_value) tests.test_jsonpath_rw_ext.TestJsonPath.test_index_value ... ok test_parent_value (tests.test_jsonpath_rw_ext.TestJsonPath.test_parent_value) tests.test_jsonpath_rw_ext.TestJsonPath.test_parent_value ... ok test_root_auto_id (tests.test_jsonpath_rw_ext.TestJsonPath.test_root_auto_id) tests.test_jsonpath_rw_ext.TestJsonPath.test_root_auto_id ... ok test_root_paths (tests.test_jsonpath_rw_ext.TestJsonPath.test_root_paths) tests.test_jsonpath_rw_ext.TestJsonPath.test_root_paths ... ok test_root_value (tests.test_jsonpath_rw_ext.TestJsonPath.test_root_value) tests.test_jsonpath_rw_ext.TestJsonPath.test_root_value ... ok test_slice_auto_id (tests.test_jsonpath_rw_ext.TestJsonPath.test_slice_auto_id) tests.test_jsonpath_rw_ext.TestJsonPath.test_slice_auto_id ... ok test_slice_paths (tests.test_jsonpath_rw_ext.TestJsonPath.test_slice_paths) tests.test_jsonpath_rw_ext.TestJsonPath.test_slice_paths ... ok test_slice_value (tests.test_jsonpath_rw_ext.TestJsonPath.test_slice_value) tests.test_jsonpath_rw_ext.TestJsonPath.test_slice_value ... ok test_this_auto_id (tests.test_jsonpath_rw_ext.TestJsonPath.test_this_auto_id) tests.test_jsonpath_rw_ext.TestJsonPath.test_this_auto_id ... ok test_this_paths (tests.test_jsonpath_rw_ext.TestJsonPath.test_this_paths) tests.test_jsonpath_rw_ext.TestJsonPath.test_this_paths ... ok test_this_value (tests.test_jsonpath_rw_ext.TestJsonPath.test_this_value) tests.test_jsonpath_rw_ext.TestJsonPath.test_this_value ... ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sorted_list) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sorted_list_indexed) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sorted_dict) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sorted_dict_indexed) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (len_list) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (len_dict) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (len_str) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (filter_exists_syntax1) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (filter_exists_syntax2) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (filter_exists_syntax3) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (filter_exists_syntax4) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (filter_eq1) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (filter_eq2) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (filter_eq3) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (filter_gt) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (filter_and) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (filter_float_gt) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sort1) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sort1_indexed) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sort2) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sort2_indexed) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sort3) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sort3_indexed) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sort4) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sort4_indexed) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sort5_twofields) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sort5_indexed) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_number_only) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_mul1) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_mul2) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_mul3) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_mul4) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_mul5) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_mul6) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_mul7) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_str0) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_str1) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_str2) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_str3) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_str4) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_list1) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_list2) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_list_err1) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_err1) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_err2) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (real_life_example1) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (real_life_example2) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (real_life_example3) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (real_life_example4) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sub1) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sub2) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (str1) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (split1) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (split2) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (bug-#2-correct) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (bug-#2-wrong) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (boolean-filter-true) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (boolean-filter-false) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (boolean-filter-other-datatypes-involved) ... 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (boolean-filter-string-true-string-literal) ... 50 shift/reduce conflicts ok test_filename_mode (tests.bin.test_jsonpath.TestJsonPathScript.test_filename_mode) ... ok test_stdin_mode (tests.bin.test_jsonpath.TestJsonPathScript.test_stdin_mode) ... ok test_basic_errors (tests.test_lexer.TestLexer.test_basic_errors) ... lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' ok test_simple_inputs (tests.test_lexer.TestLexer.test_simple_inputs) ... lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' ok test_DatumInContext_in_context (tests.test_jsonpath.TestDatumInContext.test_DatumInContext_in_context) ... ok test_DatumInContext_init (tests.test_jsonpath.TestDatumInContext.test_DatumInContext_init) ... ok test_child_auto_id (tests.test_jsonpath.TestJsonPath.test_child_auto_id) ... ok test_child_paths (tests.test_jsonpath.TestJsonPath.test_child_paths) ... ok test_child_value (tests.test_jsonpath.TestJsonPath.test_child_value) ... ok test_descendants_auto_id (tests.test_jsonpath.TestJsonPath.test_descendants_auto_id) ... ok test_descendants_paths (tests.test_jsonpath.TestJsonPath.test_descendants_paths) ... ok test_descendants_value (tests.test_jsonpath.TestJsonPath.test_descendants_value) ... ok test_fields_auto_id (tests.test_jsonpath.TestJsonPath.test_fields_auto_id) ... ok test_fields_paths (tests.test_jsonpath.TestJsonPath.test_fields_paths) ... ok test_fields_value (tests.test_jsonpath.TestJsonPath.test_fields_value) ... ok test_hyphen_key (tests.test_jsonpath.TestJsonPath.test_hyphen_key) ... ok test_index_auto_id (tests.test_jsonpath.TestJsonPath.test_index_auto_id) ... ok test_index_paths (tests.test_jsonpath.TestJsonPath.test_index_paths) ... ok test_index_value (tests.test_jsonpath.TestJsonPath.test_index_value) ... ok foo =?= foo * =?= * baz,bizzle =?= baz,bizzle [1] =?= [1] [1:] =?= [1] [:] =?= [*] [*] =?= [*] [:2] =?= [:2] [1:2] =?= [1:2] [5:-2] =?= [5:-2] foo.baz =?= foo.baz foo.baz,bizzle =?= foo.baz,bizzle foo where baz =?= foo where baz foo..baz =?= foo..baz foo..baz.bing =?= foo..baz.bing parse("foo.baz.id").find({'foo': {'baz': 3}}) =?= ['foo.baz'] parse("foo.baz.id").find({'foo': {'baz': [3]}}) =?= ['foo.baz'] parse("foo.baz.id").find({'foo': {'id': 'bizzle', 'baz': 3}}) =?= ['bizzle.baz'] parse("foo.baz.id").find({'foo': {'baz': {'id': 'hi'}}}) =?= ['foo.hi'] parse("foo.baz.bizzle.id").find({'foo': {'baz': {'bizzle': 5}}}) =?= ['foo.baz.bizzle'] parse("foo.baz").find({'foo': {'baz': 3}}).paths =?= ['foo.baz'] parse("foo.baz").find({'foo': {'baz': [3]}}).paths =?= ['foo.baz'] parse("foo.baz.bizzle").find({'foo': {'baz': {'bizzle': 5}}}).paths =?= ['foo.baz.bizzle'] parse("foo.baz").find({'foo': {'baz': 3}}) =?= [3] parse("foo.baz").find({'foo': {'baz': [3]}}) =?= [[3]] parse("foo.baz.bizzle").find({'foo': {'baz': {'bizzle': 5}}}) =?= [5] parse("foo..baz.id").find({'foo': {'baz': 1, 'bing': {'baz': 2}}}) =?= ['foo.baz', 'foo.bing.baz'] parse("foo..baz").find({'foo': {'baz': 1, 'bing': {'baz': 2}}}).paths =?= ['foo.baz', 'foo.bing.baz'] parse("foo..baz").find({'foo': {'baz': 1, 'bing': {'baz': 2}}}) =?= [1, 2] parse("foo..baz").find({'foo': [{'baz': 1}, {'baz': 2}]}) =?= [1, 2] parse("foo.id").find({'foo': 'baz'}) =?= ['foo'] parse("foo.id").find({'foo': {'id': 'baz'}}) =?= ['baz'] parse("foo,baz.id").find({'foo': 1, 'baz': 2}) =?= ['foo', 'baz'] parse("*.id").find({'foo': {'id': 1}, 'baz': 2}) =?= {'1', 'baz'} parse("foo").find({'foo': 'baz'}).paths =?= ['foo'] parse("foo,baz").find({'foo': 1, 'baz': 2}).paths =?= ['foo', 'baz'] parse("*").find({'foo': 1, 'baz': 2}).paths =?= {'foo', 'baz'} parse("*").find({'foo': 1, 'baz': 2}).paths =?= {'id', 'foo', 'baz'} parse("foo").find({'foo': 'baz'}) =?= ['baz'] parse("foo,baz").find({'foo': 1, 'baz': 2}) =?= [1, 2] parse("@foo").find({'@foo': 1}) =?= [1] parse("*").find({'foo': 1, 'baz': 2}) =?= {1, 2} parse("*").find({'foo': 1, 'baz': 2}) =?= {1, 2, '`this`'} parse("foo."bar-baz"").find({'foo': {'bar-baz': 3}}) =?= [3] parse("foo.["bar-baz","blah-blah"]").find({'foo': {'bar-baz': 3, 'blah-blah': 5}}) =?= [3, 5] parse("[0].id").find([42]) =?= ['[0]'] parse("[2].id").find([34, 65, 29, 59]) =?= ['[2]'] parse("[0]").find([42]).paths =?= ['[0]'] parse("[2]").find([34, 65, 29, 59]).paths =?= ['[2]'] parse("[0]").find([42]) =?= [42] parse("[5]").find([42]) =?= [] parse("[2]").find([34, 65, 29, 59]) =?= [29] parse("foo.baz.`parent`").find({'foo': {'baz': 3}}) =?= [{'baz': 3}] parse("foo.`parent`.foo.baz.`parent`.baz.bizzle").find({'foo': {'baz': {'bizzle': 5}}}) =?= [5] parse("$.id").find({'foo': 'baz'}) =?= ['$'] parse("foo.$.id").find({'foo': 'baz', 'id': 'bizzle'}) =?= ['bizzle'] parse("foo.$.baz.id").find({'foo': 4, 'baz': 3}) =?= ['baz'] parse("$").find({'foo': 'baz'}).paths =?= ['$'] parse("foo.$").find({'foo': 'baz'}).paths =?= ['$'] parse("foo.$.foo").find({'foo': 'baz'}).paths =?= ['foo'] parse("$").find({'foo': 'baz'}) =?= [{'foo': 'baz'}] parse("foo.$").find({'foo': 'baz'}) =?= [{'foo': 'baz'}] parse("foo.$.foo").find({'foo': 'baz'}) =?= ['baz'] parse("[*].id").find([1, 2, 3]) =?= ['[0]', '[1]', '[2]'] parse("[1:].id").find([1, 2, 3, 4]) =?= ['[1]', '[2]', '[3]'] parse("[*]").find([1, 2, 3]).paths =?= ['[0]', '[1]', '[2]'] parse("[1:]").find([1, 2, 3, 4]).paths =?= ['[1]', '[2]', '[3]'] parse("[*]").find([1, 2, 3]) =?= [1, 2, 3] parse("[*]").find(range(1, 4)) =?= [1, 2, 3] parse("[1:]").find([1, 2, 3, 4]) =?= [2, 3, 4] parse("[:2]").find([1, 2, 3, 4]) =?= [1, 2] parse("[*]").find(1) =?= [1] parse("[0:]").find(1) =?= [1] parse("[*]").find({'foo': 1}) =?= [{'foo': 1}] parse("[*].foo").find({'foo': 1}) =?= [1] parse("id").find({'foo': 'baz'}) =?= ['`this`'] parse("foo.`this`.id").find({'foo': 'baz'}) =?= ['foo'] parse("foo.`this`.baz.id").find({'foo': {'baz': 3}}) =?= ['foo.baz'] parse("`this`").find({'foo': 'baz'}).paths =?= ['`this`'] parse("foo.`this`").find({'foo': 'baz'}).paths =?= ['foo'] parse("foo.`this`.baz").find({'foo': {'baz': 3}}).paths =?= ['foo.baz'] parse("`this`").find({'foo': 'baz'}) =?= [{'foo': 'baz'}] parse("foo.`this`").find({'foo': 'baz'}) =?= ['baz'] parse("foo.`this`.baz").find({'foo': {'baz': 3}}) =?= [3] LexToken($,'$',1,0) LexToken($,'$',-1,-1) LexToken(ID,'hello',1,6) LexToken(ID,'hello',-1,-1) LexToken(ID,'goodbye',1,8) LexToken(ID,'goodbye',-1,-1) LexToken(ID,'doublequote"',1,13) LexToken(ID,'doublequote"',-1,-1) LexToken(ID,'doublequote"',1,14) LexToken(ID,'doublequote"',-1,-1) LexToken(ID,"singlequote'",1,14) LexToken(ID,"singlequote'",-1,-1) LexToken(ID,"singlequote'",1,13) LexToken(ID,"singlequote'",-1,-1) LexToken(ID,'fuzz',1,0) LexToken(ID,'fuzz',-1,-1) LexToken(NUMBER,1,1,0) LexToken(NUMBER,1,-1,-1) LexToken(NUMBER,45,1,0) LexToken(NUMBER,45,-1,-1) LexToken(NUMBER,-1,1,0) LexToken(NUMBER,-1,-1,-1) LexToken(NUMBER,-13,1,1) LexToken(NUMBER,-13,-1,-1) LexToken(ID,'fuzz.bang',1,10) LexToken(ID,'fuzz.bang',-1,-1) LexToken(ID,'fuzz',1,0) LexToken(ID,'fuzz',-1,-1) LexToken(.,'.',1,4) LexToken(.,'.',-1,-1) LexToken(ID,'bang',1,5) LexToken(ID,'bang',-1,-1) LexToken(ID,'fuzz',1,0) LexToken(ID,'fuzz',-1,-1) LexToken(.,'.',1,4) LexToken(.,'.',-1,-1) LexToken(*,'*',1,5) LexToken(*,'*',-1,-1) LexToken(ID,'fuzz',1,0) LexToken(ID,'fuzz',-1,-1) LexToken(DOUBLEDOT,'..',1,4) LexToken(DOUBLEDOT,'..',-1,-1) LexToken(ID,'bang',1,6) LexToken(ID,'bang',-1,-1) LexToken(&,'&',1,0) LexToken(&,'&',-1,-1) LexToken(ID,'@',1,0) LexToken(ID,'@',-1,-1) LexToken(NAMED_OPERATOR,'this',1,5) LexToken(NAMED_OPERATOR,'this',-1,-1) LexToken(|,'|',1,0) LexToken(|,'|',-1,-1) LexToken(WHERE,'where',1,0) LexToken(WHERE,'where',-1,-1) parse("foo.baz.id").find({'foo': {'baz': 3}}) =?= ['foo.baz'] parse("foo.baz.id").find({'foo': {'baz': [3]}}) =?= ['foo.baz'] parse("foo.baz.id").find({'foo': {'id': 'bizzle', 'baz': 3}}) =?= ['bizzle.baz'] parse("foo.baz.id").find({'foo': {'baz': {'id': 'hi'}}}) =?= ['foo.hi'] parse("foo.baz.bizzle.id").find({'foo': {'baz': {'bizzle': 5}}}) =?= ['foo.baz.bizzle'] parse("foo.baz").find({'foo': {'baz': 3}}).paths =?= ['foo.baz'] parse("foo.baz").find({'foo': {'baz': [3]}}).paths =?= ['foo.baz'] parse("foo.baz.bizzle").find({'foo': {'baz': {'bizzle': 5}}}).paths =?= ['foo.baz.bizzle'] parse("foo.baz").find({'foo': {'baz': 3}}) =?= [3] parse("foo.baz").find({'foo': {'baz': [3]}}) =?= [[3]] parse("foo.baz.bizzle").find({'foo': {'baz': {'bizzle': 5}}}) =?= [5] parse("foo..baz.id").find({'foo': {'baz': 1, 'bing': {'baz': 2}}}) =?= ['foo.baz', 'foo.bing.baz'] parse("foo..baz").find({'foo': {'baz': 1, 'bing': {'baz': 2}}}).paths =?= ['foo.baz', 'foo.bing.baz'] parse("foo..baz").find({'foo': {'baz': 1, 'bing': {'baz': 2}}}) =?= [1, 2] parse("foo..baz").find({'foo': [{'baz': 1}, {'baz': 2}]}) =?= [1, 2] parse("foo.id").find({'foo': 'baz'}) =?= ['foo'] parse("foo.id").find({'foo': {'id': 'baz'}}) =?= ['baz'] parse("foo,baz.id").find({'foo': 1, 'baz': 2}) =?= ['foo', 'baz'] parse("*.id").find({'foo': {'id': 1}, 'baz': 2}) =?= {'1', 'baz'} parse("foo").find({'foo': 'baz'}).paths =?= ['foo'] parse("foo,baz").find({'foo': 1, 'baz': 2}).paths =?= ['foo', 'baz'] parse("*").find({'foo': 1, 'baz': 2}).paths =?= {'foo', 'baz'} parse("*").find({'foo': 1, 'baz': 2}).paths =?= {'id', 'foo', 'baz'} parse("foo").find({'foo': 'baz'}) =?= ['baz'] parse("foo,baz").find({'foo': 1, 'baz': 2}) =?= [1, 2] parse("@foo").find({'@foo': 1}) =?= [1] parse("*").find({'foo': 1, 'baz': 2}) =?= {1, 2} parse("*").find({'foo': 1, 'baz': 2}) =?= {1, 2, '`this`'} parse("foo.bar-baz").find({'foo': {'bar-baz': 3}}) =?= [3] parse("foo.[bar-baz,blah-blah]").find({'foo': {'bar-baz': 3, 'blah-blah': 5}}) =?= [3, 5] parse("foo.-baz").find({'foo': {'-baz': 8}}) =?= [8] parse("[0].id").find([42]) =?= ['[0]'] parse("[2].id").find([34, 65, 29, 59]) =?= ['[2]'] parse("[0]").find([42]).paths =?= ['[0]'] parse("[2]").find([34, 65, 29, 59]).paths =?= ['[2]'] parse("[0]").find([42]) =?= [42] parse("[5]").find([42]) =?= [] parse("[2]").find([34, 65, 29, 59]) =?= [29] parse("[0]").find(None) =?= [] parse("foo.baz.`parent`").find({'foo': {'baz': 3}}) =?= [{'baz': 3}]test_parent_value (tests.test_jsonpath.TestJsonPath.test_parent_value) ... ok test_root_auto_id (tests.test_jsonpath.TestJsonPath.test_root_auto_id) ... ok test_root_paths (tests.test_jsonpath.TestJsonPath.test_root_paths) ... ok test_root_value (tests.test_jsonpath.TestJsonPath.test_root_value) ... ok test_slice_auto_id (tests.test_jsonpath.TestJsonPath.test_slice_auto_id) ... ok test_slice_paths (tests.test_jsonpath.TestJsonPath.test_slice_paths) ... ok test_slice_value (tests.test_jsonpath.TestJsonPath.test_slice_value) ... ok test_this_auto_id (tests.test_jsonpath.TestJsonPath.test_this_auto_id) ... ok test_this_paths (tests.test_jsonpath.TestJsonPath.test_this_paths) ... ok test_this_value (tests.test_jsonpath.TestJsonPath.test_this_value) ... ok test_update_child (tests.test_jsonpath.TestJsonPath.test_update_child) ... ok test_update_descendants (tests.test_jsonpath.TestJsonPath.test_update_descendants) ... ok test_update_descendants_where (tests.test_jsonpath.TestJsonPath.test_update_descendants_where) ... ok test_update_fields (tests.test_jsonpath.TestJsonPath.test_update_fields) ... ok test_update_index (tests.test_jsonpath.TestJsonPath.test_update_index) ... ok test_update_root (tests.test_jsonpath.TestJsonPath.test_update_root) ... ok test_update_slice (tests.test_jsonpath.TestJsonPath.test_update_slice) ... ok test_update_this (tests.test_jsonpath.TestJsonPath.test_update_this) ... ok test_update_where (tests.test_jsonpath.TestJsonPath.test_update_where) ... ok ---------------------------------------------------------------------- Ran 123 tests in 6.289s OK parse("foo.`parent`.foo.baz.`parent`.baz.bizzle").find({'foo': {'baz': {'bizzle': 5}}}) =?= [5] parse("$.id").find({'foo': 'baz'}) =?= ['$'] parse("foo.$.id").find({'foo': 'baz', 'id': 'bizzle'}) =?= ['bizzle'] parse("foo.$.baz.id").find({'foo': 4, 'baz': 3}) =?= ['baz'] parse("$").find({'foo': 'baz'}).paths =?= ['$'] parse("foo.$").find({'foo': 'baz'}).paths =?= ['$'] parse("foo.$.foo").find({'foo': 'baz'}).paths =?= ['foo'] parse("$").find({'foo': 'baz'}) =?= [{'foo': 'baz'}] parse("foo.$").find({'foo': 'baz'}) =?= [{'foo': 'baz'}] parse("foo.$.foo").find({'foo': 'baz'}) =?= ['baz'] parse("[*].id").find([1, 2, 3]) =?= ['[0]', '[1]', '[2]'] parse("[1:].id").find([1, 2, 3, 4]) =?= ['[1]', '[2]', '[3]'] parse("[*]").find([1, 2, 3]).paths =?= ['[0]', '[1]', '[2]'] parse("[1:]").find([1, 2, 3, 4]).paths =?= ['[1]', '[2]', '[3]'] parse("[*]").find([1, 2, 3]) =?= [1, 2, 3] parse("[*]").find(range(1, 4)) =?= [1, 2, 3] parse("[1:]").find([1, 2, 3, 4]) =?= [2, 3, 4] parse("[:2]").find([1, 2, 3, 4]) =?= [1, 2] parse("[*]").find(1) =?= [1] parse("[0:]").find(1) =?= [1] parse("[*]").find({'foo': 1}) =?= [{'foo': 1}] parse("[*].foo").find({'foo': 1}) =?= [1] parse("id").find({'foo': 'baz'}) =?= ['`this`'] parse("foo.`this`.id").find({'foo': 'baz'}) =?= ['foo'] parse("foo.`this`.baz.id").find({'foo': {'baz': 3}}) =?= ['foo.baz'] parse("`this`").find({'foo': 'baz'}).paths =?= ['`this`'] parse("foo.`this`").find({'foo': 'baz'}).paths =?= ['foo'] parse("foo.`this`.baz").find({'foo': {'baz': 3}}).paths =?= ['foo.baz'] parse("`this`").find({'foo': 'baz'}) =?= [{'foo': 'baz'}] parse("foo.`this`").find({'foo': 'baz'}) =?= ['baz'] parse("foo.`this`.baz").find({'foo': {'baz': 3}}) =?= [3] parse('$.foo').update({'foo': 'bar'}, 'baz') =?= {'foo': 'baz'} parse('foo.bar').update({'foo': {'bar': 1}}, 'baz') =?= {'foo': {'bar': 'baz'}} parse('$..somefield').update({'somefield': 1}, 42) =?= {'somefield': 42} parse('$..nestedfield').update({'outer': {'nestedfield': 1}}, 42) =?= {'outer': {'nestedfield': 42}} parse('$..bar').update({'outs': {'bar': 1, 'ins': {'bar': 9}}, 'outs2': {'bar': 2}}, 42) =?= {'outs': {'bar': 42, 'ins': {'bar': 42}}, 'outs2': {'bar': 42}} parse('(* where flag) .. bar').update({'foo': {'bar': 1, 'flag': 1}, 'baz': {'bar': 2}}, 3) =?= {'foo': {'bar': 3, 'flag': 1}, 'baz': {'bar': 2}} parse('foo').update({'foo': 1}, 5) =?= {'foo': 5} parse('$.*').update({'foo': 1, 'bar': 2}, 3) =?= {'foo': 3, 'bar': 3} parse('[0]').update(['foo', 'bar', 'baz'], 'test') =?= ['test', 'bar', 'baz'] parse('$').update('foo', 'bar') =?= 'bar' parse('[0:2]').update(['foo', 'bar', 'baz'], 'test') =?= ['test', 'test', 'baz'] parse('`this`').update('foo', 'bar') =?= 'bar' parse('*.bar where baz').update({'foo': {'bar': {'baz': 1}}, 'bar': {'baz': 2}}, 5) =?= {'foo': {'bar': 5}, 'bar': {'baz': 2}} + RPM_EC=0 ++ jobs -p + exit 0 Processing files: python3-jsonpath-ng-1.5.1-14.fc41.noarch Executing(%doc): /bin/sh -e /var/tmp/rpm-tmp.t231CP + umask 022 + cd /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build + cd jsonpath-ng-1.5.1 + DOCDIR=/builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/share/doc/python3-jsonpath-ng + export LC_ALL=C.UTF-8 + LC_ALL=C.UTF-8 + export DOCDIR + /usr/bin/mkdir -p /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/share/doc/python3-jsonpath-ng + cp -pr /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/jsonpath-ng-1.5.1/README.rst /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT/usr/share/doc/python3-jsonpath-ng + RPM_EC=0 ++ jobs -p + exit 0 Provides: python-jsonpath-ng = 1.5.1-14.fc41 python3-jsonpath-ng = 1.5.1-14.fc41 python3.13-jsonpath-ng = 1.5.1-14.fc41 python3.13dist(jsonpath-ng) = 1.5.1 python3dist(jsonpath-ng) = 1.5.1 Requires(rpmlib): rpmlib(CompressedFileNames) <= 3.0.4-1 rpmlib(FileDigests) <= 4.6.0-1 rpmlib(PartialHardlinkSets) <= 4.0.4-1 rpmlib(PayloadFilesHavePrefix) <= 4.0-1 Requires: /usr/bin/python3 python(abi) = 3.13 python3.13dist(decorator) python3.13dist(ply) python3.13dist(six) Obsoletes: python-jsonpath-ng < 1.5.1-14.fc41 Checking for unpackaged file(s): /usr/lib/rpm/check-files /builddir/build/BUILD/python-jsonpath-ng-1.5.1-build/BUILDROOT Wrote: /builddir/build/RPMS/python3-jsonpath-ng-1.5.1-14.fc41.noarch.rpm Child return code was: 0