Mock Version: 5.6 Mock Version: 5.6 Mock Version: 5.6 ENTER ['do_with_status'](['bash', '--login', '-c', '/usr/bin/rpmbuild -bs --noclean --target noarch --nodeps /builddir/build/SPECS/python-azure-synapse-spark.spec'], chrootPath='/var/lib/mock/f42-build-54938778-6531220/root'env={'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'}shell=Falselogger=timeout=201600uid=1000gid=425user='mockbuild'nspawn_args=['--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.d8p1gbj2:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11']unshare_net=TrueprintOutput=False) Using nspawn with args ['--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.d8p1gbj2:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11'] Executing command: ['/usr/bin/systemd-nspawn', '-q', '-M', 'db754f99b3574cc18372c5163705e856', '-D', '/var/lib/mock/f42-build-54938778-6531220/root', '-a', '-u', 'mockbuild', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.d8p1gbj2:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11', '--console=pipe', '--setenv=TERM=vt100', '--setenv=SHELL=/bin/bash', '--setenv=HOME=/builddir', '--setenv=HOSTNAME=mock', '--setenv=PATH=/usr/bin:/bin:/usr/sbin:/sbin', '--setenv=PROMPT_COMMAND=printf "\\033]0;\\007"', '--setenv=PS1= \\s-\\v\\$ ', '--setenv=LANG=C.UTF-8', '--resolv-conf=off', 'bash', '--login', '-c', '/usr/bin/rpmbuild -bs --noclean --target noarch --nodeps /builddir/build/SPECS/python-azure-synapse-spark.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8', 'SYSTEMD_NSPAWN_TMPFS_TMP': '0', 'SYSTEMD_SECCOMP': '0'} and shell False Building target platforms: noarch Building for target noarch setting SOURCE_DATE_EPOCH=1721347200 Wrote: /builddir/build/SRPMS/python-azure-synapse-spark-0.2.0-17.fc42.src.rpm Child return code was: 0 ENTER ['do_with_status'](['bash', '--login', '-c', '/usr/bin/rpmbuild -br --noclean --target noarch --nodeps /builddir/build/SPECS/python-azure-synapse-spark.spec'], chrootPath='/var/lib/mock/f42-build-54938778-6531220/root'env={'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'}shell=Falselogger=timeout=201600uid=1000gid=425user='mockbuild'nspawn_args=['--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.d8p1gbj2:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11']unshare_net=TrueraiseExc=FalseprintOutput=False) Using nspawn with args ['--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.d8p1gbj2:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11'] Executing command: ['/usr/bin/systemd-nspawn', '-q', '-M', 'e3a83f8b42544d24b8817eb24528f756', '-D', '/var/lib/mock/f42-build-54938778-6531220/root', '-a', '-u', 'mockbuild', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.d8p1gbj2:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11', '--console=pipe', '--setenv=TERM=vt100', '--setenv=SHELL=/bin/bash', '--setenv=HOME=/builddir', '--setenv=HOSTNAME=mock', '--setenv=PATH=/usr/bin:/bin:/usr/sbin:/sbin', '--setenv=PROMPT_COMMAND=printf "\\033]0;\\007"', '--setenv=PS1= \\s-\\v\\$ ', '--setenv=LANG=C.UTF-8', '--resolv-conf=off', 'bash', '--login', '-c', '/usr/bin/rpmbuild -br --noclean --target noarch --nodeps /builddir/build/SPECS/python-azure-synapse-spark.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8', 'SYSTEMD_NSPAWN_TMPFS_TMP': '0', 'SYSTEMD_SECCOMP': '0'} and shell False Building target platforms: noarch Building for target noarch setting SOURCE_DATE_EPOCH=1721347200 Executing(%mkbuilddir): /bin/sh -e /var/tmp/rpm-tmp.LIdFmy + umask 022 + cd /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build + test -d /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build + /usr/bin/chmod -Rf a+rX,u+w,g-w,o-w /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build + /usr/bin/rm -rf /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build + /usr/bin/mkdir -p /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build + /usr/bin/mkdir -p /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/SPECPARTS + RPM_EC=0 ++ jobs -p + exit 0 Executing(%prep): /bin/sh -e /var/tmp/rpm-tmp.dLDPOP + umask 022 + cd /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build + cd /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build + rm -rf azure-synapse-spark-0.2.0 + /usr/lib/rpm/rpmuncompress -x /builddir/build/SOURCES/azure-synapse-spark-0.2.0.zip + STATUS=0 + '[' 0 -ne 0 ']' + cd azure-synapse-spark-0.2.0 + /usr/bin/chmod -Rf a+rX,u+w,g-w,o-w . + RPM_EC=0 ++ jobs -p + exit 0 Executing(%generate_buildrequires): /bin/sh -e /var/tmp/rpm-tmp.QeNeDz + umask 022 + cd /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build + cd azure-synapse-spark-0.2.0 + echo pyproject-rpm-macros + echo python3-devel + echo 'python3dist(packaging)' + echo 'python3dist(pip) >= 19' + '[' -f pyproject.toml ']' + '[' -f setup.py ']' + echo 'python3dist(setuptools) >= 40.8' + rm -rfv '*.dist-info/' + '[' -f /usr/bin/python3 ']' + mkdir -p /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/.pyproject-builddir + echo -n + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer ' + CXXFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer ' + FFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer -I/usr/lib/gfortran/modules ' + FCFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer -I/usr/lib/gfortran/modules ' + VALAFLAGS=-g + RUSTFLAGS='-Copt-level=3 -Cdebuginfo=2 -Ccodegen-units=1 -Cstrip=none -Cforce-frame-pointers=yes --cap-lints=warn' + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -Wl,--build-id=sha1 ' + LT_SYS_LIBRARY_PATH=/usr/lib: + CC=gcc + CXX=g++ + TMPDIR=/builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/.pyproject-builddir + RPM_TOXENV=py313 + HOSTNAME=rpmbuild + /usr/bin/python3 -Bs /usr/lib/rpm/redhat/pyproject_buildrequires.py --generate-extras --python3_pkgversion 3 --wheeldir /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/pyproject-wheeldir --output /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/python-azure-synapse-spark-0.2.0-17.fc42.noarch-pyproject-buildrequires Handling setuptools >= 40.8 from default build backend Requirement not satisfied: setuptools >= 40.8 Exiting dependency generation pass: build backend + cat /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/python-azure-synapse-spark-0.2.0-17.fc42.noarch-pyproject-buildrequires + rm -rfv '*.dist-info/' + RPM_EC=0 ++ jobs -p + exit 0 Wrote: /builddir/build/SRPMS/python-azure-synapse-spark-0.2.0-17.fc42.buildreqs.nosrc.rpm Child return code was: 11 Dynamic buildrequires detected Going to install missing buildrequires. See root.log for details. ENTER ['do_with_status'](['bash', '--login', '-c', '/usr/bin/rpmbuild -br --noprep --noclean --target noarch --nodeps /builddir/build/SPECS/python-azure-synapse-spark.spec'], chrootPath='/var/lib/mock/f42-build-54938778-6531220/root'env={'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'}shell=Falselogger=timeout=201600uid=1000gid=425user='mockbuild'nspawn_args=['--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.d8p1gbj2:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11']unshare_net=TrueraiseExc=FalseprintOutput=False) Using nspawn with args ['--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.d8p1gbj2:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11'] Executing command: ['/usr/bin/systemd-nspawn', '-q', '-M', 'f3d89de5904d4f518bfb11a6fdfdde40', '-D', '/var/lib/mock/f42-build-54938778-6531220/root', '-a', '-u', 'mockbuild', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.d8p1gbj2:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11', '--console=pipe', '--setenv=TERM=vt100', '--setenv=SHELL=/bin/bash', '--setenv=HOME=/builddir', '--setenv=HOSTNAME=mock', '--setenv=PATH=/usr/bin:/bin:/usr/sbin:/sbin', '--setenv=PROMPT_COMMAND=printf "\\033]0;\\007"', '--setenv=PS1= \\s-\\v\\$ ', '--setenv=LANG=C.UTF-8', '--resolv-conf=off', 'bash', '--login', '-c', '/usr/bin/rpmbuild -br --noprep --noclean --target noarch --nodeps /builddir/build/SPECS/python-azure-synapse-spark.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8', 'SYSTEMD_NSPAWN_TMPFS_TMP': '0', 'SYSTEMD_SECCOMP': '0'} and shell False Building target platforms: noarch Building for target noarch setting SOURCE_DATE_EPOCH=1721347200 Executing(%generate_buildrequires): /bin/sh -e /var/tmp/rpm-tmp.ImRRKV + umask 022 + cd /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build + cd azure-synapse-spark-0.2.0 + echo pyproject-rpm-macros + echo python3-devel + echo 'python3dist(packaging)' + echo 'python3dist(pip) >= 19' + '[' -f pyproject.toml ']' + '[' -f setup.py ']' + echo 'python3dist(setuptools) >= 40.8' + rm -rfv '*.dist-info/' + '[' -f /usr/bin/python3 ']' + mkdir -p /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/.pyproject-builddir + echo -n + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer ' + CXXFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer ' + FFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer -I/usr/lib/gfortran/modules ' + FCFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer -I/usr/lib/gfortran/modules ' + VALAFLAGS=-g + RUSTFLAGS='-Copt-level=3 -Cdebuginfo=2 -Ccodegen-units=1 -Cstrip=none -Cforce-frame-pointers=yes --cap-lints=warn' + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -Wl,--build-id=sha1 ' + LT_SYS_LIBRARY_PATH=/usr/lib: + CC=gcc + CXX=g++ + TMPDIR=/builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/.pyproject-builddir + RPM_TOXENV=py313 + HOSTNAME=rpmbuild + /usr/bin/python3 -Bs /usr/lib/rpm/redhat/pyproject_buildrequires.py --generate-extras --python3_pkgversion 3 --wheeldir /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/pyproject-wheeldir --output /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/python-azure-synapse-spark-0.2.0-17.fc42.noarch-pyproject-buildrequires Handling setuptools >= 40.8 from default build backend Requirement satisfied: setuptools >= 40.8 (installed: setuptools 74.1.3) running egg_info writing azure_synapse_spark.egg-info/PKG-INFO writing dependency_links to azure_synapse_spark.egg-info/dependency_links.txt writing requirements to azure_synapse_spark.egg-info/requires.txt writing top-level names to azure_synapse_spark.egg-info/top_level.txt reading manifest file 'azure_synapse_spark.egg-info/SOURCES.txt' reading manifest template 'MANIFEST.in' warning: no files found matching '*.py' under directory 'tests' warning: no files found matching '*.yaml' under directory 'tests' writing manifest file 'azure_synapse_spark.egg-info/SOURCES.txt' running dist_info writing azure_synapse_spark.egg-info/PKG-INFO writing dependency_links to azure_synapse_spark.egg-info/dependency_links.txt writing requirements to azure_synapse_spark.egg-info/requires.txt writing top-level names to azure_synapse_spark.egg-info/top_level.txt reading manifest file 'azure_synapse_spark.egg-info/SOURCES.txt' reading manifest template 'MANIFEST.in' warning: no files found matching '*.py' under directory 'tests' warning: no files found matching '*.yaml' under directory 'tests' writing manifest file 'azure_synapse_spark.egg-info/SOURCES.txt' creating '/builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/azure_synapse_spark-0.2.0.dist-info' Handling msrest >=0.5.0 from hook generated metadata: Requires-Dist (azure-synapse-spark) Requirement not satisfied: msrest >=0.5.0 Handling azure-common ~=1.1 from hook generated metadata: Requires-Dist (azure-synapse-spark) Requirement not satisfied: azure-common ~=1.1 Handling azure-core <2.0.0,>=1.6.0 from hook generated metadata: Requires-Dist (azure-synapse-spark) Requirement not satisfied: azure-core <2.0.0,>=1.6.0 Handling azure-synapse-nspkg ; python_version<'3.0' from hook generated metadata: Requires-Dist (azure-synapse-spark) Ignoring alien requirement: azure-synapse-nspkg ; python_version<'3.0' + cat /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/python-azure-synapse-spark-0.2.0-17.fc42.noarch-pyproject-buildrequires + rm -rfv azure_synapse_spark-0.2.0.dist-info/ removed 'azure_synapse_spark-0.2.0.dist-info/top_level.txt' removed 'azure_synapse_spark-0.2.0.dist-info/METADATA' removed directory 'azure_synapse_spark-0.2.0.dist-info/' + RPM_EC=0 ++ jobs -p + exit 0 Wrote: /builddir/build/SRPMS/python-azure-synapse-spark-0.2.0-17.fc42.buildreqs.nosrc.rpm Child return code was: 11 Dynamic buildrequires detected Going to install missing buildrequires. See root.log for details. ENTER ['do_with_status'](['bash', '--login', '-c', '/usr/bin/rpmbuild -br --noprep --noclean --target noarch --nodeps /builddir/build/SPECS/python-azure-synapse-spark.spec'], chrootPath='/var/lib/mock/f42-build-54938778-6531220/root'env={'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'}shell=Falselogger=timeout=201600uid=1000gid=425user='mockbuild'nspawn_args=['--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.d8p1gbj2:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11']unshare_net=TrueraiseExc=FalseprintOutput=False) Using nspawn with args ['--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.d8p1gbj2:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11'] Executing command: ['/usr/bin/systemd-nspawn', '-q', '-M', 'f7bc8355fb754228a2982e688e8b48f7', '-D', '/var/lib/mock/f42-build-54938778-6531220/root', '-a', '-u', 'mockbuild', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.d8p1gbj2:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11', '--console=pipe', '--setenv=TERM=vt100', '--setenv=SHELL=/bin/bash', '--setenv=HOME=/builddir', '--setenv=HOSTNAME=mock', '--setenv=PATH=/usr/bin:/bin:/usr/sbin:/sbin', '--setenv=PROMPT_COMMAND=printf "\\033]0;\\007"', '--setenv=PS1= \\s-\\v\\$ ', '--setenv=LANG=C.UTF-8', '--resolv-conf=off', 'bash', '--login', '-c', '/usr/bin/rpmbuild -br --noprep --noclean --target noarch --nodeps /builddir/build/SPECS/python-azure-synapse-spark.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8', 'SYSTEMD_NSPAWN_TMPFS_TMP': '0', 'SYSTEMD_SECCOMP': '0'} and shell False Building target platforms: noarch Building for target noarch setting SOURCE_DATE_EPOCH=1721347200 Executing(%generate_buildrequires): /bin/sh -e /var/tmp/rpm-tmp.CrSQ6R + umask 022 + cd /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build + cd azure-synapse-spark-0.2.0 + echo pyproject-rpm-macros + echo python3-devel + echo 'python3dist(packaging)' + echo 'python3dist(pip) >= 19' + '[' -f pyproject.toml ']' + '[' -f setup.py ']' + echo 'python3dist(setuptools) >= 40.8' + rm -rfv '*.dist-info/' + '[' -f /usr/bin/python3 ']' + mkdir -p /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/.pyproject-builddir + echo -n + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer ' + CXXFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer ' + FFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer -I/usr/lib/gfortran/modules ' + FCFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer -I/usr/lib/gfortran/modules ' + VALAFLAGS=-g + RUSTFLAGS='-Copt-level=3 -Cdebuginfo=2 -Ccodegen-units=1 -Cstrip=none -Cforce-frame-pointers=yes --cap-lints=warn' + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -Wl,--build-id=sha1 ' + LT_SYS_LIBRARY_PATH=/usr/lib: + CC=gcc + CXX=g++ + TMPDIR=/builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/.pyproject-builddir + RPM_TOXENV=py313 + HOSTNAME=rpmbuild + /usr/bin/python3 -Bs /usr/lib/rpm/redhat/pyproject_buildrequires.py --generate-extras --python3_pkgversion 3 --wheeldir /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/pyproject-wheeldir --output /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/python-azure-synapse-spark-0.2.0-17.fc42.noarch-pyproject-buildrequires Handling setuptools >= 40.8 from default build backend Requirement satisfied: setuptools >= 40.8 (installed: setuptools 74.1.3) running egg_info writing azure_synapse_spark.egg-info/PKG-INFO writing dependency_links to azure_synapse_spark.egg-info/dependency_links.txt writing requirements to azure_synapse_spark.egg-info/requires.txt writing top-level names to azure_synapse_spark.egg-info/top_level.txt reading manifest file 'azure_synapse_spark.egg-info/SOURCES.txt' reading manifest template 'MANIFEST.in' warning: no files found matching '*.py' under directory 'tests' warning: no files found matching '*.yaml' under directory 'tests' writing manifest file 'azure_synapse_spark.egg-info/SOURCES.txt' running dist_info writing azure_synapse_spark.egg-info/PKG-INFO writing dependency_links to azure_synapse_spark.egg-info/dependency_links.txt writing requirements to azure_synapse_spark.egg-info/requires.txt writing top-level names to azure_synapse_spark.egg-info/top_level.txt reading manifest file 'azure_synapse_spark.egg-info/SOURCES.txt' reading manifest template 'MANIFEST.in' warning: no files found matching '*.py' under directory 'tests' warning: no files found matching '*.yaml' under directory 'tests' writing manifest file 'azure_synapse_spark.egg-info/SOURCES.txt' creating '/builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/azure_synapse_spark-0.2.0.dist-info' Handling msrest >=0.5.0 from hook generated metadata: Requires-Dist (azure-synapse-spark) Requirement satisfied: msrest >=0.5.0 (installed: msrest 0.7.1) Handling azure-common ~=1.1 from hook generated metadata: Requires-Dist (azure-synapse-spark) Requirement satisfied: azure-common ~=1.1 (installed: azure-common 1.1.28) Handling azure-core <2.0.0,>=1.6.0 from hook generated metadata: Requires-Dist (azure-synapse-spark) Requirement satisfied: azure-core <2.0.0,>=1.6.0 (installed: azure-core 1.30.2) Handling azure-synapse-nspkg ; python_version<'3.0' from hook generated metadata: Requires-Dist (azure-synapse-spark) Ignoring alien requirement: azure-synapse-nspkg ; python_version<'3.0' + cat /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/python-azure-synapse-spark-0.2.0-17.fc42.noarch-pyproject-buildrequires + rm -rfv azure_synapse_spark-0.2.0.dist-info/ removed 'azure_synapse_spark-0.2.0.dist-info/top_level.txt' removed 'azure_synapse_spark-0.2.0.dist-info/METADATA' removed directory 'azure_synapse_spark-0.2.0.dist-info/' + RPM_EC=0 ++ jobs -p + exit 0 Wrote: /builddir/build/SRPMS/python-azure-synapse-spark-0.2.0-17.fc42.buildreqs.nosrc.rpm Child return code was: 11 Dynamic buildrequires detected Going to install missing buildrequires. See root.log for details. ENTER ['do_with_status'](['bash', '--login', '-c', '/usr/bin/rpmbuild -ba --noprep --noclean --target noarch /builddir/build/SPECS/python-azure-synapse-spark.spec'], chrootPath='/var/lib/mock/f42-build-54938778-6531220/root'env={'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'}shell=Falselogger=timeout=201600uid=1000gid=425user='mockbuild'nspawn_args=['--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.d8p1gbj2:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11']unshare_net=TrueprintOutput=False) Using nspawn with args ['--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.d8p1gbj2:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11'] Executing command: ['/usr/bin/systemd-nspawn', '-q', '-M', '8412e813bf6547fd88a900af1997ac0d', '-D', '/var/lib/mock/f42-build-54938778-6531220/root', '-a', '-u', 'mockbuild', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.d8p1gbj2:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11', '--console=pipe', '--setenv=TERM=vt100', '--setenv=SHELL=/bin/bash', '--setenv=HOME=/builddir', '--setenv=HOSTNAME=mock', '--setenv=PATH=/usr/bin:/bin:/usr/sbin:/sbin', '--setenv=PROMPT_COMMAND=printf "\\033]0;\\007"', '--setenv=PS1= \\s-\\v\\$ ', '--setenv=LANG=C.UTF-8', '--resolv-conf=off', 'bash', '--login', '-c', '/usr/bin/rpmbuild -ba --noprep --noclean --target noarch /builddir/build/SPECS/python-azure-synapse-spark.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8', 'SYSTEMD_NSPAWN_TMPFS_TMP': '0', 'SYSTEMD_SECCOMP': '0'} and shell False Building target platforms: noarch Building for target noarch setting SOURCE_DATE_EPOCH=1721347200 Executing(%generate_buildrequires): /bin/sh -e /var/tmp/rpm-tmp.kGHoUr + umask 022 + cd /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build + cd azure-synapse-spark-0.2.0 + echo pyproject-rpm-macros + echo python3-devel + echo 'python3dist(packaging)' + echo 'python3dist(pip) >= 19' + '[' -f pyproject.toml ']' + '[' -f setup.py ']' + echo 'python3dist(setuptools) >= 40.8' + rm -rfv '*.dist-info/' + '[' -f /usr/bin/python3 ']' + mkdir -p /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/.pyproject-builddir + echo -n + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer ' + CXXFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer ' + FFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer -I/usr/lib/gfortran/modules ' + FCFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer -I/usr/lib/gfortran/modules ' + VALAFLAGS=-g + RUSTFLAGS='-Copt-level=3 -Cdebuginfo=2 -Ccodegen-units=1 -Cstrip=none -Cforce-frame-pointers=yes --cap-lints=warn' + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -Wl,--build-id=sha1 ' + LT_SYS_LIBRARY_PATH=/usr/lib: + CC=gcc + CXX=g++ + TMPDIR=/builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/.pyproject-builddir + RPM_TOXENV=py313 + HOSTNAME=rpmbuild + /usr/bin/python3 -Bs /usr/lib/rpm/redhat/pyproject_buildrequires.py --generate-extras --python3_pkgversion 3 --wheeldir /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/pyproject-wheeldir --output /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/python-azure-synapse-spark-0.2.0-17.fc42.noarch-pyproject-buildrequires Handling setuptools >= 40.8 from default build backend Requirement satisfied: setuptools >= 40.8 (installed: setuptools 74.1.3) running egg_info writing azure_synapse_spark.egg-info/PKG-INFO writing dependency_links to azure_synapse_spark.egg-info/dependency_links.txt writing requirements to azure_synapse_spark.egg-info/requires.txt writing top-level names to azure_synapse_spark.egg-info/top_level.txt reading manifest file 'azure_synapse_spark.egg-info/SOURCES.txt' reading manifest template 'MANIFEST.in' warning: no files found matching '*.py' under directory 'tests' warning: no files found matching '*.yaml' under directory 'tests' writing manifest file 'azure_synapse_spark.egg-info/SOURCES.txt' running dist_info writing azure_synapse_spark.egg-info/PKG-INFO writing dependency_links to azure_synapse_spark.egg-info/dependency_links.txt writing requirements to azure_synapse_spark.egg-info/requires.txt writing top-level names to azure_synapse_spark.egg-info/top_level.txt reading manifest file 'azure_synapse_spark.egg-info/SOURCES.txt' reading manifest template 'MANIFEST.in' warning: no files found matching '*.py' under directory 'tests' warning: no files found matching '*.yaml' under directory 'tests' writing manifest file 'azure_synapse_spark.egg-info/SOURCES.txt' creating '/builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/azure_synapse_spark-0.2.0.dist-info' Handling msrest >=0.5.0 from hook generated metadata: Requires-Dist (azure-synapse-spark) Requirement satisfied: msrest >=0.5.0 (installed: msrest 0.7.1) Handling azure-common ~=1.1 from hook generated metadata: Requires-Dist (azure-synapse-spark) Requirement satisfied: azure-common ~=1.1 (installed: azure-common 1.1.28) Handling azure-core <2.0.0,>=1.6.0 from hook generated metadata: Requires-Dist (azure-synapse-spark) Requirement satisfied: azure-core <2.0.0,>=1.6.0 (installed: azure-core 1.30.2) Handling azure-synapse-nspkg ; python_version<'3.0' from hook generated metadata: Requires-Dist (azure-synapse-spark) Ignoring alien requirement: azure-synapse-nspkg ; python_version<'3.0' + cat /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/python-azure-synapse-spark-0.2.0-17.fc42.noarch-pyproject-buildrequires + rm -rfv azure_synapse_spark-0.2.0.dist-info/ removed 'azure_synapse_spark-0.2.0.dist-info/top_level.txt' removed 'azure_synapse_spark-0.2.0.dist-info/METADATA' removed directory 'azure_synapse_spark-0.2.0.dist-info/' + RPM_EC=0 ++ jobs -p + exit 0 Executing(%build): /bin/sh -e /var/tmp/rpm-tmp.Epescw + umask 022 + cd /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer ' + export CFLAGS + CXXFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer ' + export CXXFLAGS + FFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer -I/usr/lib/gfortran/modules ' + export FFLAGS + FCFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer -I/usr/lib/gfortran/modules ' + export FCFLAGS + VALAFLAGS=-g + export VALAFLAGS + RUSTFLAGS='-Copt-level=3 -Cdebuginfo=2 -Ccodegen-units=1 -Cstrip=none -Cforce-frame-pointers=yes --cap-lints=warn' + export RUSTFLAGS + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -Wl,--build-id=sha1 ' + export LDFLAGS + LT_SYS_LIBRARY_PATH=/usr/lib: + export LT_SYS_LIBRARY_PATH + CC=gcc + export CC + CXX=g++ + export CXX + cd azure-synapse-spark-0.2.0 + mkdir -p /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/.pyproject-builddir + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer ' + CXXFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer ' + FFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer -I/usr/lib/gfortran/modules ' + FCFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer -I/usr/lib/gfortran/modules ' + VALAFLAGS=-g + RUSTFLAGS='-Copt-level=3 -Cdebuginfo=2 -Ccodegen-units=1 -Cstrip=none -Cforce-frame-pointers=yes --cap-lints=warn' + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -Wl,--build-id=sha1 ' + LT_SYS_LIBRARY_PATH=/usr/lib: + CC=gcc + CXX=g++ + TMPDIR=/builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/.pyproject-builddir + /usr/bin/python3 -Bs /usr/lib/rpm/redhat/pyproject_wheel.py /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/pyproject-wheeldir Processing /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0 Preparing metadata (pyproject.toml): started Running command Preparing metadata (pyproject.toml) running dist_info creating /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/.pyproject-builddir/pip-modern-metadata-jhb4cnto/azure_synapse_spark.egg-info writing /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/.pyproject-builddir/pip-modern-metadata-jhb4cnto/azure_synapse_spark.egg-info/PKG-INFO writing dependency_links to /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/.pyproject-builddir/pip-modern-metadata-jhb4cnto/azure_synapse_spark.egg-info/dependency_links.txt writing requirements to /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/.pyproject-builddir/pip-modern-metadata-jhb4cnto/azure_synapse_spark.egg-info/requires.txt writing top-level names to /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/.pyproject-builddir/pip-modern-metadata-jhb4cnto/azure_synapse_spark.egg-info/top_level.txt writing manifest file '/builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/.pyproject-builddir/pip-modern-metadata-jhb4cnto/azure_synapse_spark.egg-info/SOURCES.txt' reading manifest file '/builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/.pyproject-builddir/pip-modern-metadata-jhb4cnto/azure_synapse_spark.egg-info/SOURCES.txt' reading manifest template 'MANIFEST.in' warning: no files found matching '*.py' under directory 'tests' warning: no files found matching '*.yaml' under directory 'tests' writing manifest file '/builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/.pyproject-builddir/pip-modern-metadata-jhb4cnto/azure_synapse_spark.egg-info/SOURCES.txt' creating '/builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/.pyproject-builddir/pip-modern-metadata-jhb4cnto/azure_synapse_spark-0.2.0.dist-info' Preparing metadata (pyproject.toml): finished with status 'done' Building wheels for collected packages: azure-synapse-spark Building wheel for azure-synapse-spark (pyproject.toml): started Running command Building wheel for azure-synapse-spark (pyproject.toml) running bdist_wheel running build running build_py creating build creating build/lib creating build/lib/azure creating build/lib/azure/synapse creating build/lib/azure/synapse/spark copying azure/synapse/spark/__init__.py -> build/lib/azure/synapse/spark copying azure/synapse/spark/_spark_client.py -> build/lib/azure/synapse/spark copying azure/synapse/spark/_version.py -> build/lib/azure/synapse/spark copying azure/synapse/spark/_configuration.py -> build/lib/azure/synapse/spark creating build/lib/azure/synapse/spark/operations copying azure/synapse/spark/operations/_spark_session_operations.py -> build/lib/azure/synapse/spark/operations copying azure/synapse/spark/operations/__init__.py -> build/lib/azure/synapse/spark/operations copying azure/synapse/spark/operations/_spark_batch_operations.py -> build/lib/azure/synapse/spark/operations creating build/lib/azure/synapse/spark/aio copying azure/synapse/spark/aio/_spark_client_async.py -> build/lib/azure/synapse/spark/aio copying azure/synapse/spark/aio/__init__.py -> build/lib/azure/synapse/spark/aio copying azure/synapse/spark/aio/_configuration_async.py -> build/lib/azure/synapse/spark/aio creating build/lib/azure/synapse/spark/models copying azure/synapse/spark/models/_spark_client_enums.py -> build/lib/azure/synapse/spark/models copying azure/synapse/spark/models/__init__.py -> build/lib/azure/synapse/spark/models copying azure/synapse/spark/models/_models_py3.py -> build/lib/azure/synapse/spark/models copying azure/synapse/spark/models/_models.py -> build/lib/azure/synapse/spark/models creating build/lib/azure/synapse/spark/aio/operations_async copying azure/synapse/spark/aio/operations_async/_spark_session_operations_async.py -> build/lib/azure/synapse/spark/aio/operations_async copying azure/synapse/spark/aio/operations_async/__init__.py -> build/lib/azure/synapse/spark/aio/operations_async copying azure/synapse/spark/aio/operations_async/_spark_batch_operations_async.py -> build/lib/azure/synapse/spark/aio/operations_async installing to build/bdist.linux-aarch64/wheel running install running install_lib creating build/bdist.linux-aarch64 creating build/bdist.linux-aarch64/wheel creating build/bdist.linux-aarch64/wheel/azure creating build/bdist.linux-aarch64/wheel/azure/synapse creating build/bdist.linux-aarch64/wheel/azure/synapse/spark copying build/lib/azure/synapse/spark/__init__.py -> build/bdist.linux-aarch64/wheel/./azure/synapse/spark copying build/lib/azure/synapse/spark/_spark_client.py -> build/bdist.linux-aarch64/wheel/./azure/synapse/spark copying build/lib/azure/synapse/spark/_version.py -> build/bdist.linux-aarch64/wheel/./azure/synapse/spark copying build/lib/azure/synapse/spark/_configuration.py -> build/bdist.linux-aarch64/wheel/./azure/synapse/spark creating build/bdist.linux-aarch64/wheel/azure/synapse/spark/operations copying build/lib/azure/synapse/spark/operations/_spark_session_operations.py -> build/bdist.linux-aarch64/wheel/./azure/synapse/spark/operations copying build/lib/azure/synapse/spark/operations/__init__.py -> build/bdist.linux-aarch64/wheel/./azure/synapse/spark/operations copying build/lib/azure/synapse/spark/operations/_spark_batch_operations.py -> build/bdist.linux-aarch64/wheel/./azure/synapse/spark/operations creating build/bdist.linux-aarch64/wheel/azure/synapse/spark/aio copying build/lib/azure/synapse/spark/aio/_spark_client_async.py -> build/bdist.linux-aarch64/wheel/./azure/synapse/spark/aio copying build/lib/azure/synapse/spark/aio/__init__.py -> build/bdist.linux-aarch64/wheel/./azure/synapse/spark/aio copying build/lib/azure/synapse/spark/aio/_configuration_async.py -> build/bdist.linux-aarch64/wheel/./azure/synapse/spark/aio creating build/bdist.linux-aarch64/wheel/azure/synapse/spark/aio/operations_async copying build/lib/azure/synapse/spark/aio/operations_async/_spark_session_operations_async.py -> build/bdist.linux-aarch64/wheel/./azure/synapse/spark/aio/operations_async copying build/lib/azure/synapse/spark/aio/operations_async/__init__.py -> build/bdist.linux-aarch64/wheel/./azure/synapse/spark/aio/operations_async copying build/lib/azure/synapse/spark/aio/operations_async/_spark_batch_operations_async.py -> build/bdist.linux-aarch64/wheel/./azure/synapse/spark/aio/operations_async creating build/bdist.linux-aarch64/wheel/azure/synapse/spark/models copying build/lib/azure/synapse/spark/models/_spark_client_enums.py -> build/bdist.linux-aarch64/wheel/./azure/synapse/spark/models copying build/lib/azure/synapse/spark/models/__init__.py -> build/bdist.linux-aarch64/wheel/./azure/synapse/spark/models copying build/lib/azure/synapse/spark/models/_models_py3.py -> build/bdist.linux-aarch64/wheel/./azure/synapse/spark/models copying build/lib/azure/synapse/spark/models/_models.py -> build/bdist.linux-aarch64/wheel/./azure/synapse/spark/models running install_egg_info running egg_info writing azure_synapse_spark.egg-info/PKG-INFO writing dependency_links to azure_synapse_spark.egg-info/dependency_links.txt writing requirements to azure_synapse_spark.egg-info/requires.txt writing top-level names to azure_synapse_spark.egg-info/top_level.txt reading manifest file 'azure_synapse_spark.egg-info/SOURCES.txt' reading manifest template 'MANIFEST.in' warning: no files found matching '*.py' under directory 'tests' warning: no files found matching '*.yaml' under directory 'tests' writing manifest file 'azure_synapse_spark.egg-info/SOURCES.txt' Copying azure_synapse_spark.egg-info to build/bdist.linux-aarch64/wheel/./azure_synapse_spark-0.2.0-py3.13.egg-info running install_scripts creating build/bdist.linux-aarch64/wheel/azure_synapse_spark-0.2.0.dist-info/WHEEL creating '/builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/.pyproject-builddir/pip-wheel-12z3eopv/.tmp-clzy2n9y/azure_synapse_spark-0.2.0-py2.py3-none-any.whl' and adding 'build/bdist.linux-aarch64/wheel' to it adding 'azure/synapse/spark/__init__.py' adding 'azure/synapse/spark/_configuration.py' adding 'azure/synapse/spark/_spark_client.py' adding 'azure/synapse/spark/_version.py' adding 'azure/synapse/spark/aio/__init__.py' adding 'azure/synapse/spark/aio/_configuration_async.py' adding 'azure/synapse/spark/aio/_spark_client_async.py' adding 'azure/synapse/spark/aio/operations_async/__init__.py' adding 'azure/synapse/spark/aio/operations_async/_spark_batch_operations_async.py' adding 'azure/synapse/spark/aio/operations_async/_spark_session_operations_async.py' adding 'azure/synapse/spark/models/__init__.py' adding 'azure/synapse/spark/models/_models.py' adding 'azure/synapse/spark/models/_models_py3.py' adding 'azure/synapse/spark/models/_spark_client_enums.py' adding 'azure/synapse/spark/operations/__init__.py' adding 'azure/synapse/spark/operations/_spark_batch_operations.py' adding 'azure/synapse/spark/operations/_spark_session_operations.py' adding 'azure_synapse_spark-0.2.0.dist-info/METADATA' adding 'azure_synapse_spark-0.2.0.dist-info/WHEEL' adding 'azure_synapse_spark-0.2.0.dist-info/top_level.txt' adding 'azure_synapse_spark-0.2.0.dist-info/RECORD' removing build/bdist.linux-aarch64/wheel Building wheel for azure-synapse-spark (pyproject.toml): finished with status 'done' Created wheel for azure-synapse-spark: filename=azure_synapse_spark-0.2.0-py2.py3-none-any.whl size=29869 sha256=bbbb04fbc9fb58f683de5f4616403a33e9a9c728ad626e689703c4eb0dbb018d Stored in directory: /builddir/.cache/pip/wheels/c6/3b/80/f4d8469d0a0384a5f6baa5c3ad0f55b3db09c0b221c412c02c Successfully built azure-synapse-spark + RPM_EC=0 ++ jobs -p + exit 0 Executing(%install): /bin/sh -e /var/tmp/rpm-tmp.Zc32Qb + umask 022 + cd /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build + '[' /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT '!=' / ']' + rm -rf /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT ++ dirname /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT + mkdir -p /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build + mkdir /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer ' + export CFLAGS + CXXFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer ' + export CXXFLAGS + FFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer -I/usr/lib/gfortran/modules ' + export FFLAGS + FCFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer -I/usr/lib/gfortran/modules ' + export FCFLAGS + VALAFLAGS=-g + export VALAFLAGS + RUSTFLAGS='-Copt-level=3 -Cdebuginfo=2 -Ccodegen-units=1 -Cstrip=none -Cforce-frame-pointers=yes --cap-lints=warn' + export RUSTFLAGS + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -Wl,--build-id=sha1 ' + export LDFLAGS + LT_SYS_LIBRARY_PATH=/usr/lib: + export LT_SYS_LIBRARY_PATH + CC=gcc + export CC + CXX=g++ + export CXX + cd azure-synapse-spark-0.2.0 ++ ls /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/pyproject-wheeldir/azure_synapse_spark-0.2.0-py2.py3-none-any.whl ++ xargs basename --multiple ++ sed -E 's/([^-]+)-([^-]+)-.+\.whl/\1==\2/' + specifier=azure_synapse_spark==0.2.0 + '[' -z azure_synapse_spark==0.2.0 ']' + TMPDIR=/builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/.pyproject-builddir + /usr/bin/python3 -m pip install --root /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT --prefix /usr --no-deps --disable-pip-version-check --progress-bar off --verbose --ignore-installed --no-warn-script-location --no-index --no-cache-dir --find-links /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/pyproject-wheeldir azure_synapse_spark==0.2.0 Using pip 24.3.1 from /usr/lib/python3.13/site-packages/pip (python 3.13) Looking in links: /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/pyproject-wheeldir Processing ./pyproject-wheeldir/azure_synapse_spark-0.2.0-py2.py3-none-any.whl Installing collected packages: azure_synapse_spark Successfully installed azure_synapse_spark-0.2.0 + '[' -d /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/bin ']' + rm -f /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/python-azure-synapse-spark-0.2.0-17.fc42.noarch-pyproject-ghost-distinfo + site_dirs=() + '[' -d /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/lib/python3.13/site-packages ']' + site_dirs+=("/usr/lib/python3.13/site-packages") + '[' /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/lib64/python3.13/site-packages '!=' /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/lib/python3.13/site-packages ']' + '[' -d /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/lib64/python3.13/site-packages ']' + for site_dir in ${site_dirs[@]} + for distinfo in /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT$site_dir/*.dist-info + echo '%ghost /usr/lib/python3.13/site-packages/azure_synapse_spark-0.2.0.dist-info' + sed -i s/pip/rpm/ /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/lib/python3.13/site-packages/azure_synapse_spark-0.2.0.dist-info/INSTALLER + PYTHONPATH=/usr/lib/rpm/redhat + /usr/bin/python3 -B /usr/lib/rpm/redhat/pyproject_preprocess_record.py --buildroot /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT --record /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/lib/python3.13/site-packages/azure_synapse_spark-0.2.0.dist-info/RECORD --output /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/python-azure-synapse-spark-0.2.0-17.fc42.noarch-pyproject-record + rm -fv /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/lib/python3.13/site-packages/azure_synapse_spark-0.2.0.dist-info/RECORD removed '/builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/lib/python3.13/site-packages/azure_synapse_spark-0.2.0.dist-info/RECORD' + rm -fv /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/lib/python3.13/site-packages/azure_synapse_spark-0.2.0.dist-info/REQUESTED removed '/builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/lib/python3.13/site-packages/azure_synapse_spark-0.2.0.dist-info/REQUESTED' ++ wc -l /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/python-azure-synapse-spark-0.2.0-17.fc42.noarch-pyproject-ghost-distinfo ++ cut -f1 '-d ' + lines=1 + '[' 1 -ne 1 ']' + RPM_FILES_ESCAPE=4.19 + /usr/bin/python3 /usr/lib/rpm/redhat/pyproject_save_files.py --output-files /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/python-azure-synapse-spark-0.2.0-17.fc42.noarch-pyproject-files --output-modules /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/python-azure-synapse-spark-0.2.0-17.fc42.noarch-pyproject-modules --buildroot /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT --sitelib /usr/lib/python3.13/site-packages --sitearch /usr/lib64/python3.13/site-packages --python-version 3.13 --pyproject-record /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/python-azure-synapse-spark-0.2.0-17.fc42.noarch-pyproject-record --prefix /usr azure + /usr/lib/rpm/check-buildroot + /usr/lib/rpm/redhat/brp-ldconfig + /usr/lib/rpm/brp-compress + /usr/lib/rpm/brp-strip /usr/bin/strip + /usr/lib/rpm/brp-strip-comment-note /usr/bin/strip /usr/bin/objdump + /usr/lib/rpm/redhat/brp-strip-lto /usr/bin/strip + /usr/lib/rpm/brp-strip-static-archive /usr/bin/strip + /usr/lib/rpm/check-rpaths + /usr/lib/rpm/redhat/brp-mangle-shebangs + /usr/lib/rpm/brp-remove-la-files + env /usr/lib/rpm/redhat/brp-python-bytecompile '' 1 0 -j12 Bytecompiling .py files below /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/lib/python3.13 using python3.13 + /usr/lib/rpm/redhat/brp-python-hardlink + /usr/bin/add-determinism --brp -j12 /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/lib/python3.13/site-packages/azure/synapse/spark/aio/operations_async/__pycache__/__init__.cpython-313.pyc: rewriting with normalized contents /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/lib/python3.13/site-packages/azure/synapse/spark/aio/__pycache__/__init__.cpython-313.pyc: rewriting with normalized contents /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/lib/python3.13/site-packages/azure/synapse/spark/operations/__pycache__/__init__.cpython-313.pyc: rewriting with normalized contents /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/lib/python3.13/site-packages/azure/synapse/spark/models/__pycache__/__init__.cpython-313.pyc: rewriting with normalized contents /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/lib/python3.13/site-packages/azure/synapse/spark/aio/__pycache__/_configuration_async.cpython-313.pyc: rewriting with normalized contents /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/lib/python3.13/site-packages/azure/synapse/spark/models/__pycache__/_spark_client_enums.cpython-313.pyc: rewriting with normalized contents /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/lib/python3.13/site-packages/azure/synapse/spark/aio/__pycache__/_spark_client_async.cpython-313.pyc: rewriting with normalized contents /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/lib/python3.13/site-packages/azure/synapse/spark/__pycache__/__init__.cpython-313.pyc: rewriting with normalized contents /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/lib/python3.13/site-packages/azure/synapse/spark/aio/operations_async/__pycache__/_spark_batch_operations_async.cpython-313.pyc: rewriting with normalized contents /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/lib/python3.13/site-packages/azure/synapse/spark/__pycache__/_configuration.cpython-313.pyc: rewriting with normalized contents /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/lib/python3.13/site-packages/azure/synapse/spark/operations/__pycache__/_spark_batch_operations.cpython-313.pyc: rewriting with normalized contents /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/lib/python3.13/site-packages/azure/synapse/spark/__pycache__/_version.cpython-313.pyc: rewriting with normalized contents /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/lib/python3.13/site-packages/azure/synapse/spark/__pycache__/_spark_client.cpython-313.pyc: rewriting with normalized contents /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/lib/python3.13/site-packages/azure/synapse/spark/aio/operations_async/__pycache__/_spark_session_operations_async.cpython-313.pyc: rewriting with normalized contents /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/lib/python3.13/site-packages/azure/synapse/spark/operations/__pycache__/_spark_session_operations.cpython-313.pyc: rewriting with normalized contents /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/lib/python3.13/site-packages/azure/synapse/spark/models/__pycache__/_models.cpython-313.pyc: rewriting with normalized contents /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/lib/python3.13/site-packages/azure/synapse/spark/models/__pycache__/_models_py3.cpython-313.pyc: rewriting with normalized contents Scanned 18 directories and 55 files, processed 17 inodes, 17 modified (0 replaced + 17 rewritten), 0 unsupported format, 0 errors Executing(%check): /bin/sh -e /var/tmp/rpm-tmp.aNaDqe + umask 022 + cd /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer ' + export CFLAGS + CXXFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer ' + export CXXFLAGS + FFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer -I/usr/lib/gfortran/modules ' + export FFLAGS + FCFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -mbranch-protection=standard -fasynchronous-unwind-tables -fstack-clash-protection -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer -I/usr/lib/gfortran/modules ' + export FCFLAGS + VALAFLAGS=-g + export VALAFLAGS + RUSTFLAGS='-Copt-level=3 -Cdebuginfo=2 -Ccodegen-units=1 -Cstrip=none -Cforce-frame-pointers=yes --cap-lints=warn' + export RUSTFLAGS + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -Wl,--build-id=sha1 ' + export LDFLAGS + LT_SYS_LIBRARY_PATH=/usr/lib: + export LT_SYS_LIBRARY_PATH + CC=gcc + export CC + CXX=g++ + export CXX + cd azure-synapse-spark-0.2.0 + '[' '!' -f /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/python-azure-synapse-spark-0.2.0-17.fc42.noarch-pyproject-modules ']' + PATH=/builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/bin:/usr/bin:/bin:/usr/sbin:/sbin:/usr/local/sbin + PYTHONPATH=/builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/lib64/python3.13/site-packages:/builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/lib/python3.13/site-packages + _PYTHONSITE=/builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/lib64/python3.13/site-packages:/builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/lib/python3.13/site-packages + PYTHONDONTWRITEBYTECODE=1 + /usr/bin/python3 -sP /usr/lib/rpm/redhat/import_all_modules.py -f /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/python-azure-synapse-spark-0.2.0-17.fc42.noarch-pyproject-modules Check import: azure Check import: azure.synapse Check import: azure.synapse.spark Check import: azure.synapse.spark.aio Check import: azure.synapse.spark.aio.operations_async Check import: azure.synapse.spark.models Check import: azure.synapse.spark.operations + RPM_EC=0 ++ jobs -p + exit 0 Processing files: python3-azure-synapse-spark-0.2.0-17.fc42.noarch Executing(%doc): /bin/sh -e /var/tmp/rpm-tmp.H2nYKy + umask 022 + cd /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build + cd azure-synapse-spark-0.2.0 + DOCDIR=/builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/share/doc/python3-azure-synapse-spark + export LC_ALL=C.UTF-8 + LC_ALL=C.UTF-8 + export DOCDIR + /usr/bin/mkdir -p /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/share/doc/python3-azure-synapse-spark + cp -pr /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/azure-synapse-spark-0.2.0/README.md /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT/usr/share/doc/python3-azure-synapse-spark + RPM_EC=0 ++ jobs -p + exit 0 Provides: python-azure-synapse-spark = 1:0.2.0-17.fc42 python3-azure-synapse-spark = 1:0.2.0-17.fc42 python3.13-azure-synapse-spark = 1:0.2.0-17.fc42 python3.13dist(azure-synapse-spark) = 0.2 python3dist(azure-synapse-spark) = 0.2 Requires(rpmlib): rpmlib(CompressedFileNames) <= 3.0.4-1 rpmlib(FileDigests) <= 4.6.0-1 rpmlib(PartialHardlinkSets) <= 4.0.4-1 rpmlib(PayloadFilesHavePrefix) <= 4.0-1 Requires: (python3.13dist(azure-common) >= 1.1 with python3.13dist(azure-common) < 2) (python3.13dist(azure-core) < 2~~ with python3.13dist(azure-core) >= 1.6) python(abi) = 3.13 python3.13dist(msrest) >= 0.5 Checking for unpackaged file(s): /usr/lib/rpm/check-files /builddir/build/BUILD/python-azure-synapse-spark-0.2.0-build/BUILDROOT Wrote: /builddir/build/SRPMS/python-azure-synapse-spark-0.2.0-17.fc42.src.rpm Wrote: /builddir/build/RPMS/python3-azure-synapse-spark-0.2.0-17.fc42.noarch.rpm Child return code was: 0