Mock Version: 3.5 ENTER ['do_with_status'](['bash', '--login', '-c', '/usr/bin/rpmbuild -bs --target x86_64 --nodeps /builddir/build/SPECS/python-spark-etl.spec'], chrootPath='/var/lib/mock/openeuler-22.03_LTS_SP1-x86_64-1686399912.615285/root'env={'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'}shell=Falselogger=timeout=0uid=1000gid=135user='mockbuild'nspawn_args=[]unshare_net=FalseprintOutput=True) Executing command: ['bash', '--login', '-c', '/usr/bin/rpmbuild -bs --target x86_64 --nodeps /builddir/build/SPECS/python-spark-etl.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'} and shell False /etc/profile.d/system-info.sh: line 55: ps: command not found Building target platforms: x86_64 Building for target x86_64 Wrote: /builddir/build/SRPMS/python-spark-etl-0.0.130-1.src.rpm Child return code was: 0 Mock Version: 3.5 ENTER ['do_with_status'](['bash', '--login', '-c', '/usr/bin/rpmbuild -bs --target x86_64 --nodeps /builddir/build/SPECS/python-spark-etl.spec'], chrootPath='/var/lib/mock/openeuler-22.03_LTS_SP1-x86_64-1686399912.615285/root'env={'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'}shell=Falselogger=timeout=0uid=1000gid=135user='mockbuild'nspawn_args=[]unshare_net=FalseprintOutput=True) Executing command: ['bash', '--login', '-c', '/usr/bin/rpmbuild -bs --target x86_64 --nodeps /builddir/build/SPECS/python-spark-etl.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'} and shell False /etc/profile.d/system-info.sh: line 55: ps: command not found Building target platforms: x86_64 Building for target x86_64 Wrote: /builddir/build/SRPMS/python-spark-etl-0.0.130-1.src.rpm Child return code was: 0 ENTER ['do_with_status'](['bash', '--login', '-c', '/usr/bin/rpmbuild -bb --target x86_64 --nodeps /builddir/build/SPECS/python-spark-etl.spec'], chrootPath='/var/lib/mock/openeuler-22.03_LTS_SP1-x86_64-1686399912.615285/root'env={'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'}shell=Falselogger=timeout=0uid=1000gid=135user='mockbuild'nspawn_args=[]unshare_net=FalseprintOutput=True) Executing command: ['bash', '--login', '-c', '/usr/bin/rpmbuild -bb --target x86_64 --nodeps /builddir/build/SPECS/python-spark-etl.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'} and shell False /etc/profile.d/system-info.sh: line 55: ps: command not found Building target platforms: x86_64 Building for target x86_64 Executing(%prep): /bin/sh -e /var/tmp/rpm-tmp.4dLgVV + umask 022 + cd /builddir/build/BUILD + cd /builddir/build/BUILD + rm -rf spark-etl-0.0.130 + /usr/bin/gzip -dc /builddir/build/SOURCES/spark-etl-0.0.130.tar.gz + /usr/bin/tar -xof - + STATUS=0 + '[' 0 -ne 0 ']' + cd spark-etl-0.0.130 + /usr/bin/chmod -Rf a+rX,u+w,g-w,o-w . + RPM_EC=0 ++ jobs -p + exit 0 Executing(%build): /bin/sh -e /var/tmp/rpm-tmp.I6Ik8J + umask 022 + cd /builddir/build/BUILD + cd spark-etl-0.0.130 + CFLAGS='-O2 -g -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -grecord-gcc-switches -specs=/usr/lib/rpm/generic-hardened-cc1 -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection ' + LDFLAGS='-Wl,-z,relro -Wl,-z,now -specs=/usr/lib/rpm/generic-hardened-ld' + /usr/bin/python3 setup.py build '--executable=/usr/bin/python3 -s' running build running build_py creating build creating build/lib creating build/lib/spark_etl copying src/spark_etl/__init__.py -> build/lib/spark_etl copying src/spark_etl/utils.py -> build/lib/spark_etl copying src/spark_etl/application.py -> build/lib/spark_etl copying src/spark_etl/template.py -> build/lib/spark_etl copying src/spark_etl/ssh_config.py -> build/lib/spark_etl copying src/spark_etl/build.py -> build/lib/spark_etl copying src/spark_etl/misc_tools.py -> build/lib/spark_etl copying src/spark_etl/version.py -> build/lib/spark_etl copying src/spark_etl/cmds.py -> build/lib/spark_etl copying src/spark_etl/exceptions.py -> build/lib/spark_etl creating build/lib/spark_etl/job_submitters copying src/spark_etl/job_submitters/__init__.py -> build/lib/spark_etl/job_submitters copying src/spark_etl/job_submitters/abstract_job_submitter.py -> build/lib/spark_etl/job_submitters copying src/spark_etl/job_submitters/livy_job_submitter.py -> build/lib/spark_etl/job_submitters creating build/lib/spark_etl/vendors copying src/spark_etl/vendors/__init__.py -> build/lib/spark_etl/vendors creating build/lib/spark_etl/core copying src/spark_etl/core/__init__.py -> build/lib/spark_etl/core copying src/spark_etl/core/s3_client.py -> build/lib/spark_etl/core copying src/spark_etl/core/main.py -> build/lib/spark_etl/core copying src/spark_etl/core/hdfs_client.py -> build/lib/spark_etl/core copying src/spark_etl/core/local_client.py -> build/lib/spark_etl/core creating build/lib/spark_etl/deployers copying src/spark_etl/deployers/__init__.py -> build/lib/spark_etl/deployers copying src/spark_etl/deployers/hdfs_deployer.py -> build/lib/spark_etl/deployers copying src/spark_etl/deployers/s3_deployer.py -> build/lib/spark_etl/deployers copying src/spark_etl/deployers/abstract_deployer.py -> build/lib/spark_etl/deployers copying src/spark_etl/deployers/job_loader.py -> build/lib/spark_etl/deployers creating build/lib/spark_etl/vendors/oracle copying src/spark_etl/vendors/oracle/__init__.py -> build/lib/spark_etl/vendors/oracle copying src/spark_etl/vendors/oracle/job_loader.py -> build/lib/spark_etl/vendors/oracle copying src/spark_etl/vendors/oracle/dataflow_job_submitter.py -> build/lib/spark_etl/vendors/oracle copying src/spark_etl/vendors/oracle/tools.py -> build/lib/spark_etl/vendors/oracle copying src/spark_etl/vendors/oracle/dataflow_deployer.py -> build/lib/spark_etl/vendors/oracle creating build/lib/spark_etl/vendors/local copying src/spark_etl/vendors/local/__init__.py -> build/lib/spark_etl/vendors/local copying src/spark_etl/vendors/local/pyspark_job_submitter.py -> build/lib/spark_etl/vendors/local copying src/spark_etl/vendors/local/local_deployer.py -> build/lib/spark_etl/vendors/local running egg_info writing src/spark_etl.egg-info/PKG-INFO writing dependency_links to src/spark_etl.egg-info/dependency_links.txt writing entry points to src/spark_etl.egg-info/entry_points.txt writing requirements to src/spark_etl.egg-info/requires.txt writing top-level names to src/spark_etl.egg-info/top_level.txt reading manifest file 'src/spark_etl.egg-info/SOURCES.txt' adding license file 'LICENSE' writing manifest file 'src/spark_etl.egg-info/SOURCES.txt' creating build/lib/spark_etl/core/loader_util creating build/lib/spark_etl/core/loader_util/resources copying src/spark_etl/core/loader_util/resources/job_loader.py -> build/lib/spark_etl/core/loader_util/resources + sleep 1 + RPM_EC=0 ++ jobs -p + exit 0 Executing(%install): /bin/sh -e /var/tmp/rpm-tmp.pFR77a + umask 022 + cd /builddir/build/BUILD + '[' /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64 '!=' / ']' + rm -rf /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64 ++ dirname /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64 + mkdir -p /builddir/build/BUILDROOT + mkdir /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64 + cd spark-etl-0.0.130 + CFLAGS='-O2 -g -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -grecord-gcc-switches -specs=/usr/lib/rpm/generic-hardened-cc1 -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection ' + LDFLAGS='-Wl,-z,relro -Wl,-z,now -specs=/usr/lib/rpm/generic-hardened-ld' + /usr/bin/python3 setup.py install -O1 --skip-build --root /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64 running install /usr/lib/python3.9/site-packages/setuptools/command/install.py:34: SetuptoolsDeprecationWarning: setup.py install is deprecated. Use build and pip and other standards-based tools. warnings.warn( running install_lib creating /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr creating /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib creating /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9 creating /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages creating /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl copying build/lib/spark_etl/__init__.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl copying build/lib/spark_etl/utils.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl copying build/lib/spark_etl/application.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl creating /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/job_submitters copying build/lib/spark_etl/job_submitters/__init__.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/job_submitters copying build/lib/spark_etl/job_submitters/abstract_job_submitter.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/job_submitters copying build/lib/spark_etl/job_submitters/livy_job_submitter.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/job_submitters copying build/lib/spark_etl/template.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl copying build/lib/spark_etl/ssh_config.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl copying build/lib/spark_etl/build.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl copying build/lib/spark_etl/misc_tools.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl creating /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/vendors copying build/lib/spark_etl/vendors/__init__.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/vendors creating /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/vendors/oracle copying build/lib/spark_etl/vendors/oracle/__init__.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/vendors/oracle copying build/lib/spark_etl/vendors/oracle/job_loader.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/vendors/oracle copying build/lib/spark_etl/vendors/oracle/dataflow_job_submitter.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/vendors/oracle copying build/lib/spark_etl/vendors/oracle/tools.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/vendors/oracle copying build/lib/spark_etl/vendors/oracle/dataflow_deployer.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/vendors/oracle creating /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/vendors/local copying build/lib/spark_etl/vendors/local/__init__.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/vendors/local copying build/lib/spark_etl/vendors/local/pyspark_job_submitter.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/vendors/local copying build/lib/spark_etl/vendors/local/local_deployer.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/vendors/local copying build/lib/spark_etl/version.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl creating /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/core copying build/lib/spark_etl/core/__init__.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/core copying build/lib/spark_etl/core/s3_client.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/core copying build/lib/spark_etl/core/main.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/core creating /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/core/loader_util creating /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/core/loader_util/resources copying build/lib/spark_etl/core/loader_util/resources/job_loader.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/core/loader_util/resources copying build/lib/spark_etl/core/hdfs_client.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/core copying build/lib/spark_etl/core/local_client.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/core creating /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/deployers copying build/lib/spark_etl/deployers/__init__.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/deployers copying build/lib/spark_etl/deployers/hdfs_deployer.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/deployers copying build/lib/spark_etl/deployers/s3_deployer.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/deployers copying build/lib/spark_etl/deployers/abstract_deployer.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/deployers copying build/lib/spark_etl/deployers/job_loader.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/deployers copying build/lib/spark_etl/cmds.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl copying build/lib/spark_etl/exceptions.py -> /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/__init__.py to __init__.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/utils.py to utils.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/application.py to application.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/job_submitters/__init__.py to __init__.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/job_submitters/abstract_job_submitter.py to abstract_job_submitter.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/job_submitters/livy_job_submitter.py to livy_job_submitter.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/template.py to template.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/ssh_config.py to ssh_config.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/build.py to build.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/misc_tools.py to misc_tools.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/vendors/__init__.py to __init__.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/vendors/oracle/__init__.py to __init__.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/vendors/oracle/job_loader.py to job_loader.cpython-39.pyc File "/usr/lib/python3.9/site-packages/spark_etl/vendors/oracle/job_loader.py", line 16 {% if use_instance_principle %} ^ SyntaxError: invalid syntax byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/vendors/oracle/dataflow_job_submitter.py to dataflow_job_submitter.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/vendors/oracle/tools.py to tools.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/vendors/oracle/dataflow_deployer.py to dataflow_deployer.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/vendors/local/__init__.py to __init__.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/vendors/local/pyspark_job_submitter.py to pyspark_job_submitter.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/vendors/local/local_deployer.py to local_deployer.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/version.py to version.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/core/__init__.py to __init__.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/core/s3_client.py to s3_client.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/core/main.py to main.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/core/loader_util/resources/job_loader.py to job_loader.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/core/hdfs_client.py to hdfs_client.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/core/local_client.py to local_client.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/deployers/__init__.py to __init__.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/deployers/hdfs_deployer.py to hdfs_deployer.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/deployers/s3_deployer.py to s3_deployer.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/deployers/abstract_deployer.py to abstract_deployer.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/deployers/job_loader.py to job_loader.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/cmds.py to cmds.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/exceptions.py to exceptions.cpython-39.pyc writing byte-compilation script '/tmp/tmpelq0g1rf.py' /usr/bin/python3 /tmp/tmpelq0g1rf.py File "/usr/lib/python3.9/site-packages/spark_etl/vendors/oracle/job_loader.py", line 16 {% if use_instance_principle %} ^ SyntaxError: invalid syntax removing /tmp/tmpelq0g1rf.py running install_egg_info running egg_info writing src/spark_etl.egg-info/PKG-INFO writing dependency_links to src/spark_etl.egg-info/dependency_links.txt writing entry points to src/spark_etl.egg-info/entry_points.txt writing requirements to src/spark_etl.egg-info/requires.txt writing top-level names to src/spark_etl.egg-info/top_level.txt reading manifest file 'src/spark_etl.egg-info/SOURCES.txt' adding license file 'LICENSE' writing manifest file 'src/spark_etl.egg-info/SOURCES.txt' Copying src/spark_etl.egg-info to /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl-0.0.130-py3.9.egg-info running install_scripts Installing etl script to /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/bin + install -d -m755 /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64//usr/share/doc/python-spark-etl + '[' -d doc ']' ~/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64 ~/build/BUILD/spark-etl-0.0.130 + '[' -d docs ']' + '[' -d example ']' + '[' -d examples ']' + pushd /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64 + '[' -d usr/lib ']' + find usr/lib -type f -printf '"/%h/%f"\n' + '[' -d usr/lib64 ']' + '[' -d usr/bin ']' + find usr/bin -type f -printf '"/%h/%f"\n' + '[' -d usr/sbin ']' + touch doclist.lst + '[' -d usr/share/man ']' + popd ~/build/BUILD/spark-etl-0.0.130 + mv /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/filelist.lst . + mv /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/doclist.lst . + /usr/bin/find-debuginfo -j4 --strict-build-id -i --build-id-seed 0.0.130-1 --unique-debug-suffix -0.0.130-1.x86_64 --unique-debug-src-base python-spark-etl-0.0.130-1.x86_64 -S debugsourcefiles.list /builddir/build/BUILD/spark-etl-0.0.130 find: 'debug': No such file or directory + /usr/lib/rpm/check-buildroot + /usr/lib/rpm/brp-ldconfig + /usr/lib/rpm/brp-compress + /usr/lib/rpm/brp-strip-static-archive /usr/bin/strip + /usr/lib/rpm/brp-python-bytecompile /usr/bin/python 1 1 Bytecompiling .py files below /builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9 using /usr/bin/python3.9 *** Error compiling '/builddir/build/BUILDROOT/python-spark-etl-0.0.130-1.x86_64/usr/lib/python3.9/site-packages/spark_etl/vendors/oracle/job_loader.py'... File "/usr/lib/python3.9/site-packages/spark_etl/vendors/oracle/job_loader.py", line 16 {% if use_instance_principle %} ^ SyntaxError: invalid syntax error: Bad exit status from /var/tmp/rpm-tmp.pFR77a (%install) Bad exit status from /var/tmp/rpm-tmp.pFR77a (%install) RPM build errors: Child return code was: 1 EXCEPTION: [Error('Command failed: \n # bash --login -c /usr/bin/rpmbuild -bb --target x86_64 --nodeps /builddir/build/SPECS/python-spark-etl.spec\n', 1)] Traceback (most recent call last): File "/usr/lib/python3.9/site-packages/mockbuild/trace_decorator.py", line 93, in trace result = func(*args, **kw) File "/usr/lib/python3.9/site-packages/mockbuild/util.py", line 598, in do_with_status raise exception.Error("Command failed: \n # %s\n%s" % (command, output), child.returncode) mockbuild.exception.Error: Command failed: # bash --login -c /usr/bin/rpmbuild -bb --target x86_64 --nodeps /builddir/build/SPECS/python-spark-etl.spec