Mock Version: 3.5 ENTER ['do_with_status'](['bash', '--login', '-c', '/usr/bin/rpmbuild -bs --target x86_64 --nodeps /builddir/build/SPECS/python-tokenizer-tools.spec'], chrootPath='/var/lib/mock/openeuler-22.03_LTS_SP1-x86_64-1686488637.609806/root'env={'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'}shell=Falselogger=timeout=0uid=1000gid=135user='mockbuild'nspawn_args=[]unshare_net=FalseprintOutput=True) Executing command: ['bash', '--login', '-c', '/usr/bin/rpmbuild -bs --target x86_64 --nodeps /builddir/build/SPECS/python-tokenizer-tools.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'} and shell False /etc/profile.d/system-info.sh: line 55: ps: command not found Building target platforms: x86_64 Building for target x86_64 Wrote: /builddir/build/SRPMS/python-tokenizer-tools-0.46.1-1.src.rpm Child return code was: 0 Mock Version: 3.5 ENTER ['do_with_status'](['bash', '--login', '-c', '/usr/bin/rpmbuild -bs --target x86_64 --nodeps /builddir/build/SPECS/python-tokenizer-tools.spec'], chrootPath='/var/lib/mock/openeuler-22.03_LTS_SP1-x86_64-1686488637.609806/root'env={'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'}shell=Falselogger=timeout=0uid=1000gid=135user='mockbuild'nspawn_args=[]unshare_net=FalseprintOutput=True) Executing command: ['bash', '--login', '-c', '/usr/bin/rpmbuild -bs --target x86_64 --nodeps /builddir/build/SPECS/python-tokenizer-tools.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'} and shell False /etc/profile.d/system-info.sh: line 55: ps: command not found Building target platforms: x86_64 Building for target x86_64 Wrote: /builddir/build/SRPMS/python-tokenizer-tools-0.46.1-1.src.rpm Child return code was: 0 ENTER ['do_with_status'](['bash', '--login', '-c', '/usr/bin/rpmbuild -bb --target x86_64 --nodeps /builddir/build/SPECS/python-tokenizer-tools.spec'], chrootPath='/var/lib/mock/openeuler-22.03_LTS_SP1-x86_64-1686488637.609806/root'env={'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'}shell=Falselogger=timeout=0uid=1000gid=135user='mockbuild'nspawn_args=[]unshare_net=FalseprintOutput=True) Executing command: ['bash', '--login', '-c', '/usr/bin/rpmbuild -bb --target x86_64 --nodeps /builddir/build/SPECS/python-tokenizer-tools.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'} and shell False /etc/profile.d/system-info.sh: line 55: ps: command not found Building target platforms: x86_64 Building for target x86_64 Executing(%prep): /bin/sh -e /var/tmp/rpm-tmp.huuZdA + umask 022 + cd /builddir/build/BUILD + cd /builddir/build/BUILD + rm -rf tokenizer_tools-0.46.1 + /usr/bin/gzip -dc /builddir/build/SOURCES/tokenizer_tools-0.46.1.tar.gz + /usr/bin/tar -xof - + STATUS=0 + '[' 0 -ne 0 ']' + cd tokenizer_tools-0.46.1 + /usr/bin/chmod -Rf a+rX,u+w,g-w,o-w . + RPM_EC=0 ++ jobs -p + exit 0 Executing(%build): /bin/sh -e /var/tmp/rpm-tmp.Tf4ZUX + umask 022 + cd /builddir/build/BUILD + cd tokenizer_tools-0.46.1 + CFLAGS='-O2 -g -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -grecord-gcc-switches -specs=/usr/lib/rpm/generic-hardened-cc1 -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection ' + LDFLAGS='-Wl,-z,relro -Wl,-z,now -specs=/usr/lib/rpm/generic-hardened-ld' + /usr/bin/python3 setup.py build '--executable=/usr/bin/python3 -s' /usr/lib/python3.9/site-packages/setuptools/installer.py:27: SetuptoolsDeprecationWarning: setuptools.installer is deprecated. Requirements should be satisfied by a PEP 517 installer. warnings.warn( WARNING: The wheel package is not available. /usr/lib64/python3.9/distutils/dist.py:274: UserWarning: Unknown distribution option: 'extra_require' warnings.warn(msg) running build running build_py creating build creating build/lib creating build/lib/tokenizer_tools copying tokenizer_tools/split_data.py -> build/lib/tokenizer_tools copying tokenizer_tools/hooks.py -> build/lib/tokenizer_tools copying tokenizer_tools/cli.py -> build/lib/tokenizer_tools copying tokenizer_tools/__init__.py -> build/lib/tokenizer_tools copying tokenizer_tools/tokenizer_tools.py -> build/lib/tokenizer_tools creating build/lib/tokenizer_tools/converter copying tokenizer_tools/converter/conllx_to_rasa.py -> build/lib/tokenizer_tools/converter copying tokenizer_tools/converter/offset_to_sentence.py -> build/lib/tokenizer_tools/converter copying tokenizer_tools/converter/sentence_to_offset.py -> build/lib/tokenizer_tools/converter copying tokenizer_tools/converter/conllx_to_offset.py -> build/lib/tokenizer_tools/converter copying tokenizer_tools/converter/__init__.py -> build/lib/tokenizer_tools/converter creating build/lib/tokenizer_tools/evaluator copying tokenizer_tools/evaluator/offset_evaluator.py -> build/lib/tokenizer_tools/evaluator copying tokenizer_tools/evaluator/tag_level_evaluator.py -> build/lib/tokenizer_tools/evaluator copying tokenizer_tools/evaluator/__init__.py -> build/lib/tokenizer_tools/evaluator creating build/lib/tokenizer_tools/conll copying tokenizer_tools/conll/reader.py -> build/lib/tokenizer_tools/conll copying tokenizer_tools/conll/iterator_reader.py -> build/lib/tokenizer_tools/conll copying tokenizer_tools/conll/entity_vocabulary.py -> build/lib/tokenizer_tools/conll copying tokenizer_tools/conll/writer.py -> build/lib/tokenizer_tools/conll copying tokenizer_tools/conll/__init__.py -> build/lib/tokenizer_tools/conll copying tokenizer_tools/conll/tag_collector.py -> build/lib/tokenizer_tools/conll creating build/lib/tokenizer_tools/utils copying tokenizer_tools/utils/batch_iterator.py -> build/lib/tokenizer_tools/utils copying tokenizer_tools/utils/deprecated_api.py -> build/lib/tokenizer_tools/utils copying tokenizer_tools/utils/generate_tagset.py -> build/lib/tokenizer_tools/utils copying tokenizer_tools/utils/deprecated_support.py -> build/lib/tokenizer_tools/utils copying tokenizer_tools/utils/__init__.py -> build/lib/tokenizer_tools/utils copying tokenizer_tools/utils/util_list.py -> build/lib/tokenizer_tools/utils creating build/lib/tokenizer_tools/conllz copying tokenizer_tools/conllz/reader.py -> build/lib/tokenizer_tools/conllz copying tokenizer_tools/conllz/iterator_reader.py -> build/lib/tokenizer_tools/conllz copying tokenizer_tools/conllz/sentence.py -> build/lib/tokenizer_tools/conllz copying tokenizer_tools/conllz/multiprocess_reader.py -> build/lib/tokenizer_tools/conllz copying tokenizer_tools/conllz/writer.py -> build/lib/tokenizer_tools/conllz copying tokenizer_tools/conllz/__init__.py -> build/lib/tokenizer_tools/conllz creating build/lib/tokenizer_tools/tagset copying tokenizer_tools/tagset/exceptions.py -> build/lib/tokenizer_tools/tagset copying tokenizer_tools/tagset/BMES.py -> build/lib/tokenizer_tools/tagset copying tokenizer_tools/tagset/__init__.py -> build/lib/tokenizer_tools/tagset creating build/lib/tokenizer_tools/format_converter copying tokenizer_tools/format_converter/corpus_to_rasa_json.py -> build/lib/tokenizer_tools/format_converter copying tokenizer_tools/format_converter/__init__.py -> build/lib/tokenizer_tools/format_converter creating build/lib/tokenizer_tools/evaluator/token copying tokenizer_tools/evaluator/token/tag_level.py -> build/lib/tokenizer_tools/evaluator/token copying tokenizer_tools/evaluator/token/token_level.py -> build/lib/tokenizer_tools/evaluator/token copying tokenizer_tools/evaluator/token/base_evaluator.py -> build/lib/tokenizer_tools/evaluator/token copying tokenizer_tools/evaluator/token/__init__.py -> build/lib/tokenizer_tools/evaluator/token creating build/lib/tokenizer_tools/tagset/converter copying tokenizer_tools/tagset/converter/__init__.py -> build/lib/tokenizer_tools/tagset/converter copying tokenizer_tools/tagset/converter/offset_to_biluo.py -> build/lib/tokenizer_tools/tagset/converter creating build/lib/tokenizer_tools/tagset/NER copying tokenizer_tools/tagset/NER/BILUO.py -> build/lib/tokenizer_tools/tagset/NER copying tokenizer_tools/tagset/NER/IOB.py -> build/lib/tokenizer_tools/tagset/NER copying tokenizer_tools/tagset/NER/__init__.py -> build/lib/tokenizer_tools/tagset/NER copying tokenizer_tools/tagset/NER/base_tagset.py -> build/lib/tokenizer_tools/tagset/NER creating build/lib/tokenizer_tools/tagset/offset copying tokenizer_tools/tagset/offset/corpus_diff.py -> build/lib/tokenizer_tools/tagset/offset copying tokenizer_tools/tagset/offset/plottable_counter.py -> build/lib/tokenizer_tools/tagset/offset copying tokenizer_tools/tagset/offset/exceptions.py -> build/lib/tokenizer_tools/tagset/offset copying tokenizer_tools/tagset/offset/span_set.py -> build/lib/tokenizer_tools/tagset/offset copying tokenizer_tools/tagset/offset/corpus.py -> build/lib/tokenizer_tools/tagset/offset copying tokenizer_tools/tagset/offset/corpus_pb2.py -> build/lib/tokenizer_tools/tagset/offset copying tokenizer_tools/tagset/offset/corpus_statistics.py -> build/lib/tokenizer_tools/tagset/offset copying tokenizer_tools/tagset/offset/span.py -> build/lib/tokenizer_tools/tagset/offset copying tokenizer_tools/tagset/offset/sequence.py -> build/lib/tokenizer_tools/tagset/offset copying tokenizer_tools/tagset/offset/corpus_metric.py -> build/lib/tokenizer_tools/tagset/offset copying tokenizer_tools/tagset/offset/corpus_stream_writer.py -> build/lib/tokenizer_tools/tagset/offset copying tokenizer_tools/tagset/offset/transform.py -> build/lib/tokenizer_tools/tagset/offset copying tokenizer_tools/tagset/offset/language_model.py -> build/lib/tokenizer_tools/tagset/offset copying tokenizer_tools/tagset/offset/corpus_stream_reader.py -> build/lib/tokenizer_tools/tagset/offset copying tokenizer_tools/tagset/offset/corpus_protobuf.py -> build/lib/tokenizer_tools/tagset/offset copying tokenizer_tools/tagset/offset/document_compare_ways.py -> build/lib/tokenizer_tools/tagset/offset copying tokenizer_tools/tagset/offset/document_block_view.py -> build/lib/tokenizer_tools/tagset/offset copying tokenizer_tools/tagset/offset/corpus_set.py -> build/lib/tokenizer_tools/tagset/offset copying tokenizer_tools/tagset/offset/seqeval.py -> build/lib/tokenizer_tools/tagset/offset copying tokenizer_tools/tagset/offset/__init__.py -> build/lib/tokenizer_tools/tagset/offset copying tokenizer_tools/tagset/offset/document.py -> build/lib/tokenizer_tools/tagset/offset copying tokenizer_tools/tagset/offset/speed_compare_between_pb_and_conllx.py -> build/lib/tokenizer_tools/tagset/offset creating build/lib/tokenizer_tools/tagset/offset/analysis copying tokenizer_tools/tagset/offset/analysis/entity_placeholder.py -> build/lib/tokenizer_tools/tagset/offset/analysis copying tokenizer_tools/tagset/offset/analysis/read_markdown.py -> build/lib/tokenizer_tools/tagset/offset/analysis copying tokenizer_tools/tagset/offset/analysis/express_pattern.py -> build/lib/tokenizer_tools/tagset/offset/analysis copying tokenizer_tools/tagset/offset/analysis/document_pattern.py -> build/lib/tokenizer_tools/tagset/offset/analysis copying tokenizer_tools/tagset/offset/analysis/corpus_pattern.py -> build/lib/tokenizer_tools/tagset/offset/analysis copying tokenizer_tools/tagset/offset/analysis/plain_text.py -> build/lib/tokenizer_tools/tagset/offset/analysis copying tokenizer_tools/tagset/offset/analysis/__init__.py -> build/lib/tokenizer_tools/tagset/offset/analysis creating build/lib/tokenizer_tools/tagset/offset/diagnose copying tokenizer_tools/tagset/offset/diagnose/entity_disagreement.py -> build/lib/tokenizer_tools/tagset/offset/diagnose copying tokenizer_tools/tagset/offset/diagnose/__init__.py -> build/lib/tokenizer_tools/tagset/offset/diagnose creating build/lib/tokenizer_tools/tagset/offset/dashboard copying tokenizer_tools/tagset/offset/dashboard/main.py -> build/lib/tokenizer_tools/tagset/offset/dashboard copying tokenizer_tools/tagset/offset/dashboard/es_query.py -> build/lib/tokenizer_tools/tagset/offset/dashboard copying tokenizer_tools/tagset/offset/dashboard/text_perplexity.py -> build/lib/tokenizer_tools/tagset/offset/dashboard copying tokenizer_tools/tagset/offset/dashboard/__init__.py -> build/lib/tokenizer_tools/tagset/offset/dashboard copying tokenizer_tools/tagset/offset/dashboard/es_create.py -> build/lib/tokenizer_tools/tagset/offset/dashboard copying tokenizer_tools/tagset/offset/dashboard/mistagged_span.py -> build/lib/tokenizer_tools/tagset/offset/dashboard running egg_info writing tokenizer_tools.egg-info/PKG-INFO writing dependency_links to tokenizer_tools.egg-info/dependency_links.txt writing entry points to tokenizer_tools.egg-info/entry_points.txt writing requirements to tokenizer_tools.egg-info/requires.txt writing top-level names to tokenizer_tools.egg-info/top_level.txt reading manifest file 'tokenizer_tools.egg-info/SOURCES.txt' reading manifest template 'MANIFEST.in' warning: no previously-included files matching '__pycache__' found under directory '*' warning: no previously-included files matching '*.py[co]' found under directory '*' warning: no files found matching '*.jpg' under directory 'docs' warning: no files found matching '*.png' under directory 'docs' warning: no files found matching '*.gif' under directory 'docs' adding license file 'LICENSE' adding license file 'AUTHORS.rst' writing manifest file 'tokenizer_tools.egg-info/SOURCES.txt' + sleep 1 + RPM_EC=0 ++ jobs -p + exit 0 Executing(%install): /bin/sh -e /var/tmp/rpm-tmp.buTT9x + umask 022 + cd /builddir/build/BUILD + '[' /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64 '!=' / ']' + rm -rf /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64 ++ dirname /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64 + mkdir -p /builddir/build/BUILDROOT + mkdir /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64 + cd tokenizer_tools-0.46.1 + CFLAGS='-O2 -g -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -grecord-gcc-switches -specs=/usr/lib/rpm/generic-hardened-cc1 -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection ' + LDFLAGS='-Wl,-z,relro -Wl,-z,now -specs=/usr/lib/rpm/generic-hardened-ld' + /usr/bin/python3 setup.py install -O1 --skip-build --root /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64 /usr/lib/python3.9/site-packages/setuptools/installer.py:27: SetuptoolsDeprecationWarning: setuptools.installer is deprecated. Requirements should be satisfied by a PEP 517 installer. warnings.warn( WARNING: The wheel package is not available. /usr/lib64/python3.9/distutils/dist.py:274: UserWarning: Unknown distribution option: 'extra_require' warnings.warn(msg) running install /usr/lib/python3.9/site-packages/setuptools/command/install.py:34: SetuptoolsDeprecationWarning: setup.py install is deprecated. Use build and pip and other standards-based tools. warnings.warn( running install_lib creating /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr creating /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib creating /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9 creating /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages creating /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools creating /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/converter copying build/lib/tokenizer_tools/converter/conllx_to_rasa.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/converter copying build/lib/tokenizer_tools/converter/offset_to_sentence.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/converter copying build/lib/tokenizer_tools/converter/sentence_to_offset.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/converter copying build/lib/tokenizer_tools/converter/conllx_to_offset.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/converter copying build/lib/tokenizer_tools/converter/__init__.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/converter copying build/lib/tokenizer_tools/split_data.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools creating /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/evaluator creating /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/evaluator/token copying build/lib/tokenizer_tools/evaluator/token/tag_level.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/evaluator/token copying build/lib/tokenizer_tools/evaluator/token/token_level.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/evaluator/token copying build/lib/tokenizer_tools/evaluator/token/base_evaluator.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/evaluator/token copying build/lib/tokenizer_tools/evaluator/token/__init__.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/evaluator/token copying build/lib/tokenizer_tools/evaluator/offset_evaluator.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/evaluator copying build/lib/tokenizer_tools/evaluator/tag_level_evaluator.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/evaluator copying build/lib/tokenizer_tools/evaluator/__init__.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/evaluator creating /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/conll copying build/lib/tokenizer_tools/conll/reader.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/conll copying build/lib/tokenizer_tools/conll/iterator_reader.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/conll copying build/lib/tokenizer_tools/conll/entity_vocabulary.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/conll copying build/lib/tokenizer_tools/conll/writer.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/conll copying build/lib/tokenizer_tools/conll/__init__.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/conll copying build/lib/tokenizer_tools/conll/tag_collector.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/conll creating /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/utils copying build/lib/tokenizer_tools/utils/batch_iterator.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/utils copying build/lib/tokenizer_tools/utils/deprecated_api.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/utils copying build/lib/tokenizer_tools/utils/generate_tagset.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/utils copying build/lib/tokenizer_tools/utils/deprecated_support.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/utils copying build/lib/tokenizer_tools/utils/__init__.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/utils copying build/lib/tokenizer_tools/utils/util_list.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/utils creating /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/conllz copying build/lib/tokenizer_tools/conllz/reader.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/conllz copying build/lib/tokenizer_tools/conllz/iterator_reader.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/conllz copying build/lib/tokenizer_tools/conllz/sentence.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/conllz copying build/lib/tokenizer_tools/conllz/multiprocess_reader.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/conllz copying build/lib/tokenizer_tools/conllz/writer.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/conllz copying build/lib/tokenizer_tools/conllz/__init__.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/conllz copying build/lib/tokenizer_tools/hooks.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools copying build/lib/tokenizer_tools/cli.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools creating /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset copying build/lib/tokenizer_tools/tagset/exceptions.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset creating /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/converter copying build/lib/tokenizer_tools/tagset/converter/__init__.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/converter copying build/lib/tokenizer_tools/tagset/converter/offset_to_biluo.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/converter creating /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/NER copying build/lib/tokenizer_tools/tagset/NER/BILUO.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/NER copying build/lib/tokenizer_tools/tagset/NER/IOB.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/NER copying build/lib/tokenizer_tools/tagset/NER/__init__.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/NER copying build/lib/tokenizer_tools/tagset/NER/base_tagset.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/NER creating /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset copying build/lib/tokenizer_tools/tagset/offset/corpus_diff.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset copying build/lib/tokenizer_tools/tagset/offset/plottable_counter.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset copying build/lib/tokenizer_tools/tagset/offset/exceptions.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset copying build/lib/tokenizer_tools/tagset/offset/span_set.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset copying build/lib/tokenizer_tools/tagset/offset/corpus.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset copying build/lib/tokenizer_tools/tagset/offset/corpus_pb2.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset copying build/lib/tokenizer_tools/tagset/offset/corpus_statistics.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset copying build/lib/tokenizer_tools/tagset/offset/span.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset copying build/lib/tokenizer_tools/tagset/offset/sequence.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset copying build/lib/tokenizer_tools/tagset/offset/corpus_metric.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset copying build/lib/tokenizer_tools/tagset/offset/corpus_stream_writer.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset creating /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/analysis copying build/lib/tokenizer_tools/tagset/offset/analysis/entity_placeholder.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/analysis copying build/lib/tokenizer_tools/tagset/offset/analysis/read_markdown.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/analysis copying build/lib/tokenizer_tools/tagset/offset/analysis/express_pattern.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/analysis copying build/lib/tokenizer_tools/tagset/offset/analysis/document_pattern.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/analysis copying build/lib/tokenizer_tools/tagset/offset/analysis/corpus_pattern.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/analysis copying build/lib/tokenizer_tools/tagset/offset/analysis/plain_text.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/analysis copying build/lib/tokenizer_tools/tagset/offset/analysis/__init__.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/analysis creating /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/diagnose copying build/lib/tokenizer_tools/tagset/offset/diagnose/entity_disagreement.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/diagnose copying build/lib/tokenizer_tools/tagset/offset/diagnose/__init__.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/diagnose copying build/lib/tokenizer_tools/tagset/offset/transform.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset copying build/lib/tokenizer_tools/tagset/offset/language_model.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset copying build/lib/tokenizer_tools/tagset/offset/corpus_stream_reader.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset copying build/lib/tokenizer_tools/tagset/offset/corpus_protobuf.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset copying build/lib/tokenizer_tools/tagset/offset/document_compare_ways.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset copying build/lib/tokenizer_tools/tagset/offset/document_block_view.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset creating /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/dashboard copying build/lib/tokenizer_tools/tagset/offset/dashboard/main.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/dashboard copying build/lib/tokenizer_tools/tagset/offset/dashboard/es_query.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/dashboard copying build/lib/tokenizer_tools/tagset/offset/dashboard/text_perplexity.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/dashboard copying build/lib/tokenizer_tools/tagset/offset/dashboard/__init__.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/dashboard copying build/lib/tokenizer_tools/tagset/offset/dashboard/es_create.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/dashboard copying build/lib/tokenizer_tools/tagset/offset/dashboard/mistagged_span.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/dashboard copying build/lib/tokenizer_tools/tagset/offset/corpus_set.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset copying build/lib/tokenizer_tools/tagset/offset/seqeval.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset copying build/lib/tokenizer_tools/tagset/offset/__init__.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset copying build/lib/tokenizer_tools/tagset/offset/document.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset copying build/lib/tokenizer_tools/tagset/offset/speed_compare_between_pb_and_conllx.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset copying build/lib/tokenizer_tools/tagset/BMES.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset copying build/lib/tokenizer_tools/tagset/__init__.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset copying build/lib/tokenizer_tools/__init__.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools copying build/lib/tokenizer_tools/tokenizer_tools.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools creating /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/format_converter copying build/lib/tokenizer_tools/format_converter/corpus_to_rasa_json.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/format_converter copying build/lib/tokenizer_tools/format_converter/__init__.py -> /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/format_converter byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/converter/conllx_to_rasa.py to conllx_to_rasa.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/converter/offset_to_sentence.py to offset_to_sentence.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/converter/sentence_to_offset.py to sentence_to_offset.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/converter/conllx_to_offset.py to conllx_to_offset.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/converter/__init__.py to __init__.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/split_data.py to split_data.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/evaluator/token/tag_level.py to tag_level.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/evaluator/token/token_level.py to token_level.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/evaluator/token/base_evaluator.py to base_evaluator.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/evaluator/token/__init__.py to __init__.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/evaluator/offset_evaluator.py to offset_evaluator.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/evaluator/tag_level_evaluator.py to tag_level_evaluator.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/evaluator/__init__.py to __init__.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/conll/reader.py to reader.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/conll/iterator_reader.py to iterator_reader.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/conll/entity_vocabulary.py to entity_vocabulary.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/conll/writer.py to writer.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/conll/__init__.py to __init__.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/conll/tag_collector.py to tag_collector.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/utils/batch_iterator.py to batch_iterator.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/utils/deprecated_api.py to deprecated_api.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/utils/generate_tagset.py to generate_tagset.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/utils/deprecated_support.py to deprecated_support.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/utils/__init__.py to __init__.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/utils/util_list.py to util_list.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/conllz/reader.py to reader.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/conllz/iterator_reader.py to iterator_reader.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/conllz/sentence.py to sentence.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/conllz/multiprocess_reader.py to multiprocess_reader.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/conllz/writer.py to writer.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/conllz/__init__.py to __init__.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/hooks.py to hooks.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/cli.py to cli.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/exceptions.py to exceptions.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/converter/__init__.py to __init__.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/converter/offset_to_biluo.py to offset_to_biluo.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/NER/BILUO.py to BILUO.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/NER/IOB.py to IOB.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/NER/__init__.py to __init__.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/NER/base_tagset.py to base_tagset.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/corpus_diff.py to corpus_diff.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/plottable_counter.py to plottable_counter.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/exceptions.py to exceptions.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/span_set.py to span_set.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/corpus.py to corpus.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/corpus_pb2.py to corpus_pb2.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/corpus_statistics.py to corpus_statistics.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/span.py to span.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/sequence.py to sequence.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/corpus_metric.py to corpus_metric.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/corpus_stream_writer.py to corpus_stream_writer.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/analysis/entity_placeholder.py to entity_placeholder.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/analysis/read_markdown.py to read_markdown.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/analysis/express_pattern.py to express_pattern.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/analysis/document_pattern.py to document_pattern.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/analysis/corpus_pattern.py to corpus_pattern.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/analysis/plain_text.py to plain_text.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/analysis/__init__.py to __init__.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/diagnose/entity_disagreement.py to entity_disagreement.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/diagnose/__init__.py to __init__.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/transform.py to transform.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/language_model.py to language_model.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/corpus_stream_reader.py to corpus_stream_reader.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/corpus_protobuf.py to corpus_protobuf.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/document_compare_ways.py to document_compare_ways.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/document_block_view.py to document_block_view.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/dashboard/main.py to main.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/dashboard/es_query.py to es_query.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/dashboard/text_perplexity.py to text_perplexity.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/dashboard/__init__.py to __init__.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/dashboard/es_create.py to es_create.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/dashboard/mistagged_span.py to mistagged_span.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/corpus_set.py to corpus_set.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/seqeval.py to seqeval.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/__init__.py to __init__.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/document.py to document.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/offset/speed_compare_between_pb_and_conllx.py to speed_compare_between_pb_and_conllx.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/BMES.py to BMES.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tagset/__init__.py to __init__.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/__init__.py to __init__.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/tokenizer_tools.py to tokenizer_tools.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/format_converter/corpus_to_rasa_json.py to corpus_to_rasa_json.cpython-39.pyc byte-compiling /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools/format_converter/__init__.py to __init__.cpython-39.pyc writing byte-compilation script '/tmp/tmpsegq4v30.py' /usr/bin/python3 /tmp/tmpsegq4v30.py removing /tmp/tmpsegq4v30.py running install_egg_info running egg_info writing tokenizer_tools.egg-info/PKG-INFO writing dependency_links to tokenizer_tools.egg-info/dependency_links.txt writing entry points to tokenizer_tools.egg-info/entry_points.txt writing requirements to tokenizer_tools.egg-info/requires.txt writing top-level names to tokenizer_tools.egg-info/top_level.txt reading manifest file 'tokenizer_tools.egg-info/SOURCES.txt' reading manifest template 'MANIFEST.in' warning: no previously-included files matching '__pycache__' found under directory '*' warning: no previously-included files matching '*.py[co]' found under directory '*' warning: no files found matching '*.jpg' under directory 'docs' warning: no files found matching '*.png' under directory 'docs' warning: no files found matching '*.gif' under directory 'docs' adding license file 'LICENSE' adding license file 'AUTHORS.rst' writing manifest file 'tokenizer_tools.egg-info/SOURCES.txt' Copying tokenizer_tools.egg-info to /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9/site-packages/tokenizer_tools-0.46.1-py3.9.egg-info running install_scripts Installing tokenizer_tools script to /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/bin + install -d -m755 /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64//usr/share/doc/python-tokenizer-tools + '[' -d doc ']' + '[' -d docs ']' + cp -arf docs /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64//usr/share/doc/python-tokenizer-tools + '[' -d example ']' ~/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64 ~/build/BUILD/tokenizer_tools-0.46.1 + '[' -d examples ']' + pushd /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64 + '[' -d usr/lib ']' + find usr/lib -type f -printf '"/%h/%f"\n' + '[' -d usr/lib64 ']' + '[' -d usr/bin ']' + find usr/bin -type f -printf '"/%h/%f"\n' + '[' -d usr/sbin ']' + touch doclist.lst + '[' -d usr/share/man ']' + popd ~/build/BUILD/tokenizer_tools-0.46.1 + mv /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/filelist.lst . + mv /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/doclist.lst . + /usr/bin/find-debuginfo -j4 --strict-build-id -i --build-id-seed 0.46.1-1 --unique-debug-suffix -0.46.1-1.x86_64 --unique-debug-src-base python-tokenizer-tools-0.46.1-1.x86_64 -S debugsourcefiles.list /builddir/build/BUILD/tokenizer_tools-0.46.1 find: 'debug': No such file or directory + /usr/lib/rpm/check-buildroot + /usr/lib/rpm/brp-ldconfig + /usr/lib/rpm/brp-compress + /usr/lib/rpm/brp-strip-static-archive /usr/bin/strip + /usr/lib/rpm/brp-python-bytecompile /usr/bin/python 1 1 Bytecompiling .py files below /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64/usr/lib/python3.9 using /usr/bin/python3.9 + /usr/lib/rpm/brp-python-hardlink Processing files: python3-tokenizer-tools-0.46.1-1.noarch Provides: python-tokenizer-tools python3-tokenizer-tools = 0.46.1-1 python3.9dist(tokenizer-tools) = 0.46.1 python3dist(tokenizer-tools) = 0.46.1 Requires(rpmlib): rpmlib(CompressedFileNames) <= 3.0.4-1 rpmlib(FileDigests) <= 4.6.0-1 rpmlib(PartialHardlinkSets) <= 4.0.4-1 rpmlib(PayloadFilesHavePrefix) <= 4.0-1 Requires: /usr/bin/python3 python(abi) = 3.9 python3.9dist(click) >= 6 python3.9dist(elasticsearch-dsl) python3.9dist(flask) python3.9dist(fuzzywuzzy) python3.9dist(mistletoe) python3.9dist(pandas) python3.9dist(plotly) python3.9dist(pyahocorasick) python3.9dist(scikit-learn) python3.9dist(scipy) python3.9dist(setuptools) python3.9dist(tqdm) Processing files: python-tokenizer-tools-help-0.46.1-1.noarch warning: Empty %files file /builddir/build/BUILD/tokenizer_tools-0.46.1/doclist.lst Provides: python-tokenizer-tools-help = 0.46.1-1 python3-tokenizer-tools-doc Requires(rpmlib): rpmlib(CompressedFileNames) <= 3.0.4-1 rpmlib(FileDigests) <= 4.6.0-1 rpmlib(PayloadFilesHavePrefix) <= 4.0-1 Checking for unpackaged file(s): /usr/lib/rpm/check-files /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64 Wrote: /builddir/build/RPMS/python-tokenizer-tools-help-0.46.1-1.noarch.rpm Wrote: /builddir/build/RPMS/python3-tokenizer-tools-0.46.1-1.noarch.rpm Executing(%clean): /bin/sh -e /var/tmp/rpm-tmp.m3cPZJ + umask 022 + cd /builddir/build/BUILD + cd tokenizer_tools-0.46.1 + /usr/bin/rm -rf /builddir/build/BUILDROOT/python-tokenizer-tools-0.46.1-1.x86_64 + RPM_EC=0 ++ jobs -p + exit 0 Child return code was: 0