%global _empty_manifest_terminate_build 0 Name: python-pytest-elk-reporter Version: 0.2.2 Release: 1 Summary: A simple plugin to use with pytest License: MIT URL: https://github.com/fruch/pytest-elk-reporter Source0: https://mirrors.nju.edu.cn/pypi/web/packages/73/d4/bdfedae7f96cec9cbe6c28b0c7fb3e469e8a54d8d2f9021280f95e972497/pytest-elk-reporter-0.2.2.tar.gz BuildArch: noarch Requires: python3-pytest Requires: python3-requests Requires: python3-six %description # pytest-elk-reporter [![PyPI version](https://img.shields.io/pypi/v/pytest-elk-reporter.svg?style=flat)](https://pypi.org/project/pytest-elk-reporter) [![Python versions](https://img.shields.io/pypi/pyversions/pytest-elk-reporter.svg?style=flat)](https://pypi.org/project/pytest-elk-reporter) [![.github/workflows/tests.yml](https://github.com/fruch/pytest-elk-reporter/workflows/.github/workflows/tests.yml/badge.svg)](https://github.com/fruch/pytest-elk-reporter/actions?query=branch%3Amaster) [![Libraries.io dependency status for GitHub repo](https://img.shields.io/librariesio/github/fruch/pytest-elk-reporter.svg?style=flat)](https://libraries.io/github/fruch/pytest-elk-reporter) [![Using Black](https://img.shields.io/badge/code%20style-black-000000.svg)](https://github.com/python/black) [![Codecov Reports](https://codecov.io/gh/fruch/pytest-elk-reporter/branch/master/graph/badge.svg)](https://codecov.io/gh/fruch/pytest-elk-reporter) A plugin to send pytest test results to [ELK] stack, with extra context data ## Features * Reporting into Elasticsearch each test result, as the test finish * Automatically append context data to each test: * git information such as `branch` or `last commit` and more * all of CI env variables * Jenkins * Travis * Circle CI * Github Actions * username if available * Report a test summery to Elastic for each session with all the context data * can append any user data into the context sent to elastic ## Requirements * having [pytest] tests written ## Installation You can install "pytest-elk-reporter" via [pip] from [PyPI] ``` bash pip install pytest-elk-reporter ``` ### ElasticSearch configuration We need this auto_create_index enable for the indexes that are going to be used, since we don't have code to create the indexes, this is the default ```bash curl -X PUT "localhost:9200/_cluster/settings" -H 'Content-Type: application/json' -d' { "persistent": { "action.auto_create_index": "true" } } ' ``` For more info on this elasticsearch feature check their [index documention](https://www.elastic.co/guide/en/elasticsearch/reference/current/docs-index_.html#index-creation) ## Usage ### Run and configure from command line ```bash pytest --es-address 127.0.0.1:9200 # or if you need user/password to authenticate pytest --es-address my-elk-server.io:9200 --es-username fruch --es-password 'passwordsarenicetohave' ``` ### Configure from code (ideally in conftest.py) ```python from pytest_elk_reporter import ElkReporter def pytest_plugin_registered(plugin, manager): if isinstance(plugin, ElkReporter): # TODO: get credentials in more secure fashion programmatically, maybe AWS secrets or the likes # or put them in plain-text in the code... what can ever go wrong... plugin.es_address = "my-elk-server.io:9200" plugin.es_user = 'fruch' plugin.es_password = 'passwordsarenicetohave' plugin.es_index_name = 'test_data' ``` ### Configure from pytest ini file ```ini # put this in pytest.ini / tox.ini / setup.cfg [pytest] es_address = my-elk-server.io:9200 es_user = fruch es_password = passwordsarenicetohave es_index_name = test_data ``` see [pytest docs](https://docs.pytest.org/en/latest/customize.html) for more about how to configure using .ini files ### Collect context data for the whole session For example, with this I'll be able to build a dash board per version ```python import pytest @pytest.fixture(scope="session", autouse=True) def report_formal_version_to_elk(request): """ Append my own data specific, for example which of the code uner test is used """ # TODO: take it programticly of of the code under test... my_data = {"formal_version": "1.0.0-rc2" } elk = request.config.pluginmanager.get_plugin("elk-reporter-runtime") elk.session_data.update(**my_data) ``` ### Collect data for specific tests ```python import requests def test_my_service_and_collect_timings(request, elk_reporter): response = requests.get("http://my-server.io/api/do_something") assert response.status_code == 200 elk_reporter.append_test_data(request, {"do_something_response_time": response.elapsed.total_seconds() }) # now doing response time per version dashboard quite easy # and yeah, it's not exactly real usable metric, it's just an example... ``` Or via `record_property` built-in fixture (that is normally used to collect data into the junitxml): ```python import requests def test_my_service_and_collect_timings(record_property): response = requests.get("http://my-server.io/api/do_something") assert response.status_code == 200 record_property("do_something_response_time", response.elapsed.total_seconds()) # now doing response time per version dashboard quite easy # and yeah, it's not exactly real usable metric, it's just an example... ``` ## Split tests base on history Cool thing that can be done now that you have history of the tests is to split the test base on their actually runtime when passing. for integration test which might be long (minutes to hours), this would be priceless. In this example we going to split the run in maximum 4min slices while any test that doesn't have history information would be assumed to be 60sec long ```bash # pytest --collect-only --es-splice --es-max-splice-time=4 --es-default-test-time=60 ... 0: 0:04:00 - 3 - ['test_history_slices.py::test_should_pass_1', 'test_history_slices.py::test_should_pass_2', 'test_history_slices.py::test_should_pass_3'] 1: 0:04:00 - 2 - ['test_history_slices.py::test_with_history_data', 'test_history_slices.py::test_that_failed'] ... # cat include000.txt test_history_slices.py::test_should_pass_1 test_history_slices.py::test_should_pass_2 test_history_slices.py::test_should_pass_3 # cat include000.txt test_history_slices.py::test_with_history_data test_history_slices.py::test_that_failed ### now we can run the each slice on it's own machines ### on machine1 # pytest $(cat include000.txt) ### on machine2 # pytest $(cat include001.txt) ``` ## Contributing Contributions are very welcome. Tests can be run with [`tox`][tox], please ensure the coverage at least stays the same before you submit a pull request. ## License Distributed under the terms of the [MIT][MIT] license, "pytest-elk-reporter" is free and open source software ## Issues If you encounter any problems, please [file an issue] along with a detailed description. ## Thanks This [pytest] plugin was generated with [Cookiecutter] along with [@hackebrot]'s [cookiecutter-pytest-plugin] template. [ELK]: https://www.elastic.co/elk-stack [Cookiecutter]: https://github.com/audreyr/cookiecutter [@hackebrot]: https://github.com/hackebrot [MIT]: http://opensource.org/licenses/MIT [cookiecutter-pytest-plugin]: https://github.com/pytest-dev/cookiecutter-pytest-plugin [file an issue]: https://github.com/fruch/pytest-elk-reporter/issues [pytest]: https://github.com/pytest-dev/pytest [tox]: https://tox.readthedocs.io/en/latest/ [pip]: https://pypi.org/project/pip/ [PyPI]: https://pypi.org/project %package -n python3-pytest-elk-reporter Summary: A simple plugin to use with pytest Provides: python-pytest-elk-reporter BuildRequires: python3-devel BuildRequires: python3-setuptools BuildRequires: python3-pip %description -n python3-pytest-elk-reporter # pytest-elk-reporter [![PyPI version](https://img.shields.io/pypi/v/pytest-elk-reporter.svg?style=flat)](https://pypi.org/project/pytest-elk-reporter) [![Python versions](https://img.shields.io/pypi/pyversions/pytest-elk-reporter.svg?style=flat)](https://pypi.org/project/pytest-elk-reporter) [![.github/workflows/tests.yml](https://github.com/fruch/pytest-elk-reporter/workflows/.github/workflows/tests.yml/badge.svg)](https://github.com/fruch/pytest-elk-reporter/actions?query=branch%3Amaster) [![Libraries.io dependency status for GitHub repo](https://img.shields.io/librariesio/github/fruch/pytest-elk-reporter.svg?style=flat)](https://libraries.io/github/fruch/pytest-elk-reporter) [![Using Black](https://img.shields.io/badge/code%20style-black-000000.svg)](https://github.com/python/black) [![Codecov Reports](https://codecov.io/gh/fruch/pytest-elk-reporter/branch/master/graph/badge.svg)](https://codecov.io/gh/fruch/pytest-elk-reporter) A plugin to send pytest test results to [ELK] stack, with extra context data ## Features * Reporting into Elasticsearch each test result, as the test finish * Automatically append context data to each test: * git information such as `branch` or `last commit` and more * all of CI env variables * Jenkins * Travis * Circle CI * Github Actions * username if available * Report a test summery to Elastic for each session with all the context data * can append any user data into the context sent to elastic ## Requirements * having [pytest] tests written ## Installation You can install "pytest-elk-reporter" via [pip] from [PyPI] ``` bash pip install pytest-elk-reporter ``` ### ElasticSearch configuration We need this auto_create_index enable for the indexes that are going to be used, since we don't have code to create the indexes, this is the default ```bash curl -X PUT "localhost:9200/_cluster/settings" -H 'Content-Type: application/json' -d' { "persistent": { "action.auto_create_index": "true" } } ' ``` For more info on this elasticsearch feature check their [index documention](https://www.elastic.co/guide/en/elasticsearch/reference/current/docs-index_.html#index-creation) ## Usage ### Run and configure from command line ```bash pytest --es-address 127.0.0.1:9200 # or if you need user/password to authenticate pytest --es-address my-elk-server.io:9200 --es-username fruch --es-password 'passwordsarenicetohave' ``` ### Configure from code (ideally in conftest.py) ```python from pytest_elk_reporter import ElkReporter def pytest_plugin_registered(plugin, manager): if isinstance(plugin, ElkReporter): # TODO: get credentials in more secure fashion programmatically, maybe AWS secrets or the likes # or put them in plain-text in the code... what can ever go wrong... plugin.es_address = "my-elk-server.io:9200" plugin.es_user = 'fruch' plugin.es_password = 'passwordsarenicetohave' plugin.es_index_name = 'test_data' ``` ### Configure from pytest ini file ```ini # put this in pytest.ini / tox.ini / setup.cfg [pytest] es_address = my-elk-server.io:9200 es_user = fruch es_password = passwordsarenicetohave es_index_name = test_data ``` see [pytest docs](https://docs.pytest.org/en/latest/customize.html) for more about how to configure using .ini files ### Collect context data for the whole session For example, with this I'll be able to build a dash board per version ```python import pytest @pytest.fixture(scope="session", autouse=True) def report_formal_version_to_elk(request): """ Append my own data specific, for example which of the code uner test is used """ # TODO: take it programticly of of the code under test... my_data = {"formal_version": "1.0.0-rc2" } elk = request.config.pluginmanager.get_plugin("elk-reporter-runtime") elk.session_data.update(**my_data) ``` ### Collect data for specific tests ```python import requests def test_my_service_and_collect_timings(request, elk_reporter): response = requests.get("http://my-server.io/api/do_something") assert response.status_code == 200 elk_reporter.append_test_data(request, {"do_something_response_time": response.elapsed.total_seconds() }) # now doing response time per version dashboard quite easy # and yeah, it's not exactly real usable metric, it's just an example... ``` Or via `record_property` built-in fixture (that is normally used to collect data into the junitxml): ```python import requests def test_my_service_and_collect_timings(record_property): response = requests.get("http://my-server.io/api/do_something") assert response.status_code == 200 record_property("do_something_response_time", response.elapsed.total_seconds()) # now doing response time per version dashboard quite easy # and yeah, it's not exactly real usable metric, it's just an example... ``` ## Split tests base on history Cool thing that can be done now that you have history of the tests is to split the test base on their actually runtime when passing. for integration test which might be long (minutes to hours), this would be priceless. In this example we going to split the run in maximum 4min slices while any test that doesn't have history information would be assumed to be 60sec long ```bash # pytest --collect-only --es-splice --es-max-splice-time=4 --es-default-test-time=60 ... 0: 0:04:00 - 3 - ['test_history_slices.py::test_should_pass_1', 'test_history_slices.py::test_should_pass_2', 'test_history_slices.py::test_should_pass_3'] 1: 0:04:00 - 2 - ['test_history_slices.py::test_with_history_data', 'test_history_slices.py::test_that_failed'] ... # cat include000.txt test_history_slices.py::test_should_pass_1 test_history_slices.py::test_should_pass_2 test_history_slices.py::test_should_pass_3 # cat include000.txt test_history_slices.py::test_with_history_data test_history_slices.py::test_that_failed ### now we can run the each slice on it's own machines ### on machine1 # pytest $(cat include000.txt) ### on machine2 # pytest $(cat include001.txt) ``` ## Contributing Contributions are very welcome. Tests can be run with [`tox`][tox], please ensure the coverage at least stays the same before you submit a pull request. ## License Distributed under the terms of the [MIT][MIT] license, "pytest-elk-reporter" is free and open source software ## Issues If you encounter any problems, please [file an issue] along with a detailed description. ## Thanks This [pytest] plugin was generated with [Cookiecutter] along with [@hackebrot]'s [cookiecutter-pytest-plugin] template. [ELK]: https://www.elastic.co/elk-stack [Cookiecutter]: https://github.com/audreyr/cookiecutter [@hackebrot]: https://github.com/hackebrot [MIT]: http://opensource.org/licenses/MIT [cookiecutter-pytest-plugin]: https://github.com/pytest-dev/cookiecutter-pytest-plugin [file an issue]: https://github.com/fruch/pytest-elk-reporter/issues [pytest]: https://github.com/pytest-dev/pytest [tox]: https://tox.readthedocs.io/en/latest/ [pip]: https://pypi.org/project/pip/ [PyPI]: https://pypi.org/project %package help Summary: Development documents and examples for pytest-elk-reporter Provides: python3-pytest-elk-reporter-doc %description help # pytest-elk-reporter [![PyPI version](https://img.shields.io/pypi/v/pytest-elk-reporter.svg?style=flat)](https://pypi.org/project/pytest-elk-reporter) [![Python versions](https://img.shields.io/pypi/pyversions/pytest-elk-reporter.svg?style=flat)](https://pypi.org/project/pytest-elk-reporter) [![.github/workflows/tests.yml](https://github.com/fruch/pytest-elk-reporter/workflows/.github/workflows/tests.yml/badge.svg)](https://github.com/fruch/pytest-elk-reporter/actions?query=branch%3Amaster) [![Libraries.io dependency status for GitHub repo](https://img.shields.io/librariesio/github/fruch/pytest-elk-reporter.svg?style=flat)](https://libraries.io/github/fruch/pytest-elk-reporter) [![Using Black](https://img.shields.io/badge/code%20style-black-000000.svg)](https://github.com/python/black) [![Codecov Reports](https://codecov.io/gh/fruch/pytest-elk-reporter/branch/master/graph/badge.svg)](https://codecov.io/gh/fruch/pytest-elk-reporter) A plugin to send pytest test results to [ELK] stack, with extra context data ## Features * Reporting into Elasticsearch each test result, as the test finish * Automatically append context data to each test: * git information such as `branch` or `last commit` and more * all of CI env variables * Jenkins * Travis * Circle CI * Github Actions * username if available * Report a test summery to Elastic for each session with all the context data * can append any user data into the context sent to elastic ## Requirements * having [pytest] tests written ## Installation You can install "pytest-elk-reporter" via [pip] from [PyPI] ``` bash pip install pytest-elk-reporter ``` ### ElasticSearch configuration We need this auto_create_index enable for the indexes that are going to be used, since we don't have code to create the indexes, this is the default ```bash curl -X PUT "localhost:9200/_cluster/settings" -H 'Content-Type: application/json' -d' { "persistent": { "action.auto_create_index": "true" } } ' ``` For more info on this elasticsearch feature check their [index documention](https://www.elastic.co/guide/en/elasticsearch/reference/current/docs-index_.html#index-creation) ## Usage ### Run and configure from command line ```bash pytest --es-address 127.0.0.1:9200 # or if you need user/password to authenticate pytest --es-address my-elk-server.io:9200 --es-username fruch --es-password 'passwordsarenicetohave' ``` ### Configure from code (ideally in conftest.py) ```python from pytest_elk_reporter import ElkReporter def pytest_plugin_registered(plugin, manager): if isinstance(plugin, ElkReporter): # TODO: get credentials in more secure fashion programmatically, maybe AWS secrets or the likes # or put them in plain-text in the code... what can ever go wrong... plugin.es_address = "my-elk-server.io:9200" plugin.es_user = 'fruch' plugin.es_password = 'passwordsarenicetohave' plugin.es_index_name = 'test_data' ``` ### Configure from pytest ini file ```ini # put this in pytest.ini / tox.ini / setup.cfg [pytest] es_address = my-elk-server.io:9200 es_user = fruch es_password = passwordsarenicetohave es_index_name = test_data ``` see [pytest docs](https://docs.pytest.org/en/latest/customize.html) for more about how to configure using .ini files ### Collect context data for the whole session For example, with this I'll be able to build a dash board per version ```python import pytest @pytest.fixture(scope="session", autouse=True) def report_formal_version_to_elk(request): """ Append my own data specific, for example which of the code uner test is used """ # TODO: take it programticly of of the code under test... my_data = {"formal_version": "1.0.0-rc2" } elk = request.config.pluginmanager.get_plugin("elk-reporter-runtime") elk.session_data.update(**my_data) ``` ### Collect data for specific tests ```python import requests def test_my_service_and_collect_timings(request, elk_reporter): response = requests.get("http://my-server.io/api/do_something") assert response.status_code == 200 elk_reporter.append_test_data(request, {"do_something_response_time": response.elapsed.total_seconds() }) # now doing response time per version dashboard quite easy # and yeah, it's not exactly real usable metric, it's just an example... ``` Or via `record_property` built-in fixture (that is normally used to collect data into the junitxml): ```python import requests def test_my_service_and_collect_timings(record_property): response = requests.get("http://my-server.io/api/do_something") assert response.status_code == 200 record_property("do_something_response_time", response.elapsed.total_seconds()) # now doing response time per version dashboard quite easy # and yeah, it's not exactly real usable metric, it's just an example... ``` ## Split tests base on history Cool thing that can be done now that you have history of the tests is to split the test base on their actually runtime when passing. for integration test which might be long (minutes to hours), this would be priceless. In this example we going to split the run in maximum 4min slices while any test that doesn't have history information would be assumed to be 60sec long ```bash # pytest --collect-only --es-splice --es-max-splice-time=4 --es-default-test-time=60 ... 0: 0:04:00 - 3 - ['test_history_slices.py::test_should_pass_1', 'test_history_slices.py::test_should_pass_2', 'test_history_slices.py::test_should_pass_3'] 1: 0:04:00 - 2 - ['test_history_slices.py::test_with_history_data', 'test_history_slices.py::test_that_failed'] ... # cat include000.txt test_history_slices.py::test_should_pass_1 test_history_slices.py::test_should_pass_2 test_history_slices.py::test_should_pass_3 # cat include000.txt test_history_slices.py::test_with_history_data test_history_slices.py::test_that_failed ### now we can run the each slice on it's own machines ### on machine1 # pytest $(cat include000.txt) ### on machine2 # pytest $(cat include001.txt) ``` ## Contributing Contributions are very welcome. Tests can be run with [`tox`][tox], please ensure the coverage at least stays the same before you submit a pull request. ## License Distributed under the terms of the [MIT][MIT] license, "pytest-elk-reporter" is free and open source software ## Issues If you encounter any problems, please [file an issue] along with a detailed description. ## Thanks This [pytest] plugin was generated with [Cookiecutter] along with [@hackebrot]'s [cookiecutter-pytest-plugin] template. [ELK]: https://www.elastic.co/elk-stack [Cookiecutter]: https://github.com/audreyr/cookiecutter [@hackebrot]: https://github.com/hackebrot [MIT]: http://opensource.org/licenses/MIT [cookiecutter-pytest-plugin]: https://github.com/pytest-dev/cookiecutter-pytest-plugin [file an issue]: https://github.com/fruch/pytest-elk-reporter/issues [pytest]: https://github.com/pytest-dev/pytest [tox]: https://tox.readthedocs.io/en/latest/ [pip]: https://pypi.org/project/pip/ [PyPI]: https://pypi.org/project %prep %autosetup -n pytest-elk-reporter-0.2.2 %build %py3_build %install %py3_install install -d -m755 %{buildroot}/%{_pkgdocdir} if [ -d doc ]; then cp -arf doc %{buildroot}/%{_pkgdocdir}; fi if [ -d docs ]; then cp -arf docs %{buildroot}/%{_pkgdocdir}; fi if [ -d example ]; then cp -arf example %{buildroot}/%{_pkgdocdir}; fi if [ -d examples ]; then cp -arf examples %{buildroot}/%{_pkgdocdir}; fi pushd %{buildroot} if [ -d usr/lib ]; then find usr/lib -type f -printf "/%h/%f\n" >> filelist.lst fi if [ -d usr/lib64 ]; then find usr/lib64 -type f -printf "/%h/%f\n" >> filelist.lst fi if [ -d usr/bin ]; then find usr/bin -type f -printf "/%h/%f\n" >> filelist.lst fi if [ -d usr/sbin ]; then find usr/sbin -type f -printf "/%h/%f\n" >> filelist.lst fi touch doclist.lst if [ -d usr/share/man ]; then find usr/share/man -type f -printf "/%h/%f.gz\n" >> doclist.lst fi popd mv %{buildroot}/filelist.lst . mv %{buildroot}/doclist.lst . %files -n python3-pytest-elk-reporter -f filelist.lst %dir %{python3_sitelib}/* %files help -f doclist.lst %{_docdir}/* %changelog * Wed May 10 2023 Python_Bot - 0.2.2-1 - Package Spec generated