diff options
| author | CoprDistGit <infra@openeuler.org> | 2023-05-15 03:33:42 +0000 |
|---|---|---|
| committer | CoprDistGit <infra@openeuler.org> | 2023-05-15 03:33:42 +0000 |
| commit | 4d9a7d442f41d1eb883a7402b1febde399405e92 (patch) | |
| tree | a742f7249323ba97b05120feb3190ad228feca71 /python-tqdm-multiprocess.spec | |
| parent | aef06cc18b3bcf489a6f3aef98b799356074b1a5 (diff) | |
automatic import of python-tqdm-multiprocess
Diffstat (limited to 'python-tqdm-multiprocess.spec')
| -rw-r--r-- | python-tqdm-multiprocess.spec | 372 |
1 files changed, 372 insertions, 0 deletions
diff --git a/python-tqdm-multiprocess.spec b/python-tqdm-multiprocess.spec new file mode 100644 index 0000000..4b0caa9 --- /dev/null +++ b/python-tqdm-multiprocess.spec @@ -0,0 +1,372 @@ +%global _empty_manifest_terminate_build 0 +Name: python-tqdm-multiprocess +Version: 0.0.11 +Release: 1 +Summary: Easy multiprocessing with tqdm and logging redirected to main process. +License: MIT License +URL: https://github.com/EleutherAI/tqdm-multiprocess +Source0: https://mirrors.nju.edu.cn/pypi/web/packages/b4/1e/de81bd0f6cb2b61d6ee7ccbf304d99a42a0f53879481536dfb3288ee9a87/tqdm-multiprocess-0.0.11.tar.gz +BuildArch: noarch + +Requires: python3-tqdm +Requires: python3-colorama +Requires: python3-twine + +%description +# tqdm-multiprocess +Using queues, tqdm-multiprocess supports multiple worker processes, each with multiple tqdm progress bars, displaying them cleanly through the main process. The worker processes also have access to a single global tqdm for aggregate progress monitoring. + +Logging is also redirected from the subprocesses to the root logger in the main process. + +Currently doesn't support tqdm(iterator), you will need to intialize your worker tqdms with a total and update manually. + +Due to the performance limits of the default Python multiprocess queue you need to update your global and worker process tqdms infrequently to avoid flooding the main process. I will attempt to implement a lock free ringbuffer at some point to see if things can be improved. + +## Installation + +```bash +pip install tqdm-multiprocess +``` + +## Usage + +*TqdmMultiProcessPool* creates a standard python multiprocessing pool with the desired number of processes. Under the hood it uses async_apply with an event loop to monitor a tqdm and logging queue, allowing the worker processes to redirect both their tqdm objects and logging messages to your main process. There is also a queue for the workers to update the single global tqdm. + +As shown below, you create a list of tasks containing their function and a tuple with your parameters. The functions you pass in will need the extra arguments on the end "tqdm_func, global_tqdm". You must use tqdm_func when initializing your tqdms for the redirection to work. As mentioned above, passing iterators into the tqdm function is currently not supported, so set total=total_steps when setting up your tqdm, and then update the progress manually with the update() method. All other arguments to tqdm should work fine. + +Once you have your task list, call the map() method on your pool, passing in the process count, global_tqdm (or None), task list, as well as error and done callback functions. The error callback will be trigerred if your task functions return anything evaluating as False (if not task_result in the source code). The done callback will be called when the task succesfully completes. + +The map method returns a list containing the returned results for all your tasks in original order. + +### examples/basic_example.py + +```python +from time import sleep +import multiprocessing +import tqdm + +import logging +from tqdm_multiprocess.logger import setup_logger_tqdm +logger = logging.getLogger(__name__) + +from tqdm_multiprocess import TqdmMultiProcessPool + +iterations1 = 100 +iterations2 = 5 +iterations3 = 2 +def some_other_function(tqdm_func, global_tqdm): + + total_iterations = iterations1 * iterations2 * iterations3 + with tqdm_func(total=total_iterations, dynamic_ncols=True) as progress3: + progress3.set_description("outer") + for i in range(iterations3): + logger.info("outer") + total_iterations = iterations1 * iterations2 + with tqdm_func(total=total_iterations, dynamic_ncols=True) as progress2: + progress2.set_description("middle") + for j in range(iterations2): + logger.info("middle") + #for k in tqdm_func(range(iterations1), dynamic_ncols=True, desc="inner"): + with tqdm_func(total=iterations1, dynamic_ncols=True) as progress1: + for j in range(iterations1): + # logger.info("inner") # Spam slows down tqdm too much + progress1.set_description("innert") + sleep(0.01) + progress1.update() + progress2.update() + progress3.update() + global_tqdm.update() + + logger.warning(f"Warning test message. {multiprocessing.current_process().name}") + logger.error(f"Error test message. {multiprocessing.current_process().name}") + + +# Multiprocessed +def example_multiprocessing_function(some_input, tqdm_func, global_tqdm): + logger.debug(f"Debug test message - I won't show up in console. {multiprocessing.current_process().name}") + logger.info(f"Info test message. {multiprocessing.current_process().name}") + some_other_function(tqdm_func, global_tqdm) + return True + +def error_callback(result): + print("Error!") + +def done_callback(result): + print("Done. Result: ", result) + +def example(): + pool = TqdmMultiProcessPool() + process_count = 4 + task_count = 10 + initial_tasks = [(example_multiprocessing_function, (i,)) for i in range(task_count)] + total_iterations = iterations1 * iterations2 * iterations3 * task_count + with tqdm.tqdm(total=total_iterations, dynamic_ncols=True) as global_progress: + global_progress.set_description("global") + results = pool.map(process_count, global_progress, initial_tasks, error_callback, done_callback) + print(results) + +if __name__ == '__main__': + logfile_path = "tqdm_multiprocessing_example.log" + setup_logger_tqdm(logfile_path) # Logger will write messages using tqdm.write + example() +``` + + + + +%package -n python3-tqdm-multiprocess +Summary: Easy multiprocessing with tqdm and logging redirected to main process. +Provides: python-tqdm-multiprocess +BuildRequires: python3-devel +BuildRequires: python3-setuptools +BuildRequires: python3-pip +%description -n python3-tqdm-multiprocess +# tqdm-multiprocess +Using queues, tqdm-multiprocess supports multiple worker processes, each with multiple tqdm progress bars, displaying them cleanly through the main process. The worker processes also have access to a single global tqdm for aggregate progress monitoring. + +Logging is also redirected from the subprocesses to the root logger in the main process. + +Currently doesn't support tqdm(iterator), you will need to intialize your worker tqdms with a total and update manually. + +Due to the performance limits of the default Python multiprocess queue you need to update your global and worker process tqdms infrequently to avoid flooding the main process. I will attempt to implement a lock free ringbuffer at some point to see if things can be improved. + +## Installation + +```bash +pip install tqdm-multiprocess +``` + +## Usage + +*TqdmMultiProcessPool* creates a standard python multiprocessing pool with the desired number of processes. Under the hood it uses async_apply with an event loop to monitor a tqdm and logging queue, allowing the worker processes to redirect both their tqdm objects and logging messages to your main process. There is also a queue for the workers to update the single global tqdm. + +As shown below, you create a list of tasks containing their function and a tuple with your parameters. The functions you pass in will need the extra arguments on the end "tqdm_func, global_tqdm". You must use tqdm_func when initializing your tqdms for the redirection to work. As mentioned above, passing iterators into the tqdm function is currently not supported, so set total=total_steps when setting up your tqdm, and then update the progress manually with the update() method. All other arguments to tqdm should work fine. + +Once you have your task list, call the map() method on your pool, passing in the process count, global_tqdm (or None), task list, as well as error and done callback functions. The error callback will be trigerred if your task functions return anything evaluating as False (if not task_result in the source code). The done callback will be called when the task succesfully completes. + +The map method returns a list containing the returned results for all your tasks in original order. + +### examples/basic_example.py + +```python +from time import sleep +import multiprocessing +import tqdm + +import logging +from tqdm_multiprocess.logger import setup_logger_tqdm +logger = logging.getLogger(__name__) + +from tqdm_multiprocess import TqdmMultiProcessPool + +iterations1 = 100 +iterations2 = 5 +iterations3 = 2 +def some_other_function(tqdm_func, global_tqdm): + + total_iterations = iterations1 * iterations2 * iterations3 + with tqdm_func(total=total_iterations, dynamic_ncols=True) as progress3: + progress3.set_description("outer") + for i in range(iterations3): + logger.info("outer") + total_iterations = iterations1 * iterations2 + with tqdm_func(total=total_iterations, dynamic_ncols=True) as progress2: + progress2.set_description("middle") + for j in range(iterations2): + logger.info("middle") + #for k in tqdm_func(range(iterations1), dynamic_ncols=True, desc="inner"): + with tqdm_func(total=iterations1, dynamic_ncols=True) as progress1: + for j in range(iterations1): + # logger.info("inner") # Spam slows down tqdm too much + progress1.set_description("innert") + sleep(0.01) + progress1.update() + progress2.update() + progress3.update() + global_tqdm.update() + + logger.warning(f"Warning test message. {multiprocessing.current_process().name}") + logger.error(f"Error test message. {multiprocessing.current_process().name}") + + +# Multiprocessed +def example_multiprocessing_function(some_input, tqdm_func, global_tqdm): + logger.debug(f"Debug test message - I won't show up in console. {multiprocessing.current_process().name}") + logger.info(f"Info test message. {multiprocessing.current_process().name}") + some_other_function(tqdm_func, global_tqdm) + return True + +def error_callback(result): + print("Error!") + +def done_callback(result): + print("Done. Result: ", result) + +def example(): + pool = TqdmMultiProcessPool() + process_count = 4 + task_count = 10 + initial_tasks = [(example_multiprocessing_function, (i,)) for i in range(task_count)] + total_iterations = iterations1 * iterations2 * iterations3 * task_count + with tqdm.tqdm(total=total_iterations, dynamic_ncols=True) as global_progress: + global_progress.set_description("global") + results = pool.map(process_count, global_progress, initial_tasks, error_callback, done_callback) + print(results) + +if __name__ == '__main__': + logfile_path = "tqdm_multiprocessing_example.log" + setup_logger_tqdm(logfile_path) # Logger will write messages using tqdm.write + example() +``` + + + + +%package help +Summary: Development documents and examples for tqdm-multiprocess +Provides: python3-tqdm-multiprocess-doc +%description help +# tqdm-multiprocess +Using queues, tqdm-multiprocess supports multiple worker processes, each with multiple tqdm progress bars, displaying them cleanly through the main process. The worker processes also have access to a single global tqdm for aggregate progress monitoring. + +Logging is also redirected from the subprocesses to the root logger in the main process. + +Currently doesn't support tqdm(iterator), you will need to intialize your worker tqdms with a total and update manually. + +Due to the performance limits of the default Python multiprocess queue you need to update your global and worker process tqdms infrequently to avoid flooding the main process. I will attempt to implement a lock free ringbuffer at some point to see if things can be improved. + +## Installation + +```bash +pip install tqdm-multiprocess +``` + +## Usage + +*TqdmMultiProcessPool* creates a standard python multiprocessing pool with the desired number of processes. Under the hood it uses async_apply with an event loop to monitor a tqdm and logging queue, allowing the worker processes to redirect both their tqdm objects and logging messages to your main process. There is also a queue for the workers to update the single global tqdm. + +As shown below, you create a list of tasks containing their function and a tuple with your parameters. The functions you pass in will need the extra arguments on the end "tqdm_func, global_tqdm". You must use tqdm_func when initializing your tqdms for the redirection to work. As mentioned above, passing iterators into the tqdm function is currently not supported, so set total=total_steps when setting up your tqdm, and then update the progress manually with the update() method. All other arguments to tqdm should work fine. + +Once you have your task list, call the map() method on your pool, passing in the process count, global_tqdm (or None), task list, as well as error and done callback functions. The error callback will be trigerred if your task functions return anything evaluating as False (if not task_result in the source code). The done callback will be called when the task succesfully completes. + +The map method returns a list containing the returned results for all your tasks in original order. + +### examples/basic_example.py + +```python +from time import sleep +import multiprocessing +import tqdm + +import logging +from tqdm_multiprocess.logger import setup_logger_tqdm +logger = logging.getLogger(__name__) + +from tqdm_multiprocess import TqdmMultiProcessPool + +iterations1 = 100 +iterations2 = 5 +iterations3 = 2 +def some_other_function(tqdm_func, global_tqdm): + + total_iterations = iterations1 * iterations2 * iterations3 + with tqdm_func(total=total_iterations, dynamic_ncols=True) as progress3: + progress3.set_description("outer") + for i in range(iterations3): + logger.info("outer") + total_iterations = iterations1 * iterations2 + with tqdm_func(total=total_iterations, dynamic_ncols=True) as progress2: + progress2.set_description("middle") + for j in range(iterations2): + logger.info("middle") + #for k in tqdm_func(range(iterations1), dynamic_ncols=True, desc="inner"): + with tqdm_func(total=iterations1, dynamic_ncols=True) as progress1: + for j in range(iterations1): + # logger.info("inner") # Spam slows down tqdm too much + progress1.set_description("innert") + sleep(0.01) + progress1.update() + progress2.update() + progress3.update() + global_tqdm.update() + + logger.warning(f"Warning test message. {multiprocessing.current_process().name}") + logger.error(f"Error test message. {multiprocessing.current_process().name}") + + +# Multiprocessed +def example_multiprocessing_function(some_input, tqdm_func, global_tqdm): + logger.debug(f"Debug test message - I won't show up in console. {multiprocessing.current_process().name}") + logger.info(f"Info test message. {multiprocessing.current_process().name}") + some_other_function(tqdm_func, global_tqdm) + return True + +def error_callback(result): + print("Error!") + +def done_callback(result): + print("Done. Result: ", result) + +def example(): + pool = TqdmMultiProcessPool() + process_count = 4 + task_count = 10 + initial_tasks = [(example_multiprocessing_function, (i,)) for i in range(task_count)] + total_iterations = iterations1 * iterations2 * iterations3 * task_count + with tqdm.tqdm(total=total_iterations, dynamic_ncols=True) as global_progress: + global_progress.set_description("global") + results = pool.map(process_count, global_progress, initial_tasks, error_callback, done_callback) + print(results) + +if __name__ == '__main__': + logfile_path = "tqdm_multiprocessing_example.log" + setup_logger_tqdm(logfile_path) # Logger will write messages using tqdm.write + example() +``` + + + + +%prep +%autosetup -n tqdm-multiprocess-0.0.11 + +%build +%py3_build + +%install +%py3_install +install -d -m755 %{buildroot}/%{_pkgdocdir} +if [ -d doc ]; then cp -arf doc %{buildroot}/%{_pkgdocdir}; fi +if [ -d docs ]; then cp -arf docs %{buildroot}/%{_pkgdocdir}; fi +if [ -d example ]; then cp -arf example %{buildroot}/%{_pkgdocdir}; fi +if [ -d examples ]; then cp -arf examples %{buildroot}/%{_pkgdocdir}; fi +pushd %{buildroot} +if [ -d usr/lib ]; then + find usr/lib -type f -printf "/%h/%f\n" >> filelist.lst +fi +if [ -d usr/lib64 ]; then + find usr/lib64 -type f -printf "/%h/%f\n" >> filelist.lst +fi +if [ -d usr/bin ]; then + find usr/bin -type f -printf "/%h/%f\n" >> filelist.lst +fi +if [ -d usr/sbin ]; then + find usr/sbin -type f -printf "/%h/%f\n" >> filelist.lst +fi +touch doclist.lst +if [ -d usr/share/man ]; then + find usr/share/man -type f -printf "/%h/%f.gz\n" >> doclist.lst +fi +popd +mv %{buildroot}/filelist.lst . +mv %{buildroot}/doclist.lst . + +%files -n python3-tqdm-multiprocess -f filelist.lst +%dir %{python3_sitelib}/* + +%files help -f doclist.lst +%{_docdir}/* + +%changelog +* Mon May 15 2023 Python_Bot <Python_Bot@openeuler.org> - 0.0.11-1 +- Package Spec generated |
