summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorCoprDistGit <infra@openeuler.org>2023-04-10 18:33:53 +0000
committerCoprDistGit <infra@openeuler.org>2023-04-10 18:33:53 +0000
commit9fb74223cf3d3b43c0a638fa4e963b0ad83c6410 (patch)
tree433b82c09034d770e76910861b641798a1b44d63
parenta7758d51c69344cb9d827e016ead008f52871c66 (diff)
automatic import of python-mrjob
-rw-r--r--.gitignore1
-rw-r--r--python-mrjob.spec150
-rw-r--r--sources1
3 files changed, 152 insertions, 0 deletions
diff --git a/.gitignore b/.gitignore
index e69de29..3133009 100644
--- a/.gitignore
+++ b/.gitignore
@@ -0,0 +1 @@
+/mrjob-0.7.4.tar.gz
diff --git a/python-mrjob.spec b/python-mrjob.spec
new file mode 100644
index 0000000..b348b57
--- /dev/null
+++ b/python-mrjob.spec
@@ -0,0 +1,150 @@
+%global _empty_manifest_terminate_build 0
+Name: python-mrjob
+Version: 0.7.4
+Release: 1
+Summary: Python MapReduce framework
+License: Apache
+URL: http://github.com/Yelp/mrjob
+Source0: https://mirrors.nju.edu.cn/pypi/web/packages/2c/ed/207853d1ebc6b549551d12db35e471289d26cac2cdae363419357294d3c5/mrjob-0.7.4.tar.gz
+BuildArch: noarch
+
+Requires: python3-PyYAML
+Requires: python3-boto3
+Requires: python3-botocore
+Requires: python3-google-cloud-dataproc
+Requires: python3-google-cloud-logging
+Requires: python3-google-cloud-storage
+Requires: python3-rapidjson
+Requires: python3-simplejson
+Requires: python3-ujson
+
+%description
+mrjob is a Python 2.7/3.4+ package that helps you write and run Hadoop
+Streaming jobs.
+`Stable version (v0.7.4) documentation <http://mrjob.readthedocs.org/en/stable/>`_
+`Development version documentation <http://mrjob.readthedocs.org/en/latest/>`_
+mrjob fully supports Amazon's Elastic MapReduce (EMR) service, which allows you
+to buy time on a Hadoop cluster on an hourly basis. mrjob has basic support for Google Cloud Dataproc (Dataproc)
+which allows you to buy time on a Hadoop cluster on a minute-by-minute basis. It also works with your own
+Hadoop cluster.
+Some important features:
+* Run jobs on EMR, Google Cloud Dataproc, your own Hadoop cluster, or locally (for testing).
+* Write multi-step jobs (one map-reduce step feeds into the next)
+* Easily launch Spark jobs on EMR or your own Hadoop cluster
+* Duplicate your production environment inside Hadoop
+ * Upload your source tree and put it in your job's ``$PYTHONPATH``
+ * Run make and other setup scripts
+ * Set environment variables (e.g. ``$TZ``)
+ * Easily install python packages from tarballs (EMR only)
+ * Setup handled transparently by ``mrjob.conf`` config file
+* Automatically interpret error logs
+* SSH tunnel to hadoop job tracker (EMR only)
+* Minimal setup
+ * To run on EMR, set ``$AWS_ACCESS_KEY_ID`` and ``$AWS_SECRET_ACCESS_KEY``
+ * To run on Dataproc, set ``$GOOGLE_APPLICATION_CREDENTIALS``
+ * No setup needed to use mrjob on your own Hadoop cluster
+
+%package -n python3-mrjob
+Summary: Python MapReduce framework
+Provides: python-mrjob
+BuildRequires: python3-devel
+BuildRequires: python3-setuptools
+BuildRequires: python3-pip
+%description -n python3-mrjob
+mrjob is a Python 2.7/3.4+ package that helps you write and run Hadoop
+Streaming jobs.
+`Stable version (v0.7.4) documentation <http://mrjob.readthedocs.org/en/stable/>`_
+`Development version documentation <http://mrjob.readthedocs.org/en/latest/>`_
+mrjob fully supports Amazon's Elastic MapReduce (EMR) service, which allows you
+to buy time on a Hadoop cluster on an hourly basis. mrjob has basic support for Google Cloud Dataproc (Dataproc)
+which allows you to buy time on a Hadoop cluster on a minute-by-minute basis. It also works with your own
+Hadoop cluster.
+Some important features:
+* Run jobs on EMR, Google Cloud Dataproc, your own Hadoop cluster, or locally (for testing).
+* Write multi-step jobs (one map-reduce step feeds into the next)
+* Easily launch Spark jobs on EMR or your own Hadoop cluster
+* Duplicate your production environment inside Hadoop
+ * Upload your source tree and put it in your job's ``$PYTHONPATH``
+ * Run make and other setup scripts
+ * Set environment variables (e.g. ``$TZ``)
+ * Easily install python packages from tarballs (EMR only)
+ * Setup handled transparently by ``mrjob.conf`` config file
+* Automatically interpret error logs
+* SSH tunnel to hadoop job tracker (EMR only)
+* Minimal setup
+ * To run on EMR, set ``$AWS_ACCESS_KEY_ID`` and ``$AWS_SECRET_ACCESS_KEY``
+ * To run on Dataproc, set ``$GOOGLE_APPLICATION_CREDENTIALS``
+ * No setup needed to use mrjob on your own Hadoop cluster
+
+%package help
+Summary: Development documents and examples for mrjob
+Provides: python3-mrjob-doc
+%description help
+mrjob is a Python 2.7/3.4+ package that helps you write and run Hadoop
+Streaming jobs.
+`Stable version (v0.7.4) documentation <http://mrjob.readthedocs.org/en/stable/>`_
+`Development version documentation <http://mrjob.readthedocs.org/en/latest/>`_
+mrjob fully supports Amazon's Elastic MapReduce (EMR) service, which allows you
+to buy time on a Hadoop cluster on an hourly basis. mrjob has basic support for Google Cloud Dataproc (Dataproc)
+which allows you to buy time on a Hadoop cluster on a minute-by-minute basis. It also works with your own
+Hadoop cluster.
+Some important features:
+* Run jobs on EMR, Google Cloud Dataproc, your own Hadoop cluster, or locally (for testing).
+* Write multi-step jobs (one map-reduce step feeds into the next)
+* Easily launch Spark jobs on EMR or your own Hadoop cluster
+* Duplicate your production environment inside Hadoop
+ * Upload your source tree and put it in your job's ``$PYTHONPATH``
+ * Run make and other setup scripts
+ * Set environment variables (e.g. ``$TZ``)
+ * Easily install python packages from tarballs (EMR only)
+ * Setup handled transparently by ``mrjob.conf`` config file
+* Automatically interpret error logs
+* SSH tunnel to hadoop job tracker (EMR only)
+* Minimal setup
+ * To run on EMR, set ``$AWS_ACCESS_KEY_ID`` and ``$AWS_SECRET_ACCESS_KEY``
+ * To run on Dataproc, set ``$GOOGLE_APPLICATION_CREDENTIALS``
+ * No setup needed to use mrjob on your own Hadoop cluster
+
+%prep
+%autosetup -n mrjob-0.7.4
+
+%build
+%py3_build
+
+%install
+%py3_install
+install -d -m755 %{buildroot}/%{_pkgdocdir}
+if [ -d doc ]; then cp -arf doc %{buildroot}/%{_pkgdocdir}; fi
+if [ -d docs ]; then cp -arf docs %{buildroot}/%{_pkgdocdir}; fi
+if [ -d example ]; then cp -arf example %{buildroot}/%{_pkgdocdir}; fi
+if [ -d examples ]; then cp -arf examples %{buildroot}/%{_pkgdocdir}; fi
+pushd %{buildroot}
+if [ -d usr/lib ]; then
+ find usr/lib -type f -printf "/%h/%f\n" >> filelist.lst
+fi
+if [ -d usr/lib64 ]; then
+ find usr/lib64 -type f -printf "/%h/%f\n" >> filelist.lst
+fi
+if [ -d usr/bin ]; then
+ find usr/bin -type f -printf "/%h/%f\n" >> filelist.lst
+fi
+if [ -d usr/sbin ]; then
+ find usr/sbin -type f -printf "/%h/%f\n" >> filelist.lst
+fi
+touch doclist.lst
+if [ -d usr/share/man ]; then
+ find usr/share/man -type f -printf "/%h/%f.gz\n" >> doclist.lst
+fi
+popd
+mv %{buildroot}/filelist.lst .
+mv %{buildroot}/doclist.lst .
+
+%files -n python3-mrjob -f filelist.lst
+%dir %{python3_sitelib}/*
+
+%files help -f doclist.lst
+%{_docdir}/*
+
+%changelog
+* Mon Apr 10 2023 Python_Bot <Python_Bot@openeuler.org> - 0.7.4-1
+- Package Spec generated
diff --git a/sources b/sources
new file mode 100644
index 0000000..b7375a3
--- /dev/null
+++ b/sources
@@ -0,0 +1 @@
+690655e16432de9cbadbb4fc3afd6b87 mrjob-0.7.4.tar.gz