1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
|
%global _empty_manifest_terminate_build 0
Name: python-hdx-python-scraper
Version: 2.1.5
Release: 1
Summary: HDX Python scraper utilities to assemble data from multiple sources
License: MIT
URL: https://github.com/OCHA-DAP/hdx-python-scraper
Source0: https://mirrors.nju.edu.cn/pypi/web/packages/4a/c2/4dba3339e0187f2cd5a34d1bcdc7801684684d4e03657b1d677ff064e8e9/hdx-python-scraper-2.1.5.tar.gz
BuildArch: noarch
Requires: python3-hdx-python-api
Requires: python3-gspread
Requires: python3-regex
Requires: python3-pandas
%description
[](https://github.com/OCHA-DAP/hdx-python-scraper/actions/workflows/run-python-tests.yml)
[](https://codecov.io/gh/OCHA-DAP/hdx-python-scraper)
[](https://github.com/psf/black)
[](https://pycqa.github.io/isort/)
The HDX Python Scraper Library is designed to enable you to easily develop code that
assembles data from one or more tabular sources that can be csv, xls, xlsx or JSON. It
uses a YAML file that specifies for each source what needs to be read and allows some
transformations to be performed on the data. The output is written to JSON, Google sheets
and/or Excel and includes the addition of
[Humanitarian Exchange Language (HXL)](https://hxlstandard.org/) hashtags specified in
the YAML file. Custom Python scrapers can also be written that conform to a defined
specification and the framework handles the execution of both configurable and custom
scrapers.
For more information, please read the
[documentation](https://hdx-python-scraper.readthedocs.io/en/latest/).
This library is part of the
[Humanitarian Data Exchange](https://data.humdata.org/) (HDX) project. If you have
humanitarian related data, please upload your datasets to HDX.
%package -n python3-hdx-python-scraper
Summary: HDX Python scraper utilities to assemble data from multiple sources
Provides: python-hdx-python-scraper
BuildRequires: python3-devel
BuildRequires: python3-setuptools
BuildRequires: python3-pip
%description -n python3-hdx-python-scraper
[](https://github.com/OCHA-DAP/hdx-python-scraper/actions/workflows/run-python-tests.yml)
[](https://codecov.io/gh/OCHA-DAP/hdx-python-scraper)
[](https://github.com/psf/black)
[](https://pycqa.github.io/isort/)
The HDX Python Scraper Library is designed to enable you to easily develop code that
assembles data from one or more tabular sources that can be csv, xls, xlsx or JSON. It
uses a YAML file that specifies for each source what needs to be read and allows some
transformations to be performed on the data. The output is written to JSON, Google sheets
and/or Excel and includes the addition of
[Humanitarian Exchange Language (HXL)](https://hxlstandard.org/) hashtags specified in
the YAML file. Custom Python scrapers can also be written that conform to a defined
specification and the framework handles the execution of both configurable and custom
scrapers.
For more information, please read the
[documentation](https://hdx-python-scraper.readthedocs.io/en/latest/).
This library is part of the
[Humanitarian Data Exchange](https://data.humdata.org/) (HDX) project. If you have
humanitarian related data, please upload your datasets to HDX.
%package help
Summary: Development documents and examples for hdx-python-scraper
Provides: python3-hdx-python-scraper-doc
%description help
[](https://github.com/OCHA-DAP/hdx-python-scraper/actions/workflows/run-python-tests.yml)
[](https://codecov.io/gh/OCHA-DAP/hdx-python-scraper)
[](https://github.com/psf/black)
[](https://pycqa.github.io/isort/)
The HDX Python Scraper Library is designed to enable you to easily develop code that
assembles data from one or more tabular sources that can be csv, xls, xlsx or JSON. It
uses a YAML file that specifies for each source what needs to be read and allows some
transformations to be performed on the data. The output is written to JSON, Google sheets
and/or Excel and includes the addition of
[Humanitarian Exchange Language (HXL)](https://hxlstandard.org/) hashtags specified in
the YAML file. Custom Python scrapers can also be written that conform to a defined
specification and the framework handles the execution of both configurable and custom
scrapers.
For more information, please read the
[documentation](https://hdx-python-scraper.readthedocs.io/en/latest/).
This library is part of the
[Humanitarian Data Exchange](https://data.humdata.org/) (HDX) project. If you have
humanitarian related data, please upload your datasets to HDX.
%prep
%autosetup -n hdx-python-scraper-2.1.5
%build
%py3_build
%install
%py3_install
install -d -m755 %{buildroot}/%{_pkgdocdir}
if [ -d doc ]; then cp -arf doc %{buildroot}/%{_pkgdocdir}; fi
if [ -d docs ]; then cp -arf docs %{buildroot}/%{_pkgdocdir}; fi
if [ -d example ]; then cp -arf example %{buildroot}/%{_pkgdocdir}; fi
if [ -d examples ]; then cp -arf examples %{buildroot}/%{_pkgdocdir}; fi
pushd %{buildroot}
if [ -d usr/lib ]; then
find usr/lib -type f -printf "/%h/%f\n" >> filelist.lst
fi
if [ -d usr/lib64 ]; then
find usr/lib64 -type f -printf "/%h/%f\n" >> filelist.lst
fi
if [ -d usr/bin ]; then
find usr/bin -type f -printf "/%h/%f\n" >> filelist.lst
fi
if [ -d usr/sbin ]; then
find usr/sbin -type f -printf "/%h/%f\n" >> filelist.lst
fi
touch doclist.lst
if [ -d usr/share/man ]; then
find usr/share/man -type f -printf "/%h/%f.gz\n" >> doclist.lst
fi
popd
mv %{buildroot}/filelist.lst .
mv %{buildroot}/doclist.lst .
%files -n python3-hdx-python-scraper -f filelist.lst
%dir %{python3_sitelib}/*
%files help -f doclist.lst
%{_docdir}/*
%changelog
* Tue May 30 2023 Python_Bot <Python_Bot@openeuler.org> - 2.1.5-1
- Package Spec generated
|