<?xml version="1.0" encoding="UTF-8"?>
<metadata xmlns="http://linux.duke.edu/metadata/common" xmlns:rpm="http://linux.duke.edu/metadata/rpm" packages="13">
<package type="rpm">
  <name>ollama</name>
  <arch>aarch64</arch>
  <version epoch="0" ver="0.2.5" rel="1"/>
  <checksum type="sha256" pkgid="YES">3cb7ec1c383f9a75068f92e18a0a2e9837ccb2a7d5d4e075c13b62b2dcecd176</checksum>
  <summary>Get up and running with Llama 3, Mistral, Gemma 2, and other large language models.</summary>
  <description>Get up and running with Llama 3, Mistral, Gemma 2, and other large language models.</description>
  <packager></packager>
  <url>https://github.com/ollama/ollama</url>
  <time file="1721048267" build="1721047958"/>
  <size package="7618537" installed="24355616" archive="0"/>
  <location href="00094507-ollama/ollama-0.2.5-1.aarch64.rpm"/>
  <format>
    <rpm:license>MIT license</rpm:license>
    <rpm:vendor>openEuler Copr - user Flysky</rpm:vendor>
    <rpm:group>Unspecified</rpm:group>
    <rpm:buildhost>eur-prod-workerlocal-aarch64-normal-prod-00214508-20240715-1245</rpm:buildhost>
    <rpm:sourcerpm>ollama-0.2.5-1.src.rpm</rpm:sourcerpm>
    <rpm:header-range start="944" end="4677"/>
    <rpm:provides>
      <rpm:entry name="ollama" flags="EQ" epoch="0" ver="0.2.5" rel="1"/>
      <rpm:entry name="ollama(aarch-64)" flags="EQ" epoch="0" ver="0.2.5" rel="1"/>
    </rpm:provides>
    <rpm:requires>
      <rpm:entry name="libgcc_s.so.1()(64bit)"/>
      <rpm:entry name="libgcc_s.so.1(GCC_3.0)(64bit)"/>
      <rpm:entry name="libm.so.6()(64bit)"/>
      <rpm:entry name="libm.so.6(GLIBC_2.17)(64bit)"/>
      <rpm:entry name="libm.so.6(GLIBC_2.27)(64bit)"/>
      <rpm:entry name="libm.so.6(GLIBC_2.29)(64bit)"/>
      <rpm:entry name="libresolv.so.2()(64bit)"/>
      <rpm:entry name="libstdc++.so.6()(64bit)"/>
      <rpm:entry name="libstdc++.so.6(CXXABI_1.3)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(CXXABI_1.3.11)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(CXXABI_1.3.13)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(CXXABI_1.3.2)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(CXXABI_1.3.3)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(CXXABI_1.3.5)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.11)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.14)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.15)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.17)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.18)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.20)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.21)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.22)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.29)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.9)(64bit)"/>
      <rpm:entry name="rtld(GNU_HASH)"/>
      <rpm:entry name="libc.so.6(GLIBC_2.38)(64bit)"/>
    </rpm:requires>
    <file>/usr/bin/ollama</file>
  </format>
</package>
<package type="rpm">
  <name>ollama</name>
  <arch>src</arch>
  <version epoch="0" ver="0.2.5" rel="1"/>
  <checksum type="sha256" pkgid="YES">3e3f275f6240dcd5d478d1d51bda63ccb4632ee6bcdb5c133f68256677cd9b44</checksum>
  <summary>Get up and running with Llama 3, Mistral, Gemma 2, and other large language models.</summary>
  <description>Get up and running with Llama 3, Mistral, Gemma 2, and other large language models.</description>
  <packager></packager>
  <url>https://github.com/ollama/ollama</url>
  <time file="1721048267" build="1721047944"/>
  <size package="5223" installed="1078" archive="0"/>
  <location href="00094507-ollama/ollama-0.2.5-1.src.rpm"/>
  <format>
    <rpm:license>MIT license</rpm:license>
    <rpm:vendor>openEuler Copr - user Flysky</rpm:vendor>
    <rpm:group>Unspecified</rpm:group>
    <rpm:buildhost>eur-prod-workerlocal-aarch64-normal-prod-00214508-20240715-1245</rpm:buildhost>
    <rpm:sourcerpm></rpm:sourcerpm>
    <rpm:header-range start="944" end="4595"/>
    <rpm:provides>
      <rpm:entry name="ollama" flags="EQ" epoch="0" ver="0.2.5" rel="1"/>
      <rpm:entry name="ollama-debuginfo" flags="EQ" epoch="0" ver="0.2.5" rel="1"/>
      <rpm:entry name="ollama-debugsource" flags="EQ" epoch="0" ver="0.2.5" rel="1"/>
    </rpm:provides>
    <rpm:requires>
      <rpm:entry name="cmake"/>
      <rpm:entry name="g++"/>
      <rpm:entry name="gcc"/>
      <rpm:entry name="git"/>
      <rpm:entry name="golang"/>
    </rpm:requires>
  </format>
</package>
<package type="rpm">
  <name>ollama</name>
  <arch>aarch64</arch>
  <version epoch="0" ver="0.2.7" rel="1"/>
  <checksum type="sha256" pkgid="YES">bd270f36d14c8cbf7d3aff18bfead93b40250fd3319d8e054677386dcf65e976</checksum>
  <summary>Get up and running with Llama 3, Mistral, Gemma 2, and other large language models.</summary>
  <description>Ollama is a relatively new but powerful framework designed for
serving machine learning models. It's designed to be efficient,
scalable, and easy to use, making it an attractive option for
developers and organizations looking to deploy their AI models
into production.</description>
  <packager></packager>
  <url>https://ollama.com</url>
  <time file="1721743430" build="1721743086"/>
  <size package="8125317" installed="27576866" archive="0"/>
  <location href="00094538-ollama/ollama-0.2.7-1.aarch64.rpm"/>
  <format>
    <rpm:license>MIT</rpm:license>
    <rpm:vendor>openEuler Copr - user Flysky</rpm:vendor>
    <rpm:group>Unspecified</rpm:group>
    <rpm:buildhost>eur-prod-workerlocal-aarch64-normal-prod-00214642-20240723-1346</rpm:buildhost>
    <rpm:sourcerpm>ollama-0.2.7-1.src.rpm</rpm:sourcerpm>
    <rpm:header-range start="944" end="6953"/>
    <rpm:provides>
      <rpm:entry name="ollama" flags="EQ" epoch="0" ver="0.2.7" rel="1"/>
      <rpm:entry name="ollama(aarch-64)" flags="EQ" epoch="0" ver="0.2.7" rel="1"/>
    </rpm:provides>
    <rpm:requires>
      <rpm:entry name="/bin/sh" pre="1"/>
      <rpm:entry name="/bin/sh"/>
      <rpm:entry name="libgcc_s.so.1()(64bit)"/>
      <rpm:entry name="libgcc_s.so.1(GCC_3.0)(64bit)"/>
      <rpm:entry name="libm.so.6()(64bit)"/>
      <rpm:entry name="libm.so.6(GLIBC_2.17)(64bit)"/>
      <rpm:entry name="libm.so.6(GLIBC_2.27)(64bit)"/>
      <rpm:entry name="libm.so.6(GLIBC_2.29)(64bit)"/>
      <rpm:entry name="libresolv.so.2()(64bit)"/>
      <rpm:entry name="libstdc++.so.6()(64bit)"/>
      <rpm:entry name="libstdc++.so.6(CXXABI_1.3)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(CXXABI_1.3.11)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(CXXABI_1.3.13)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(CXXABI_1.3.2)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(CXXABI_1.3.3)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(CXXABI_1.3.5)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(CXXABI_1.3.9)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.11)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.14)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.15)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.17)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.18)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.20)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.21)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.22)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.29)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.9)(64bit)"/>
      <rpm:entry name="rtld(GNU_HASH)"/>
      <rpm:entry name="libc.so.6(GLIBC_2.38)(64bit)"/>
    </rpm:requires>
    <file>/usr/bin/ollama</file>
  </format>
</package>
<package type="rpm">
  <name>ollama</name>
  <arch>src</arch>
  <version epoch="0" ver="0.2.7" rel="1"/>
  <checksum type="sha256" pkgid="YES">13d6a6b8e3cc42abe9a3a16c263c27e39b97bc48e4159b6e1733c9e23cebeb17</checksum>
  <summary>Get up and running with Llama 3, Mistral, Gemma 2, and other large language models.</summary>
  <description>Ollama is a relatively new but powerful framework designed for
serving machine learning models. It's designed to be efficient,
scalable, and easy to use, making it an attractive option for
developers and organizations looking to deploy their AI models
into production.</description>
  <packager></packager>
  <url>https://ollama.com</url>
  <time file="1721743430" build="1721743070"/>
  <size package="30828703" installed="31160866" archive="0"/>
  <location href="00094538-ollama/ollama-0.2.7-1.src.rpm"/>
  <format>
    <rpm:license>MIT</rpm:license>
    <rpm:vendor>openEuler Copr - user Flysky</rpm:vendor>
    <rpm:group>Unspecified</rpm:group>
    <rpm:buildhost>eur-prod-workerlocal-aarch64-normal-prod-00214642-20240723-1346</rpm:buildhost>
    <rpm:sourcerpm></rpm:sourcerpm>
    <rpm:header-range start="944" end="7654"/>
    <rpm:provides>
      <rpm:entry name="ollama" flags="EQ" epoch="0" ver="0.2.7" rel="1"/>
    </rpm:provides>
    <rpm:requires>
      <rpm:entry name="cmake"/>
      <rpm:entry name="gcc-c++"/>
      <rpm:entry name="git"/>
      <rpm:entry name="golang"/>
    </rpm:requires>
  </format>
</package>
<package type="rpm">
  <name>ollama</name>
  <arch>aarch64</arch>
  <version epoch="0" ver="0.3.0" rel="1"/>
  <checksum type="sha256" pkgid="YES">ac3903522ad70077ba743254e9d0d499f6a7c74f0dd8947c6b8a5a3bab9f0dc3</checksum>
  <summary>Get up and running with Llama 3, Mistral, Gemma 2, and other large language models.</summary>
  <description>Ollama is a relatively new but powerful framework designed for
serving machine learning models. It's designed to be efficient,
scalable, and easy to use, making it an attractive option for
developers and organizations looking to deploy their AI models
into production.</description>
  <packager></packager>
  <url>https://ollama.com</url>
  <time file="1722068949" build="1722068606"/>
  <size package="7825421" installed="24830986" archive="0"/>
  <location href="00094561-ollama/ollama-0.3.0-1.aarch64.rpm"/>
  <format>
    <rpm:license>MIT</rpm:license>
    <rpm:vendor>openEuler Copr - user Flysky</rpm:vendor>
    <rpm:group>Unspecified</rpm:group>
    <rpm:buildhost>eur-prod-workerlocal-aarch64-normal-prod-00214710-20240727-0725</rpm:buildhost>
    <rpm:sourcerpm>ollama-0.3.0-1.src.rpm</rpm:sourcerpm>
    <rpm:header-range start="944" end="7009"/>
    <rpm:provides>
      <rpm:entry name="ollama" flags="EQ" epoch="0" ver="0.3.0" rel="1"/>
      <rpm:entry name="ollama(aarch-64)" flags="EQ" epoch="0" ver="0.3.0" rel="1"/>
    </rpm:provides>
    <rpm:requires>
      <rpm:entry name="/bin/sh" pre="1"/>
      <rpm:entry name="/bin/sh"/>
      <rpm:entry name="libgcc_s.so.1()(64bit)"/>
      <rpm:entry name="libgcc_s.so.1(GCC_3.0)(64bit)"/>
      <rpm:entry name="libm.so.6()(64bit)"/>
      <rpm:entry name="libm.so.6(GLIBC_2.17)(64bit)"/>
      <rpm:entry name="libm.so.6(GLIBC_2.27)(64bit)"/>
      <rpm:entry name="libm.so.6(GLIBC_2.29)(64bit)"/>
      <rpm:entry name="libresolv.so.2()(64bit)"/>
      <rpm:entry name="libstdc++.so.6()(64bit)"/>
      <rpm:entry name="libstdc++.so.6(CXXABI_1.3)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(CXXABI_1.3.11)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(CXXABI_1.3.13)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(CXXABI_1.3.2)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(CXXABI_1.3.3)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(CXXABI_1.3.5)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(CXXABI_1.3.9)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.11)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.14)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.15)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.17)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.18)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.20)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.21)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.22)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.29)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.9)(64bit)"/>
      <rpm:entry name="rtld(GNU_HASH)"/>
      <rpm:entry name="libc.so.6(GLIBC_2.38)(64bit)"/>
    </rpm:requires>
    <file>/usr/bin/ollama</file>
  </format>
</package>
<package type="rpm">
  <name>ollama</name>
  <arch>src</arch>
  <version epoch="0" ver="0.3.0" rel="1"/>
  <checksum type="sha256" pkgid="YES">21cd9cfcae85a9547ca67148f878839f76258bf40618a782b0ee2acd0baa373e</checksum>
  <summary>Get up and running with Llama 3, Mistral, Gemma 2, and other large language models.</summary>
  <description>Ollama is a relatively new but powerful framework designed for
serving machine learning models. It's designed to be efficient,
scalable, and easy to use, making it an attractive option for
developers and organizations looking to deploy their AI models
into production.</description>
  <packager></packager>
  <url>https://ollama.com</url>
  <time file="1722068950" build="1722068591"/>
  <size package="29097873" installed="29399732" archive="0"/>
  <location href="00094561-ollama/ollama-0.3.0-1.src.rpm"/>
  <format>
    <rpm:license>MIT</rpm:license>
    <rpm:vendor>openEuler Copr - user Flysky</rpm:vendor>
    <rpm:group>Unspecified</rpm:group>
    <rpm:buildhost>eur-prod-workerlocal-aarch64-normal-prod-00214710-20240727-0725</rpm:buildhost>
    <rpm:sourcerpm></rpm:sourcerpm>
    <rpm:header-range start="944" end="7895"/>
    <rpm:provides>
      <rpm:entry name="ollama" flags="EQ" epoch="0" ver="0.3.0" rel="1"/>
    </rpm:provides>
    <rpm:requires>
      <rpm:entry name="cmake"/>
      <rpm:entry name="gcc-c++"/>
      <rpm:entry name="git"/>
      <rpm:entry name="golang"/>
    </rpm:requires>
  </format>
</package>
<package type="rpm">
  <name>ollama</name>
  <arch>aarch64</arch>
  <version epoch="0" ver="0.3.6" rel="1"/>
  <checksum type="sha256" pkgid="YES">e6d3406078f0326a75e985987f37857b465b5191b79dedfdbcd79e2a38458f5c</checksum>
  <summary>Get up and running with Llama 3, Mistral, Gemma 2, and other large language models.</summary>
  <description>Ollama is a relatively new but powerful framework designed for
serving machine learning models. It's designed to be efficient,
scalable, and easy to use, making it an attractive option for
developers and organizations looking to deploy their AI models
into production.</description>
  <packager></packager>
  <url>https://ollama.com</url>
  <time file="1724140190" build="1724139866"/>
  <size package="8126349" installed="25668205" archive="0"/>
  <location href="00096886-ollama/ollama-0.3.6-1.aarch64.rpm"/>
  <format>
    <rpm:license>MIT</rpm:license>
    <rpm:vendor>openEuler Copr - user Flysky</rpm:vendor>
    <rpm:group>Unspecified</rpm:group>
    <rpm:buildhost>eur-prod-workerlocal-aarch64-normal-prod-00436037-20240820-0049</rpm:buildhost>
    <rpm:sourcerpm>ollama-0.3.6-1.src.rpm</rpm:sourcerpm>
    <rpm:header-range start="944" end="13437"/>
    <rpm:provides>
      <rpm:entry name="ollama" flags="EQ" epoch="0" ver="0.3.6" rel="1"/>
      <rpm:entry name="ollama(aarch-64)" flags="EQ" epoch="0" ver="0.3.6" rel="1"/>
    </rpm:provides>
    <rpm:requires>
      <rpm:entry name="/bin/bash"/>
      <rpm:entry name="/bin/sh" pre="1"/>
      <rpm:entry name="/bin/sh"/>
      <rpm:entry name="libgcc_s.so.1()(64bit)"/>
      <rpm:entry name="libgcc_s.so.1(GCC_3.0)(64bit)"/>
      <rpm:entry name="libm.so.6()(64bit)"/>
      <rpm:entry name="libm.so.6(GLIBC_2.17)(64bit)"/>
      <rpm:entry name="libm.so.6(GLIBC_2.27)(64bit)"/>
      <rpm:entry name="libm.so.6(GLIBC_2.29)(64bit)"/>
      <rpm:entry name="libresolv.so.2()(64bit)"/>
      <rpm:entry name="libstdc++.so.6()(64bit)"/>
      <rpm:entry name="libstdc++.so.6(CXXABI_1.3)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(CXXABI_1.3.11)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(CXXABI_1.3.13)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(CXXABI_1.3.2)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(CXXABI_1.3.3)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(CXXABI_1.3.5)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(CXXABI_1.3.9)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.11)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.14)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.15)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.17)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.18)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.20)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.21)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.22)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.29)(64bit)"/>
      <rpm:entry name="libstdc++.so.6(GLIBCXX_3.4.9)(64bit)"/>
      <rpm:entry name="python3"/>
      <rpm:entry name="rtld(GNU_HASH)"/>
      <rpm:entry name="libc.so.6(GLIBC_2.38)(64bit)"/>
    </rpm:requires>
    <file>/usr/bin/ollama</file>
  </format>
</package>
<package type="rpm">
  <name>ollama</name>
  <arch>src</arch>
  <version epoch="0" ver="0.3.6" rel="1"/>
  <checksum type="sha256" pkgid="YES">a374f279d3420cbd60624fbd67c27bceedcd3c6e87fd2a4d6ae8d24be3c046bf</checksum>
  <summary>Get up and running with Llama 3, Mistral, Gemma 2, and other large language models.</summary>
  <description>Ollama is a relatively new but powerful framework designed for
serving machine learning models. It's designed to be efficient,
scalable, and easy to use, making it an attractive option for
developers and organizations looking to deploy their AI models
into production.</description>
  <packager></packager>
  <url>https://ollama.com</url>
  <time file="1724140190" build="1724139850"/>
  <size package="29447392" installed="29748605" archive="0"/>
  <location href="00096886-ollama/ollama-0.3.6-1.src.rpm"/>
  <format>
    <rpm:license>MIT</rpm:license>
    <rpm:vendor>openEuler Copr - user Flysky</rpm:vendor>
    <rpm:group>Unspecified</rpm:group>
    <rpm:buildhost>eur-prod-workerlocal-aarch64-normal-prod-00436037-20240820-0049</rpm:buildhost>
    <rpm:sourcerpm></rpm:sourcerpm>
    <rpm:header-range start="944" end="9148"/>
    <rpm:provides>
      <rpm:entry name="ollama" flags="EQ" epoch="0" ver="0.3.6" rel="1"/>
    </rpm:provides>
    <rpm:requires>
      <rpm:entry name="cmake"/>
      <rpm:entry name="gcc-c++"/>
      <rpm:entry name="git"/>
      <rpm:entry name="golang"/>
    </rpm:requires>
  </format>
</package>
<package type="rpm">
  <name>ollama</name>
  <arch>src</arch>
  <version epoch="0" ver="0.3.6" rel="1"/>
  <checksum type="sha256" pkgid="YES">68bbe1fb9d94b937efbcb1dcf0e4c6bf9ad5f1b8f8425cb5c828d3366f5ee4e9</checksum>
  <summary>Get up and running with Llama 3, Mistral, Gemma 2, and other large language models.</summary>
  <description>Ollama is a relatively new but powerful framework designed for
serving machine learning models. It's designed to be efficient,
scalable, and easy to use, making it an attractive option for
developers and organizations looking to deploy their AI models
into production.</description>
  <packager></packager>
  <url>https://ollama.com</url>
  <time file="1724147376" build="1724147375"/>
  <size package="29456665" installed="29747926" archive="29748940"/>
  <location href="00096891-ollama/ollama-0.3.6-1.src.rpm"/>
  <format>
    <rpm:license>MIT</rpm:license>
    <rpm:vendor>openEuler Copr - user Flysky</rpm:vendor>
    <rpm:group>Unspecified</rpm:group>
    <rpm:buildhost>eur-prod-workerlocal-aarch64-normal-prod-00436051-20240820-0857</rpm:buildhost>
    <rpm:sourcerpm></rpm:sourcerpm>
    <rpm:header-range start="4504" end="12153"/>
    <rpm:provides>
      <rpm:entry name="ollama" flags="EQ" epoch="0" ver="0.3.6" rel="1"/>
    </rpm:provides>
    <rpm:requires>
      <rpm:entry name="cmake"/>
      <rpm:entry name="gcc-c++"/>
      <rpm:entry name="git"/>
      <rpm:entry name="golang"/>
    </rpm:requires>
  </format>
</package>
<package type="rpm">
  <name>ollama</name>
  <arch>src</arch>
  <version epoch="0" ver="0.3.6" rel="1"/>
  <checksum type="sha256" pkgid="YES">51ac24d1ff0fcd887a2b686b762c33935169b4fdac10b3374f553aa1e2a59404</checksum>
  <summary>Get up and running with Llama 3, Mistral, Gemma 2, and other large language models.</summary>
  <description>Ollama is a relatively new but powerful framework designed for
serving machine learning models. It's designed to be efficient,
scalable, and easy to use, making it an attractive option for
developers and organizations looking to deploy their AI models
into production.</description>
  <packager></packager>
  <url>https://ollama.com</url>
  <time file="1728562554" build="1724150768"/>
  <size package="29452816" installed="29747927" archive="0"/>
  <location href="00096892-ollama/ollama-0.3.6-1.src.rpm"/>
  <format>
    <rpm:license>MIT</rpm:license>
    <rpm:vendor>openEuler Copr - user Flysky</rpm:vendor>
    <rpm:group>Unspecified</rpm:group>
    <rpm:buildhost>eur-prod-workerlocal-aarch64-normal-prod-00436054-20240820-0947</rpm:buildhost>
    <rpm:sourcerpm></rpm:sourcerpm>
    <rpm:header-range start="640" end="8290"/>
    <rpm:provides>
      <rpm:entry name="ollama" flags="EQ" epoch="0" ver="0.3.6" rel="1"/>
    </rpm:provides>
    <rpm:requires>
      <rpm:entry name="cmake"/>
      <rpm:entry name="gcc-c++"/>
      <rpm:entry name="git"/>
      <rpm:entry name="golang"/>
    </rpm:requires>
  </format>
</package>
<package type="rpm">
  <name>ollama</name>
  <arch>src</arch>
  <version epoch="0" ver="0.3.6" rel="1"/>
  <checksum type="sha256" pkgid="YES">637d7b7b4f560da8edc4f6be3f6153b924da44fe425b1d0a8ace5cef9cb680ed</checksum>
  <summary>Get up and running with Llama 3, Mistral, Gemma 2, and other large language models.</summary>
  <description>Ollama is a relatively new but powerful framework designed for
serving machine learning models. It's designed to be efficient,
scalable, and easy to use, making it an attractive option for
developers and organizations looking to deploy their AI models
into production.</description>
  <packager></packager>
  <url>https://ollama.com</url>
  <time file="1724157693" build="1724157692"/>
  <size package="29456666" installed="29747929" archive="29748944"/>
  <location href="00096900-ollama/ollama-0.3.6-1.src.rpm"/>
  <format>
    <rpm:license>MIT</rpm:license>
    <rpm:vendor>openEuler Copr - user Flysky</rpm:vendor>
    <rpm:group>Unspecified</rpm:group>
    <rpm:buildhost>eur-prod-workerlocal-aarch64-normal-prod-00436078-20240820-1238</rpm:buildhost>
    <rpm:sourcerpm></rpm:sourcerpm>
    <rpm:header-range start="4504" end="12156"/>
    <rpm:provides>
      <rpm:entry name="ollama" flags="EQ" epoch="0" ver="0.3.6" rel="1"/>
    </rpm:provides>
    <rpm:requires>
      <rpm:entry name="cmake"/>
      <rpm:entry name="gcc-c++"/>
      <rpm:entry name="git"/>
      <rpm:entry name="golang"/>
    </rpm:requires>
  </format>
</package>
<package type="rpm">
  <name>ollama-debuginfo</name>
  <arch>aarch64</arch>
  <version epoch="0" ver="0.2.5" rel="1"/>
  <checksum type="sha256" pkgid="YES">1b85dc4d9dc55cfe6f53e06c25f69863dfc92e8c87c60ed48079e02ecc07cfb2</checksum>
  <summary>Debug information for package ollama</summary>
  <description>This package provides debug information for package ollama.
Debug information is useful when developing applications that use this
package or when debugging this package.</description>
  <packager></packager>
  <url>https://github.com/ollama/ollama</url>
  <time file="1721048267" build="1721047958"/>
  <size package="6402449" installed="40092976" archive="0"/>
  <location href="00094507-ollama/ollama-debuginfo-0.2.5-1.aarch64.rpm"/>
  <format>
    <rpm:license>MIT license</rpm:license>
    <rpm:vendor>openEuler Copr - user Flysky</rpm:vendor>
    <rpm:group>Development/Debug</rpm:group>
    <rpm:buildhost>eur-prod-workerlocal-aarch64-normal-prod-00214508-20240715-1245</rpm:buildhost>
    <rpm:sourcerpm>ollama-0.2.5-1.src.rpm</rpm:sourcerpm>
    <rpm:header-range start="944" end="3709"/>
    <rpm:provides>
      <rpm:entry name="ollama-debuginfo" flags="EQ" epoch="0" ver="0.2.5" rel="1"/>
      <rpm:entry name="ollama-debuginfo(aarch-64)" flags="EQ" epoch="0" ver="0.2.5" rel="1"/>
    </rpm:provides>
    <rpm:recommends>
      <rpm:entry name="ollama-debugsource(aarch-64)" flags="EQ" epoch="0" ver="0.2.5" rel="1"/>
    </rpm:recommends>
    <file>/usr/lib/debug/usr/bin/ollama-0.2.5-1.aarch64.debug</file>
  </format>
</package>
<package type="rpm">
  <name>ollama-debugsource</name>
  <arch>aarch64</arch>
  <version epoch="0" ver="0.2.5" rel="1"/>
  <checksum type="sha256" pkgid="YES">6fc5a11a4a701ea158f4885b0062d8d966e0d291f2229344973e15bc4bfc6156</checksum>
  <summary>Debug sources for package ollama</summary>
  <description>This package provides debug sources for package ollama.
Debug sources are useful when developing applications that use this
package or when debugging this package.</description>
  <packager></packager>
  <url>https://github.com/ollama/ollama</url>
  <time file="1721048267" build="1721047958"/>
  <size package="163677" installed="699757" archive="0"/>
  <location href="00094507-ollama/ollama-debugsource-0.2.5-1.aarch64.rpm"/>
  <format>
    <rpm:license>MIT license</rpm:license>
    <rpm:vendor>openEuler Copr - user Flysky</rpm:vendor>
    <rpm:group>Development/Debug</rpm:group>
    <rpm:buildhost>eur-prod-workerlocal-aarch64-normal-prod-00214508-20240715-1245</rpm:buildhost>
    <rpm:sourcerpm>ollama-0.2.5-1.src.rpm</rpm:sourcerpm>
    <rpm:header-range start="944" end="16157"/>
    <rpm:provides>
      <rpm:entry name="ollama-debugsource" flags="EQ" epoch="0" ver="0.2.5" rel="1"/>
      <rpm:entry name="ollama-debugsource(aarch-64)" flags="EQ" epoch="0" ver="0.2.5" rel="1"/>
    </rpm:provides>
  </format>
</package>
</metadata>