Mock Version: 3.5 ENTER ['do_with_status'](['bash', '--login', '-c', '/usr/bin/rpmbuild -bs --target x86_64 --nodeps /builddir/build/SPECS/ollama.spec'], chrootPath='/var/lib/mock/openeuler-24.03_LTS-x86_64-1724139810.736560/root'env={'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'}shell=Falselogger=timeout=0uid=1000gid=135user='mockbuild'nspawn_args=[]unshare_net=FalseprintOutput=True) Executing command: ['bash', '--login', '-c', '/usr/bin/rpmbuild -bs --target x86_64 --nodeps /builddir/build/SPECS/ollama.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'} and shell False Building target platforms: x86_64 Building for target x86_64 Wrote: /builddir/build/SRPMS/ollama-0.3.6-1.src.rpm Child return code was: 0 Mock Version: 3.5 ENTER ['do_with_status'](['bash', '--login', '-c', '/usr/bin/rpmbuild -bs --target x86_64 --nodeps /builddir/build/SPECS/ollama.spec'], chrootPath='/var/lib/mock/openeuler-24.03_LTS-x86_64-1724139810.736560/root'env={'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'}shell=Falselogger=timeout=0uid=1000gid=135user='mockbuild'nspawn_args=[]unshare_net=FalseprintOutput=True) Executing command: ['bash', '--login', '-c', '/usr/bin/rpmbuild -bs --target x86_64 --nodeps /builddir/build/SPECS/ollama.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'} and shell False Building target platforms: x86_64 Building for target x86_64 Wrote: /builddir/build/SRPMS/ollama-0.3.6-1.src.rpm Child return code was: 0 ENTER ['do_with_status'](['bash', '--login', '-c', '/usr/bin/rpmbuild -bb --target x86_64 --nodeps /builddir/build/SPECS/ollama.spec'], chrootPath='/var/lib/mock/openeuler-24.03_LTS-x86_64-1724139810.736560/root'env={'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'}shell=Falselogger=timeout=0uid=1000gid=135user='mockbuild'nspawn_args=[]unshare_net=FalseprintOutput=True) Executing command: ['bash', '--login', '-c', '/usr/bin/rpmbuild -bb --target x86_64 --nodeps /builddir/build/SPECS/ollama.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'} and shell False Building target platforms: x86_64 Building for target x86_64 Executing(%prep): /bin/sh -e /var/tmp/rpm-tmp.13p3fm + umask 022 + cd /builddir/build/BUILD + cd /builddir/build/BUILD + rm -rf ollama-0.3.6 + /usr/lib/rpm/rpmuncompress -x /builddir/build/SOURCES/v0.3.6.tar.gz + STATUS=0 + '[' 0 -ne 0 ']' + cd ollama-0.3.6 + /usr/bin/chmod -Rf a+rX,u+w,g-w,o-w . + tar -xzf /builddir/build/SOURCES/vendor.tar.gz + cd llm + tar -xzf /builddir/build/SOURCES/b3536.tar.gz + mv llama.cpp-b3536/AUTHORS llama.cpp-b3536/CMakeLists.txt llama.cpp-b3536/CMakePresets.json llama.cpp-b3536/CONTRIBUTING.md llama.cpp-b3536/LICENSE llama.cpp-b3536/Makefile llama.cpp-b3536/Package.swift llama.cpp-b3536/README.md llama.cpp-b3536/SECURITY.md llama.cpp-b3536/ci llama.cpp-b3536/cmake llama.cpp-b3536/common llama.cpp-b3536/convert_hf_to_gguf.py llama.cpp-b3536/convert_hf_to_gguf_update.py llama.cpp-b3536/convert_llama_ggml_to_gguf.py llama.cpp-b3536/convert_lora_to_gguf.py llama.cpp-b3536/docs llama.cpp-b3536/examples llama.cpp-b3536/flake.lock llama.cpp-b3536/flake.nix llama.cpp-b3536/ggml llama.cpp-b3536/gguf-py llama.cpp-b3536/grammars llama.cpp-b3536/include llama.cpp-b3536/media llama.cpp-b3536/models llama.cpp-b3536/mypy.ini llama.cpp-b3536/pocs llama.cpp-b3536/poetry.lock llama.cpp-b3536/prompts llama.cpp-b3536/pyproject.toml llama.cpp-b3536/pyrightconfig.json llama.cpp-b3536/requirements llama.cpp-b3536/requirements.txt llama.cpp-b3536/scripts llama.cpp-b3536/spm-headers llama.cpp-b3536/src llama.cpp-b3536/tests llama.cpp/ + rm -rf llama.cpp-b3536/ + cd llama.cpp + git init hint: Using 'master' as the name for the initial branch. This default branch name hint: is subject to change. To configure the initial branch name to use in all hint: of your new repositories, which will suppress this warning, call: hint: hint: git config --global init.defaultBranch hint: hint: Names commonly chosen instead of 'master' are 'main', 'trunk' and hint: 'development'. The just-created branch can be renamed via this command: hint: hint: git branch -m Initialized empty Git repository in /builddir/build/BUILD/ollama-0.3.6/llm/llama.cpp/.git/ + git config user.email tmp@example.com + git config user.name tmp + git add . + git commit -a -s -m init [master (root-commit) 2125b11] init 884 files changed, 303188 insertions(+) create mode 100644 AUTHORS create mode 100644 CMakeLists.txt create mode 100644 CMakePresets.json create mode 100644 CONTRIBUTING.md create mode 100644 LICENSE create mode 100644 Makefile create mode 100644 Package.swift create mode 100644 README.md create mode 100644 SECURITY.md create mode 100644 ci/README.md create mode 100755 ci/run.sh create mode 100644 cmake/arm64-windows-llvm.cmake create mode 100644 cmake/arm64-windows-msvc.cmake create mode 100644 cmake/build-info.cmake create mode 100644 cmake/git-vars.cmake create mode 100644 cmake/llama-config.cmake.in create mode 100644 cmake/llama.pc.in create mode 100644 common/CMakeLists.txt create mode 100644 common/base64.hpp create mode 100644 common/build-info.cpp.in create mode 100644 common/cmake/build-info-gen-cpp.cmake create mode 100644 common/common.cpp create mode 100644 common/common.h create mode 100644 common/console.cpp create mode 100644 common/console.h create mode 100644 common/grammar-parser.cpp create mode 100644 common/grammar-parser.h create mode 100644 common/json-schema-to-grammar.cpp create mode 100644 common/json-schema-to-grammar.h create mode 100644 common/json.hpp create mode 100644 common/log.h create mode 100644 common/ngram-cache.cpp create mode 100644 common/ngram-cache.h create mode 100644 common/sampling.cpp create mode 100644 common/sampling.h create mode 100644 common/stb_image.h create mode 100644 common/train.cpp create mode 100644 common/train.h create mode 100755 convert_hf_to_gguf.py create mode 100755 convert_hf_to_gguf_update.py create mode 100755 convert_llama_ggml_to_gguf.py create mode 100755 convert_lora_to_gguf.py create mode 100644 docs/android.md create mode 100644 docs/backend/BLIS.md create mode 100644 docs/backend/SYCL.md create mode 100644 docs/build.md create mode 100644 docs/development/HOWTO-add-model.md create mode 100644 docs/development/debugging-tests.md create mode 100755 docs/development/llama-star/idea-arch.key create mode 100644 docs/development/llama-star/idea-arch.pdf create mode 100644 docs/development/token_generation_performance_tips.md create mode 100644 docs/docker.md create mode 100644 docs/install.md create mode 100644 examples/CMakeLists.txt create mode 100755 examples/Miku.sh create mode 100644 examples/baby-llama/CMakeLists.txt create mode 100644 examples/baby-llama/baby-llama.cpp create mode 100755 examples/base-translate.sh create mode 100644 examples/batched-bench/CMakeLists.txt create mode 100644 examples/batched-bench/README.md create mode 100644 examples/batched-bench/batched-bench.cpp create mode 100644 examples/batched.swift/.gitignore create mode 100755 examples/batched.swift/Makefile create mode 100644 examples/batched.swift/Package.swift create mode 100644 examples/batched.swift/README.md create mode 100644 examples/batched.swift/Sources/main.swift create mode 100644 examples/batched/CMakeLists.txt create mode 100644 examples/batched/README.md create mode 100644 examples/batched/batched.cpp create mode 100644 examples/benchmark/CMakeLists.txt create mode 100644 examples/benchmark/benchmark-matmult.cpp create mode 100644 examples/chat-13B.bat create mode 100755 examples/chat-13B.sh create mode 100755 examples/chat-persistent.sh create mode 100755 examples/chat-vicuna.sh create mode 100755 examples/chat.sh create mode 100644 examples/convert-llama2c-to-ggml/CMakeLists.txt create mode 100644 examples/convert-llama2c-to-ggml/README.md create mode 100644 examples/convert-llama2c-to-ggml/convert-llama2c-to-ggml.cpp create mode 100755 examples/convert_legacy_llama.py create mode 100644 examples/cvector-generator/CMakeLists.txt create mode 100644 examples/cvector-generator/README.md create mode 100644 examples/cvector-generator/completions.txt create mode 100644 examples/cvector-generator/cvector-generator.cpp create mode 100644 examples/cvector-generator/mean.hpp create mode 100644 examples/cvector-generator/negative.txt create mode 100644 examples/cvector-generator/pca.hpp create mode 100644 examples/cvector-generator/positive.txt create mode 100644 examples/deprecation-warning/README.md create mode 100644 examples/deprecation-warning/deprecation-warning.cpp create mode 100644 examples/embedding/CMakeLists.txt create mode 100644 examples/embedding/README.md create mode 100644 examples/embedding/embedding.cpp create mode 100644 examples/eval-callback/CMakeLists.txt create mode 100644 examples/eval-callback/README.md create mode 100644 examples/eval-callback/eval-callback.cpp create mode 100644 examples/export-lora/CMakeLists.txt create mode 100644 examples/export-lora/README.md create mode 100644 examples/export-lora/export-lora.cpp create mode 100644 examples/gbnf-validator/CMakeLists.txt create mode 100644 examples/gbnf-validator/gbnf-validator.cpp create mode 100644 examples/gguf-hash/CMakeLists.txt create mode 100644 examples/gguf-hash/README.md create mode 100644 examples/gguf-hash/deps/rotate-bits/package.json create mode 100644 examples/gguf-hash/deps/rotate-bits/rotate-bits.h create mode 100644 examples/gguf-hash/deps/sha1/package.json create mode 100644 examples/gguf-hash/deps/sha1/sha1.c create mode 100644 examples/gguf-hash/deps/sha1/sha1.h create mode 100644 examples/gguf-hash/deps/sha256/package.json create mode 100644 examples/gguf-hash/deps/sha256/sha256.c create mode 100644 examples/gguf-hash/deps/sha256/sha256.h create mode 100644 examples/gguf-hash/deps/xxhash/clib.json create mode 100644 examples/gguf-hash/deps/xxhash/xxhash.c create mode 100644 examples/gguf-hash/deps/xxhash/xxhash.h create mode 100644 examples/gguf-hash/gguf-hash.cpp create mode 100644 examples/gguf-split/CMakeLists.txt create mode 100644 examples/gguf-split/README.md create mode 100644 examples/gguf-split/gguf-split.cpp create mode 100755 examples/gguf-split/tests.sh create mode 100644 examples/gguf/CMakeLists.txt create mode 100644 examples/gguf/gguf.cpp create mode 100644 examples/gritlm/CMakeLists.txt create mode 100644 examples/gritlm/README.md create mode 100644 examples/gritlm/gritlm.cpp create mode 100644 examples/imatrix/CMakeLists.txt create mode 100644 examples/imatrix/README.md create mode 100644 examples/imatrix/imatrix.cpp create mode 100644 examples/infill/CMakeLists.txt create mode 100644 examples/infill/README.md create mode 100644 examples/infill/infill.cpp create mode 100644 examples/jeopardy/README.md create mode 100755 examples/jeopardy/graph.py create mode 100755 examples/jeopardy/jeopardy.sh create mode 100644 examples/jeopardy/qasheet.csv create mode 100644 examples/jeopardy/questions.txt create mode 100644 examples/json_schema_pydantic_example.py create mode 100755 examples/json_schema_to_grammar.py create mode 100644 examples/llama-bench/CMakeLists.txt create mode 100644 examples/llama-bench/README.md create mode 100644 examples/llama-bench/llama-bench.cpp create mode 100644 examples/llama.android/.gitignore create mode 100644 examples/llama.android/README.md create mode 100644 examples/llama.android/app/.gitignore create mode 100644 examples/llama.android/app/build.gradle.kts create mode 100644 examples/llama.android/app/proguard-rules.pro create mode 100644 examples/llama.android/app/src/main/AndroidManifest.xml create mode 100644 examples/llama.android/app/src/main/java/com/example/llama/Downloadable.kt create mode 100644 examples/llama.android/app/src/main/java/com/example/llama/MainActivity.kt create mode 100644 examples/llama.android/app/src/main/java/com/example/llama/MainViewModel.kt create mode 100644 examples/llama.android/app/src/main/java/com/example/llama/ui/theme/Color.kt create mode 100644 examples/llama.android/app/src/main/java/com/example/llama/ui/theme/Theme.kt create mode 100644 examples/llama.android/app/src/main/java/com/example/llama/ui/theme/Type.kt create mode 100644 examples/llama.android/app/src/main/res/drawable/ic_launcher_background.xml create mode 100644 examples/llama.android/app/src/main/res/drawable/ic_launcher_foreground.xml create mode 100644 examples/llama.android/app/src/main/res/mipmap-anydpi/ic_launcher.xml create mode 100644 examples/llama.android/app/src/main/res/mipmap-anydpi/ic_launcher_round.xml create mode 100644 examples/llama.android/app/src/main/res/mipmap-hdpi/ic_launcher.webp create mode 100644 examples/llama.android/app/src/main/res/mipmap-hdpi/ic_launcher_round.webp create mode 100644 examples/llama.android/app/src/main/res/mipmap-mdpi/ic_launcher.webp create mode 100644 examples/llama.android/app/src/main/res/mipmap-mdpi/ic_launcher_round.webp create mode 100644 examples/llama.android/app/src/main/res/mipmap-xhdpi/ic_launcher.webp create mode 100644 examples/llama.android/app/src/main/res/mipmap-xhdpi/ic_launcher_round.webp create mode 100644 examples/llama.android/app/src/main/res/mipmap-xxhdpi/ic_launcher.webp create mode 100644 examples/llama.android/app/src/main/res/mipmap-xxhdpi/ic_launcher_round.webp create mode 100644 examples/llama.android/app/src/main/res/mipmap-xxxhdpi/ic_launcher.webp create mode 100644 examples/llama.android/app/src/main/res/mipmap-xxxhdpi/ic_launcher_round.webp create mode 100644 examples/llama.android/app/src/main/res/values/colors.xml create mode 100644 examples/llama.android/app/src/main/res/values/strings.xml create mode 100644 examples/llama.android/app/src/main/res/values/themes.xml create mode 100644 examples/llama.android/app/src/main/res/xml/backup_rules.xml create mode 100644 examples/llama.android/app/src/main/res/xml/data_extraction_rules.xml create mode 100644 examples/llama.android/build.gradle.kts create mode 100644 examples/llama.android/gradle.properties create mode 100644 examples/llama.android/gradle/wrapper/gradle-wrapper.jar create mode 100644 examples/llama.android/gradle/wrapper/gradle-wrapper.properties create mode 100755 examples/llama.android/gradlew create mode 100644 examples/llama.android/llama/.gitignore create mode 100644 examples/llama.android/llama/build.gradle.kts create mode 100644 examples/llama.android/llama/consumer-rules.pro create mode 100644 examples/llama.android/llama/proguard-rules.pro create mode 100644 examples/llama.android/llama/src/androidTest/java/android/llama/cpp/ExampleInstrumentedTest.kt create mode 100644 examples/llama.android/llama/src/main/AndroidManifest.xml create mode 100644 examples/llama.android/llama/src/main/cpp/CMakeLists.txt create mode 100644 examples/llama.android/llama/src/main/cpp/llama-android.cpp create mode 100644 examples/llama.android/llama/src/main/java/android/llama/cpp/LLamaAndroid.kt create mode 100644 examples/llama.android/llama/src/test/java/android/llama/cpp/ExampleUnitTest.kt create mode 100644 examples/llama.android/settings.gradle.kts create mode 100644 examples/llama.swiftui/.gitignore create mode 100644 examples/llama.swiftui/README.md create mode 100644 examples/llama.swiftui/llama.cpp.swift/LibLlama.swift create mode 100644 examples/llama.swiftui/llama.swiftui.xcodeproj/project.pbxproj create mode 100644 examples/llama.swiftui/llama.swiftui.xcodeproj/project.xcworkspace/contents.xcworkspacedata create mode 100644 examples/llama.swiftui/llama.swiftui/Assets.xcassets/AppIcon.appiconset/Contents.json create mode 100644 examples/llama.swiftui/llama.swiftui/Assets.xcassets/Contents.json create mode 100644 examples/llama.swiftui/llama.swiftui/Models/LlamaState.swift create mode 100644 examples/llama.swiftui/llama.swiftui/Resources/models/.gitignore create mode 100644 examples/llama.swiftui/llama.swiftui/UI/ContentView.swift create mode 100644 examples/llama.swiftui/llama.swiftui/UI/DownloadButton.swift create mode 100644 examples/llama.swiftui/llama.swiftui/UI/InputButton.swift create mode 100644 examples/llama.swiftui/llama.swiftui/UI/LoadCustomButton.swift create mode 100644 examples/llama.swiftui/llama.swiftui/llama_swiftuiApp.swift create mode 100644 examples/llama.vim create mode 100644 examples/llava/CMakeLists.txt create mode 100644 examples/llava/MobileVLM-README.md create mode 100644 examples/llava/README.md create mode 100755 examples/llava/android/adb_run.sh create mode 100755 examples/llava/android/build_64.sh create mode 100644 examples/llava/clip.cpp create mode 100644 examples/llava/clip.h create mode 100644 examples/llava/convert_image_encoder_to_gguf.py create mode 100644 examples/llava/llava-cli.cpp create mode 100644 examples/llava/llava.cpp create mode 100644 examples/llava/llava.h create mode 100644 examples/llava/llava_surgery.py create mode 100644 examples/llava/llava_surgery_v2.py create mode 100644 examples/llava/requirements.txt create mode 100644 examples/llm.vim create mode 100644 examples/lookahead/CMakeLists.txt create mode 100644 examples/lookahead/README.md create mode 100644 examples/lookahead/lookahead.cpp create mode 100644 examples/lookup/CMakeLists.txt create mode 100644 examples/lookup/README.md create mode 100644 examples/lookup/lookup-create.cpp create mode 100644 examples/lookup/lookup-merge.cpp create mode 100644 examples/lookup/lookup-stats.cpp create mode 100644 examples/lookup/lookup.cpp create mode 100644 examples/main-cmake-pkg/.gitignore create mode 100644 examples/main-cmake-pkg/CMakeLists.txt create mode 100644 examples/main-cmake-pkg/README.md create mode 100644 examples/main/CMakeLists.txt create mode 100644 examples/main/README.md create mode 100644 examples/main/main.cpp create mode 100644 examples/parallel/CMakeLists.txt create mode 100644 examples/parallel/README.md create mode 100644 examples/parallel/parallel.cpp create mode 100644 examples/passkey/CMakeLists.txt create mode 100644 examples/passkey/README.md create mode 100644 examples/passkey/passkey.cpp create mode 100644 examples/perplexity/CMakeLists.txt create mode 100644 examples/perplexity/README.md create mode 100644 examples/perplexity/perplexity.cpp create mode 100644 examples/pydantic_models_to_grammar.py create mode 100755 examples/pydantic_models_to_grammar_examples.py create mode 100644 examples/quantize-stats/CMakeLists.txt create mode 100644 examples/quantize-stats/quantize-stats.cpp create mode 100644 examples/quantize/CMakeLists.txt create mode 100644 examples/quantize/README.md create mode 100644 examples/quantize/quantize.cpp create mode 100644 examples/quantize/tests.sh create mode 100755 examples/reason-act.sh create mode 100644 examples/regex_to_grammar.py create mode 100644 examples/retrieval/CMakeLists.txt create mode 100644 examples/retrieval/README.md create mode 100644 examples/retrieval/retrieval.cpp create mode 100644 examples/rpc/CMakeLists.txt create mode 100644 examples/rpc/README.md create mode 100644 examples/rpc/rpc-server.cpp create mode 100644 examples/save-load-state/CMakeLists.txt create mode 100644 examples/save-load-state/save-load-state.cpp create mode 100755 examples/server-llama2-13B.sh create mode 100644 examples/server/CMakeLists.txt create mode 100644 examples/server/README.md create mode 100644 examples/server/bench/README.md create mode 100644 examples/server/bench/bench.py create mode 100644 examples/server/bench/prometheus.yml create mode 100644 examples/server/bench/requirements.txt create mode 100644 examples/server/bench/script.js create mode 100755 examples/server/chat-llama2.sh create mode 100644 examples/server/chat.mjs create mode 100755 examples/server/chat.sh create mode 100755 examples/server/deps.sh create mode 100644 examples/server/httplib.h create mode 100755 examples/server/public/colorthemes.css create mode 100644 examples/server/public/completion.js create mode 100644 examples/server/public/favicon.ico create mode 100644 examples/server/public/index-new.html create mode 100644 examples/server/public/index.html create mode 100644 examples/server/public/index.js create mode 100644 examples/server/public/json-schema-to-grammar.mjs create mode 100644 examples/server/public/prompt-formats.js create mode 100755 examples/server/public/style.css create mode 100644 examples/server/public/system-prompts.js create mode 100755 examples/server/public/theme-beeninorder.css create mode 100755 examples/server/public/theme-ketivah.css create mode 100755 examples/server/public/theme-mangotango.css create mode 100755 examples/server/public/theme-playground.css create mode 100755 examples/server/public/theme-polarnight.css create mode 100755 examples/server/public/theme-snowstorm.css create mode 100644 examples/server/public_simplechat/datautils.mjs create mode 100644 examples/server/public_simplechat/index.html create mode 100644 examples/server/public_simplechat/readme.md create mode 100644 examples/server/public_simplechat/simplechat.css create mode 100644 examples/server/public_simplechat/simplechat.js create mode 100644 examples/server/public_simplechat/simplechat_screens.webp create mode 100644 examples/server/public_simplechat/ui.mjs create mode 100644 examples/server/server.cpp create mode 100644 examples/server/tests/README.md create mode 100644 examples/server/tests/features/embeddings.feature create mode 100644 examples/server/tests/features/environment.py create mode 100644 examples/server/tests/features/issues.feature create mode 100644 examples/server/tests/features/lora.feature create mode 100644 examples/server/tests/features/parallel.feature create mode 100644 examples/server/tests/features/passkey.feature create mode 100644 examples/server/tests/features/results.feature create mode 100644 examples/server/tests/features/security.feature create mode 100644 examples/server/tests/features/server.feature create mode 100644 examples/server/tests/features/slotsave.feature create mode 100644 examples/server/tests/features/steps/steps.py create mode 100644 examples/server/tests/features/wrong_usages.feature create mode 100644 examples/server/tests/requirements.txt create mode 100755 examples/server/tests/tests.sh create mode 100644 examples/server/themes/README.md create mode 100644 examples/server/themes/buttons-top/README.md create mode 100644 examples/server/themes/buttons-top/buttons_top.png create mode 100644 examples/server/themes/buttons-top/favicon.ico create mode 100644 examples/server/themes/buttons-top/index.html create mode 100644 examples/server/themes/wild/README.md create mode 100644 examples/server/themes/wild/favicon.ico create mode 100644 examples/server/themes/wild/index.html create mode 100644 examples/server/themes/wild/llama_cpp.png create mode 100644 examples/server/themes/wild/llamapattern.png create mode 100644 examples/server/themes/wild/wild.png create mode 100644 examples/server/utils.hpp create mode 100644 examples/server_embd.py create mode 100644 examples/simple/CMakeLists.txt create mode 100644 examples/simple/README.md create mode 100644 examples/simple/simple.cpp create mode 100644 examples/speculative/CMakeLists.txt create mode 100644 examples/speculative/README.md create mode 100644 examples/speculative/speculative.cpp create mode 100644 examples/sycl/CMakeLists.txt create mode 100644 examples/sycl/README.md create mode 100755 examples/sycl/build.sh create mode 100644 examples/sycl/ls-sycl-device.cpp create mode 100755 examples/sycl/run-llama2.sh create mode 100644 examples/sycl/win-build-sycl.bat create mode 100644 examples/sycl/win-run-llama2.bat create mode 100644 examples/tokenize/CMakeLists.txt create mode 100644 examples/tokenize/tokenize.cpp create mode 100755 examples/ts-type-to-grammar.sh create mode 100644 flake.lock create mode 100644 flake.nix create mode 100644 ggml/.gitignore create mode 100644 ggml/CMakeLists.txt create mode 100644 ggml/cmake/FindSIMD.cmake create mode 100644 ggml/include/ggml-alloc.h create mode 100644 ggml/include/ggml-backend.h create mode 100644 ggml/include/ggml-blas.h create mode 100644 ggml/include/ggml-cann.h create mode 100644 ggml/include/ggml-cuda.h create mode 100644 ggml/include/ggml-kompute.h create mode 100644 ggml/include/ggml-metal.h create mode 100644 ggml/include/ggml-rpc.h create mode 100644 ggml/include/ggml-sycl.h create mode 100644 ggml/include/ggml-vulkan.h create mode 100644 ggml/include/ggml.h create mode 100644 ggml/src/CMakeLists.txt create mode 100644 ggml/src/ggml-aarch64.c create mode 100644 ggml/src/ggml-aarch64.h create mode 100644 ggml/src/ggml-alloc.c create mode 100644 ggml/src/ggml-backend-impl.h create mode 100644 ggml/src/ggml-backend.c create mode 100644 ggml/src/ggml-blas.cpp create mode 100644 ggml/src/ggml-cann.cpp create mode 100644 ggml/src/ggml-cann/.clang-format create mode 100644 ggml/src/ggml-cann/Doxyfile create mode 100644 ggml/src/ggml-cann/acl_tensor.cpp create mode 100644 ggml/src/ggml-cann/acl_tensor.h create mode 100644 ggml/src/ggml-cann/aclnn_ops.cpp create mode 100644 ggml/src/ggml-cann/aclnn_ops.h create mode 100644 ggml/src/ggml-cann/common.h create mode 100644 ggml/src/ggml-cann/kernels/CMakeLists.txt create mode 100644 ggml/src/ggml-cann/kernels/ascendc_kernels.h create mode 100644 ggml/src/ggml-cann/kernels/dup.cpp create mode 100644 ggml/src/ggml-cann/kernels/get_row_f16.cpp create mode 100644 ggml/src/ggml-cann/kernels/get_row_f32.cpp create mode 100644 ggml/src/ggml-cann/kernels/get_row_q4_0.cpp create mode 100644 ggml/src/ggml-cann/kernels/get_row_q8_0.cpp create mode 100644 ggml/src/ggml-cann/kernels/quantize_f16_q8_0.cpp create mode 100644 ggml/src/ggml-cann/kernels/quantize_f32_q8_0.cpp create mode 100644 ggml/src/ggml-cann/kernels/quantize_float_to_q4_0.cpp create mode 100644 ggml/src/ggml-common.h create mode 100644 ggml/src/ggml-cuda.cu create mode 100644 ggml/src/ggml-cuda/acc.cu create mode 100644 ggml/src/ggml-cuda/acc.cuh create mode 100644 ggml/src/ggml-cuda/arange.cu create mode 100644 ggml/src/ggml-cuda/arange.cuh create mode 100644 ggml/src/ggml-cuda/argsort.cu create mode 100644 ggml/src/ggml-cuda/argsort.cuh create mode 100644 ggml/src/ggml-cuda/binbcast.cu create mode 100644 ggml/src/ggml-cuda/binbcast.cuh create mode 100644 ggml/src/ggml-cuda/clamp.cu create mode 100644 ggml/src/ggml-cuda/clamp.cuh create mode 100644 ggml/src/ggml-cuda/common.cuh create mode 100644 ggml/src/ggml-cuda/concat.cu create mode 100644 ggml/src/ggml-cuda/concat.cuh create mode 100644 ggml/src/ggml-cuda/conv-transpose-1d.cu create mode 100644 ggml/src/ggml-cuda/conv-transpose-1d.cuh create mode 100644 ggml/src/ggml-cuda/convert.cu create mode 100644 ggml/src/ggml-cuda/convert.cuh create mode 100644 ggml/src/ggml-cuda/cpy.cu create mode 100644 ggml/src/ggml-cuda/cpy.cuh create mode 100644 ggml/src/ggml-cuda/dequantize.cuh create mode 100644 ggml/src/ggml-cuda/diagmask.cu create mode 100644 ggml/src/ggml-cuda/diagmask.cuh create mode 100644 ggml/src/ggml-cuda/dmmv.cu create mode 100644 ggml/src/ggml-cuda/dmmv.cuh create mode 100644 ggml/src/ggml-cuda/fattn-common.cuh create mode 100644 ggml/src/ggml-cuda/fattn-tile-f16.cu create mode 100644 ggml/src/ggml-cuda/fattn-tile-f16.cuh create mode 100644 ggml/src/ggml-cuda/fattn-tile-f32.cu create mode 100644 ggml/src/ggml-cuda/fattn-tile-f32.cuh create mode 100644 ggml/src/ggml-cuda/fattn-vec-f16.cuh create mode 100644 ggml/src/ggml-cuda/fattn-vec-f32.cuh create mode 100644 ggml/src/ggml-cuda/fattn-wmma-f16.cuh create mode 100644 ggml/src/ggml-cuda/fattn.cu create mode 100644 ggml/src/ggml-cuda/fattn.cuh create mode 100644 ggml/src/ggml-cuda/getrows.cu create mode 100644 ggml/src/ggml-cuda/getrows.cuh create mode 100644 ggml/src/ggml-cuda/im2col.cu create mode 100644 ggml/src/ggml-cuda/im2col.cuh create mode 100644 ggml/src/ggml-cuda/mma.cuh create mode 100644 ggml/src/ggml-cuda/mmq.cu create mode 100644 ggml/src/ggml-cuda/mmq.cuh create mode 100644 ggml/src/ggml-cuda/mmvq.cu create mode 100644 ggml/src/ggml-cuda/mmvq.cuh create mode 100644 ggml/src/ggml-cuda/norm.cu create mode 100644 ggml/src/ggml-cuda/norm.cuh create mode 100644 ggml/src/ggml-cuda/pad.cu create mode 100644 ggml/src/ggml-cuda/pad.cuh create mode 100644 ggml/src/ggml-cuda/pool2d.cu create mode 100644 ggml/src/ggml-cuda/pool2d.cuh create mode 100644 ggml/src/ggml-cuda/quantize.cu create mode 100644 ggml/src/ggml-cuda/quantize.cuh create mode 100644 ggml/src/ggml-cuda/rope.cu create mode 100644 ggml/src/ggml-cuda/rope.cuh create mode 100644 ggml/src/ggml-cuda/scale.cu create mode 100644 ggml/src/ggml-cuda/scale.cuh create mode 100644 ggml/src/ggml-cuda/softmax.cu create mode 100644 ggml/src/ggml-cuda/softmax.cuh create mode 100644 ggml/src/ggml-cuda/sumrows.cu create mode 100644 ggml/src/ggml-cuda/sumrows.cuh create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-f16-f16.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-f16-q4_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-f16-q4_1.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-f16-q5_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-f16-q5_1.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-f16-q8_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-q4_0-f16.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-q4_0-q4_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-q4_0-q4_1.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-q4_0-q5_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-q4_0-q5_1.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-q4_0-q8_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-q4_1-f16.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-q4_1-q4_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-q4_1-q4_1.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-q4_1-q5_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-q4_1-q5_1.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-q4_1-q8_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-q5_0-f16.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-q5_0-q4_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-q5_0-q4_1.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-q5_0-q5_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-q5_0-q5_1.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-q5_0-q8_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-q5_1-f16.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-q5_1-q4_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-q5_1-q4_1.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-q5_1-q5_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-q5_1-q5_1.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-q5_1-q8_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-q8_0-f16.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-q8_0-q4_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-q8_0-q4_1.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-q8_0-q5_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-q8_0-q5_1.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs128-q8_0-q8_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs256-f16-f16.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs64-f16-f16.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs64-f16-q4_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs64-f16-q4_1.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs64-f16-q5_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs64-f16-q5_1.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f16-instance-hs64-f16-q8_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-f16-f16.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-f16-q4_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-f16-q4_1.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-f16-q5_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-f16-q5_1.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-f16-q8_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-q4_0-f16.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-q4_0-q4_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-q4_0-q4_1.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-q4_0-q5_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-q4_0-q5_1.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-q4_0-q8_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-q4_1-f16.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-q4_1-q4_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-q4_1-q4_1.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-q4_1-q5_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-q4_1-q5_1.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-q4_1-q8_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-q5_0-f16.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-q5_0-q4_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-q5_0-q4_1.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-q5_0-q5_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-q5_0-q5_1.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-q5_0-q8_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-q5_1-f16.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-q5_1-q4_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-q5_1-q4_1.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-q5_1-q5_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-q5_1-q5_1.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-q5_1-q8_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-q8_0-f16.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-q8_0-q4_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-q8_0-q4_1.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-q8_0-q5_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-q8_0-q5_1.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs128-q8_0-q8_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs256-f16-f16.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs64-f16-f16.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs64-f16-q4_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs64-f16-q4_1.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs64-f16-q5_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs64-f16-q5_1.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-vec-f32-instance-hs64-f16-q8_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-wmma-f16-instance-kqfloat-cpb16.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-wmma-f16-instance-kqfloat-cpb32.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-wmma-f16-instance-kqhalf-cpb16.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-wmma-f16-instance-kqhalf-cpb32.cu create mode 100644 ggml/src/ggml-cuda/template-instances/fattn-wmma-f16-instance-kqhalf-cpb8.cu create mode 100755 ggml/src/ggml-cuda/template-instances/generate_cu_files.py create mode 100644 ggml/src/ggml-cuda/template-instances/mmq-instance-iq1_s.cu create mode 100644 ggml/src/ggml-cuda/template-instances/mmq-instance-iq2_s.cu create mode 100644 ggml/src/ggml-cuda/template-instances/mmq-instance-iq2_xs.cu create mode 100644 ggml/src/ggml-cuda/template-instances/mmq-instance-iq2_xxs.cu create mode 100644 ggml/src/ggml-cuda/template-instances/mmq-instance-iq3_s.cu create mode 100644 ggml/src/ggml-cuda/template-instances/mmq-instance-iq3_xxs.cu create mode 100644 ggml/src/ggml-cuda/template-instances/mmq-instance-iq4_nl.cu create mode 100644 ggml/src/ggml-cuda/template-instances/mmq-instance-iq4_xs.cu create mode 100644 ggml/src/ggml-cuda/template-instances/mmq-instance-q2_k.cu create mode 100644 ggml/src/ggml-cuda/template-instances/mmq-instance-q3_k.cu create mode 100644 ggml/src/ggml-cuda/template-instances/mmq-instance-q4_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/mmq-instance-q4_1.cu create mode 100644 ggml/src/ggml-cuda/template-instances/mmq-instance-q4_k.cu create mode 100644 ggml/src/ggml-cuda/template-instances/mmq-instance-q5_0.cu create mode 100644 ggml/src/ggml-cuda/template-instances/mmq-instance-q5_1.cu create mode 100644 ggml/src/ggml-cuda/template-instances/mmq-instance-q5_k.cu create mode 100644 ggml/src/ggml-cuda/template-instances/mmq-instance-q6_k.cu create mode 100644 ggml/src/ggml-cuda/template-instances/mmq-instance-q8_0.cu create mode 100644 ggml/src/ggml-cuda/tsembd.cu create mode 100644 ggml/src/ggml-cuda/tsembd.cuh create mode 100644 ggml/src/ggml-cuda/unary.cu create mode 100644 ggml/src/ggml-cuda/unary.cuh create mode 100644 ggml/src/ggml-cuda/upscale.cu create mode 100644 ggml/src/ggml-cuda/upscale.cuh create mode 100644 ggml/src/ggml-cuda/vecdotq.cuh create mode 100644 ggml/src/ggml-cuda/vendors/cuda.h create mode 100644 ggml/src/ggml-cuda/vendors/hip.h create mode 100644 ggml/src/ggml-cuda/vendors/musa.h create mode 100644 ggml/src/ggml-impl.h create mode 100644 ggml/src/ggml-kompute.cpp create mode 100644 ggml/src/ggml-metal.m create mode 100644 ggml/src/ggml-metal.metal create mode 100644 ggml/src/ggml-quants.c create mode 100644 ggml/src/ggml-quants.h create mode 100644 ggml/src/ggml-rpc.cpp create mode 100644 ggml/src/ggml-sycl.cpp create mode 100644 ggml/src/ggml-sycl/backend.hpp create mode 100644 ggml/src/ggml-sycl/common.cpp create mode 100644 ggml/src/ggml-sycl/common.hpp create mode 100644 ggml/src/ggml-sycl/concat.cpp create mode 100644 ggml/src/ggml-sycl/concat.hpp create mode 100644 ggml/src/ggml-sycl/conv.cpp create mode 100644 ggml/src/ggml-sycl/conv.hpp create mode 100644 ggml/src/ggml-sycl/convert.cpp create mode 100644 ggml/src/ggml-sycl/convert.hpp create mode 100644 ggml/src/ggml-sycl/dequantize.hpp create mode 100644 ggml/src/ggml-sycl/dmmv.cpp create mode 100644 ggml/src/ggml-sycl/dmmv.hpp create mode 100644 ggml/src/ggml-sycl/dpct/helper.hpp create mode 100644 ggml/src/ggml-sycl/mmq.cpp create mode 100644 ggml/src/ggml-sycl/mmq.hpp create mode 100644 ggml/src/ggml-sycl/mmvq.cpp create mode 100644 ggml/src/ggml-sycl/mmvq.hpp create mode 100644 ggml/src/ggml-sycl/norm.cpp create mode 100644 ggml/src/ggml-sycl/norm.hpp create mode 100644 ggml/src/ggml-sycl/presets.hpp create mode 100644 ggml/src/ggml-sycl/rope.cpp create mode 100644 ggml/src/ggml-sycl/rope.hpp create mode 100644 ggml/src/ggml-sycl/softmax.cpp create mode 100644 ggml/src/ggml-sycl/softmax.hpp create mode 100644 ggml/src/ggml-sycl/tsembd.cpp create mode 100644 ggml/src/ggml-sycl/tsembd.hpp create mode 100644 ggml/src/ggml-sycl/vecdotq.hpp create mode 100644 ggml/src/ggml-vulkan.cpp create mode 100644 ggml/src/ggml.c create mode 100644 ggml/src/kompute-shaders/common.comp create mode 100644 ggml/src/kompute-shaders/op_add.comp create mode 100644 ggml/src/kompute-shaders/op_addrow.comp create mode 100644 ggml/src/kompute-shaders/op_cpy_f16_f16.comp create mode 100644 ggml/src/kompute-shaders/op_cpy_f16_f32.comp create mode 100644 ggml/src/kompute-shaders/op_cpy_f32_f16.comp create mode 100644 ggml/src/kompute-shaders/op_cpy_f32_f32.comp create mode 100644 ggml/src/kompute-shaders/op_diagmask.comp create mode 100644 ggml/src/kompute-shaders/op_gelu.comp create mode 100644 ggml/src/kompute-shaders/op_getrows.comp create mode 100644 ggml/src/kompute-shaders/op_getrows_f16.comp create mode 100644 ggml/src/kompute-shaders/op_getrows_f32.comp create mode 100644 ggml/src/kompute-shaders/op_getrows_q4_0.comp create mode 100644 ggml/src/kompute-shaders/op_getrows_q4_1.comp create mode 100644 ggml/src/kompute-shaders/op_getrows_q6_k.comp create mode 100644 ggml/src/kompute-shaders/op_mul.comp create mode 100644 ggml/src/kompute-shaders/op_mul_mat_f16.comp create mode 100644 ggml/src/kompute-shaders/op_mul_mat_mat_f32.comp create mode 100644 ggml/src/kompute-shaders/op_mul_mat_q4_0.comp create mode 100644 ggml/src/kompute-shaders/op_mul_mat_q4_1.comp create mode 100644 ggml/src/kompute-shaders/op_mul_mat_q6_k.comp create mode 100644 ggml/src/kompute-shaders/op_mul_mat_q8_0.comp create mode 100644 ggml/src/kompute-shaders/op_mul_mv_q_n.comp create mode 100644 ggml/src/kompute-shaders/op_mul_mv_q_n_pre.comp create mode 100644 ggml/src/kompute-shaders/op_norm.comp create mode 100644 ggml/src/kompute-shaders/op_relu.comp create mode 100644 ggml/src/kompute-shaders/op_rmsnorm.comp create mode 100644 ggml/src/kompute-shaders/op_rope_f16.comp create mode 100644 ggml/src/kompute-shaders/op_rope_f32.comp create mode 100644 ggml/src/kompute-shaders/op_scale.comp create mode 100644 ggml/src/kompute-shaders/op_scale_8.comp create mode 100644 ggml/src/kompute-shaders/op_silu.comp create mode 100644 ggml/src/kompute-shaders/op_softmax.comp create mode 100644 ggml/src/kompute-shaders/rope_common.comp create mode 100644 ggml/src/llamafile/sgemm.cpp create mode 100644 ggml/src/llamafile/sgemm.h create mode 100644 ggml/src/vulkan-shaders/CMakeLists.txt create mode 100644 ggml/src/vulkan-shaders/add.comp create mode 100644 ggml/src/vulkan-shaders/argsort.comp create mode 100644 ggml/src/vulkan-shaders/clamp.comp create mode 100644 ggml/src/vulkan-shaders/concat.comp create mode 100644 ggml/src/vulkan-shaders/copy.comp create mode 100644 ggml/src/vulkan-shaders/dequant_f32.comp create mode 100644 ggml/src/vulkan-shaders/dequant_funcs.comp create mode 100644 ggml/src/vulkan-shaders/dequant_head.comp create mode 100644 ggml/src/vulkan-shaders/dequant_iq4_nl.comp create mode 100644 ggml/src/vulkan-shaders/dequant_q2_k.comp create mode 100644 ggml/src/vulkan-shaders/dequant_q3_k.comp create mode 100644 ggml/src/vulkan-shaders/dequant_q4_0.comp create mode 100644 ggml/src/vulkan-shaders/dequant_q4_1.comp create mode 100644 ggml/src/vulkan-shaders/dequant_q4_k.comp create mode 100644 ggml/src/vulkan-shaders/dequant_q5_0.comp create mode 100644 ggml/src/vulkan-shaders/dequant_q5_1.comp create mode 100644 ggml/src/vulkan-shaders/dequant_q5_k.comp create mode 100644 ggml/src/vulkan-shaders/dequant_q6_k.comp create mode 100644 ggml/src/vulkan-shaders/dequant_q8_0.comp create mode 100644 ggml/src/vulkan-shaders/diag_mask_inf.comp create mode 100644 ggml/src/vulkan-shaders/div.comp create mode 100644 ggml/src/vulkan-shaders/gelu.comp create mode 100644 ggml/src/vulkan-shaders/gelu_quick.comp create mode 100644 ggml/src/vulkan-shaders/generic_binary_head.comp create mode 100644 ggml/src/vulkan-shaders/generic_head.comp create mode 100644 ggml/src/vulkan-shaders/generic_unary_head.comp create mode 100644 ggml/src/vulkan-shaders/get_rows.comp create mode 100644 ggml/src/vulkan-shaders/get_rows_quant.comp create mode 100644 ggml/src/vulkan-shaders/group_norm.comp create mode 100644 ggml/src/vulkan-shaders/im2col.comp create mode 100644 ggml/src/vulkan-shaders/leaky_relu.comp create mode 100644 ggml/src/vulkan-shaders/mul.comp create mode 100644 ggml/src/vulkan-shaders/mul_mat_split_k_reduce.comp create mode 100644 ggml/src/vulkan-shaders/mul_mat_vec.comp create mode 100644 ggml/src/vulkan-shaders/mul_mat_vec_base.comp create mode 100644 ggml/src/vulkan-shaders/mul_mat_vec_nc.comp create mode 100644 ggml/src/vulkan-shaders/mul_mat_vec_p021.comp create mode 100644 ggml/src/vulkan-shaders/mul_mat_vec_q2_k.comp create mode 100644 ggml/src/vulkan-shaders/mul_mat_vec_q3_k.comp create mode 100644 ggml/src/vulkan-shaders/mul_mat_vec_q4_k.comp create mode 100644 ggml/src/vulkan-shaders/mul_mat_vec_q5_k.comp create mode 100644 ggml/src/vulkan-shaders/mul_mat_vec_q6_k.comp create mode 100644 ggml/src/vulkan-shaders/mul_mm.comp create mode 100644 ggml/src/vulkan-shaders/norm.comp create mode 100644 ggml/src/vulkan-shaders/pad.comp create mode 100644 ggml/src/vulkan-shaders/relu.comp create mode 100644 ggml/src/vulkan-shaders/rms_norm.comp create mode 100644 ggml/src/vulkan-shaders/rope_head.comp create mode 100644 ggml/src/vulkan-shaders/rope_neox.comp create mode 100644 ggml/src/vulkan-shaders/rope_norm.comp create mode 100644 ggml/src/vulkan-shaders/scale.comp create mode 100644 ggml/src/vulkan-shaders/silu.comp create mode 100644 ggml/src/vulkan-shaders/soft_max.comp create mode 100644 ggml/src/vulkan-shaders/square.comp create mode 100644 ggml/src/vulkan-shaders/sum_rows.comp create mode 100644 ggml/src/vulkan-shaders/tanh.comp create mode 100644 ggml/src/vulkan-shaders/timestep_embedding.comp create mode 100644 ggml/src/vulkan-shaders/types.comp create mode 100644 ggml/src/vulkan-shaders/upscale.comp create mode 100644 ggml/src/vulkan-shaders/vulkan-shaders-gen.cpp create mode 100644 gguf-py/LICENSE create mode 100644 gguf-py/README.md create mode 100644 gguf-py/examples/reader.py create mode 100755 gguf-py/examples/writer.py create mode 100644 gguf-py/gguf/__init__.py create mode 100644 gguf-py/gguf/constants.py create mode 100644 gguf-py/gguf/gguf.py create mode 100644 gguf-py/gguf/gguf_reader.py create mode 100644 gguf-py/gguf/gguf_writer.py create mode 100644 gguf-py/gguf/lazy.py create mode 100644 gguf-py/gguf/metadata.py create mode 100644 gguf-py/gguf/py.typed create mode 100644 gguf-py/gguf/quants.py create mode 100644 gguf-py/gguf/tensor_mapping.py create mode 100644 gguf-py/gguf/utility.py create mode 100644 gguf-py/gguf/vocab.py create mode 100644 gguf-py/pyproject.toml create mode 100644 gguf-py/scripts/__init__.py create mode 100755 gguf-py/scripts/gguf_convert_endian.py create mode 100755 gguf-py/scripts/gguf_dump.py create mode 100755 gguf-py/scripts/gguf_hash.py create mode 100755 gguf-py/scripts/gguf_new_metadata.py create mode 100755 gguf-py/scripts/gguf_set_metadata.py create mode 100644 gguf-py/tests/__init__.py create mode 100755 gguf-py/tests/test_metadata.py create mode 100644 grammars/README.md create mode 100644 grammars/arithmetic.gbnf create mode 100644 grammars/c.gbnf create mode 100644 grammars/chess.gbnf create mode 100644 grammars/japanese.gbnf create mode 100644 grammars/json.gbnf create mode 100644 grammars/json_arr.gbnf create mode 100644 grammars/list.gbnf create mode 100644 include/llama.h create mode 100644 media/llama-leader.jpeg create mode 100644 media/llama0-banner.png create mode 100644 media/llama0-logo.png create mode 100644 media/llama1-banner.png create mode 100644 media/llama1-logo.png create mode 100644 media/matmul.png create mode 100644 media/matmul.svg create mode 100644 models/.editorconfig create mode 100644 models/ggml-vocab-aquila.gguf create mode 100644 models/ggml-vocab-baichuan.gguf create mode 100644 models/ggml-vocab-bert-bge.gguf create mode 100644 models/ggml-vocab-bert-bge.gguf.inp create mode 100644 models/ggml-vocab-bert-bge.gguf.out create mode 100644 models/ggml-vocab-command-r.gguf create mode 100644 models/ggml-vocab-command-r.gguf.inp create mode 100644 models/ggml-vocab-command-r.gguf.out create mode 100644 models/ggml-vocab-deepseek-coder.gguf create mode 100644 models/ggml-vocab-deepseek-coder.gguf.inp create mode 100644 models/ggml-vocab-deepseek-coder.gguf.out create mode 100644 models/ggml-vocab-deepseek-llm.gguf create mode 100644 models/ggml-vocab-deepseek-llm.gguf.inp create mode 100644 models/ggml-vocab-deepseek-llm.gguf.out create mode 100644 models/ggml-vocab-falcon.gguf create mode 100644 models/ggml-vocab-falcon.gguf.inp create mode 100644 models/ggml-vocab-falcon.gguf.out create mode 100644 models/ggml-vocab-gpt-2.gguf create mode 100644 models/ggml-vocab-gpt-2.gguf.inp create mode 100644 models/ggml-vocab-gpt-2.gguf.out create mode 100644 models/ggml-vocab-gpt-neox.gguf create mode 100644 models/ggml-vocab-llama-bpe.gguf create mode 100644 models/ggml-vocab-llama-bpe.gguf.inp create mode 100644 models/ggml-vocab-llama-bpe.gguf.out create mode 100644 models/ggml-vocab-llama-spm.gguf create mode 100644 models/ggml-vocab-llama-spm.gguf.inp create mode 100644 models/ggml-vocab-llama-spm.gguf.out create mode 100644 models/ggml-vocab-mpt.gguf create mode 100644 models/ggml-vocab-mpt.gguf.inp create mode 100644 models/ggml-vocab-mpt.gguf.out create mode 100644 models/ggml-vocab-phi-3.gguf create mode 100644 models/ggml-vocab-phi-3.gguf.inp create mode 100644 models/ggml-vocab-phi-3.gguf.out create mode 100644 models/ggml-vocab-qwen2.gguf create mode 100644 models/ggml-vocab-qwen2.gguf.inp create mode 100644 models/ggml-vocab-qwen2.gguf.out create mode 100644 models/ggml-vocab-refact.gguf create mode 100644 models/ggml-vocab-refact.gguf.inp create mode 100644 models/ggml-vocab-refact.gguf.out create mode 100644 models/ggml-vocab-starcoder.gguf create mode 100644 models/ggml-vocab-starcoder.gguf.inp create mode 100644 models/ggml-vocab-starcoder.gguf.out create mode 100644 mypy.ini create mode 100644 pocs/CMakeLists.txt create mode 100644 pocs/vdot/CMakeLists.txt create mode 100644 pocs/vdot/q8dot.cpp create mode 100644 pocs/vdot/vdot.cpp create mode 100644 poetry.lock create mode 100644 prompts/LLM-questions.txt create mode 100644 prompts/alpaca.txt create mode 100644 prompts/assistant.txt create mode 100644 prompts/chat-with-baichuan.txt create mode 100644 prompts/chat-with-bob.txt create mode 100644 prompts/chat-with-qwen.txt create mode 100644 prompts/chat-with-vicuna-v0.txt create mode 100644 prompts/chat-with-vicuna-v1.txt create mode 100644 prompts/chat.txt create mode 100644 prompts/dan-modified.txt create mode 100644 prompts/dan.txt create mode 100644 prompts/mnemonics.txt create mode 100644 prompts/parallel-questions.txt create mode 100644 prompts/reason-act.txt create mode 100644 pyproject.toml create mode 100644 pyrightconfig.json create mode 100644 requirements.txt create mode 100644 requirements/requirements-all.txt create mode 100644 requirements/requirements-compare-llama-bench.txt create mode 100644 requirements/requirements-convert_hf_to_gguf.txt create mode 100644 requirements/requirements-convert_hf_to_gguf_update.txt create mode 100644 requirements/requirements-convert_legacy_llama.txt create mode 100644 requirements/requirements-convert_llama_ggml_to_gguf.txt create mode 100644 requirements/requirements-convert_lora_to_gguf.txt create mode 100644 requirements/requirements-pydantic.txt create mode 100644 requirements/requirements-test-tokenizer-random.txt create mode 100755 scripts/build-info.sh create mode 100755 scripts/check-requirements.sh create mode 100755 scripts/ci-run.sh create mode 100755 scripts/compare-commits.sh create mode 100755 scripts/compare-llama-bench.py create mode 100755 scripts/debug-test.sh create mode 100755 scripts/gen-authors.sh create mode 100644 scripts/gen-unicode-data.py create mode 100644 scripts/get-flags.mk create mode 100755 scripts/get-hellaswag.sh create mode 100755 scripts/get-pg.sh create mode 100755 scripts/get-wikitext-103.sh create mode 100755 scripts/get-wikitext-2.sh create mode 100755 scripts/get-winogrande.sh create mode 100755 scripts/hf.sh create mode 100644 scripts/install-oneapi.bat create mode 100644 scripts/pod-llama.sh create mode 100755 scripts/qnt-all.sh create mode 100755 scripts/run-all-perf.sh create mode 100755 scripts/run-all-ppl.sh create mode 100755 scripts/run-with-preset.py create mode 100644 scripts/server-llm.sh create mode 100755 scripts/sync-ggml-am.sh create mode 100644 scripts/sync-ggml.last create mode 100755 scripts/sync-ggml.sh create mode 100755 scripts/verify-checksum-models.py create mode 100644 scripts/xxd.cmake create mode 120000 spm-headers/ggml-alloc.h create mode 120000 spm-headers/ggml-backend.h create mode 120000 spm-headers/ggml-metal.h create mode 120000 spm-headers/ggml.h create mode 120000 spm-headers/llama.h create mode 100644 src/CMakeLists.txt create mode 100644 src/llama-grammar.cpp create mode 100644 src/llama-grammar.h create mode 100644 src/llama-impl.h create mode 100644 src/llama-sampling.cpp create mode 100644 src/llama-sampling.h create mode 100644 src/llama-vocab.cpp create mode 100644 src/llama-vocab.h create mode 100644 src/llama.cpp create mode 100644 src/unicode-data.cpp create mode 100644 src/unicode-data.h create mode 100644 src/unicode.cpp create mode 100644 src/unicode.h create mode 100644 tests/.gitignore create mode 100644 tests/CMakeLists.txt create mode 100644 tests/get-model.cpp create mode 100644 tests/get-model.h create mode 100644 tests/run-json-schema-to-grammar.mjs create mode 100644 tests/test-autorelease.cpp create mode 100644 tests/test-backend-ops.cpp create mode 100644 tests/test-c.c create mode 100644 tests/test-chat-template.cpp create mode 100644 tests/test-double-float.cpp create mode 100644 tests/test-grad0.cpp create mode 100644 tests/test-grammar-integration.cpp create mode 100644 tests/test-grammar-parser.cpp create mode 100755 tests/test-json-schema-to-grammar.cpp create mode 100644 tests/test-llama-grammar.cpp create mode 100644 tests/test-model-load-cancel.cpp create mode 100644 tests/test-opt.cpp create mode 100644 tests/test-quantize-fns.cpp create mode 100644 tests/test-quantize-perf.cpp create mode 100644 tests/test-rope.cpp create mode 100644 tests/test-sampling.cpp create mode 100644 tests/test-tokenizer-0.cpp create mode 100644 tests/test-tokenizer-0.py create mode 100755 tests/test-tokenizer-0.sh create mode 100644 tests/test-tokenizer-1-bpe.cpp create mode 100644 tests/test-tokenizer-1-spm.cpp create mode 100644 tests/test-tokenizer-random.py + cd ../../ + tar -xzf /builddir/build/SOURCES/main.tar.gz + mv ollama-ui-main ollama-ui + RPM_EC=0 ++ jobs -p + exit 0 + umask 022 + cd /builddir/build/BUILD + cd ollama-0.3.6 + export GOPROXY=https://goproxy.cn + GOPROXY=https://goproxy.cn + export OLLAMA_SKIP_PATCHING=1 + OLLAMA_SKIP_PATCHING=1 + export 'GOFLAGS='\''-ldflags=-w -s "-X=github.com/ollama/ollama/version.Version=0.3.6" "-X=github.com/ollama/ollama/server.mode=release"'\''' + GOFLAGS=''\''-ldflags=-w -s "-X=github.com/ollama/ollama/version.Version=0.3.6" "-X=github.com/ollama/ollama/server.mode=release"'\''' + go generate ./... go: downloading go1.22.5 (linux/amd64) Executing(%build): /bin/sh -e /var/tmp/rpm-tmp.W3AyPY + set -o pipefail Starting linux generate script + echo 'Starting linux generate script' + '[' -z '' ']' + '[' -x /usr/local/cuda/bin/nvcc ']' ++ command -v nvcc + export CUDACXX= + CUDACXX= + COMMON_CMAKE_DEFS='-DBUILD_SHARED_LIBS=off -DCMAKE_POSITION_INDEPENDENT_CODE=on -DGGML_NATIVE=off -DGGML_AVX=on -DGGML_AVX2=off -DGGML_AVX512=off -DGGML_FMA=off -DGGML_F16C=off -DGGML_OPENMP=off' ++ dirname ./gen_linux.sh + source ./gen_common.sh + init_vars + case "${GOARCH}" in + ARCH=x86_64 + LLAMACPP_DIR=../llama.cpp + CMAKE_DEFS= + CMAKE_TARGETS='--target ollama_llama_server' + echo '' + grep -- -g + CMAKE_DEFS='-DCMAKE_BUILD_TYPE=Release -DLLAMA_SERVER_VERBOSE=off ' + case $(uname -s) in ++ uname -s + LIB_EXT=so + WHOLE_ARCHIVE=-Wl,--whole-archive + NO_WHOLE_ARCHIVE=-Wl,--no-whole-archive + GCC_ARCH= Skipping submodule initialization + '[' -z '' ']' + CMAKE_CUDA_ARCHITECTURES='50;52;61;70;75;80' + git_module_setup + '[' -n 1 ']' + echo 'Skipping submodule initialization' + return + apply_patches + grep ollama ../llama.cpp/CMakeLists.txt + echo 'add_subdirectory(../ext_server ext_server) # ollama' ++ ls -A ../patches/01-load-progress.diff ../patches/02-clip-log.diff ../patches/03-load_exception.diff ../patches/04-metal.diff ../patches/05-default-pretokenizer.diff ../patches/06-embeddings.diff ../patches/07-clip-unicode.diff ../patches/08-pooling.diff ../patches/09-lora.diff ../patches/11-phi3-sliding-window.diff + '[' -n '../patches/01-load-progress.diff ../patches/02-clip-log.diff ../patches/03-load_exception.diff ../patches/04-metal.diff ../patches/05-default-pretokenizer.diff ../patches/06-embeddings.diff ../patches/07-clip-unicode.diff ../patches/08-pooling.diff ../patches/09-lora.diff ../patches/11-phi3-sliding-window.diff' ']' + for patch in ../patches/*.diff ++ grep '^+++ ' ../patches/01-load-progress.diff ++ cut -f2 '-d ' ++ cut -f2- -d/ + for file in $(grep "^+++ " ${patch} | cut -f2 -d' ' | cut -f2- -d/) + cd ../llama.cpp + git checkout common/common.cpp Updated 0 paths from the index + for file in $(grep "^+++ " ${patch} | cut -f2 -d' ' | cut -f2- -d/) + cd ../llama.cpp + git checkout common/common.h Updated 0 paths from the index + for patch in ../patches/*.diff ++ grep '^+++ ' ../patches/02-clip-log.diff ++ cut -f2 '-d ' ++ cut -f2- -d/ + for file in $(grep "^+++ " ${patch} | cut -f2 -d' ' | cut -f2- -d/) + cd ../llama.cpp + git checkout examples/llava/clip.cpp Updated 0 paths from the index + for patch in ../patches/*.diff ++ grep '^+++ ' ../patches/03-load_exception.diff ++ cut -f2 '-d ' ++ cut -f2- -d/ + for file in $(grep "^+++ " ${patch} | cut -f2 -d' ' | cut -f2- -d/) + cd ../llama.cpp + git checkout src/llama.cpp Updated 0 paths from the index + for patch in ../patches/*.diff ++ grep '^+++ ' ../patches/04-metal.diff ++ cut -f2 '-d ' ++ cut -f2- -d/ + for file in $(grep "^+++ " ${patch} | cut -f2 -d' ' | cut -f2- -d/) + cd ../llama.cpp + git checkout ggml/src/ggml-metal.m Updated 0 paths from the index + for patch in ../patches/*.diff ++ grep '^+++ ' ../patches/05-default-pretokenizer.diff ++ cut -f2 '-d ' ++ cut -f2- -d/ + for file in $(grep "^+++ " ${patch} | cut -f2 -d' ' | cut -f2- -d/) + cd ../llama.cpp + git checkout src/llama.cpp Updated 0 paths from the index + for patch in ../patches/*.diff ++ grep '^+++ ' ../patches/06-embeddings.diff ++ cut -f2 '-d ' ++ cut -f2- -d/ + for file in $(grep "^+++ " ${patch} | cut -f2 -d' ' | cut -f2- -d/) + cd ../llama.cpp + git checkout src/llama.cpp Updated 0 paths from the index + for patch in ../patches/*.diff ++ grep '^+++ ' ../patches/07-clip-unicode.diff ++ cut -f2 '-d ' ++ cut -f2- -d/ + for file in $(grep "^+++ " ${patch} | cut -f2 -d' ' | cut -f2- -d/) + cd ../llama.cpp + git checkout examples/llava/clip.cpp Updated 0 paths from the index + for patch in ../patches/*.diff ++ grep '^+++ ' ../patches/08-pooling.diff ++ cut -f2 '-d ' ++ cut -f2- -d/ + for file in $(grep "^+++ " ${patch} | cut -f2 -d' ' | cut -f2- -d/) + cd ../llama.cpp + git checkout src/llama.cpp Updated 0 paths from the index + for patch in ../patches/*.diff ++ grep '^+++ ' ../patches/09-lora.diff ++ cut -f2 '-d ' ++ cut -f2- -d/ + for file in $(grep "^+++ " ${patch} | cut -f2 -d' ' | cut -f2- -d/) + cd ../llama.cpp + git checkout common/common.cpp Updated 0 paths from the index + for file in $(grep "^+++ " ${patch} | cut -f2 -d' ' | cut -f2- -d/) + cd ../llama.cpp + git checkout include/llama.h Updated 0 paths from the index + for file in $(grep "^+++ " ${patch} | cut -f2 -d' ' | cut -f2- -d/) + cd ../llama.cpp + git checkout src/llama.cpp Updated 0 paths from the index + for patch in ../patches/*.diff ++ grep '^+++ ' ../patches/11-phi3-sliding-window.diff ++ cut -f2 '-d ' ++ cut -f2- -d/ + for file in $(grep "^+++ " ${patch} | cut -f2 -d' ' | cut -f2- -d/) + cd ../llama.cpp + git checkout src/llama.cpp Updated 0 paths from the index + for patch in ../patches/*.diff + cd ../llama.cpp + git apply ../patches/01-load-progress.diff + for patch in ../patches/*.diff + cd ../llama.cpp + git apply ../patches/02-clip-log.diff + for patch in ../patches/*.diff + cd ../llama.cpp + git apply ../patches/03-load_exception.diff + for patch in ../patches/*.diff + cd ../llama.cpp + git apply ../patches/04-metal.diff + for patch in ../patches/*.diff + cd ../llama.cpp + git apply ../patches/05-default-pretokenizer.diff + for patch in ../patches/*.diff + cd ../llama.cpp + git apply ../patches/06-embeddings.diff + for patch in ../patches/*.diff + cd ../llama.cpp + git apply ../patches/07-clip-unicode.diff + for patch in ../patches/*.diff + cd ../llama.cpp + git apply ../patches/08-pooling.diff + for patch in ../patches/*.diff + cd ../llama.cpp + git apply ../patches/09-lora.diff + for patch in ../patches/*.diff + cd ../llama.cpp + git apply ../patches/11-phi3-sliding-window.diff + init_vars + case "${GOARCH}" in + ARCH=x86_64 + LLAMACPP_DIR=../llama.cpp + CMAKE_DEFS= + CMAKE_TARGETS='--target ollama_llama_server' + echo '' + grep -- -g + CMAKE_DEFS='-DCMAKE_BUILD_TYPE=Release -DLLAMA_SERVER_VERBOSE=off ' + case $(uname -s) in ++ uname -s + LIB_EXT=so + WHOLE_ARCHIVE=-Wl,--whole-archive + NO_WHOLE_ARCHIVE=-Wl,--no-whole-archive + GCC_ARCH= + '[' -z '50;52;61;70;75;80' ']' + '[' -z '' -o '' = static ']' + init_vars + case "${GOARCH}" in + ARCH=x86_64 + LLAMACPP_DIR=../llama.cpp + CMAKE_DEFS= + CMAKE_TARGETS='--target ollama_llama_server' + echo '' + grep -- -g + CMAKE_DEFS='-DCMAKE_BUILD_TYPE=Release -DLLAMA_SERVER_VERBOSE=off ' + case $(uname -s) in ++ uname -s + LIB_EXT=so + WHOLE_ARCHIVE=-Wl,--whole-archive + NO_WHOLE_ARCHIVE=-Wl,--no-whole-archive + GCC_ARCH= Building static library + '[' -z '50;52;61;70;75;80' ']' + CMAKE_TARGETS='--target llama --target ggml' + CMAKE_DEFS='-DBUILD_SHARED_LIBS=off -DGGML_NATIVE=off -DGGML_AVX=off -DGGML_AVX2=off -DGGML_AVX512=off -DGGML_FMA=off -DGGML_F16C=off -DGGML_OPENMP=off -DCMAKE_BUILD_TYPE=Release -DLLAMA_SERVER_VERBOSE=off ' + BUILD_DIR=../build/linux/x86_64_static + echo 'Building static library' + build + cmake -S ../llama.cpp -B ../build/linux/x86_64_static -DBUILD_SHARED_LIBS=off -DGGML_NATIVE=off -DGGML_AVX=off -DGGML_AVX2=off -DGGML_AVX512=off -DGGML_FMA=off -DGGML_F16C=off -DGGML_OPENMP=off -DCMAKE_BUILD_TYPE=Release -DLLAMA_SERVER_VERBOSE=off -- The C compiler identification is GNU 12.3.1 -- The CXX compiler identification is GNU 12.3.1 -- Detecting C compiler ABI info -- Detecting C compiler ABI info - done -- Check for working C compiler: /usr/bin/cc - skipped -- Detecting C compile features -- Detecting C compile features - done -- Detecting CXX compiler ABI info -- Detecting CXX compiler ABI info - done -- Check for working CXX compiler: /usr/bin/c++ - skipped -- Detecting CXX compile features -- Detecting CXX compile features - done -- Found Git: /usr/bin/git (found version "2.43.0") -- Performing Test CMAKE_HAVE_LIBC_PTHREAD -- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Success -- Found Threads: TRUE -- Using llamafile -- Warning: ccache not found - consider installing it for faster compilation or disable this warning with GGML_CCACHE=OFF -- CMAKE_SYSTEM_PROCESSOR: x86_64 -- x86 detected -- Configuring done (0.6s) -- Generating done (0.1s) -- Build files have been written to: /builddir/build/BUILD/ollama-0.3.6/llm/build/linux/x86_64_static + cmake --build ../build/linux/x86_64_static --target llama --target ggml -j8 [ 11%] Building C object ggml/src/CMakeFiles/ggml.dir/ggml.c.o [ 11%] Building C object ggml/src/CMakeFiles/ggml.dir/ggml-backend.c.o [ 33%] Building C object ggml/src/CMakeFiles/ggml.dir/ggml-alloc.c.o [ 33%] Building C object ggml/src/CMakeFiles/ggml.dir/ggml-quants.c.o [ 33%] Building CXX object ggml/src/CMakeFiles/ggml.dir/llamafile/sgemm.cpp.o [ 44%] Building C object ggml/src/CMakeFiles/ggml.dir/ggml-aarch64.c.o /builddir/build/BUILD/ollama-0.3.6/llm/llama.cpp/ggml/src/ggml.c: In function 'ggml_vec_mad_f16': /builddir/build/BUILD/ollama-0.3.6/llm/llama.cpp/ggml/src/ggml.c:2174:45: warning: passing argument 1 of '__sse_f16x4_load' discards 'const' qualifier from pointer target type [-Wdiscarded-qualifiers] 2174 | ax[j] = GGML_F16_VEC_LOAD(x + i + j*GGML_F16_EPR, j); | ^ /builddir/build/BUILD/ollama-0.3.6/llm/llama.cpp/ggml/src/ggml.c:1626:50: note: in definition of macro 'GGML_F32Cx4_LOAD' 1626 | #define GGML_F32Cx4_LOAD(x) __sse_f16x4_load(x) | ^ /builddir/build/BUILD/ollama-0.3.6/llm/llama.cpp/ggml/src/ggml.c:2174:21: note: in expansion of macro 'GGML_F16_VEC_LOAD' 2174 | ax[j] = GGML_F16_VEC_LOAD(x + i + j*GGML_F16_EPR, j); | ^~~~~~~~~~~~~~~~~ /builddir/build/BUILD/ollama-0.3.6/llm/llama.cpp/ggml/src/ggml.c:1601:52: note: expected 'ggml_fp16_t *' {aka 'short unsigned int *'} but argument is of type 'const ggml_fp16_t *' {aka 'const short unsigned int *'} 1601 | static inline __m128 __sse_f16x4_load(ggml_fp16_t *x) { | ~~~~~~~~~~~~~^ [ 55%] Linking CXX static library libggml.a [ 55%] Built target ggml [ 55%] Building CXX object src/CMakeFiles/llama.dir/llama.cpp.o [ 55%] Building CXX object src/CMakeFiles/llama.dir/llama-grammar.cpp.o [ 66%] Building CXX object src/CMakeFiles/llama.dir/llama-vocab.cpp.o [ 77%] Building CXX object src/CMakeFiles/llama.dir/llama-sampling.cpp.o [ 88%] Building CXX object src/CMakeFiles/llama.dir/unicode.cpp.o [ 88%] Building CXX object src/CMakeFiles/llama.dir/unicode-data.cpp.o [100%] Linking CXX static library libllama.a [100%] Built target llama [100%] Built target ggml + init_vars + case "${GOARCH}" in + ARCH=x86_64 + LLAMACPP_DIR=../llama.cpp + CMAKE_DEFS= + CMAKE_TARGETS='--target ollama_llama_server' + echo '' + grep -- -g + CMAKE_DEFS='-DCMAKE_BUILD_TYPE=Release -DLLAMA_SERVER_VERBOSE=off ' + case $(uname -s) in ++ uname -s + LIB_EXT=so + WHOLE_ARCHIVE=-Wl,--whole-archive + NO_WHOLE_ARCHIVE=-Wl,--no-whole-archive + GCC_ARCH= + '[' -z '50;52;61;70;75;80' ']' + '[' -z '' ']' + '[' -n '' ']' + COMMON_CPU_DEFS='-DBUILD_SHARED_LIBS=off -DCMAKE_POSITION_INDEPENDENT_CODE=on -DGGML_NATIVE=off -DGGML_OPENMP=off' + '[' -z '' -o '' = cpu ']' + init_vars + case "${GOARCH}" in + ARCH=x86_64 + LLAMACPP_DIR=../llama.cpp + CMAKE_DEFS= + CMAKE_TARGETS='--target ollama_llama_server' + echo '' + grep -- -g + CMAKE_DEFS='-DCMAKE_BUILD_TYPE=Release -DLLAMA_SERVER_VERBOSE=off ' + case $(uname -s) in ++ uname -s + LIB_EXT=so Building LCD CPU + WHOLE_ARCHIVE=-Wl,--whole-archive + NO_WHOLE_ARCHIVE=-Wl,--no-whole-archive + GCC_ARCH= + '[' -z '50;52;61;70;75;80' ']' + CMAKE_DEFS='-DBUILD_SHARED_LIBS=off -DCMAKE_POSITION_INDEPENDENT_CODE=on -DGGML_NATIVE=off -DGGML_OPENMP=off -DGGML_AVX=off -DGGML_AVX2=off -DGGML_AVX512=off -DGGML_FMA=off -DGGML_F16C=off -DCMAKE_BUILD_TYPE=Release -DLLAMA_SERVER_VERBOSE=off ' + BUILD_DIR=../build/linux/x86_64/cpu + echo 'Building LCD CPU' + build + cmake -S ../llama.cpp -B ../build/linux/x86_64/cpu -DBUILD_SHARED_LIBS=off -DCMAKE_POSITION_INDEPENDENT_CODE=on -DGGML_NATIVE=off -DGGML_OPENMP=off -DGGML_AVX=off -DGGML_AVX2=off -DGGML_AVX512=off -DGGML_FMA=off -DGGML_F16C=off -DCMAKE_BUILD_TYPE=Release -DLLAMA_SERVER_VERBOSE=off -- The C compiler identification is GNU 12.3.1 -- The CXX compiler identification is GNU 12.3.1 -- Detecting C compiler ABI info -- Detecting C compiler ABI info - done -- Check for working C compiler: /usr/bin/cc - skipped -- Detecting C compile features -- Detecting C compile features - done -- Detecting CXX compiler ABI info -- Detecting CXX compiler ABI info - done -- Check for working CXX compiler: /usr/bin/c++ - skipped -- Detecting CXX compile features -- Detecting CXX compile features - done -- Found Git: /usr/bin/git (found version "2.43.0") -- Performing Test CMAKE_HAVE_LIBC_PTHREAD -- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Success -- Found Threads: TRUE -- Using llamafile -- Warning: ccache not found - consider installing it for faster compilation or disable this warning with GGML_CCACHE=OFF -- CMAKE_SYSTEM_PROCESSOR: x86_64 -- x86 detected -- Configuring done (0.6s) -- Generating done (0.1s) -- Build files have been written to: /builddir/build/BUILD/ollama-0.3.6/llm/build/linux/x86_64/cpu + cmake --build ../build/linux/x86_64/cpu --target ollama_llama_server -j8 [ 0%] Generating build details from Git [ 6%] Building C object ggml/src/CMakeFiles/ggml.dir/ggml.c.o [ 12%] Building C object ggml/src/CMakeFiles/ggml.dir/ggml-alloc.c.o -- Found Git: /usr/bin/git (found version "2.43.0") [ 12%] Building C object ggml/src/CMakeFiles/ggml.dir/ggml-backend.c.o [ 18%] Building C object ggml/src/CMakeFiles/ggml.dir/ggml-quants.c.o [ 18%] Building CXX object ggml/src/CMakeFiles/ggml.dir/llamafile/sgemm.cpp.o [ 25%] Building C object ggml/src/CMakeFiles/ggml.dir/ggml-aarch64.c.o [ 31%] Building CXX object common/CMakeFiles/build_info.dir/build-info.cpp.o [ 31%] Built target build_info /builddir/build/BUILD/ollama-0.3.6/llm/llama.cpp/ggml/src/ggml.c: In function 'ggml_vec_mad_f16': /builddir/build/BUILD/ollama-0.3.6/llm/llama.cpp/ggml/src/ggml.c:2174:45: warning: passing argument 1 of '__sse_f16x4_load' discards 'const' qualifier from pointer target type [-Wdiscarded-qualifiers] 2174 | ax[j] = GGML_F16_VEC_LOAD(x + i + j*GGML_F16_EPR, j); | ^ /builddir/build/BUILD/ollama-0.3.6/llm/llama.cpp/ggml/src/ggml.c:1626:50: note: in definition of macro 'GGML_F32Cx4_LOAD' 1626 | #define GGML_F32Cx4_LOAD(x) __sse_f16x4_load(x) | ^ /builddir/build/BUILD/ollama-0.3.6/llm/llama.cpp/ggml/src/ggml.c:2174:21: note: in expansion of macro 'GGML_F16_VEC_LOAD' 2174 | ax[j] = GGML_F16_VEC_LOAD(x + i + j*GGML_F16_EPR, j); | ^~~~~~~~~~~~~~~~~ /builddir/build/BUILD/ollama-0.3.6/llm/llama.cpp/ggml/src/ggml.c:1601:52: note: expected 'ggml_fp16_t *' {aka 'short unsigned int *'} but argument is of type 'const ggml_fp16_t *' {aka 'const short unsigned int *'} 1601 | static inline __m128 __sse_f16x4_load(ggml_fp16_t *x) { | ~~~~~~~~~~~~~^ [ 37%] Linking CXX static library libggml.a [ 37%] Built target ggml [ 37%] Building CXX object src/CMakeFiles/llama.dir/llama.cpp.o [ 43%] Building CXX object src/CMakeFiles/llama.dir/llama-grammar.cpp.o [ 43%] Building CXX object src/CMakeFiles/llama.dir/llama-vocab.cpp.o [ 50%] Building CXX object src/CMakeFiles/llama.dir/llama-sampling.cpp.o [ 56%] Building CXX object src/CMakeFiles/llama.dir/unicode.cpp.o [ 56%] Building CXX object src/CMakeFiles/llama.dir/unicode-data.cpp.o [ 62%] Linking CXX static library libllama.a [ 62%] Built target llama [ 62%] Building CXX object examples/llava/CMakeFiles/llava.dir/llava.cpp.o [ 68%] Building CXX object examples/llava/CMakeFiles/llava.dir/clip.cpp.o [ 75%] Building CXX object common/CMakeFiles/common.dir/common.cpp.o [ 75%] Building CXX object common/CMakeFiles/common.dir/sampling.cpp.o [ 75%] Building CXX object common/CMakeFiles/common.dir/console.cpp.o [ 81%] Building CXX object common/CMakeFiles/common.dir/grammar-parser.cpp.o [ 87%] Building CXX object common/CMakeFiles/common.dir/json-schema-to-grammar.cpp.o [ 87%] Building CXX object common/CMakeFiles/common.dir/train.cpp.o [ 93%] Building CXX object common/CMakeFiles/common.dir/ngram-cache.cpp.o [ 93%] Built target llava [ 93%] Linking CXX static library libcommon.a [ 93%] Built target common [ 93%] Building CXX object ext_server/CMakeFiles/ollama_llama_server.dir/server.cpp.o [100%] Linking CXX executable ../bin/ollama_llama_server [100%] Built target ollama_llama_server + compress + echo 'Compressing payloads to reduce overall binary size...' + pids= Compressing payloads to reduce overall binary size... + rm -rf '../build/linux/x86_64/cpu/bin/*.gz' + for f in ${BUILD_DIR}/bin/* + pids+=' 3940' + '[' -d ../build/linux/x86_64/cpu/lib ']' + echo + for pid in ${pids} + wait 3940 + gzip -n --best -f ../build/linux/x86_64/cpu/bin/ollama_llama_server Finished compression + echo 'Finished compression' + '[' x86_64 == x86_64 ']' + '[' -z '' -o '' = cpu_avx ']' + init_vars + case "${GOARCH}" in + ARCH=x86_64 + LLAMACPP_DIR=../llama.cpp + CMAKE_DEFS= + CMAKE_TARGETS='--target ollama_llama_server' + echo '' + grep -- -g + CMAKE_DEFS='-DCMAKE_BUILD_TYPE=Release -DLLAMA_SERVER_VERBOSE=off ' + case $(uname -s) in ++ uname -s + LIB_EXT=so + WHOLE_ARCHIVE=-Wl,--whole-archive + NO_WHOLE_ARCHIVE=-Wl,--no-whole-archive + GCC_ARCH= Building AVX CPU + '[' -z '50;52;61;70;75;80' ']' + CMAKE_DEFS='-DBUILD_SHARED_LIBS=off -DCMAKE_POSITION_INDEPENDENT_CODE=on -DGGML_NATIVE=off -DGGML_OPENMP=off -DGGML_AVX=on -DGGML_AVX2=off -DGGML_AVX512=off -DGGML_FMA=off -DGGML_F16C=off -DCMAKE_BUILD_TYPE=Release -DLLAMA_SERVER_VERBOSE=off ' + BUILD_DIR=../build/linux/x86_64/cpu_avx + echo 'Building AVX CPU' + build + cmake -S ../llama.cpp -B ../build/linux/x86_64/cpu_avx -DBUILD_SHARED_LIBS=off -DCMAKE_POSITION_INDEPENDENT_CODE=on -DGGML_NATIVE=off -DGGML_OPENMP=off -DGGML_AVX=on -DGGML_AVX2=off -DGGML_AVX512=off -DGGML_FMA=off -DGGML_F16C=off -DCMAKE_BUILD_TYPE=Release -DLLAMA_SERVER_VERBOSE=off -- The C compiler identification is GNU 12.3.1 -- The CXX compiler identification is GNU 12.3.1 -- Detecting C compiler ABI info -- Detecting C compiler ABI info - done -- Check for working C compiler: /usr/bin/cc - skipped -- Detecting C compile features -- Detecting C compile features - done -- Detecting CXX compiler ABI info -- Detecting CXX compiler ABI info - done -- Check for working CXX compiler: /usr/bin/c++ - skipped -- Detecting CXX compile features -- Detecting CXX compile features - done -- Found Git: /usr/bin/git (found version "2.43.0") -- Performing Test CMAKE_HAVE_LIBC_PTHREAD -- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Success -- Found Threads: TRUE -- Using llamafile -- Warning: ccache not found - consider installing it for faster compilation or disable this warning with GGML_CCACHE=OFF -- CMAKE_SYSTEM_PROCESSOR: x86_64 -- x86 detected -- Configuring done (0.6s) -- Generating done (0.1s) -- Build files have been written to: /builddir/build/BUILD/ollama-0.3.6/llm/build/linux/x86_64/cpu_avx + cmake --build ../build/linux/x86_64/cpu_avx --target ollama_llama_server -j8 [ 6%] Building CXX object common/CMakeFiles/build_info.dir/build-info.cpp.o [ 12%] Building C object ggml/src/CMakeFiles/ggml.dir/ggml.c.o [ 18%] Building C object ggml/src/CMakeFiles/ggml.dir/ggml-alloc.c.o [ 18%] Building C object ggml/src/CMakeFiles/ggml.dir/ggml-backend.c.o [ 18%] Built target build_info [ 25%] Building C object ggml/src/CMakeFiles/ggml.dir/ggml-quants.c.o [ 25%] Building CXX object ggml/src/CMakeFiles/ggml.dir/llamafile/sgemm.cpp.o [ 31%] Building C object ggml/src/CMakeFiles/ggml.dir/ggml-aarch64.c.o /builddir/build/BUILD/ollama-0.3.6/llm/llama.cpp/ggml/src/ggml.c: In function 'ggml_vec_mad_f16': /builddir/build/BUILD/ollama-0.3.6/llm/llama.cpp/ggml/src/ggml.c:2174:45: warning: passing argument 1 of '__avx_f32cx8_load' discards 'const' qualifier from pointer target type [-Wdiscarded-qualifiers] 2174 | ax[j] = GGML_F16_VEC_LOAD(x + i + j*GGML_F16_EPR, j); | ^ /builddir/build/BUILD/ollama-0.3.6/llm/llama.cpp/ggml/src/ggml.c:1346:51: note: in definition of macro 'GGML_F32Cx8_LOAD' 1346 | #define GGML_F32Cx8_LOAD(x) __avx_f32cx8_load(x) | ^ /builddir/build/BUILD/ollama-0.3.6/llm/llama.cpp/ggml/src/ggml.c:2174:21: note: in expansion of macro 'GGML_F16_VEC_LOAD' 2174 | ax[j] = GGML_F16_VEC_LOAD(x + i + j*GGML_F16_EPR, j); | ^~~~~~~~~~~~~~~~~ /builddir/build/BUILD/ollama-0.3.6/llm/llama.cpp/ggml/src/ggml.c:1329:53: note: expected 'ggml_fp16_t *' {aka 'short unsigned int *'} but argument is of type 'const ggml_fp16_t *' {aka 'const short unsigned int *'} 1329 | static inline __m256 __avx_f32cx8_load(ggml_fp16_t *x) { | ~~~~~~~~~~~~~^ [ 37%] Linking CXX static library libggml.a [ 37%] Built target ggml [ 37%] Building CXX object src/CMakeFiles/llama.dir/llama.cpp.o [ 43%] Building CXX object src/CMakeFiles/llama.dir/llama-grammar.cpp.o [ 43%] Building CXX object src/CMakeFiles/llama.dir/llama-vocab.cpp.o [ 50%] Building CXX object src/CMakeFiles/llama.dir/llama-sampling.cpp.o [ 56%] Building CXX object src/CMakeFiles/llama.dir/unicode.cpp.o [ 56%] Building CXX object src/CMakeFiles/llama.dir/unicode-data.cpp.o [ 62%] Linking CXX static library libllama.a [ 62%] Built target llama [ 62%] Building CXX object common/CMakeFiles/common.dir/common.cpp.o [ 68%] Building CXX object examples/llava/CMakeFiles/llava.dir/llava.cpp.o [ 68%] Building CXX object examples/llava/CMakeFiles/llava.dir/clip.cpp.o [ 75%] Building CXX object common/CMakeFiles/common.dir/sampling.cpp.o [ 75%] Building CXX object common/CMakeFiles/common.dir/console.cpp.o [ 81%] Building CXX object common/CMakeFiles/common.dir/grammar-parser.cpp.o [ 87%] Building CXX object common/CMakeFiles/common.dir/json-schema-to-grammar.cpp.o [ 87%] Building CXX object common/CMakeFiles/common.dir/train.cpp.o [ 93%] Building CXX object common/CMakeFiles/common.dir/ngram-cache.cpp.o [ 93%] Built target llava [ 93%] Linking CXX static library libcommon.a [ 93%] Built target common [ 93%] Building CXX object ext_server/CMakeFiles/ollama_llama_server.dir/server.cpp.o [100%] Linking CXX executable ../bin/ollama_llama_server [100%] Built target ollama_llama_server + compress + echo 'Compressing payloads to reduce overall binary size...' Compressing payloads to reduce overall binary size... + pids= + rm -rf '../build/linux/x86_64/cpu_avx/bin/*.gz' + for f in ${BUILD_DIR}/bin/* + pids+=' 4175' + '[' -d ../build/linux/x86_64/cpu_avx/lib ']' + gzip -n --best -f ../build/linux/x86_64/cpu_avx/bin/ollama_llama_server + echo + for pid in ${pids} + wait 4175 Finished compression + echo 'Finished compression' + '[' -z '' -o '' = cpu_avx2 ']' + init_vars + case "${GOARCH}" in + ARCH=x86_64 + LLAMACPP_DIR=../llama.cpp + CMAKE_DEFS= + CMAKE_TARGETS='--target ollama_llama_server' + echo '' + grep -- -g + CMAKE_DEFS='-DCMAKE_BUILD_TYPE=Release -DLLAMA_SERVER_VERBOSE=off ' + case $(uname -s) in ++ uname -s + LIB_EXT=so + WHOLE_ARCHIVE=-Wl,--whole-archive + NO_WHOLE_ARCHIVE=-Wl,--no-whole-archive + GCC_ARCH= Building AVX2 CPU + '[' -z '50;52;61;70;75;80' ']' + CMAKE_DEFS='-DBUILD_SHARED_LIBS=off -DCMAKE_POSITION_INDEPENDENT_CODE=on -DGGML_NATIVE=off -DGGML_OPENMP=off -DGGML_AVX=on -DGGML_AVX2=on -DGGML_AVX512=off -DGGML_FMA=on -DGGML_F16C=on -DCMAKE_BUILD_TYPE=Release -DLLAMA_SERVER_VERBOSE=off ' + BUILD_DIR=../build/linux/x86_64/cpu_avx2 + echo 'Building AVX2 CPU' + build + cmake -S ../llama.cpp -B ../build/linux/x86_64/cpu_avx2 -DBUILD_SHARED_LIBS=off -DCMAKE_POSITION_INDEPENDENT_CODE=on -DGGML_NATIVE=off -DGGML_OPENMP=off -DGGML_AVX=on -DGGML_AVX2=on -DGGML_AVX512=off -DGGML_FMA=on -DGGML_F16C=on -DCMAKE_BUILD_TYPE=Release -DLLAMA_SERVER_VERBOSE=off -- The C compiler identification is GNU 12.3.1 -- The CXX compiler identification is GNU 12.3.1 -- Detecting C compiler ABI info -- Detecting C compiler ABI info - done -- Check for working C compiler: /usr/bin/cc - skipped -- Detecting C compile features -- Detecting C compile features - done -- Detecting CXX compiler ABI info -- Detecting CXX compiler ABI info - done -- Check for working CXX compiler: /usr/bin/c++ - skipped -- Detecting CXX compile features -- Detecting CXX compile features - done -- Found Git: /usr/bin/git (found version "2.43.0") -- Performing Test CMAKE_HAVE_LIBC_PTHREAD -- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Success -- Found Threads: TRUE -- Using llamafile -- Warning: ccache not found - consider installing it for faster compilation or disable this warning with GGML_CCACHE=OFF -- CMAKE_SYSTEM_PROCESSOR: x86_64 -- x86 detected -- Configuring done (0.6s) -- Generating done (0.1s) -- Build files have been written to: /builddir/build/BUILD/ollama-0.3.6/llm/build/linux/x86_64/cpu_avx2 + cmake --build ../build/linux/x86_64/cpu_avx2 --target ollama_llama_server -j8 [ 6%] Building C object ggml/src/CMakeFiles/ggml.dir/ggml.c.o [ 6%] Building C object ggml/src/CMakeFiles/ggml.dir/ggml-backend.c.o [ 12%] Building CXX object common/CMakeFiles/build_info.dir/build-info.cpp.o [ 18%] Building C object ggml/src/CMakeFiles/ggml.dir/ggml-alloc.c.o [ 25%] Building C object ggml/src/CMakeFiles/ggml.dir/ggml-quants.c.o [ 25%] Built target build_info [ 25%] Building CXX object ggml/src/CMakeFiles/ggml.dir/llamafile/sgemm.cpp.o [ 31%] Building C object ggml/src/CMakeFiles/ggml.dir/ggml-aarch64.c.o [ 37%] Linking CXX static library libggml.a [ 37%] Built target ggml [ 37%] Building CXX object src/CMakeFiles/llama.dir/llama.cpp.o [ 50%] Building CXX object src/CMakeFiles/llama.dir/llama-grammar.cpp.o [ 50%] Building CXX object src/CMakeFiles/llama.dir/llama-sampling.cpp.o [ 50%] Building CXX object src/CMakeFiles/llama.dir/llama-vocab.cpp.o [ 56%] Building CXX object src/CMakeFiles/llama.dir/unicode.cpp.o [ 56%] Building CXX object src/CMakeFiles/llama.dir/unicode-data.cpp.o [ 62%] Linking CXX static library libllama.a [ 62%] Built target llama [ 62%] Building CXX object examples/llava/CMakeFiles/llava.dir/llava.cpp.o [ 68%] Building CXX object common/CMakeFiles/common.dir/sampling.cpp.o [ 68%] Building CXX object common/CMakeFiles/common.dir/console.cpp.o [ 68%] Building CXX object common/CMakeFiles/common.dir/common.cpp.o [ 75%] Building CXX object examples/llava/CMakeFiles/llava.dir/clip.cpp.o [ 81%] Building CXX object common/CMakeFiles/common.dir/grammar-parser.cpp.o [ 87%] Building CXX object common/CMakeFiles/common.dir/json-schema-to-grammar.cpp.o [ 87%] Building CXX object common/CMakeFiles/common.dir/train.cpp.o [ 93%] Building CXX object common/CMakeFiles/common.dir/ngram-cache.cpp.o [ 93%] Built target llava [ 93%] Linking CXX static library libcommon.a [ 93%] Built target common [ 93%] Building CXX object ext_server/CMakeFiles/ollama_llama_server.dir/server.cpp.o [100%] Linking CXX executable ../bin/ollama_llama_server [100%] Built target ollama_llama_server + compress + echo 'Compressing payloads to reduce overall binary size...' Compressing payloads to reduce overall binary size... + pids= + rm -rf '../build/linux/x86_64/cpu_avx2/bin/*.gz' + for f in ${BUILD_DIR}/bin/* + pids+=' 4461' + '[' -d ../build/linux/x86_64/cpu_avx2/lib ']' + echo + for pid in ${pids} + wait 4461 + gzip -n --best -f ../build/linux/x86_64/cpu_avx2/bin/ollama_llama_server Finished compression + echo 'Finished compression' + '[' -z '' ']' + '[' -d /usr/local/cuda/lib64 ']' + '[' -z '' ']' + '[' -d /opt/cuda/targets/x86_64-linux/lib ']' + '[' -z '' ']' + CUDART_LIB_DIR= + '[' -z '' -a -d '' ']' + '[' -z '' ']' + ONEAPI_ROOT=/opt/intel/oneapi + '[' -z '' -a -d /opt/intel/oneapi ']' + '[' -z '' ']' + ROCM_PATH=/opt/rocm + '[' -z '' ']' + '[' -d /usr/lib/cmake/CLBlast ']' + '[' -z '' -a -d /opt/rocm ']' + cleanup + cd ../llama.cpp/ + git checkout CMakeLists.txt Updated 1 path from the index ++ ls -A ../patches/01-load-progress.diff ../patches/02-clip-log.diff ../patches/03-load_exception.diff ../patches/04-metal.diff ../patches/05-default-pretokenizer.diff ../patches/06-embeddings.diff ../patches/07-clip-unicode.diff ../patches/08-pooling.diff ../patches/09-lora.diff ../patches/11-phi3-sliding-window.diff + '[' -n '../patches/01-load-progress.diff ../patches/02-clip-log.diff ../patches/03-load_exception.diff ../patches/04-metal.diff ../patches/05-default-pretokenizer.diff ../patches/06-embeddings.diff ../patches/07-clip-unicode.diff ../patches/08-pooling.diff ../patches/09-lora.diff ../patches/11-phi3-sliding-window.diff' ']' + for patch in ../patches/*.diff ++ grep '^+++ ' ../patches/01-load-progress.diff ++ cut -f2- -d/ ++ cut -f2 '-d ' + for file in $(grep "^+++ " ${patch} | cut -f2 -d' ' | cut -f2- -d/) + cd ../llama.cpp + git checkout common/common.cpp Updated 1 path from the index + for file in $(grep "^+++ " ${patch} | cut -f2 -d' ' | cut -f2- -d/) + cd ../llama.cpp + git checkout common/common.h Updated 1 path from the index + for patch in ../patches/*.diff ++ grep '^+++ ' ../patches/02-clip-log.diff ++ cut -f2 '-d ' ++ cut -f2- -d/ + for file in $(grep "^+++ " ${patch} | cut -f2 -d' ' | cut -f2- -d/) + cd ../llama.cpp + git checkout examples/llava/clip.cpp Updated 1 path from the index + for patch in ../patches/*.diff ++ grep '^+++ ' ../patches/03-load_exception.diff ++ cut -f2 '-d ' ++ cut -f2- -d/ + for file in $(grep "^+++ " ${patch} | cut -f2 -d' ' | cut -f2- -d/) + cd ../llama.cpp + git checkout src/llama.cpp Updated 1 path from the index + for patch in ../patches/*.diff ++ grep '^+++ ' ../patches/04-metal.diff ++ cut -f2 '-d ' ++ cut -f2- -d/ + for file in $(grep "^+++ " ${patch} | cut -f2 -d' ' | cut -f2- -d/) + cd ../llama.cpp + git checkout ggml/src/ggml-metal.m Updated 1 path from the index + for patch in ../patches/*.diff ++ grep '^+++ ' ../patches/05-default-pretokenizer.diff ++ cut -f2 '-d ' ++ cut -f2- -d/ + for file in $(grep "^+++ " ${patch} | cut -f2 -d' ' | cut -f2- -d/) + cd ../llama.cpp + git checkout src/llama.cpp Updated 0 paths from the index + for patch in ../patches/*.diff ++ grep '^+++ ' ../patches/06-embeddings.diff ++ cut -f2 '-d ' ++ cut -f2- -d/ + for file in $(grep "^+++ " ${patch} | cut -f2 -d' ' | cut -f2- -d/) + cd ../llama.cpp + git checkout src/llama.cpp Updated 0 paths from the index + for patch in ../patches/*.diff ++ grep '^+++ ' ../patches/07-clip-unicode.diff ++ cut -f2 '-d ' ++ cut -f2- -d/ + for file in $(grep "^+++ " ${patch} | cut -f2 -d' ' | cut -f2- -d/) + cd ../llama.cpp + git checkout examples/llava/clip.cpp Updated 0 paths from the index + for patch in ../patches/*.diff ++ grep '^+++ ' ../patches/08-pooling.diff ++ cut -f2 '-d ' ++ cut -f2- -d/ + for file in $(grep "^+++ " ${patch} | cut -f2 -d' ' | cut -f2- -d/) + cd ../llama.cpp + git checkout src/llama.cpp Updated 0 paths from the index + for patch in ../patches/*.diff ++ grep '^+++ ' ../patches/09-lora.diff ++ cut -f2 '-d ' ++ cut -f2- -d/ + for file in $(grep "^+++ " ${patch} | cut -f2 -d' ' | cut -f2- -d/) + cd ../llama.cpp + git checkout common/common.cpp Updated 0 paths from the index + for file in $(grep "^+++ " ${patch} | cut -f2 -d' ' | cut -f2- -d/) + cd ../llama.cpp + git checkout include/llama.h Updated 1 path from the index + for file in $(grep "^+++ " ${patch} | cut -f2 -d' ' | cut -f2- -d/) + cd ../llama.cpp + git checkout src/llama.cpp Updated 0 paths from the index + for patch in ../patches/*.diff ++ grep '^+++ ' ../patches/11-phi3-sliding-window.diff ++ cut -f2 '-d ' ++ cut -f2- -d/ + for file in $(grep "^+++ " ${patch} | cut -f2 -d' ' | cut -f2- -d/) + cd ../llama.cpp + git checkout src/llama.cpp Updated 0 paths from the index ++ cd ../build/linux/x86_64/cpu_avx2/.. ++ echo cpu cpu_avx cpu_avx2 + echo 'go generate completed. LLM runners: cpu cpu_avx cpu_avx2' go generate completed. LLM runners: cpu cpu_avx cpu_avx2 + go build . + RPM_EC=0 ++ jobs -p + exit 0 Executing(%install): /bin/sh -e /var/tmp/rpm-tmp.25ydkF + umask 022 + cd /builddir/build/BUILD + '[' /builddir/build/BUILDROOT/ollama-0.3.6-1.x86_64 '!=' / ']' + rm -rf /builddir/build/BUILDROOT/ollama-0.3.6-1.x86_64 ++ dirname /builddir/build/BUILDROOT/ollama-0.3.6-1.x86_64 + mkdir -p /builddir/build/BUILDROOT + mkdir /builddir/build/BUILDROOT/ollama-0.3.6-1.x86_64 + cd ollama-0.3.6 + install -D -m 0755 ollama /builddir/build/BUILDROOT/ollama-0.3.6-1.x86_64/usr/bin/ollama + install -D -m 0644 /builddir/build/SOURCES/ollama.service /builddir/build/BUILDROOT/ollama-0.3.6-1.x86_64/usr/lib/systemd/system/ollama.service + install -D -m 0644 /builddir/build/SOURCES/ollama-ui.service /builddir/build/BUILDROOT/ollama-0.3.6-1.x86_64/usr/lib/systemd/system/ollama-ui.service + mkdir -p /builddir/build/BUILDROOT/ollama-0.3.6-1.x86_64/usr/share/ollama/ + cp -r ollama-ui/ /builddir/build/BUILDROOT/ollama-0.3.6-1.x86_64/usr/share/ollama/ + /usr/lib/rpm/check-buildroot + /usr/lib/rpm/brp-ldconfig + /usr/lib/rpm/brp-compress + /usr/lib/rpm/brp-strip /usr/bin/strip + /usr/lib/rpm/brp-strip-comment-note /usr/bin/strip /usr/bin/objdump + /usr/lib/rpm/brp-strip-static-archive /usr/bin/strip + /usr/lib/rpm/brp-python-bytecompile /usr/bin/python 1 1 + /usr/lib/rpm/brp-python-hardlink Processing files: ollama-0.3.6-1.x86_64 Executing(%doc): /bin/sh -e /var/tmp/rpm-tmp.ZbE3iy + umask 022 + cd /builddir/build/BUILD + cd ollama-0.3.6 + DOCDIR=/builddir/build/BUILDROOT/ollama-0.3.6-1.x86_64/usr/share/doc/ollama + export LC_ALL=C + LC_ALL=C + export DOCDIR + /usr/bin/mkdir -p /builddir/build/BUILDROOT/ollama-0.3.6-1.x86_64/usr/share/doc/ollama + cp -pr README.md /builddir/build/BUILDROOT/ollama-0.3.6-1.x86_64/usr/share/doc/ollama + RPM_EC=0 ++ jobs -p + exit 0 Executing(%license): /bin/sh -e /var/tmp/rpm-tmp.YtANhb + umask 022 + cd /builddir/build/BUILD + cd ollama-0.3.6 + LICENSEDIR=/builddir/build/BUILDROOT/ollama-0.3.6-1.x86_64/usr/share/licenses/ollama + export LC_ALL=C + LC_ALL=C + export LICENSEDIR + /usr/bin/mkdir -p /builddir/build/BUILDROOT/ollama-0.3.6-1.x86_64/usr/share/licenses/ollama + cp -pr LICENSE /builddir/build/BUILDROOT/ollama-0.3.6-1.x86_64/usr/share/licenses/ollama + RPM_EC=0 ++ jobs -p + exit 0 Provides: ollama = 0.3.6-1 ollama(x86-64) = 0.3.6-1 Requires(interp): /bin/sh /bin/sh /bin/sh Requires(rpmlib): rpmlib(CompressedFileNames) <= 3.0.4-1 rpmlib(FileDigests) <= 4.6.0-1 rpmlib(PayloadFilesHavePrefix) <= 4.0-1 Requires(post): /bin/sh Requires(preun): /bin/sh Requires(postun): /bin/sh Requires: /bin/bash /bin/sh libc.so.6()(64bit) libc.so.6(GLIBC_2.14)(64bit) libc.so.6(GLIBC_2.17)(64bit) libc.so.6(GLIBC_2.2.5)(64bit) libc.so.6(GLIBC_2.29)(64bit) libc.so.6(GLIBC_2.3.2)(64bit) libc.so.6(GLIBC_2.32)(64bit) libc.so.6(GLIBC_2.33)(64bit) libc.so.6(GLIBC_2.34)(64bit) libc.so.6(GLIBC_2.38)(64bit) libc.so.6(GLIBC_2.7)(64bit) libgcc_s.so.1()(64bit) libgcc_s.so.1(GCC_3.0)(64bit) libm.so.6()(64bit) libm.so.6(GLIBC_2.2.5)(64bit) libm.so.6(GLIBC_2.27)(64bit) libm.so.6(GLIBC_2.29)(64bit) libresolv.so.2()(64bit) libstdc++.so.6()(64bit) libstdc++.so.6(CXXABI_1.3)(64bit) libstdc++.so.6(CXXABI_1.3.11)(64bit) libstdc++.so.6(CXXABI_1.3.13)(64bit) libstdc++.so.6(CXXABI_1.3.2)(64bit) libstdc++.so.6(CXXABI_1.3.3)(64bit) libstdc++.so.6(CXXABI_1.3.5)(64bit) libstdc++.so.6(CXXABI_1.3.9)(64bit) libstdc++.so.6(GLIBCXX_3.4)(64bit) libstdc++.so.6(GLIBCXX_3.4.11)(64bit) libstdc++.so.6(GLIBCXX_3.4.14)(64bit) libstdc++.so.6(GLIBCXX_3.4.15)(64bit) libstdc++.so.6(GLIBCXX_3.4.17)(64bit) libstdc++.so.6(GLIBCXX_3.4.18)(64bit) libstdc++.so.6(GLIBCXX_3.4.20)(64bit) libstdc++.so.6(GLIBCXX_3.4.21)(64bit) libstdc++.so.6(GLIBCXX_3.4.22)(64bit) libstdc++.so.6(GLIBCXX_3.4.29)(64bit) libstdc++.so.6(GLIBCXX_3.4.9)(64bit) rtld(GNU_HASH) Checking for unpackaged file(s): /usr/lib/rpm/check-files /builddir/build/BUILDROOT/ollama-0.3.6-1.x86_64 Wrote: /builddir/build/RPMS/ollama-0.3.6-1.x86_64.rpm Executing(%clean): /bin/sh -e /var/tmp/rpm-tmp.wOqov2 + umask 022 + cd /builddir/build/BUILD + cd ollama-0.3.6 + /usr/bin/rm -rf /builddir/build/BUILDROOT/ollama-0.3.6-1.x86_64 + RPM_EC=0 ++ jobs -p + exit 0 Executing(rmbuild): /bin/sh -e /var/tmp/rpm-tmp.SVgEdk + umask 022 + cd /builddir/build/BUILD + rm -rf ollama-0.3.6 ollama-0.3.6.gemspec + RPM_EC=0 ++ jobs -p + exit 0 Child return code was: 0