From 93b779652080786228271c6062b547f3bf068f45 Mon Sep 17 00:00:00 2001 From: TrueCharts Bot Date: Thu, 25 Jul 2024 02:36:20 +0200 Subject: [PATCH] chore(deps): update container image docker.io/localai/localai to v2.19.2 by renovate (#24258) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit This PR contains the following updates: | Package | Update | Change | |---|---|---| | [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) | patch | `v2.19.1-aio-cpu` -> `v2.19.2-aio-cpu` | | [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) | patch | `v2.19.1-aio-gpu-nvidia-cuda-11` -> `v2.19.2-aio-gpu-nvidia-cuda-11` | | [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) | patch | `v2.19.1-aio-gpu-nvidia-cuda-12` -> `v2.19.2-aio-gpu-nvidia-cuda-12` | | [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) | patch | `v2.19.1-cublas-cuda11-ffmpeg-core` -> `v2.19.2-cublas-cuda11-ffmpeg-core` | | [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) | patch | `v2.19.1-cublas-cuda12-ffmpeg-core` -> `v2.19.2-cublas-cuda12-ffmpeg-core` | | [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) | patch | `v2.19.1-cublas-cuda12-core` -> `v2.19.2-cublas-cuda12-core` | | [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) | patch | `v2.19.1-ffmpeg-core` -> `v2.19.2-ffmpeg-core` | --- > [!WARNING] > Some dependencies could not be looked up. Check the Dependency Dashboard for more information. --- ### Release Notes
mudler/LocalAI (docker.io/localai/localai) ### [`v2.19.2`](https://togithub.com/mudler/LocalAI/releases/tag/v2.19.2) [Compare Source](https://togithub.com/mudler/LocalAI/compare/v2.19.1...v2.19.2) This release is a patch release to fix well known issues from 2.19.x ##### What's Changed ##### Bug fixes :bug: - fix: pin setuptools 69.5.1 by [@​fakezeta](https://togithub.com/fakezeta) in [https://github.com/mudler/LocalAI/pull/2949](https://togithub.com/mudler/LocalAI/pull/2949) - fix(cuda): downgrade to 12.0 to increase compatibility range by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2994](https://togithub.com/mudler/LocalAI/pull/2994) - fix(llama.cpp): do not set anymore lora_base by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2999](https://togithub.com/mudler/LocalAI/pull/2999) ##### Exciting New Features 🎉 - ci(Makefile): reduce binary size by compressing by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2947](https://togithub.com/mudler/LocalAI/pull/2947) - feat(p2p): warn the user to start with --p2p by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2993](https://togithub.com/mudler/LocalAI/pull/2993) ##### 🧠 Models - models(gallery): add tulu 8b and 70b by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2931](https://togithub.com/mudler/LocalAI/pull/2931) - models(gallery): add suzume-orpo by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2932](https://togithub.com/mudler/LocalAI/pull/2932) - models(gallery): add archangel_sft_pythia2-8b by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2933](https://togithub.com/mudler/LocalAI/pull/2933) - models(gallery): add celestev1.2 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2937](https://togithub.com/mudler/LocalAI/pull/2937) - models(gallery): add calme-2.3-phi3-4b by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2939](https://togithub.com/mudler/LocalAI/pull/2939) - models(gallery): add calme-2.8-qwen2-7b by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2940](https://togithub.com/mudler/LocalAI/pull/2940) - models(gallery): add StellarDong-72b by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2941](https://togithub.com/mudler/LocalAI/pull/2941) - models(gallery): add calme-2.4-llama3-70b by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2942](https://togithub.com/mudler/LocalAI/pull/2942) - models(gallery): add llama3.1 70b and 8b by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/3000](https://togithub.com/mudler/LocalAI/pull/3000) ##### 📖 Documentation and examples - docs: add federation by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2929](https://togithub.com/mudler/LocalAI/pull/2929) - docs: :arrow_up: update docs version mudler/LocalAI by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2935](https://togithub.com/mudler/LocalAI/pull/2935) ##### 👒 Dependencies - chore: :arrow_up: Update ggerganov/llama.cpp by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2936](https://togithub.com/mudler/LocalAI/pull/2936) - chore: :arrow_up: Update ggerganov/llama.cpp by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2943](https://togithub.com/mudler/LocalAI/pull/2943) - chore(deps): Bump grpcio from 1.64.1 to 1.65.1 in /backend/python/openvoice by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2956](https://togithub.com/mudler/LocalAI/pull/2956) - chore(deps): Bump grpcio from 1.65.0 to 1.65.1 in /backend/python/sentencetransformers by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2955](https://togithub.com/mudler/LocalAI/pull/2955) - chore(deps): Bump grpcio from 1.65.0 to 1.65.1 in /backend/python/bark by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2951](https://togithub.com/mudler/LocalAI/pull/2951) - chore(deps): Bump docs/themes/hugo-theme-relearn from `1b2e139` to `7aec99b` by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2952](https://togithub.com/mudler/LocalAI/pull/2952) - chore(deps): Bump langchain from 0.2.8 to 0.2.10 in /examples/langchain/langchainpy-localai-example by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2959](https://togithub.com/mudler/LocalAI/pull/2959) - chore(deps): Bump numpy from 1.26.4 to 2.0.1 in /examples/langchain/langchainpy-localai-example by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2958](https://togithub.com/mudler/LocalAI/pull/2958) - chore(deps): Bump sqlalchemy from 2.0.30 to 2.0.31 in /examples/langchain/langchainpy-localai-example by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2957](https://togithub.com/mudler/LocalAI/pull/2957) - chore(deps): Bump grpcio from 1.65.0 to 1.65.1 in /backend/python/vllm by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2964](https://togithub.com/mudler/LocalAI/pull/2964) - chore(deps): Bump llama-index from 0.10.55 to 0.10.56 in /examples/chainlit by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2966](https://togithub.com/mudler/LocalAI/pull/2966) - chore(deps): Bump grpcio from 1.65.0 to 1.65.1 in /backend/python/common/template by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2963](https://togithub.com/mudler/LocalAI/pull/2963) - chore(deps): Bump weaviate-client from 4.6.5 to 4.6.7 in /examples/chainlit by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2965](https://togithub.com/mudler/LocalAI/pull/2965) - chore(deps): Bump grpcio from 1.65.0 to 1.65.1 in /backend/python/transformers by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2970](https://togithub.com/mudler/LocalAI/pull/2970) - chore(deps): Bump openai from 1.35.13 to 1.37.0 in /examples/functions by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2973](https://togithub.com/mudler/LocalAI/pull/2973) - chore(deps): Bump grpcio from 1.65.0 to 1.65.1 in /backend/python/diffusers by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2969](https://togithub.com/mudler/LocalAI/pull/2969) - chore(deps): Bump grpcio from 1.65.0 to 1.65.1 in /backend/python/exllama2 by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2971](https://togithub.com/mudler/LocalAI/pull/2971) - chore(deps): Bump grpcio from 1.65.0 to 1.65.1 in /backend/python/rerankers by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2974](https://togithub.com/mudler/LocalAI/pull/2974) - chore(deps): Bump grpcio from 1.65.0 to 1.65.1 in /backend/python/coqui by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2980](https://togithub.com/mudler/LocalAI/pull/2980) - chore(deps): Bump grpcio from 1.65.0 to 1.65.1 in /backend/python/parler-tts by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2982](https://togithub.com/mudler/LocalAI/pull/2982) - chore(deps): Bump grpcio from 1.65.0 to 1.65.1 in /backend/python/vall-e-x by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2981](https://togithub.com/mudler/LocalAI/pull/2981) - chore(deps): Bump grpcio from 1.65.0 to 1.65.1 in /backend/python/transformers-musicgen by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2990](https://togithub.com/mudler/LocalAI/pull/2990) - chore(deps): Bump grpcio from 1.65.0 to 1.65.1 in /backend/python/autogptq by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2984](https://togithub.com/mudler/LocalAI/pull/2984) - chore(deps): Bump llama-index from 0.10.55 to 0.10.56 in /examples/langchain-chroma by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2986](https://togithub.com/mudler/LocalAI/pull/2986) - chore(deps): Bump grpcio from 1.65.0 to 1.65.1 in /backend/python/mamba by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2989](https://togithub.com/mudler/LocalAI/pull/2989) - chore: :arrow_up: Update ggerganov/llama.cpp by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2992](https://togithub.com/mudler/LocalAI/pull/2992) - chore(deps): Bump langchain-community from 0.2.7 to 0.2.9 in /examples/langchain/langchainpy-localai-example by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2960](https://togithub.com/mudler/LocalAI/pull/2960) - chore(deps): Bump openai from 1.35.13 to 1.37.0 in /examples/langchain/langchainpy-localai-example by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2961](https://togithub.com/mudler/LocalAI/pull/2961) - chore(deps): Bump langchain from 0.2.8 to 0.2.10 in /examples/functions by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2975](https://togithub.com/mudler/LocalAI/pull/2975) - chore(deps): Bump openai from 1.35.13 to 1.37.0 in /examples/langchain-chroma by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2988](https://togithub.com/mudler/LocalAI/pull/2988) - chore(deps): Bump langchain from 0.2.8 to 0.2.10 in /examples/langchain-chroma by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2987](https://togithub.com/mudler/LocalAI/pull/2987) - chore: :arrow_up: Update ggerganov/llama.cpp by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2995](https://togithub.com/mudler/LocalAI/pull/2995) ##### Other Changes - ci(Makefile): enable p2p on cross-arm64 builds by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2928](https://togithub.com/mudler/LocalAI/pull/2928) **Full Changelog**: https://github.com/mudler/LocalAI/compare/v2.19.1...v2.19.2
--- ### Configuration 📅 **Schedule**: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined). 🚦 **Automerge**: Enabled. ♻ **Rebasing**: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox. 🔕 **Ignore**: Close this PR and you won't be reminded about these updates again. --- - [ ] If you want to rebase/retry this PR, check this box --- This PR has been generated by [Renovate Bot](https://togithub.com/renovatebot/renovate). --- charts/stable/local-ai/Chart.yaml | 2 +- charts/stable/local-ai/values.yaml | 14 +++++++------- 2 files changed, 8 insertions(+), 8 deletions(-) diff --git a/charts/stable/local-ai/Chart.yaml b/charts/stable/local-ai/Chart.yaml index 70b7392d563..1eb060d9bb0 100644 --- a/charts/stable/local-ai/Chart.yaml +++ b/charts/stable/local-ai/Chart.yaml @@ -33,4 +33,4 @@ sources: - https://github.com/truecharts/charts/tree/master/charts/stable/local-ai - https://hub.docker.com/r/localai/localai type: application -version: 11.11.3 +version: 11.11.4 diff --git a/charts/stable/local-ai/values.yaml b/charts/stable/local-ai/values.yaml index 200f791f0bc..b66b9a9a263 100644 --- a/charts/stable/local-ai/values.yaml +++ b/charts/stable/local-ai/values.yaml @@ -5,15 +5,15 @@ image: ffmpegImage: repository: docker.io/localai/localai pullPolicy: IfNotPresent - tag: v2.19.1-ffmpeg-core@sha256:c63afdd8ede2b29b8a8fa15dc3b5370d33c28deab45d3334ff07064fe0682662 + tag: v2.19.2-ffmpeg-core@sha256:cedf339779f3dec9f58d6033bc76c4ddd2581adcb942eec8e3e20ad85fff70b2 cublasCuda12Image: repository: docker.io/localai/localai pullPolicy: IfNotPresent - tag: v2.19.1-cublas-cuda12-core@sha256:89d0680ec8690fe53cbe4f1608fd514277a425e6dd5e69978d9ac9ee8d9f9d1e + tag: v2.19.2-cublas-cuda12-core@sha256:368ab8bdf7a48f9d6b77720a823aa57ee7c971d64425ab98fc902c859b567a91 cublasCuda12FfmpegImage: repository: docker.io/localai/localai pullPolicy: IfNotPresent - tag: v2.19.1-cublas-cuda12-ffmpeg-core@sha256:fefb32ece781e700a43771c4b0108a3af0d5a7c0d5c1849c17d7f265e9408b3f + tag: v2.19.2-cublas-cuda12-ffmpeg-core@sha256:ffcef33926bb24f31aefe70bba84932968c509bd11a97ca21c1d178a793b8539 cublasCuda11Image: repository: docker.io/localai/localai pullPolicy: IfNotPresent @@ -21,19 +21,19 @@ cublasCuda11Image: cublasCuda11FfmpegImage: repository: docker.io/localai/localai pullPolicy: IfNotPresent - tag: v2.19.1-cublas-cuda11-ffmpeg-core@sha256:3b868811ca8ee589eb4ea86bcabf0e991794a57a2aa1aa25ab6c023e4ba1811d + tag: v2.19.2-cublas-cuda11-ffmpeg-core@sha256:9a571aa7aab8c182aa9a8f8583dbed78a29f4ff81ebc90ab7eff57a330c5b467 allInOneCuda12Image: repository: docker.io/localai/localai pullPolicy: IfNotPresent - tag: v2.19.1-aio-gpu-nvidia-cuda-12@sha256:07fc7e18e871b711f42688a6c98fdf588f5a8711a754b44c1afbda663cc2b35d + tag: v2.19.2-aio-gpu-nvidia-cuda-12@sha256:22734d5b39f10fa5463c67b69d36e75d81acf39b71cd06a5f29342ddc66f8c13 allInOneCuda11Image: repository: docker.io/localai/localai pullPolicy: IfNotPresent - tag: v2.19.1-aio-gpu-nvidia-cuda-11@sha256:78c99ca29bf1cccd3586b54ae8fe863a44f46e3e5b27e8e2e0d9b18e20e990dc + tag: v2.19.2-aio-gpu-nvidia-cuda-11@sha256:f27dcc1040654028b8314eed6c548b84b8d1e55bc2a2ff17923a15cc8e15b237 allInOneCpuImage: repository: docker.io/localai/localai pullPolicy: IfNotPresent - tag: v2.19.1-aio-cpu@sha256:f28abab3ab6a04a7e569cb824eb1e012312eeeab8cfaad4a69ef1ffe8910199c + tag: v2.19.2-aio-cpu@sha256:e272ca3b42eaa902d1a4fd521df5a01b1bbb62dd072a66ad6a5eab01e32c0b8c securityContext: container: runAsNonRoot: false