From 09cf08f8f80ed468d0f296f4207ce87bfe7d0e15 Mon Sep 17 00:00:00 2001 From: TrueCharts Bot Date: Mon, 8 Jan 2024 19:27:29 +0100 Subject: [PATCH] chore(deps): update container image docker.io/localai/localai to v2.5.0 by renovate (#17044) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit This PR contains the following updates: | Package | Update | Change | |---|---|---| | [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) | minor | `v2.4.1-cublas-cuda11-ffmpeg-core` -> `v2.5.0-cublas-cuda11-ffmpeg-core` | --- > [!WARNING] > Some dependencies could not be looked up. Check the Dependency Dashboard for more information. --- ### Release Notes
mudler/LocalAI (docker.io/localai/localai) ### [`v2.5.0`](https://togithub.com/mudler/LocalAI/releases/tag/v2.5.0) [Compare Source](https://togithub.com/mudler/LocalAI/compare/v2.4.1...v2.5.0) ##### What's Changed This release adds more embedded models, and shrink image sizes. You can run now `phi-2` ( see [here](https://localai.io/basics/getting_started/#running-popular-models-one-click) for the full list ) locally by starting localai with: docker run -ti -p 8080:8080 localai/localai:v2.5.0-ffmpeg-core phi-2 LocalAI accepts now as argument a list of short-hands models and/or URLs pointing to valid yaml file. A popular way to host those files are Github gists. For instance, you can run `llava`, by starting `local-ai` with: ```bash docker run -ti -p 8080:8080 localai/localai:v2.5.0-ffmpeg-core https://raw.githubusercontent.com/mudler/LocalAI/master/embedded/models/llava.yaml ``` ##### Exciting New Features 🎉 - feat: more embedded models, coqui fixes, add model usage and description by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/1556](https://togithub.com/mudler/LocalAI/pull/1556) ##### 👒 Dependencies - deps(conda): use transformers-env with vllm,exllama(2) by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/1554](https://togithub.com/mudler/LocalAI/pull/1554) - deps(conda): use transformers environment with autogptq by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/1555](https://togithub.com/mudler/LocalAI/pull/1555) - :arrow_up: Update ggerganov/llama.cpp by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/1558](https://togithub.com/mudler/LocalAI/pull/1558) ##### Other Changes - :arrow_up: Update docs version mudler/LocalAI by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/1557](https://togithub.com/mudler/LocalAI/pull/1557) **Full Changelog**: https://github.com/mudler/LocalAI/compare/v2.4.1...v2.5.0
--- ### Configuration 📅 **Schedule**: Branch creation - "before 10pm on monday" in timezone Europe/Amsterdam, Automerge - At any time (no schedule defined). 🚦 **Automerge**: Enabled. ♻ **Rebasing**: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox. 🔕 **Ignore**: Close this PR and you won't be reminded about this update again. --- - [ ] If you want to rebase/retry this PR, check this box --- This PR has been generated by [Renovate Bot](https://togithub.com/renovatebot/renovate). --- charts/stable/local-ai/Chart.yaml | 4 ++-- charts/stable/local-ai/values.yaml | 12 ++++++------ 2 files changed, 8 insertions(+), 8 deletions(-) diff --git a/charts/stable/local-ai/Chart.yaml b/charts/stable/local-ai/Chart.yaml index 57f7b114326..ed3dd113890 100644 --- a/charts/stable/local-ai/Chart.yaml +++ b/charts/stable/local-ai/Chart.yaml @@ -1,8 +1,8 @@ kubeVersion: '>=1.24.0-0' apiVersion: v2 name: local-ai -version: 8.8.6 -appVersion: 2.4.0 +version: 8.14.0 +appVersion: 2.5.0 description: Self-hosted, community-driven, local OpenAI-compatible API. home: https://truecharts.org/charts/stable/local-ai icon: https://truecharts.org/img/hotlink-ok/chart-icons/local-ai.png diff --git a/charts/stable/local-ai/values.yaml b/charts/stable/local-ai/values.yaml index 11b4c35694f..8b0d27efc25 100644 --- a/charts/stable/local-ai/values.yaml +++ b/charts/stable/local-ai/values.yaml @@ -1,27 +1,27 @@ image: repository: docker.io/localai/localai pullPolicy: IfNotPresent - tag: v2.4.1@sha256:9d725dbe5bf853363d81c948780f8a4e5b48a984ffc646924682db253bf806f8 + tag: v2.5.0@sha256:f936ee39751d20423734be22c98a2ed786c15674c62375c294bb29ced2c1a37c ffmpegImage: repository: docker.io/localai/localai pullPolicy: IfNotPresent - tag: v2.4.1-ffmpeg-core@sha256:478b57ac43d8d523c8c3429f84d7909463c254df85a880cc4cba10aee959017d + tag: v2.5.0-ffmpeg-core@sha256:3e0844e20158b2e3f8b45e3a4b3d0d68366314bcd68f1ca46007961e7210f547 cublasCuda12Image: repository: docker.io/localai/localai pullPolicy: IfNotPresent - tag: v2.4.1-cublas-cuda12-core@sha256:eca7d7f5b59aa884edc4155d8628ddcdcf21e18e2cd12296e8b7f36bfae3affe + tag: v2.5.0-cublas-cuda12-core@sha256:7a732963bf30a9254291f8d705cf3ad273d0546256895a7b6d8dd933f2703b5c cublasCuda12FfmpegImage: repository: docker.io/localai/localai pullPolicy: IfNotPresent - tag: v2.4.1-cublas-cuda12-ffmpeg-core@sha256:b77bdfa20e2578c450215612e70f1b77d230c59f54b5f33e00c17f20bc24fbed + tag: v2.5.0-cublas-cuda12-ffmpeg-core@sha256:eee4baae2b0c91e3b6964561d01d5307c2efb8fdae0ac45335d37945e154a8df cublasCuda11Image: repository: docker.io/localai/localai pullPolicy: IfNotPresent - tag: v2.4.1-cublas-cuda11-core@sha256:934905f2f48b190ff5ea984c2f9d4996ab2d8e2e760568ab3c3d4a085b0b30fc + tag: v2.5.0-cublas-cuda11-core@sha256:84840ad35ac9456b02c786c5a0aa56472e13b4732fcd0e22be971c0bdbb012d9 cublasCuda11FfmpegImage: repository: docker.io/localai/localai pullPolicy: IfNotPresent - tag: v2.4.1-cublas-cuda11-ffmpeg-core@sha256:b34c297fed229dd8a60e9490808a70c4c40046e074cd36e1b045b465939d9106 + tag: v2.5.0-cublas-cuda11-ffmpeg-core@sha256:8f0144c56db1b2dae3bb350ae3a88a13f8e01a2b7b6df5ab96e46e4a54b43dce securityContext: container: runAsNonRoot: false