Skip to content

Commit 921e04a

Browse files
authored
chore: bump localai v2.26.0 (#487)
Signed-off-by: Sertac Ozercan <[email protected]>
1 parent a1624c3 commit 921e04a

File tree

11 files changed

+11
-124
lines changed

11 files changed

+11
-124
lines changed

.github/workflows/test-docker.yaml

Lines changed: 0 additions & 22 deletions
Original file line numberDiff line numberDiff line change
@@ -29,13 +29,9 @@ jobs:
2929
matrix:
3030
backend:
3131
- llama
32-
- stablediffusion
3332
arch:
3433
- amd64
3534
- arm64
36-
exclude:
37-
- backend: stablediffusion
38-
arch: arm64
3935
steps:
4036
- uses: jlumbroso/free-disk-space@54081f138730dfa15788a46383842cd2f914a1be # v1.3.1
4137
with:
@@ -127,28 +123,10 @@ jobs:
127123
exit 1
128124
fi
129125
130-
- name: run stablediffusion test
131-
if: matrix.backend == 'stablediffusion'
132-
run: |
133-
result=$(curl --fail --retry 10 --retry-all-errors http://127.0.0.1:8080/v1/images/generations -H "Content-Type: application/json" -d '{
134-
"prompt": "A cute baby llama",
135-
"size": "256x256"
136-
}')
137-
echo $result
138-
139-
url=$(echo "$result" | jq '.data[0].url')
140-
if [ -z "$url" ]; then
141-
exit 1
142-
fi
143-
144126
- name: save logs
145127
if: always()
146128
run: docker logs testmodel > /tmp/docker-${{ matrix.backend }}.log
147129

148-
- name: save generated image
149-
if: matrix.backend == 'stablediffusion'
150-
run: docker cp testmodel:/tmp/generated/images /tmp
151-
152130
- name: publish test artifacts
153131
if: always()
154132
uses: actions/upload-artifact@4cec3d8aa04e39d1a68397de0c4cd6fb9dce8ec1 # v4.6.1

.github/workflows/test-podman-applesilicon.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -45,7 +45,7 @@ jobs:
4545
4646
- name: run test (gguf)
4747
run: |
48-
result=$(curl --fail --retry 10 --retry-all-errors http://127.0.0.1:8080/v1/chat/completions -H "Content-Type: application/json" -d '{
48+
result=$(curl --fail --retry 10 --retry-all-errors http://localhost:8080/v1/chat/completions -H "Content-Type: application/json" -d '{
4949
"model": "llama-3.2-1b-instruct",
5050
"messages": [{"role": "user", "content": "explain kubernetes in a sentence"}]
5151
}')

pkg/aikit/config/specs_test.go

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -24,7 +24,6 @@ apiVersion: v1alpha1
2424
runtime: cuda
2525
backends:
2626
- exllama2
27-
- stablediffusion
2827
models:
2928
- name: test
3029
source: foo
@@ -34,7 +33,6 @@ models:
3433
Runtime: utils.RuntimeNVIDIA,
3534
Backends: []string{
3635
utils.BackendExllamaV2,
37-
utils.BackendStableDiffusion,
3836
},
3937
Models: []Model{
4038
{

pkg/aikit2llb/inference/convert.go

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@ import (
1414
const (
1515
distrolessBase = "ghcr.io/sozercan/base:latest"
1616
localAIRepo = "https://github.com/mudler/LocalAI"
17-
localAIVersion = "v2.25.0"
17+
localAIVersion = "v2.26.0"
1818
cudaVersion = "12-5"
1919
)
2020

@@ -49,8 +49,6 @@ func Aikit2LLB(c *config.InferenceConfig, platform *specs.Platform) (llb.State,
4949
switch c.Backends[b] {
5050
case utils.BackendExllamaV2:
5151
merge = installExllama(state, merge)
52-
case utils.BackendStableDiffusion:
53-
merge = installOpenCV(state, merge)
5452
case utils.BackendMamba:
5553
merge = installMamba(state, merge)
5654
case utils.BackendDiffusers:

pkg/aikit2llb/inference/stablediffusion.go

Lines changed: 0 additions & 30 deletions
This file was deleted.

pkg/build/build.go

Lines changed: 1 addition & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -452,10 +452,6 @@ func validateInferenceConfig(c *config.InferenceConfig) error {
452452
return errors.New("only one backend is supported at this time")
453453
}
454454

455-
if slices.Contains(c.Backends, utils.BackendStableDiffusion) && (slices.Contains(c.Backends, utils.BackendExllamaV2)) {
456-
return errors.New("cannot specify both stablediffusion with exllama2 at this time")
457-
}
458-
459455
if (slices.Contains(c.Backends, utils.BackendExllamaV2) || slices.Contains(c.Backends, utils.BackendMamba) || slices.Contains(c.Backends, utils.BackendDiffusers)) && c.Runtime != utils.RuntimeNVIDIA {
460456
return errors.New("exllama, mamba, and diffusers backends only supports nvidia cuda runtime. please add 'runtime: cuda' to your aikitfile.yaml")
461457
}
@@ -464,7 +460,7 @@ func validateInferenceConfig(c *config.InferenceConfig) error {
464460
return errors.New("apple silicon runtime only supports the default llama-cpp backend")
465461
}
466462

467-
backends := []string{utils.BackendExllamaV2, utils.BackendStableDiffusion, utils.BackendMamba, utils.BackendDiffusers}
463+
backends := []string{utils.BackendExllamaV2, utils.BackendMamba, utils.BackendDiffusers}
468464
for _, b := range c.Backends {
469465
if !slices.Contains(backends, b) {
470466
return errors.Errorf("backend %s is not supported", b)

pkg/build/build_test.go

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -84,7 +84,7 @@ func Test_validateConfig(t *testing.T) {
8484
args: args{c: &config.InferenceConfig{
8585
APIVersion: "v1alpha1",
8686
Runtime: "cuda",
87-
Backends: []string{"exllama", "stablediffusion"},
87+
Backends: []string{"exllama", "diffusers"},
8888
Models: []config.Model{
8989
{
9090
Name: "test",

pkg/utils/const.go

Lines changed: 3 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -4,10 +4,9 @@ const (
44
RuntimeNVIDIA = "cuda"
55
RuntimeAppleSilicon = "applesilicon" // experimental apple silicon runtime with vulkan arm64 support
66

7-
BackendStableDiffusion = "stablediffusion"
8-
BackendExllamaV2 = "exllama2"
9-
BackendMamba = "mamba"
10-
BackendDiffusers = "diffusers"
7+
BackendExllamaV2 = "exllama2"
8+
BackendMamba = "mamba"
9+
BackendDiffusers = "diffusers"
1110

1211
TargetUnsloth = "unsloth"
1312

test/aikitfile-stablediffusion.yaml

Lines changed: 0 additions & 50 deletions
This file was deleted.

website/docs/diffusion.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22
title: Diffusion
33
---
44

5-
AIKit supports [`diffusers`](#diffusers) and [`stablediffusion`](#stablediffusion) backends.
5+
AIKit supports [`diffusers`](#diffusers) backend.
66

77
## diffusers
88

@@ -16,7 +16,7 @@ Please make sure to change syntax to `#syntax=ghcr.io/sozercan/aikit:latest` in
1616

1717
https://github.com/sozercan/aikit/blob/main/test/aikitfile-diffusers.yaml
1818

19-
## stablediffusion
19+
## stablediffusion NCNN
2020

2121
https://github.com/EdVince/Stable-Diffusion-NCNN
2222

@@ -25,7 +25,7 @@ This backend:
2525
- does not support CUDA runtime yet
2626

2727
:::note
28-
This is an experimental backend and it may change in the future.
28+
This has been deprecated as of `v0.18.0` release.
2929
:::
3030

3131
### Example

website/docs/specs-inference.md

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ title: Inference API Specifications
88
apiVersion: # required. only v1alpha1 is supported at the moment
99
debug: # optional. if set to true, debug logs will be printed
1010
runtime: # optional. defaults to avx. can be "avx", "avx2", "avx512", "cuda"
11-
backends: # optional. list of additional backends. can be "stablediffusion", "exllama2", "diffusers", "mamba"
11+
backends: # optional. list of additional backends. can be "exllama2", "diffusers", "mamba"
1212
models: # required. list of models to build
1313
- name: # required. name of the model
1414
source: # required. source of the model. can be a url or a local file
@@ -26,8 +26,6 @@ Example:
2626
apiVersion: v1alpha1
2727
debug: true
2828
runtime: cuda
29-
backends:
30-
- stablediffusion
3129
models:
3230
- name: llama-2-7b-chat
3331
source: https://huggingface.co/TheBloke/Llama-2-7B-Chat-GGUF/resolve/main/llama-2-7b-chat.Q4_K_M.gguf

0 commit comments

Comments
 (0)