Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

feat: more embedded models, coqui fixes, add model usage and description #1556

Merged
merged 13 commits into from
Jan 7, 2024

Conversation

mudler
Copy link
Owner

@mudler mudler commented Jan 6, 2024

Description

This PR adds whisper, embeddings, and tts to the one-click install. I still need to test all of them - so it's still in WIP. It removes the hardcoded model gallery, but most importantly it adds a description and usage field in the model yamls so it is displayed during the model preloading. This allows for instance to inform the user on how a specific model which is loaded is meant to be used.

As not-related, improves UX and output when trying to connect to backend services. It also fixes #1549

Tested embedded models configs:

  • bert-cpp
  • rhasspy-tts
  • all-minilm-l6-v2
  • whisper-base
  • coqui
  • vall-e-x
  • rhasspy

Testing with, for example:

docker run -ti --rm -p 8080:8080 quay.io/go-skynet/local-ai:master-ffmpeg-core https://gist.githubusercontent.com/mudler/43fd618e6acf57f9b15467b0e60c1f3e/raw/edc8bc8ddd6344bd6e17747d99cc91e72147cc8b/bert-cpp.yaml

usage example is printed out during start

Copy link

netlify bot commented Jan 6, 2024

Deploy Preview for localai ready!

Name Link
🔨 Latest commit c4b2885
🔍 Latest deploy log https://app.netlify.com/sites/localai/deploys/659b1ed7c9439600086c446e
😎 Deploy Preview https://deploy-preview-1556--localai.netlify.app
📱 Preview on mobile
Toggle QR Code...

QR Code

Use your smartphone camera to open QR code link.

To edit notification comments on pull requests, go to your Netlify site configuration.

@mudler
Copy link
Owner Author

mudler commented Jan 6, 2024

I'd like to have an example for each backend in the one-click setup section, and update the docs while at it, as we miss several backends entirely

@mudler mudler changed the title feat: more embedded models, add model usage and description feat: more embedded models, coqui fixes, add model usage and description Jan 7, 2024
@mudler
Copy link
Owner Author

mudler commented Jan 7, 2024

I've tried coqui and added a fix for it in this PR (#1549). Will follow-up an e2e example as well.

@mudler
Copy link
Owner Author

mudler commented Jan 7, 2024

examples missing for:

  • stablediffusion
  • vllm
  • exllama(2)
  • autogptq

@mudler mudler merged commit e19d722 into master Jan 7, 2024
24 checks passed
@mudler mudler deleted the docs_updates_4 branch January 7, 2024 23:37
@mudler mudler added the enhancement New feature or request label Jan 8, 2024
truecharts-admin referenced this pull request in truecharts/public Jan 8, 2024
….0 by renovate (#17044)

This PR contains the following updates:

| Package | Update | Change |
|---|---|---|
| [docker.io/localai/localai](https://github.com/mudler/LocalAI) |
minor | `v2.4.1-cublas-cuda11-ffmpeg-core` ->
`v2.5.0-cublas-cuda11-ffmpeg-core` |

---

> [!WARNING]
> Some dependencies could not be looked up. Check the Dependency
Dashboard for more information.

---

### Release Notes

<details>
<summary>mudler/LocalAI (docker.io/localai/localai)</summary>

### [`v2.5.0`](https://github.com/mudler/LocalAI/releases/tag/v2.5.0)

[Compare
Source](https://github.com/mudler/LocalAI/compare/v2.4.1...v2.5.0)

<!-- Release notes generated using configuration in .github/release.yml
at master -->

##### What's Changed

This release adds more embedded models, and shrink image sizes.

You can run now `phi-2` ( see
[here](https://localai.io/basics/getting_started/#running-popular-models-one-click)
for the full list ) locally by starting localai with:

    docker run -ti -p 8080:8080 localai/localai:v2.5.0-ffmpeg-core phi-2

LocalAI accepts now as argument a list of short-hands models and/or URLs
pointing to valid yaml file. A popular way to host those files are
Github gists.

For instance, you can run `llava`, by starting `local-ai` with:

```bash
docker run -ti -p 8080:8080 localai/localai:v2.5.0-ffmpeg-core https://raw.githubusercontent.com/mudler/LocalAI/master/embedded/models/llava.yaml
```

##### Exciting New Features 🎉

- feat: more embedded models, coqui fixes, add model usage and
description by [@&#8203;mudler](https://github.com/mudler) in
[https://github.com/mudler/LocalAI/pull/1556](https://github.com/mudler/LocalAI/pull/1556)

##### 👒 Dependencies

- deps(conda): use transformers-env with vllm,exllama(2) by
[@&#8203;mudler](https://github.com/mudler) in
[https://github.com/mudler/LocalAI/pull/1554](https://github.com/mudler/LocalAI/pull/1554)
- deps(conda): use transformers environment with autogptq by
[@&#8203;mudler](https://github.com/mudler) in
[https://github.com/mudler/LocalAI/pull/1555](https://github.com/mudler/LocalAI/pull/1555)
- ⬆️ Update ggerganov/llama.cpp by
[@&#8203;localai-bot](https://github.com/localai-bot) in
[https://github.com/mudler/LocalAI/pull/1558](https://github.com/mudler/LocalAI/pull/1558)

##### Other Changes

- ⬆️ Update docs version mudler/LocalAI by
[@&#8203;localai-bot](https://github.com/localai-bot) in
[https://github.com/mudler/LocalAI/pull/1557](https://github.com/mudler/LocalAI/pull/1557)

**Full Changelog**:
mudler/LocalAI@v2.4.1...v2.5.0

</details>

---

### Configuration

📅 **Schedule**: Branch creation - "before 10pm on monday" in timezone
Europe/Amsterdam, Automerge - At any time (no schedule defined).

🚦 **Automerge**: Enabled.

♻ **Rebasing**: Whenever PR becomes conflicted, or you tick the
rebase/retry checkbox.

🔕 **Ignore**: Close this PR and you won't be reminded about this update
again.

---

- [ ] <!-- rebase-check -->If you want to rebase/retry this PR, check
this box

---

This PR has been generated by [Renovate
Bot](https://github.com/renovatebot/renovate).

<!--renovate-debug:eyJjcmVhdGVkSW5WZXIiOiIzNy4xMjcuMCIsInVwZGF0ZWRJblZlciI6IjM3LjEyNy4wIiwidGFyZ2V0QnJhbmNoIjoibWFzdGVyIn0=-->
@mudler mudler mentioned this pull request Jan 10, 2024
1 task
GabrielBarzen referenced this pull request in GabrielBarzen/charts Feb 2, 2024
….0 by renovate (truecharts#17044)

This PR contains the following updates:

| Package | Update | Change |
|---|---|---|
| [docker.io/localai/localai](https://github.com/mudler/LocalAI) |
minor | `v2.4.1-cublas-cuda11-ffmpeg-core` ->
`v2.5.0-cublas-cuda11-ffmpeg-core` |

---

> [!WARNING]
> Some dependencies could not be looked up. Check the Dependency
Dashboard for more information.

---

### Release Notes

<details>
<summary>mudler/LocalAI (docker.io/localai/localai)</summary>

### [`v2.5.0`](https://github.com/mudler/LocalAI/releases/tag/v2.5.0)

[Compare
Source](https://github.com/mudler/LocalAI/compare/v2.4.1...v2.5.0)

<!-- Release notes generated using configuration in .github/release.yml
at master -->

##### What's Changed

This release adds more embedded models, and shrink image sizes.

You can run now `phi-2` ( see
[here](https://localai.io/basics/getting_started/#running-popular-models-one-click)
for the full list ) locally by starting localai with:

    docker run -ti -p 8080:8080 localai/localai:v2.5.0-ffmpeg-core phi-2

LocalAI accepts now as argument a list of short-hands models and/or URLs
pointing to valid yaml file. A popular way to host those files are
Github gists.

For instance, you can run `llava`, by starting `local-ai` with:

```bash
docker run -ti -p 8080:8080 localai/localai:v2.5.0-ffmpeg-core https://raw.githubusercontent.com/mudler/LocalAI/master/embedded/models/llava.yaml
```

##### Exciting New Features 🎉

- feat: more embedded models, coqui fixes, add model usage and
description by [@&truecharts#8203;mudler](https://github.com/mudler) in
[https://github.com/mudler/LocalAI/pull/1556](https://github.com/mudler/LocalAI/pull/1556)

##### 👒 Dependencies

- deps(conda): use transformers-env with vllm,exllama(2) by
[@&truecharts#8203;mudler](https://github.com/mudler) in
[https://github.com/mudler/LocalAI/pull/1554](https://github.com/mudler/LocalAI/pull/1554)
- deps(conda): use transformers environment with autogptq by
[@&truecharts#8203;mudler](https://github.com/mudler) in
[https://github.com/mudler/LocalAI/pull/1555](https://github.com/mudler/LocalAI/pull/1555)
- ⬆️ Update ggerganov/llama.cpp by
[@&truecharts#8203;localai-bot](https://github.com/localai-bot) in
[https://github.com/mudler/LocalAI/pull/1558](https://github.com/mudler/LocalAI/pull/1558)

##### Other Changes

- ⬆️ Update docs version mudler/LocalAI by
[@&truecharts#8203;localai-bot](https://github.com/localai-bot) in
[https://github.com/mudler/LocalAI/pull/1557](https://github.com/mudler/LocalAI/pull/1557)

**Full Changelog**:
mudler/LocalAI@v2.4.1...v2.5.0

</details>

---

### Configuration

📅 **Schedule**: Branch creation - "before 10pm on monday" in timezone
Europe/Amsterdam, Automerge - At any time (no schedule defined).

🚦 **Automerge**: Enabled.

♻ **Rebasing**: Whenever PR becomes conflicted, or you tick the
rebase/retry checkbox.

🔕 **Ignore**: Close this PR and you won't be reminded about this update
again.

---

- [ ] <!-- rebase-check -->If you want to rebase/retry this PR, check
this box

---

This PR has been generated by [Renovate
Bot](https://github.com/renovatebot/renovate).

<!--renovate-debug:eyJjcmVhdGVkSW5WZXIiOiIzNy4xMjcuMCIsInVwZGF0ZWRJblZlciI6IjM3LjEyNy4wIiwidGFyZ2V0QnJhbmNoIjoibWFzdGVyIn0=-->
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

Successfully merging this pull request may close these issues.

TTS with coqui: Examples missing and Error 404: sendfile: file /tmp/generated/audio/piper.wav not found
1 participant