mirror of
https://github.com/charmbracelet/mods
synced 2024-10-18 07:12:18 +00:00
feat: add SambaNova Cloud (#356)
Some checks failed
build / build (ubuntu-latest) (push) Successful in 1m1s
build / snapshot (push) Failing after 0s
lint-soft / lint-soft (push) Failing after 2s
lint / lint (push) Failing after 2s
nightly / nightly (push) Failing after 0s
build / build (windows-latest) (push) Has been cancelled
build / build (macos-latest) (push) Has been cancelled
Some checks failed
build / build (ubuntu-latest) (push) Successful in 1m1s
build / snapshot (push) Failing after 0s
lint-soft / lint-soft (push) Failing after 2s
lint / lint (push) Failing after 2s
nightly / nightly (push) Failing after 0s
build / build (windows-latest) (push) Has been cancelled
build / build (macos-latest) (push) Has been cancelled
* feat[sambanova]: Add 4k models * feat[sambanova]: Add 8k models * fix: Make the 8K models default
This commit is contained in:
parent
2178cd7396
commit
cb0323ed96
|
@ -192,6 +192,29 @@ apis:
|
|||
llama3.1-70b:
|
||||
aliases: ["llama3.1-cerebras", "llama3.1-70b-cerebras"]
|
||||
max-input-chars: 24500
|
||||
sambanova:
|
||||
base-url: https://api.sambanova.ai/v1
|
||||
api-key:
|
||||
api-key-env: SAMBANOVA_API_KEY
|
||||
models: # https://community.sambanova.ai/t/supported-models/193
|
||||
Meta-Llama-3.1-8B-Instruct:
|
||||
aliases: ["llama3.1-8b-sambanova-4k", "llama3.1-instruct-8b-sambanova-4k"]
|
||||
max-input-chars: 12250
|
||||
Meta-Llama-3.1-70B-Instruct:
|
||||
aliases: ["llama3.1-70b-sambanova-4k", "llama3.1-instruct-70b-sambanova-4k"]
|
||||
max-input-chars: 12250
|
||||
Meta-Llama-3.1-405B-Instruct:
|
||||
aliases: ["llama3.1-405b-sambanova-4k", "llama3.1-instruct-405b-sambanova-4k"]
|
||||
max-input-chars: 12250
|
||||
Meta-Llama-3.1-8B-Instruct-8k:
|
||||
aliases: ["llama3.1-8b-sambanova", "llama3.1-instruct-8b-sambanova", "llama3.1-8b-sambanova-8k", "llama3.1-instruct-8b-sambanova-8k"]
|
||||
max-input-chars: 24500
|
||||
Meta-Llama-3.1-70B-Instruct-8k:
|
||||
aliases: ["llama3.1-70b-sambanova", "llama3.1-instruct-70b-sambanova", "llama3.1-70b-sambanova-8k", "llama3.1-instruct-70b-sambanova-8k"]
|
||||
max-input-chars: 24500
|
||||
Meta-Llama-3.1-405B-Instruct-8k:
|
||||
aliases: ["llama3.1-405b-sambanova", "llama3.1-instruct-405b-sambanova", "llama3.1-405b-sambanova-8k", "llama3.1-instruct-405b-sambanova-8k"]
|
||||
max-input-chars: 24500
|
||||
localai:
|
||||
# LocalAI setup instructions: https://github.com/go-skynet/LocalAI#example-use-gpt4all-j-model
|
||||
base-url: http://localhost:8080
|
||||
|
|
Loading…
Reference in a new issue