mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-29 15:23:51 +00:00
fixed comments
This commit is contained in:
parent
b9269a94b9
commit
d5034ed759
4 changed files with 3 additions and 4 deletions
|
@ -16,7 +16,7 @@ Remote vLLM inference provider for connecting to vLLM servers.
|
||||||
## Sample Configuration
|
## Sample Configuration
|
||||||
|
|
||||||
```yaml
|
```yaml
|
||||||
url: ${env.VLLM_URL:=http://localhost:8000/v1}
|
url: ${env.VLLM_URL}
|
||||||
max_tokens: ${env.VLLM_MAX_TOKENS:=4096}
|
max_tokens: ${env.VLLM_MAX_TOKENS:=4096}
|
||||||
api_token: ${env.VLLM_API_TOKEN:=fake}
|
api_token: ${env.VLLM_API_TOKEN:=fake}
|
||||||
tls_verify: ${env.VLLM_TLS_VERIFY:=true}
|
tls_verify: ${env.VLLM_TLS_VERIFY:=true}
|
||||||
|
|
|
@ -96,7 +96,6 @@ class FireworksInferenceAdapter(ModelRegistryHelper, Inference, NeedsRequestProv
|
||||||
|
|
||||||
def _get_client(self) -> Fireworks:
|
def _get_client(self) -> Fireworks:
|
||||||
fireworks_api_key = self._get_api_key()
|
fireworks_api_key = self._get_api_key()
|
||||||
print(f">>>>>> fireworks_api_key: {fireworks_api_key} <<<<<")
|
|
||||||
return Fireworks(api_key=fireworks_api_key)
|
return Fireworks(api_key=fireworks_api_key)
|
||||||
|
|
||||||
def _get_openai_client(self) -> AsyncOpenAI:
|
def _get_openai_client(self) -> AsyncOpenAI:
|
||||||
|
|
|
@ -46,7 +46,7 @@ class VLLMInferenceAdapterConfig(BaseModel):
|
||||||
@classmethod
|
@classmethod
|
||||||
def sample_run_config(
|
def sample_run_config(
|
||||||
cls,
|
cls,
|
||||||
url: str = "${env.VLLM_URL:=http://localhost:8000/v1}",
|
url: str = "${env.VLLM_URL}",
|
||||||
**kwargs,
|
**kwargs,
|
||||||
):
|
):
|
||||||
return {
|
return {
|
||||||
|
|
|
@ -26,7 +26,7 @@ providers:
|
||||||
- provider_id: ${env.ENABLE_VLLM:=__disabled__}
|
- provider_id: ${env.ENABLE_VLLM:=__disabled__}
|
||||||
provider_type: remote::vllm
|
provider_type: remote::vllm
|
||||||
config:
|
config:
|
||||||
url: ${env.VLLM_URL:=http://localhost:8000/v1}
|
url: ${env.VLLM_URL}
|
||||||
max_tokens: ${env.VLLM_MAX_TOKENS:=4096}
|
max_tokens: ${env.VLLM_MAX_TOKENS:=4096}
|
||||||
api_token: ${env.VLLM_API_TOKEN:=fake}
|
api_token: ${env.VLLM_API_TOKEN:=fake}
|
||||||
tls_verify: ${env.VLLM_TLS_VERIFY:=true}
|
tls_verify: ${env.VLLM_TLS_VERIFY:=true}
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue