llama-stack-mirror/llama_stack
Marut Pandya e2b5456e48
Add Runpod Provider + Distribution (#362)
Add Runpod as a inference provider for openAI compatible managed
endpoints.

Testing 
- Configured llama stack from scratch, set `remote::runpod` as a
inference provider.
- Added Runpod Endpoint URL and API key. 
- Started llama-stack server - llama stack run my-local-stack --port
3000
```
curl http://localhost:5000/inference/chat_completion \
-H "Content-Type: application/json" \
-d '{
	"model": "Llama3.1-8B-Instruct",
	"messages": [
		{"role": "system", "content": "You are a helpful assistant."},
		{"role": "user", "content": "Write me a 2 sentence poem about the moon"}
	],
	"sampling_params": {"temperature": 0.7, "seed": 42, "max_tokens": 512}
}' ```

---------

Signed-off-by: pandyamarut <pandyamarut@gmail.com>
2025-01-23 12:19:02 -08:00
..
apis [inference api] modify content types so they follow a more standard structure (#841) 2025-01-22 12:16:18 -08:00
cli More generic image type for OCI-compliant container technologies (#802) 2025-01-17 16:37:42 -08:00
distribution Update OpenAPI generator to output discriminator (#848) 2025-01-22 22:15:23 -08:00
providers Add Runpod Provider + Distribution (#362) 2025-01-23 12:19:02 -08:00
scripts [memory refactor][3/n] Introduce RAGToolRuntime as a specialized sub-protocol (#832) 2025-01-22 10:04:16 -08:00
templates Kill colons 2025-01-22 22:54:13 -08:00
__init__.py export LibraryClient 2024-12-13 12:08:00 -08:00