Commit graph

21 commits

Author SHA1 Message Date
Hardik Shah
171a178783 get ollama working 2024-08-07 17:53:15 -07:00
Hardik Shah
ade574a0ef minor fixes 2024-08-06 18:56:34 -07:00
Ashwin Bharambe
a0e61a3c7a Fix passthrough streaming, send headers properly not part of body :facepalm 2024-08-06 16:39:38 -07:00
Hardik Shah
039861f1c7 update inference config to take model and not model_dir 2024-08-06 15:02:47 -07:00
Ashwin Bharambe
0de5a807c7 Make each inference provider into its own subdirectory 2024-08-05 16:39:58 -07:00
Ashwin Bharambe
f64668319c Merge remote-tracking branch 'origin/main' into distros 2024-08-05 14:31:06 -07:00
Ashwin Bharambe
65a9e40174 Adapter -> Provider 2024-08-05 13:42:56 -07:00
Ashwin Bharambe
125fdb1b2a ApiSurface -> Api 2024-08-05 12:46:38 -07:00
Ashwin Bharambe
be19b22391 Bring agentic system api to toolchain
Add adapter dependencies and resolve adapters using a topological sort
2024-08-04 17:33:29 -07:00
Ashwin Bharambe
803976df26 cleanup, moving stuff to common, nuke utils 2024-08-03 20:58:00 -07:00
Ashwin Bharambe
fe582a739d add safety adapters, configuration handling, server + clients 2024-08-03 19:46:59 -07:00
Ashwin Bharambe
9dafa6ad94 implement full-passthrough in the server 2024-08-03 14:15:20 -07:00
Ashwin Bharambe
af4710c959 Improved exception handling 2024-08-02 15:52:15 -07:00
Hardik Shah
493f0d99b2 updated dependency and client model name 2024-08-02 15:37:40 -07:00
Hardik Shah
d7a4cdd70d added options to ollama inference 2024-08-02 14:44:22 -07:00
Ashwin Bharambe
d3e269fcf2 Remove inference uvicorn server entrypoint and llama inference CLI command 2024-08-02 14:18:25 -07:00
Ashwin Bharambe
2cf9915806 Distribution server now functioning 2024-08-02 13:37:40 -07:00
Ashwin Bharambe
041cafbee3 getting closer to a distro definition, distro install + configure works 2024-08-01 23:12:43 -07:00
Ashwin Bharambe
09cf3fe78b Use new definitions of Model / SKU 2024-07-31 22:44:35 -07:00
Hardik Shah
156bfa0e15
Added Ollama as an inference impl (#20)
* fix non-streaming api in inference server

* unit test for inline inference

* Added non-streaming ollama inference impl

* add streaming support for ollama inference with tests

* addressing comments

---------

Co-authored-by: Hardik Shah <hjshah@fb.com>
2024-07-31 22:08:37 -07:00
Ashwin Bharambe
5d5acc8ed5 Initial commit 2024-07-23 08:32:33 -07:00