* [1/n] migrate inference/chat_completion
* migrate inference/completion
* inference/completion
* inference regenerate openapi spec
* safety api
* migrate agentic system
* migrate apis without implementations
* re-generate openapi spec
* remove hack from openapi generator
* fix inference
* fix inference
* openapi generator rerun
* Simplified Telemetry API and tying it to logger (#57)
* Simplified Telemetry API and tying it to logger
* small update which adds a METRIC type
* move span events one level down into structured log events
---------
Co-authored-by: Ashwin Bharambe <ashwin@meta.com>
* fix api to work with openapi generator
* fix agentic calling inference
* together adapter inference
* update inference adapters
---------
Co-authored-by: Ashwin Bharambe <ashwin.bharambe@gmail.com>
Co-authored-by: Ashwin Bharambe <ashwin@meta.com>
* Use huggingface_hub inference client for TGI inference
* Update the default value for TGI URL
* Use InferenceClient.text_generation for TGI inference
* Fixes post-review and split TGI adapter into local and Inference Endpoints ones
* Update CLI reference and add typing
* Rename TGI Adapter class
* Use HfApi to get the namespace when not provide in the hf endpoint name
* Remove unecessary method argument
* Improve TGI adapter initialization condition
* Move helper into impl file + fix merging conflicts
* TGI adapter and some refactoring of other inference adapters
* Use the lower-level `generate_stream()` method for correct tool calling
---------
Co-authored-by: Ashwin Bharambe <ashwin@meta.com>