remove cuda from petals

This commit is contained in:
ishaan-jaff 2023-09-20 09:23:23 -07:00
parent 102fd83ea6
commit b81f8d2ddd
2 changed files with 2 additions and 3 deletions

View file

@ -39,7 +39,6 @@ def completion(
tokenizer = AutoTokenizer.from_pretrained(model, use_fast=False, add_bos_token=False)
model_obj = AutoDistributedModelForCausalLM.from_pretrained(model)
model_obj = model_obj.cuda()
prompt = ""
for message in messages:
@ -64,7 +63,7 @@ def completion(
)
## COMPLETION CALL
inputs = tokenizer(prompt, return_tensors="pt")["input_ids"].cuda()
inputs = tokenizer(prompt, return_tensors="pt")["input_ids"]
# optional params: max_new_tokens=1,temperature=0.9, top_p=0.6
outputs = model_obj.generate(inputs, **optional_params)

View file

@ -1,6 +1,6 @@
[tool.poetry]
name = "litellm"
version = "0.1.714"
version = "0.1.715"
description = "Library to easily interface with LLM API providers"
authors = ["BerriAI"]
license = "MIT License"