Skip to content

Commit 312e20b

Browse files
committed
openai: update after merge
typos
1 parent 7675ac6 commit 312e20b

File tree

4 files changed

+16
-15
lines changed

4 files changed

+16
-15
lines changed

examples/openai/api.py

+2-2
Original file line numberDiff line numberDiff line change
@@ -29,7 +29,7 @@ class Tool(BaseModel):
2929

3030
class ResponseFormat(BaseModel):
3131
type: Literal["json_object"]
32-
schema: Optional[Json[Any]] = None # type: ignore
32+
schema: Optional[dict[str, Any]] = None # type: ignore
3333

3434
class LlamaCppParams(BaseModel):
3535
n_predict: Optional[int] = None
@@ -67,7 +67,7 @@ class ChatCompletionRequest(LlamaCppParams):
6767
class Choice(BaseModel):
6868
index: int
6969
message: Message
70-
logprobs: Optional[Json[Any]] = None
70+
logprobs: Optional[dict[str, Any]] = None
7171
finish_reason: Union[Literal["stop"], Literal["tool_calls"]]
7272

7373
class Usage(BaseModel):

examples/openai/gguf_kvs.py

+4-5
Original file line numberDiff line numberDiff line change
@@ -8,13 +8,12 @@
88

99
class GGUFKeyValues:
1010
def __init__(self, model: Path):
11-
reader = GGUFReader(model.as_posix())
12-
self.fields = reader.fields
11+
self.reader = GGUFReader(model.as_posix())
1312
def __getitem__(self, key: str):
1413
if '{arch}' in key:
1514
key = key.replace('{arch}', self[Keys.General.ARCHITECTURE])
16-
return self.fields[key].read()
15+
return self.reader.read_field(self.reader.fields[key])
1716
def __contains__(self, key: str):
18-
return key in self.fields
17+
return key in self.reader.fields
1918
def keys(self):
20-
return self.fields.keys()
19+
return self.reader.fields.keys()

examples/openai/prompting.py

+2-2
Original file line numberDiff line numberDiff line change
@@ -180,7 +180,7 @@ def raw_render(self, messages: list[Message], add_generation_prompt: bool, omit_
180180

181181
class ChatHandlerArgs(BaseModel):
182182
chat_template: ChatTemplate
183-
response_schema: Optional[Json[Any]] = None
183+
response_schema: Optional[dict[str,Any]] = None
184184
tools: Optional[list[Tool]] = None
185185

186186
class ChatHandler(ABC):
@@ -719,7 +719,7 @@ def get_chat_handler(args: ChatHandlerArgs, parallel_calls: bool, tool_style: Op
719719
raise ValueError(f"Unsupported tool call style: {tool_style}")
720720

721721
# os.environ.get('NO_TS')
722-
def _please_respond_with_schema(schema: Json[Any]) -> str:
722+
def _please_respond_with_schema(schema: dict[str, Any]) -> str:
723723
sig = json.dumps(schema, indent=2)
724724
# _ts_converter = SchemaToTypeScriptConverter()
725725
# # _ts_converter.resolve_refs(schema, 'schema')

examples/openai/server.py

+8-6
Original file line numberDiff line numberDiff line change
@@ -37,12 +37,8 @@ def main(
3737
):
3838
import uvicorn
3939

40-
if endpoint:
41-
sys.stderr.write(f"# WARNING: Unsure which model we're talking to, fetching its chat template from HuggingFace tokenizer of {template_hf_model_id_fallback}\n")
42-
assert template_hf_model_id_fallback, "template_hf_model_id_fallback is required when using an endpoint"
43-
chat_template = ChatTemplate.from_huggingface(template_hf_model_id_fallback)
44-
45-
else:
40+
chat_template = None
41+
if model:
4642
metadata = GGUFKeyValues(Path(model))
4743

4844
if not context_length:
@@ -58,6 +54,12 @@ def main(
5854
if verbose:
5955
sys.stderr.write(f"# CHAT TEMPLATE:\n\n{chat_template}\n\n")
6056

57+
if not chat_template:
58+
sys.stderr.write(f"# WARNING: Unsure which model we're talking to, fetching its chat template from HuggingFace tokenizer of {template_hf_model_id_fallback}\n")
59+
assert template_hf_model_id_fallback or chat_template, "template_hf_model_id_fallback is required when using an endpoint without a model"
60+
chat_template = ChatTemplate.from_huggingface(template_hf_model_id_fallback)
61+
62+
if not endpoint:
6163
if verbose:
6264
sys.stderr.write(f"# Starting C++ server with model {model} on {server_host}:{server_port}\n")
6365
cmd = [

0 commit comments

Comments
 (0)