From e7774f04d1ce7a1d726d8c0192833b3acbffdd3c Mon Sep 17 00:00:00 2001 From: Olivier DEBAUCHE Date: Fri, 2 Aug 2024 18:21:47 +0200 Subject: [PATCH] Lint --- llama_cpp/server/app.py | 36 +++++++++++++++++------------------- 1 file changed, 17 insertions(+), 19 deletions(-) diff --git a/llama_cpp/server/app.py b/llama_cpp/server/app.py index cd3255176..d53ceaf9c 100644 --- a/llama_cpp/server/app.py +++ b/llama_cpp/server/app.py @@ -1,49 +1,47 @@ from __future__ import annotations import os +import contextlib import json import typing -import contextlib - -from threading import Lock from functools import partial -from typing import Iterator, List, Optional, Union, Dict +from threading import Lock -import llama_cpp +from typing import Dict, Iterator, List, Optional, Union import anyio from anyio.streams.memory import MemoryObjectSendStream -from starlette.concurrency import run_in_threadpool, iterate_in_threadpool -from fastapi import Depends, FastAPI, APIRouter, Request, HTTPException, status, Body +from fastapi import APIRouter, Body, Depends, FastAPI, HTTPException, Request, status from fastapi.middleware import Middleware from fastapi.middleware.cors import CORSMiddleware from fastapi.security import HTTPBearer from sse_starlette.sse import EventSourceResponse -from starlette_context.plugins import RequestIdPlugin # type: ignore +from starlette.concurrency import iterate_in_threadpool, run_in_threadpool from starlette_context.middleware import RawContextMiddleware - +from starlette_context.plugins import RequestIdPlugin # type: ignore + +import llama_cpp +from llama_cpp.server.errors import RouteErrorHandler from llama_cpp.server.model import ( LlamaProxy, ) from llama_cpp.server.settings import ( ConfigFileSettings, - Settings, ModelSettings, ServerSettings, + Settings, ) from llama_cpp.server.types import ( + CreateChatCompletionRequest, CreateCompletionRequest, CreateEmbeddingRequest, - CreateChatCompletionRequest, + DetokenizeInputRequest, + DetokenizeInputResponse, ModelList, + TokenizeInputCountResponse, TokenizeInputRequest, TokenizeInputResponse, - TokenizeInputCountResponse, - DetokenizeInputRequest, - DetokenizeInputResponse, ) -from llama_cpp.server.errors import RouteErrorHandler - router = APIRouter(route_class=RouteErrorHandler) @@ -150,7 +148,7 @@ def create_app( set_llama_proxy(model_settings=model_settings) if server_settings.disable_ping_events: - set_ping_message_factory(lambda: bytes()) + set_ping_message_factory(lambda: b"") return app @@ -248,7 +246,7 @@ async def authenticate( "schema": { "type": "string", "title": "Server Side Streaming response, when stream=True. " - + "See SSE format: https://developer.mozilla.org/en-US/docs/Web/API/Server-sent_events/Using_server-sent_events#Event_stream_format", # noqa: E501 + + "See SSE format: https://developer.mozilla.org/en-US/docs/Web/API/Server-sent_events/Using_server-sent_events#Event_stream_format", "example": """data: {... see CreateCompletionResponse ...} \\n\\n data: ... \\n\\n ... data: [DONE]""", } }, @@ -386,7 +384,7 @@ async def create_embedding( "schema": { "type": "string", "title": "Server Side Streaming response, when stream=True" - + "See SSE format: https://developer.mozilla.org/en-US/docs/Web/API/Server-sent_events/Using_server-sent_events#Event_stream_format", # noqa: E501 + + "See SSE format: https://developer.mozilla.org/en-US/docs/Web/API/Server-sent_events/Using_server-sent_events#Event_stream_format", "example": """data: {... see CreateChatCompletionResponse ...} \\n\\n data: ... \\n\\n ... data: [DONE]""", } },