Skip to content

Commit

Permalink
Lint
Browse files Browse the repository at this point in the history
  • Loading branch information
Smartappli authored Aug 2, 2024
1 parent e7b7fc7 commit e7774f0
Showing 1 changed file with 17 additions and 19 deletions.
36 changes: 17 additions & 19 deletions llama_cpp/server/app.py
Original file line number Diff line number Diff line change
@@ -1,49 +1,47 @@
from __future__ import annotations

import os
import contextlib
import json
import typing
import contextlib

from threading import Lock
from functools import partial
from typing import Iterator, List, Optional, Union, Dict
from threading import Lock

import llama_cpp
from typing import Dict, Iterator, List, Optional, Union

import anyio
from anyio.streams.memory import MemoryObjectSendStream
from starlette.concurrency import run_in_threadpool, iterate_in_threadpool
from fastapi import Depends, FastAPI, APIRouter, Request, HTTPException, status, Body
from fastapi import APIRouter, Body, Depends, FastAPI, HTTPException, Request, status
from fastapi.middleware import Middleware
from fastapi.middleware.cors import CORSMiddleware
from fastapi.security import HTTPBearer
from sse_starlette.sse import EventSourceResponse
from starlette_context.plugins import RequestIdPlugin # type: ignore
from starlette.concurrency import iterate_in_threadpool, run_in_threadpool
from starlette_context.middleware import RawContextMiddleware

from starlette_context.plugins import RequestIdPlugin # type: ignore

import llama_cpp
from llama_cpp.server.errors import RouteErrorHandler
from llama_cpp.server.model import (
LlamaProxy,
)
from llama_cpp.server.settings import (
ConfigFileSettings,
Settings,
ModelSettings,
ServerSettings,
Settings,
)
from llama_cpp.server.types import (
CreateChatCompletionRequest,
CreateCompletionRequest,
CreateEmbeddingRequest,
CreateChatCompletionRequest,
DetokenizeInputRequest,
DetokenizeInputResponse,
ModelList,
TokenizeInputCountResponse,
TokenizeInputRequest,
TokenizeInputResponse,
TokenizeInputCountResponse,
DetokenizeInputRequest,
DetokenizeInputResponse,
)
from llama_cpp.server.errors import RouteErrorHandler


router = APIRouter(route_class=RouteErrorHandler)

Expand Down Expand Up @@ -150,7 +148,7 @@ def create_app(
set_llama_proxy(model_settings=model_settings)

if server_settings.disable_ping_events:
set_ping_message_factory(lambda: bytes())
set_ping_message_factory(lambda: b"")

return app

Expand Down Expand Up @@ -248,7 +246,7 @@ async def authenticate(
"schema": {
"type": "string",
"title": "Server Side Streaming response, when stream=True. "
+ "See SSE format: https://developer.mozilla.org/en-US/docs/Web/API/Server-sent_events/Using_server-sent_events#Event_stream_format", # noqa: E501
+ "See SSE format: https://developer.mozilla.org/en-US/docs/Web/API/Server-sent_events/Using_server-sent_events#Event_stream_format",
"example": """data: {... see CreateCompletionResponse ...} \\n\\n data: ... \\n\\n ... data: [DONE]""",
}
},
Expand Down Expand Up @@ -386,7 +384,7 @@ async def create_embedding(
"schema": {
"type": "string",
"title": "Server Side Streaming response, when stream=True"
+ "See SSE format: https://developer.mozilla.org/en-US/docs/Web/API/Server-sent_events/Using_server-sent_events#Event_stream_format", # noqa: E501
+ "See SSE format: https://developer.mozilla.org/en-US/docs/Web/API/Server-sent_events/Using_server-sent_events#Event_stream_format",
"example": """data: {... see CreateChatCompletionResponse ...} \\n\\n data: ... \\n\\n ... data: [DONE]""",
}
},
Expand Down

0 comments on commit e7774f0

Please sign in to comment.