mirror of
https://github.com/NovaOSS/nova-api.git
synced 2024-11-25 21:33:58 +01:00
Compare commits
No commits in common. "003a7d3d716d62677281aa808aa9ddfef74af331" and "6f6f15e698ef0b2ae4dcabf0ebcfe373dae36aac" have entirely different histories.
003a7d3d71
...
6f6f15e698
2
.vscode/settings.json
vendored
2
.vscode/settings.json
vendored
|
@ -7,9 +7,9 @@
|
||||||
"**/.DS_Store": true,
|
"**/.DS_Store": true,
|
||||||
"**/Thumbs.db": true,
|
"**/Thumbs.db": true,
|
||||||
"**/__pycache__": true,
|
"**/__pycache__": true,
|
||||||
|
"**/*.css.map": true,
|
||||||
"**/.vscode": true,
|
"**/.vscode": true,
|
||||||
"**/*.map": true,
|
"**/*.map": true,
|
||||||
"**/*.css.map": true,
|
|
||||||
"tests/__pycache__": true
|
"tests/__pycache__": true
|
||||||
},
|
},
|
||||||
"hide-files.files": [
|
"hide-files.files": [
|
||||||
|
|
|
@ -6,9 +6,10 @@ costs:
|
||||||
other: 5
|
other: 5
|
||||||
|
|
||||||
chat-models:
|
chat-models:
|
||||||
gpt-4-32k: 200
|
gpt-4-32k-azure: 100
|
||||||
gpt-4: 50
|
gpt-4: 50
|
||||||
gpt-3: 10
|
gpt-4-azure: 10
|
||||||
|
gpt-3: 5
|
||||||
|
|
||||||
## Roles Explanation
|
## Roles Explanation
|
||||||
|
|
||||||
|
|
|
@ -38,10 +38,10 @@ async def count_for_messages(messages: list, model: str='gpt-3.5-turbo-0613') ->
|
||||||
tokens_per_name = -1 # if there's a name, the role is omitted
|
tokens_per_name = -1 # if there's a name, the role is omitted
|
||||||
|
|
||||||
elif 'gpt-3.5-turbo' in model:
|
elif 'gpt-3.5-turbo' in model:
|
||||||
return await count_for_messages(messages, model='gpt-3.5-turbo-0613')
|
return count_for_messages(messages, model='gpt-3.5-turbo-0613')
|
||||||
|
|
||||||
elif 'gpt-4' in model:
|
elif 'gpt-4' in model:
|
||||||
return await count_for_messages(messages, model='gpt-4-0613')
|
return count_for_messages(messages, model='gpt-4-0613')
|
||||||
|
|
||||||
else:
|
else:
|
||||||
raise NotImplementedError(f"""count_for_messages() is not implemented for model {model}.
|
raise NotImplementedError(f"""count_for_messages() is not implemented for model {model}.
|
||||||
|
|
|
@ -14,7 +14,7 @@ MODELS = [
|
||||||
]
|
]
|
||||||
# MODELS = [f'{model}-azure' for model in MODELS]
|
# MODELS = [f'{model}-azure' for model in MODELS]
|
||||||
|
|
||||||
AZURE_API = '2023-08-01-preview'
|
AZURE_API = '2023-07-01-preview'
|
||||||
|
|
||||||
async def chat_completion(**payload):
|
async def chat_completion(**payload):
|
||||||
key = await utils.random_secret_for('azure-nva1')
|
key = await utils.random_secret_for('azure-nva1')
|
||||||
|
|
|
@ -7,7 +7,6 @@ import aiohttp
|
||||||
import asyncio
|
import asyncio
|
||||||
import starlette
|
import starlette
|
||||||
|
|
||||||
from typing import Any, Coroutine, Set
|
|
||||||
from rich import print
|
from rich import print
|
||||||
from dotenv import load_dotenv
|
from dotenv import load_dotenv
|
||||||
|
|
||||||
|
@ -24,19 +23,6 @@ CRITICAL_API_ERRORS = ['invalid_api_key', 'account_deactivated']
|
||||||
|
|
||||||
keymanager = providerkeys.manager
|
keymanager = providerkeys.manager
|
||||||
|
|
||||||
background_tasks: Set[asyncio.Task[Any]] = set()
|
|
||||||
|
|
||||||
|
|
||||||
def create_background_task(coro: Coroutine[Any, Any, Any]) -> None:
|
|
||||||
"""asyncio.create_task, which prevents the task from being garbage collected.
|
|
||||||
|
|
||||||
https://docs.python.org/3/library/asyncio-task.html#asyncio.create_task
|
|
||||||
"""
|
|
||||||
task = asyncio.create_task(coro)
|
|
||||||
background_tasks.add(task)
|
|
||||||
task.add_done_callback(background_tasks.discard)
|
|
||||||
|
|
||||||
|
|
||||||
async def respond(
|
async def respond(
|
||||||
path: str='/v1/chat/completions',
|
path: str='/v1/chat/completions',
|
||||||
user: dict=None,
|
user: dict=None,
|
||||||
|
@ -63,7 +49,7 @@ async def respond(
|
||||||
'Content-Type': 'application/json'
|
'Content-Type': 'application/json'
|
||||||
}
|
}
|
||||||
|
|
||||||
for i in range(5):
|
for i in range(1):
|
||||||
try:
|
try:
|
||||||
if is_chat:
|
if is_chat:
|
||||||
target_request = await load_balancing.balance_chat_request(payload)
|
target_request = await load_balancing.balance_chat_request(payload)
|
||||||
|
@ -110,7 +96,7 @@ async def respond(
|
||||||
cookies=target_request.get('cookies'),
|
cookies=target_request.get('cookies'),
|
||||||
ssl=False,
|
ssl=False,
|
||||||
timeout=aiohttp.ClientTimeout(
|
timeout=aiohttp.ClientTimeout(
|
||||||
connect=0.75,
|
connect=1.0,
|
||||||
total=float(os.getenv('TRANSFER_TIMEOUT', '500'))
|
total=float(os.getenv('TRANSFER_TIMEOUT', '500'))
|
||||||
)
|
)
|
||||||
) as response:
|
) as response:
|
||||||
|
@ -161,33 +147,23 @@ async def respond(
|
||||||
print('[!] too many requests')
|
print('[!] too many requests')
|
||||||
continue
|
continue
|
||||||
|
|
||||||
chunk_no = 0
|
async for chunk in response.content.iter_any():
|
||||||
buffer = ''
|
chunk = chunk.decode('utf8').strip()
|
||||||
|
|
||||||
async for chunk in response.content.iter_chunked(1024):
|
|
||||||
chunk_no += 1
|
|
||||||
|
|
||||||
chunk = chunk.decode('utf8')
|
|
||||||
|
|
||||||
if 'azure' in provider_name:
|
if 'azure' in provider_name:
|
||||||
chunk = chunk.replace('data: ', '')
|
chunk = chunk.strip().replace('data: ', '')
|
||||||
|
|
||||||
if not chunk or chunk_no == 1:
|
if not chunk or 'prompt_filter_results' in chunk:
|
||||||
continue
|
continue
|
||||||
|
|
||||||
subchunks = chunk.split('\n\n')
|
yield chunk + '\n\n'
|
||||||
buffer += subchunks[0]
|
|
||||||
|
|
||||||
yield buffer + '\n\n'
|
|
||||||
buffer = subchunks[-1]
|
|
||||||
|
|
||||||
for subchunk in subchunks[1:-1]:
|
|
||||||
yield subchunk + '\n\n'
|
|
||||||
|
|
||||||
break
|
break
|
||||||
|
|
||||||
except aiohttp.client_exceptions.ServerTimeoutError:
|
except Exception as exc:
|
||||||
continue
|
print('[!] exception', exc)
|
||||||
|
# continue
|
||||||
|
raise exc
|
||||||
|
|
||||||
else:
|
else:
|
||||||
yield await errors.yield_error(500, 'Sorry, our API seems to have issues connecting to our provider(s).', 'This most likely isn\'t your fault. Please try again later.')
|
yield await errors.yield_error(500, 'Sorry, our API seems to have issues connecting to our provider(s).', 'This most likely isn\'t your fault. Please try again later.')
|
||||||
|
@ -196,7 +172,7 @@ async def respond(
|
||||||
if (not is_stream) and server_json_response:
|
if (not is_stream) and server_json_response:
|
||||||
yield json.dumps(server_json_response)
|
yield json.dumps(server_json_response)
|
||||||
|
|
||||||
create_background_task(
|
asyncio.create_task(
|
||||||
after_request.after_request(
|
after_request.after_request(
|
||||||
incoming_request=incoming_request,
|
incoming_request=incoming_request,
|
||||||
target_request=target_request,
|
target_request=target_request,
|
||||||
|
|
|
@ -216,7 +216,7 @@ async def demo():
|
||||||
raise ConnectionError('API Server is not running.')
|
raise ConnectionError('API Server is not running.')
|
||||||
|
|
||||||
for func in [
|
for func in [
|
||||||
test_chat_non_stream_gpt4,
|
# test_chat_non_stream_gpt4,
|
||||||
test_chat_stream_gpt3
|
test_chat_stream_gpt3
|
||||||
]:
|
]:
|
||||||
print(f'[*] {func.__name__}')
|
print(f'[*] {func.__name__}')
|
||||||
|
|
Loading…
Reference in a new issue