Merge pull request #4541 from oobabooga/dev

Merge dev branch
This commit is contained in:
oobabooga 2023-11-09 14:53:34 -03:00 committed by GitHub
commit 07d66e45b4
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
2 changed files with 10 additions and 2 deletions

View File

@ -74,8 +74,8 @@ def build_parameters(body, chat=False):
if str(character) == "None": if str(character) == "None":
character = "Assistant" character = "Assistant"
name1, name2, _, greeting, context, _ = load_character_memoized(character, str(body.get('your_name', shared.settings['name1'])), '', instruct=False) name1, name2, _, greeting, context, _, _ = load_character_memoized(character, str(body.get('your_name', shared.settings['name1'])), '', instruct=False)
name1_instruct, name2_instruct, _, _, context_instruct, turn_template = load_character_memoized(instruction_template, '', '', instruct=True) name1_instruct, name2_instruct, _, _, context_instruct, turn_template, _ = load_character_memoized(instruction_template, '', '', instruct=True)
generate_params.update({ generate_params.update({
'mode': str(body.get('mode', 'chat')), 'mode': str(body.get('mode', 'chat')),
'name1': str(body.get('name1', name1)), 'name1': str(body.get('name1', name1)),

View File

@ -91,6 +91,10 @@ async def openai_completions(request: Request, request_data: CompletionRequest):
async with streaming_semaphore: async with streaming_semaphore:
response = OAIcompletions.stream_completions(to_dict(request_data), is_legacy=is_legacy) response = OAIcompletions.stream_completions(to_dict(request_data), is_legacy=is_legacy)
for resp in response: for resp in response:
disconnected = await request.is_disconnected()
if disconnected:
break
yield {"data": json.dumps(resp)} yield {"data": json.dumps(resp)}
return EventSourceResponse(generator()) # SSE streaming return EventSourceResponse(generator()) # SSE streaming
@ -110,6 +114,10 @@ async def openai_chat_completions(request: Request, request_data: ChatCompletion
async with streaming_semaphore: async with streaming_semaphore:
response = OAIcompletions.stream_chat_completions(to_dict(request_data), is_legacy=is_legacy) response = OAIcompletions.stream_chat_completions(to_dict(request_data), is_legacy=is_legacy)
for resp in response: for resp in response:
disconnected = await request.is_disconnected()
if disconnected:
break
yield {"data": json.dumps(resp)} yield {"data": json.dumps(resp)}
return EventSourceResponse(generator()) # SSE streaming return EventSourceResponse(generator()) # SSE streaming