mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2024-11-22 16:17:57 +01:00
commit
07d66e45b4
@ -74,8 +74,8 @@ def build_parameters(body, chat=False):
|
|||||||
if str(character) == "None":
|
if str(character) == "None":
|
||||||
character = "Assistant"
|
character = "Assistant"
|
||||||
|
|
||||||
name1, name2, _, greeting, context, _ = load_character_memoized(character, str(body.get('your_name', shared.settings['name1'])), '', instruct=False)
|
name1, name2, _, greeting, context, _, _ = load_character_memoized(character, str(body.get('your_name', shared.settings['name1'])), '', instruct=False)
|
||||||
name1_instruct, name2_instruct, _, _, context_instruct, turn_template = load_character_memoized(instruction_template, '', '', instruct=True)
|
name1_instruct, name2_instruct, _, _, context_instruct, turn_template, _ = load_character_memoized(instruction_template, '', '', instruct=True)
|
||||||
generate_params.update({
|
generate_params.update({
|
||||||
'mode': str(body.get('mode', 'chat')),
|
'mode': str(body.get('mode', 'chat')),
|
||||||
'name1': str(body.get('name1', name1)),
|
'name1': str(body.get('name1', name1)),
|
||||||
|
@ -91,6 +91,10 @@ async def openai_completions(request: Request, request_data: CompletionRequest):
|
|||||||
async with streaming_semaphore:
|
async with streaming_semaphore:
|
||||||
response = OAIcompletions.stream_completions(to_dict(request_data), is_legacy=is_legacy)
|
response = OAIcompletions.stream_completions(to_dict(request_data), is_legacy=is_legacy)
|
||||||
for resp in response:
|
for resp in response:
|
||||||
|
disconnected = await request.is_disconnected()
|
||||||
|
if disconnected:
|
||||||
|
break
|
||||||
|
|
||||||
yield {"data": json.dumps(resp)}
|
yield {"data": json.dumps(resp)}
|
||||||
|
|
||||||
return EventSourceResponse(generator()) # SSE streaming
|
return EventSourceResponse(generator()) # SSE streaming
|
||||||
@ -110,6 +114,10 @@ async def openai_chat_completions(request: Request, request_data: ChatCompletion
|
|||||||
async with streaming_semaphore:
|
async with streaming_semaphore:
|
||||||
response = OAIcompletions.stream_chat_completions(to_dict(request_data), is_legacy=is_legacy)
|
response = OAIcompletions.stream_chat_completions(to_dict(request_data), is_legacy=is_legacy)
|
||||||
for resp in response:
|
for resp in response:
|
||||||
|
disconnected = await request.is_disconnected()
|
||||||
|
if disconnected:
|
||||||
|
break
|
||||||
|
|
||||||
yield {"data": json.dumps(resp)}
|
yield {"data": json.dumps(resp)}
|
||||||
|
|
||||||
return EventSourceResponse(generator()) # SSE streaming
|
return EventSourceResponse(generator()) # SSE streaming
|
||||||
|
Loading…
Reference in New Issue
Block a user