pycodestyle cleanup

This commit is contained in:
MB7979 2023-08-06 12:24:07 +01:00
parent 334efe987c
commit 08e6dfde35

View File

@ -421,7 +421,6 @@ def do_train(lora_name: str, always_override: bool, save_steps: int, micro_batch
del raw_text # Note: could be a gig for a large dataset, so delete redundant data as we go to be safe on RAM
train_data = Dataset.from_list(out_tokens)
del out_tokens
@ -722,7 +721,7 @@ def split_chunks(arr, size: int, step: int, max_newline_length: int, newline_tok
if split_starts[index] + size > num_tokens:
split_starts[index] = num_tokens - size + 1
if max_newline_length > 0 and newline_token in arr[split_starts[index] : split_starts[index] + max_newline_length]:
if max_newline_length > 0 and newline_token in arr[split_starts[index]:split_starts[index] + max_newline_length]:
first_newline = arr[split_starts[index]: split_starts[index] + max_newline_length].index(newline_token)
split_starts[index] += first_newline
@ -737,6 +736,7 @@ def split_chunks(arr, size: int, step: int, max_newline_length: int, newline_tok
"attention_mask": input_ids.ne(shared.tokenizer.pad_token_id),
}
def format_time(seconds: float):
if seconds < 120:
return f"`{seconds:.0f}` seconds"