fix a tps error
This commit is contained in:
parent
e1c12202aa
commit
b24a18cd3a
@ -191,7 +191,10 @@ async def eval_rwkv(
|
||||
# torch_gc()
|
||||
requests_num = requests_num - 1
|
||||
completion_end_time = time.time()
|
||||
tps = completion_tokens / (completion_end_time - completion_start_time)
|
||||
completion_interval = completion_end_time - completion_start_time
|
||||
tps = 0
|
||||
if completion_interval > 0:
|
||||
tps = completion_tokens / completion_interval
|
||||
print(f"Generation TPS: {tps:.2f}")
|
||||
|
||||
if await request.is_disconnected():
|
||||
|
@ -257,7 +257,10 @@ class AbstractRWKV(ABC):
|
||||
self.fix_tokens(self.pipeline.encode(delta_prompt))
|
||||
)
|
||||
prompt_end_time = time.time()
|
||||
tps = prompt_token_len / (prompt_end_time - prompt_start_time)
|
||||
prompt_interval = prompt_end_time - prompt_start_time
|
||||
tps = 0
|
||||
if prompt_interval > 0:
|
||||
tps = prompt_token_len / prompt_interval
|
||||
print(f"Prompt Prefill TPS: {tps:.2f}", end=" ", flush=True)
|
||||
try:
|
||||
state_cache.add_state(
|
||||
|
Loading…
Reference in New Issue
Block a user