Skip to content

Help with delay in llama-server #11005

Answered by indie-ai
indie-ai asked this question in Q&A
Dec 28, 2024 · 5 comments · 12 replies
Discussion options

You must be logged in to vote

I recently got back into my ai project and figured I would build the latestest llamacpp. After rebuilding this delay problem was gone. Just wanted to report that the problem was fixed. Just as an FYI I went back and found what commit fixed the problem:

Jan 4 b56f079 TIME=14801
Feb 1 ecef206 TIME=14939
Feb 13 e437627 TIME=15224
Feb 14 38e32eb TIME=15231
Feb 14 dbc2ec5 TIME=15180
Feb 14 94b87f8 TIME=767
Feb 15 fc1b0d0 TIME=1097
Feb 16 c2ea16f TIME=1088
Feb 17 2eea03d TIME=1081
Feb 20 0d55958 TIME=659
Mar 1 80c41dd TIME=1081
Latest TIME=1071

Seems Feb 14 commit 94b87f8 the times are back down to around a second.

So, I guess this can be marked as answered/solved?

Replies: 5 comments 12 replies

Comment options

You must be logged in to vote
1 reply
@ggerganov
Comment options

Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
11 replies
@indie-ai
Comment options

@slaren
Comment options

@indie-ai
Comment options

@indie-ai
Comment options

Answer selected by ggerganov
@ggerganov
Comment options

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
3 participants