You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The current implementation only supports the Decode function, which is a bit behind popular repos, e.g., llama.cpp.
The Prefill stage should be implemented with batched computation (GEMM) and adopt the mature communication backend e.g., OpenMPI.
This should facilitate the whole process because computing the entire prompts should be faster than decoding one by one.
The text was updated successfully, but these errors were encountered:
The current implementation only supports the Decode function, which is a bit behind popular repos, e.g.,
llama.cpp
.The Prefill stage should be implemented with batched computation (GEMM) and adopt the mature communication backend e.g., OpenMPI.
This should facilitate the whole process because computing the entire prompts should be faster than decoding one by one.
The text was updated successfully, but these errors were encountered: