[2024 Mar 8] llama_kv_cache_seq_rm() returns a bool instead of void, and new llama_n_seq_max() returns the upper limit of acceptable seq_id in batches (relevant when dealing with multiple sequences) ggerganov#5328 [2024 Mar 4] Embeddings API updated ggerganov#5796 [2024 Mar 3] ...
Max. NA's ## 0 0 0 0 0 0 38 Performance comparison Distance by pair Here are the measurements of contraction time and query time (in second) of contraction hierarchies on different graphs : Number of queries Graph Vertices Edges preprocessing algorithm 1000 2000 5000 10000 100000 1000000 ...
[2024 Mar 13] Addllama_synchronize()+llama_context_params.n_ubatchhttps://github.com/ggerganov/llama.cpp/pull/6017 [2024 Mar 8]llama_kv_cache_seq_rm()returns aboolinstead ofvoid, and newllama_n_seq_max()returns the upper limit of acceptableseq_idin batches (relevant when dealing with...
size_t maxelt, // max elts we'll consider bool all) // report all hits? { assert_eq(0, nelt_out); gws_.clear(); rands_.clear(); satpos_.clear(); eehits_.clear(); // First, count up the total number of satpos_, rands_, eehits_, and gws_ // we're goin...
代玉梅目前担任喀左福禄康纸制品有限公司、玉田县梅香服装店法定代表人,同时担任玉田县宇森农业种植有限公司监事,喀左福禄康纸制品有限公司执行董事、经理;二、代玉梅投资情况:代玉梅目前是喀左福禄康纸制品有限公司直接控股股东,持股比例为60%;目前代玉梅投资喀左福禄康纸制品有限公司最终收益股份为60%,投资玉田县梅香服装店...
Nov 4, 2024 src llama : add check for KV cache shifts (ggerganov#10401) Nov 19, 2024 tests vulkan: Optimize soft_max (ggerganov#10301) Nov 19, 2024 .clang-tidy cuda : refactor into multiple files (ggerganov#6269) Mar 25, 2024 .dockerignore ci : fix docker build number and tag ...
[2024 Mar 13] Add llama_synchronize() + llama_context_params.n_ubatch ggml-org#6017 [2024 Mar 8] llama_kv_cache_seq_rm() returns a bool instead of void, and new llama_n_seq_max() returns the upper limit of acceptable seq_id in batches (relevant when dealing with multiple sequence...
[2024 Mar 13] Add llama_synchronize() + llama_context_params.n_ubatch #6017 [2024 Mar 8] llama_kv_cache_seq_rm() returns a bool instead of void, and new llama_n_seq_max() returns the upper limit of acceptable seq_id in batches (relevant when dealing with multiple sequences) #532...
[2024 Mar 13] Add llama_synchronize() + llama_context_params.n_ubatch #6017 [2024 Mar 8] llama_kv_cache_seq_rm() returns a bool instead of void, and new llama_n_seq_max() returns the upper limit of acceptable seq_id in batches (relevant when dealing with multiple sequences) #532...
[2024 Mar 13] Add llama_synchronize() + llama_context_params.n_ubatch ggerganov#6017 [2024 Mar 8] llama_kv_cache_seq_rm() returns a bool instead of void, and new llama_n_seq_max() returns the upper limit of acceptable seq_id in batches (relevant when dealing with multiple sequence...