Admin message
为了安全,强烈建议开启2FA双因子认证:User Settings -> Account -> Enable two-factor authentication!!!
Tags
Tags give the ability to mark specific points in history as being important
b2109
e5ca3937
·
llama : do not cap thread count when MoE on CPU (#5419)
·
Feb 09, 2024
b2107
b2f87cb6
·
ggml : fix `error C2078: too many initializers` for MSVC ARM64 (#5404)
·
Feb 09, 2024
b2106
44fbe343
·
Fix Vulkan crash on APUs with very little device memory (#5424)
·
Feb 09, 2024
b2105
8e6a9d2d
·
CUDA: more warps for mmvq on NVIDIA (#5394)
·
Feb 08, 2024
b2104
41f308f5
·
llama : do not print "offloading layers" message in CPU-only builds (#5416)
·
Feb 08, 2024
b2103
6e99f2a0
·
Fix f16_sycl cpy call from Arc (#5411)
·
Feb 08, 2024
b2101
b7b74cef
·
fix trailing whitespace (#5407)
·
Feb 08, 2024
b2100
4aa43fab
·
llama : fix MiniCPM (#5392)
·
Feb 08, 2024
b2098
26d4efd1
·
sampling: fix top_k <= 0 (#5388)
·
Feb 08, 2024
b2096
c4fbb671
·
CMAKE_OSX_ARCHITECTURES for MacOS cross compilation (#5393)
·
Feb 07, 2024
b2094
b906596b
·
Add Ava in the list of llama.cpp UIs (#4362)
·
Feb 07, 2024
b2093
aa7ab99b
·
CUDA: fixed mmvq kernel for bs 2,3,4 and -sm row (#5386)
·
Feb 07, 2024
b2091
0ef46da6
·
llava-cli : always tokenize special tokens (#5382)
·
Feb 07, 2024
b2090
ee1628bd
·
Basic Vulkan Multi-GPU implementation (#5321)
·
Feb 07, 2024
b2087
316c7faf
·
llama : add MiniCPM support (#5346)
·
Feb 07, 2024
b2086
f3e2b4fa
·
server : update `/props` with "total_slots" value (#5373)
·
Feb 07, 2024
b2084
213d1439
·
server : remove model.json endpoint (#5371)
·
Feb 06, 2024
b2083
17c97fb0
·
CUDA: mul_mat_vec_q max. batch size 8 -> 4 (#5370)
·
Feb 06, 2024
b2082
b08f22c8
·
Update README.md (#5366)
·
Feb 06, 2024
b2081
f57fadc0
·
Slight quantization improvement for Q4_K and Q5_K (#5361)
·
Feb 06, 2024
1
…
105
106
107
108
109
110
111
112
113
…
178