Admin message
为了安全,强烈建议开启2FA双因子认证:User Settings -> Account -> Enable two-factor authentication!!!
Tags
Tags give the ability to mark specific points in history as being important
b2243
201294ae
·
nix: init singularity and docker images (#5056)
·
Feb 22, 2024
b2241
373ee3fb
·
Add Gemma chat template (#5665)
·
Feb 22, 2024
b2240
4cb4d8b2
·
workflows: nix: hardcode cachix ids, build unconditionally (#5663)
·
Feb 22, 2024
b2239
3a03541c
·
minor : fix trailing whitespace (#5638)
·
Feb 22, 2024
b2237
a46f5074
·
server : fallback to chatml, add AlphaMonarch chat template (#5628)
·
Feb 22, 2024
b2235
4ef245a9
·
mpt : add optional bias tensors (#5638)
·
Feb 22, 2024
b2234
973053d8
·
llama : fix loading models with shared tok_embd and output (#5651)
·
Feb 22, 2024
b2233
7c8bcc11
·
Add docs for llama_chat_apply_template (#5645)
·
Feb 22, 2024
b2232
7fe4678b
·
llama : fix session save/load with quantized KV (#5649)
·
Feb 21, 2024
b2231
ba2135cc
·
gemma : allow offloading the output tensor (#5646)
·
Feb 21, 2024
b2230
89febfed
·
examples : do not assume BOS when shifting context (#5622)
·
Feb 21, 2024
b2228
1ecea255
·
server: health: fix race condition on slots data using tasks queue (#5634)
·
Feb 21, 2024
b2226
eccd7a26
·
sync : ggml (#5633)
·
Feb 21, 2024
b2223
580111d4
·
llama : add `gemma` model (#5631)
·
Feb 21, 2024
b2222
88c46cbd
·
[SYCL] conext add name (#5624)
·
Feb 21, 2024
b2221
a14679cc
·
IQ4_NL: 4-bit non-linear quants with blocks of 32 (#5590)
·
Feb 21, 2024
b2220
6560bed3
·
server : support llava 1.6 (#5553)
·
Feb 20, 2024
b2219
06bf2cf8
·
make : fix debug build with CUDA (#5616)
·
Feb 20, 2024
b2217
9c405c9f
·
Server: use llama_chat_apply_template (#5593)
·
Feb 20, 2024
b2215
8dbbd757
·
metal : add build system support for embedded metal library (#5604)
·
Feb 20, 2024
1
…
101
102
103
104
105
106
107
108
109
…
178