🦙
🦙
I like big .vimrc and I cannot lie
- Sofia, Bulgaria
-
17:23
(UTC +02:00) - https://ggerganov.com
- @ggerganov
Sponsors
Block or Report
Block or report ggerganov
Report abuse
Contact GitHub support about this user’s behavior. Learn more about reporting abuse.
Report abusePinned
2,947 contributions in the last year
| Day of Week | December Dec | January Jan | February Feb | March Mar | April Apr | May May | June Jun | July Jul | August Aug | September Sep | October Oct | November Nov | December Dec | ||||||||||||||||||||||||||||||||||||||||
| Sunday Sun | |||||||||||||||||||||||||||||||||||||||||||||||||||||
| Monday Mon | |||||||||||||||||||||||||||||||||||||||||||||||||||||
| Tuesday Tue | |||||||||||||||||||||||||||||||||||||||||||||||||||||
| Wednesday Wed | |||||||||||||||||||||||||||||||||||||||||||||||||||||
| Thursday Thu | |||||||||||||||||||||||||||||||||||||||||||||||||||||
| Friday Fri | |||||||||||||||||||||||||||||||||||||||||||||||||||||
| Saturday Sat | |||||||||||||||||||||||||||||||||||||||||||||||||||||
Less
No contributions.
Low contributions.
Medium-low contributions.
Medium-high contributions.
High contributions.
More
Contribution activity
December 2023
Created 33 commits in 5 repositories
Created 1 repository
-
ggerganov/seamless_communication
C
This contribution was made on Dec 2
Created a pull request in ggerganov/llama.cpp that received 18 comments
llama : support quantum K cache
This PR adds support for quantizing the K cache using any of the existing quantization formats. It is applied on top of #4309 which allows per-laye…
+552
−74
lines changed
•
18
comments
Opened 12 other pull requests in 4 repositories
ggerganov/llama.cpp
7
merged
-
sync : ggml (new ops, tests, backend, etc.)
This contribution was made on Dec 7
-
server : recognize cache_prompt parameter in OAI API
This contribution was made on Dec 6
-
llama : per-layer KV cache
This contribution was made on Dec 3
-
ggml : reuse ggml_get_n_tasks() in ggml_graph_plan()
This contribution was made on Dec 3
-
ggml : fix soft max out-of-bounds access
This contribution was made on Dec 3
-
llama : fix integer overflow during quantization
This contribution was made on Dec 1
-
llama : pad KV cache size
This contribution was made on Dec 1
ggerganov/whisper.cpp
2
merged
-
sync : ggml (new ops, new backend, etc)
This contribution was made on Dec 7
-
ggml : re-enable blas for src0 != F32
This contribution was made on Dec 1
ggerganov/ggml
2
merged
-
sync : llama.cpp (fused soft max, gpu cpy ops, etc.)
This contribution was made on Dec 7
-
metal : check supported ops at runtime
This contribution was made on Dec 5
facebookresearch/seamless_communication
1
merged
-
ggml : update cmake to build on macOS
This contribution was made on Dec 2
Reviewed 26 pull requests in 3 repositories
ggerganov/llama.cpp
15 pull requests
-
build : detect host compiler and cuda compiler separately
This contribution was made on Dec 13
-
llama : add Mixtral support
This contribution was made on Dec 13
-
Adding Tests for GGUFWriter Class
This contribution was made on Dec 12
-
Windows XP: support MinGW 8.1.0
This contribution was made on Dec 12
-
Use
typosto fix comments and logs.This contribution was made on Dec 12 -
Check the full vocab for grammar only if necessary
This contribution was made on Dec 6
-
Samplers order parameters
This contribution was made on Dec 5
-
Create api_like_OAI.sh
This contribution was made on Dec 5
-
feat: Allow overriding GGUF metadata when loading model
This contribution was made on Dec 5
-
llama : per-layer KV cache
This contribution was made on Dec 3
-
build : enable libstdc++ assertions for debug builds
This contribution was made on Dec 1
-
Support attention_bias on LLaMA architecture
This contribution was made on Dec 1
-
fix get_num_physical_cores()
This contribution was made on Dec 1
-
Merge qwen to llama cpp
This contribution was made on Dec 1
-
llama : sanity checks for access to logits
This contribution was made on Dec 1
ggerganov/ggml
10 pull requests
-
ggml: new gpu kernels + extends ggml_leaky_relu + ggml_pad
This contribution was made on Dec 13
-
ggml : add support for multiple ids rows in ggml_mat_mul_id
This contribution was made on Dec 9
-
sync : llama.cpp (fused soft max, gpu cpy ops, etc.)
This contribution was made on Dec 7
-
ggml-backend : remove backend self-registration
This contribution was made on Dec 7
-
Feature: Circular Padding
This contribution was made on Dec 7
-
add more int ops
This contribution was made on Dec 7
-
test-backend-ops : add performance eval mode + improve CUDA repeat and binary broadcast ops performance
This contribution was made on Dec 6
-
test-backend-ops : initialize ggml_argsort test with unique values
This contribution was made on Dec 5
-
ggml : full broadcast in mul, add, div + ggml_mul_mat_id, ggml_argsort, ggml_top_k
This contribution was made on Dec 5
-
ggml-cuda : fix usage without CUDA devices
This contribution was made on Dec 1
ggerganov/imtui
1 pull request
-
Minor rewrite - add support for drawing other things on the terminal while ImTui is active.
This contribution was made on Dec 3
Opened 1 issue in 1 repository
ggerganov/llama.cpp
1
open
-
llama : benchmark for Apple Silicon A-series mobile chips
This contribution was made on Dec 7
Answered 1 discussion in 1 repository
ggerganov/llama.cpp
ggerganov/llama.cpp
-
Where are the architectures implemented?
This contribution was made on Dec 6






