JJ
J22
AI & ML interests
None yet
Recent Activity
liked
a model
15 days ago
NikolayKozloff/Falcon3-1B-Instruct-chatllm.cpp-q8_0-GGML
liked
a model
15 days ago
NikolayKozloff/Llama-3.1-8B-Instruct-bnb-4bit-shqip-chatllm.cpp-int8-GGML
Organizations
None yet
J22's activity
Upload tokenizer.json
1
#1 opened 2 months ago
by
J22
a horrible function in `modeling_mobilellm.py`
1
#5 opened 2 months ago
by
J22
Run this on CPU
#6 opened 4 months ago
by
J22
Run on CPU
1
#13 opened 4 months ago
by
J22
need gguf
19
#4 opened 5 months ago
by
windkkk
Best practice for tool calling with meta-llama/Meta-Llama-3.1-8B-Instruct
1
#33 opened 6 months ago
by
zzclynn
Run this on CPU and use tool calling
1
#38 opened 6 months ago
by
J22
My alternative quantizations.
5
#5 opened 6 months ago
by
ZeroWw
Tool calling is supported by ChatLLM.cpp
#36 opened 7 months ago
by
J22
can't say hello
1
#9 opened 8 months ago
by
J22
no system message?
8
#14 opened 8 months ago
by
mclassHF2023
"small" is so different from "mini" and "medium"
1
#8 opened 8 months ago
by
J22
how to set context in multi-turn QA?
6
#14 opened 8 months ago
by
J22
clarification on the usage of `short_factor` and `long_factor`?
1
#49 opened 8 months ago
by
J22
Continue the discussion: `long_factor` and `short_factor`
2
#32 opened 9 months ago
by
J22
is the '\n' after `'<|end|>'`?
1
#43 opened 9 months ago
by
J22
Is sliding window used or not?
1
#25 opened 9 months ago
by
J22
`long_factor` is never used?
2
#22 opened 9 months ago
by
J22
generate +6 min, +20GB V-ram
2
#17 opened 9 months ago
by
NickyNicky
`sliding_window` is larger than `max_position_embeddings`
1
#21 opened 9 months ago
by
J22