Junyang Lin
JustinLin610
AI & ML interests
Pretraining, NLP, CV, etc.
Organizations
Update README.md
#5 opened 4 months ago
by
naykun
Update README.md
#1 opened 4 months ago
by
naykun
fix: use tp 8 for SGLang
#1 opened 7 months ago
by
zhyncs
复杂推理进入死循环
30
#21 opened 9 months ago
by
frankgxy
Independent evaluation results
2
#1 opened about 1 year ago
by
yaronr
Have you deleted your GitHub page?
7
#10 opened over 1 year ago
by
xwzy6
The sample code could not run...
1
#16 opened over 1 year ago
by
zhiminy
fine-tuning
4
#16 opened over 1 year ago
by
SaghirAya
Maybe a silly question...
2
#18 opened over 1 year ago
by
urtuuuu
This model is Awesome
5
#20 opened over 1 year ago
by
areumtecnologia
Update tokenizer_config.json
#3 opened over 1 year ago
by
JustinLin610
请问这个版本GPU内存消耗28G与14B对比如何?
7
#7 opened over 1 year ago
by
william0014
Fine tuning this model with Proprietary Code
2
#6 opened over 1 year ago
by
vtraghu
What are the diffences of this with Qwen/CodeQwen1.5-7B
6
#5 opened over 1 year ago
by
Kalemnor
Adding Evaluation Results
#14 opened over 1 year ago
by
leaderboard-pr-bot
qwen1.5-7b-chat是不是推理起来比qwen1.5-7b快很多
3
#9 opened over 1 year ago
by
endNone
tie_word_embeddings=true ?
1
#6 opened over 1 year ago
by
salmitta
Why 72B model has different vocab size comparing with other models?
7
#1 opened almost 2 years ago
by
Mikasaka
Using llama.cpp server, responses always end with <|im_end|>
1
#2 opened over 1 year ago
by
gilankpam