Junrulu commited on
Commit
c4b9343
·
verified ·
1 Parent(s): a13eee4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -19,7 +19,7 @@ instruct_model:
19
  **Youtu-LLM** is a new, small, yet powerful LLM, contains only 1.96B parameters, supports 128k long context, and has native agentic talents. On general evaluations, Youtu-LLM significantly outperforms SOTA LLMs of similar size in terms of Commonsense, STEM, Coding and Long Context capabilities; in agent-related testing, Youtu-LLM surpasses larger-sized leaders and is truly capable of completing multiple end2end agent tasks.
20
 
21
  **Youtu-LLM** has the following features:
22
- - Type: Autoregressive Causal Language Models with Dense [MLA](https://arxiv.org/abs/2405.04434)
23
  - Release versions: [Base](https://huggingface.co/tencent/Youtu-LLM-2B-Base) and [Instruct](https://huggingface.co/tencent/Youtu-LLM-2B)
24
  - Number of Parameters: 1.96B
25
  - Number of Layers: 32
 
19
  **Youtu-LLM** is a new, small, yet powerful LLM, contains only 1.96B parameters, supports 128k long context, and has native agentic talents. On general evaluations, Youtu-LLM significantly outperforms SOTA LLMs of similar size in terms of Commonsense, STEM, Coding and Long Context capabilities; in agent-related testing, Youtu-LLM surpasses larger-sized leaders and is truly capable of completing multiple end2end agent tasks.
20
 
21
  **Youtu-LLM** has the following features:
22
+ - Type: Autoregressive Causal Language Models with Dense MLA
23
  - Release versions: [Base](https://huggingface.co/tencent/Youtu-LLM-2B-Base) and [Instruct](https://huggingface.co/tencent/Youtu-LLM-2B)
24
  - Number of Parameters: 1.96B
25
  - Number of Layers: 32