Update README.md
Browse files
README.md
CHANGED
|
@@ -19,7 +19,7 @@ instruct_model:
|
|
| 19 |
**Youtu-LLM** is a new, small, yet powerful LLM, contains only 1.96B parameters, supports 128k long context, and has native agentic talents. On general evaluations, Youtu-LLM significantly outperforms SOTA LLMs of similar size in terms of Commonsense, STEM, Coding and Long Context capabilities; in agent-related testing, Youtu-LLM surpasses larger-sized leaders and is truly capable of completing multiple end2end agent tasks.
|
| 20 |
|
| 21 |
**Youtu-LLM** has the following features:
|
| 22 |
-
- Type: Autoregressive Causal Language Models with Dense
|
| 23 |
- Release versions: [Base](https://huggingface.co/tencent/Youtu-LLM-2B-Base) and [Instruct](https://huggingface.co/tencent/Youtu-LLM-2B)
|
| 24 |
- Number of Parameters: 1.96B
|
| 25 |
- Number of Layers: 32
|
|
|
|
| 19 |
**Youtu-LLM** is a new, small, yet powerful LLM, contains only 1.96B parameters, supports 128k long context, and has native agentic talents. On general evaluations, Youtu-LLM significantly outperforms SOTA LLMs of similar size in terms of Commonsense, STEM, Coding and Long Context capabilities; in agent-related testing, Youtu-LLM surpasses larger-sized leaders and is truly capable of completing multiple end2end agent tasks.
|
| 20 |
|
| 21 |
**Youtu-LLM** has the following features:
|
| 22 |
+
- Type: Autoregressive Causal Language Models with Dense MLA
|
| 23 |
- Release versions: [Base](https://huggingface.co/tencent/Youtu-LLM-2B-Base) and [Instruct](https://huggingface.co/tencent/Youtu-LLM-2B)
|
| 24 |
- Number of Parameters: 1.96B
|
| 25 |
- Number of Layers: 32
|