•  
  •  
 

Bulletin of Chinese Academy of Sciences (Chinese Version)

Keywords

multimodal; Hunyuan large model; deep learning framework

Document Type

Give Full Play to Main Role of Enterprises in S&T Innovation

Abstract

The article discusses the emerging trends and application prospects of current large models, using Tencent’s Hunyuan large model as an example. It focuses mainly on innovations and implementations of large models in China. Companies like Google, Meta, and OpenAI have launched powerful models such as Google’s Gemini and Meta’s Llama 3, which have made significant progress in multi-modal applications and reasoning capabilities. China’s large models have significantly improved performance and efficiency by adopting the MoE (Mixture of Experts) architecture. Specifically, with its self-developed MoE trillion-parameter large model and deep learning framework, Tencent has made breakthrough advancements in large model technology and achieved exceptional performance in muti-modal applications. Moreover, Tencent has launched a one-stop AI agent creation and distribution platform. Tencent understands that industry-wide, large models are key to implementing AI and strategies, and is actively supporting the application of large models across sectors such as retail, education, finance, healthcare, media, transportation, and government, helping these industries enhance quality and efficiency.

First page

1631

Last Page

1638

Language

Chinese

Publisher

Bulletin of Chinese Academy of Sciences

References

1 Naveed H, Khan A U, Qiu S, et al. A comprehensive overview of large language models. ArXiv, 2023, doi: 10.48550/arXiv.2307.06435.

2 Makridakis S, Petropoulos F, Kang Y F. Large language models: Success and impact. forecasting, 2023, 5(3): 536-549.

3 The 13 biggest AI stories of 2023. HAI Stanford. (2023-12-14)[2024-09-02]. https://hai.stanford.edu/news/13-biggest-ai-stories-2023.

4 Zhang Z, Liu Y, et al. A bibliometric review of large language model research from 2017 to 2023. arXiv, 2023, doi: 10.48550/arXiv.2304.02020.

5 Taulli T. Large language models. (2023-06-03)[2024-09-02]. https://link.springer.com/chapter/10.1007/978-1-4842-9367-6_5.

6 Anil R, Borgeaud S, Alayrac J B, et al. Gemini: A family of highly capable multimodal models. arXiv, 2023, doi: 10.48550/arXiv.2312.11805.

7 Georgiev P, Lei V I, Burnell R, et al. Gemini 1.5: Unlocking multimodal understanding across millions of tokens of context. arXiv, 2024, doi: 10.48550/arXiv.2403.05530.

8 Akter S N , Yu Z C, Muhamed A, et al. An in-depth look at Gemini’s language abilities. arXiv, 2023, doi: 10.48550/arXiv.2312.11444.

9 TeamGemini, Google. Gemini ultra benchmark comparison. (2023-12-06)[2024-09-02]. https://assets.bwbx.io/documents/users/iqjWHBFdfxIU/r7G7RrtT6rnM/v0.

10 Google’s Gemini: Setting new benchmarks in language models. (2023-08-15)[2024-09-02]. https://www.superannotate.com/blog/googles-gemini-setting-new-benchmarks-in-language-models.

11 赵鑫. 大语言模型(LLM)综述与实用指南. (2023-04-20)[2024-09-02]. https://arthurchiao.art/posts/2023/llm-guide/. Zhao X. A comprehensive review and practical guide to large language models (LLMs). (in Chinese)

12 以ChatGPT为代表的大型语言模型研究进展. (2023-05-15)[2024-09-02]. https://www.nsfc.gov.cn/csc/20345/20348/pdf/2023/202305-714-723.pdf. Research progress of large-scale language models represented by ChatGPT. (2023-05-15)[2024-09-02]. https://www.nsfc.gov.cn/csc/20345/20348/pdf/2023/202305-714-723.pdf.(in Chinese)

Share

COinS