Mar 7, 2024 - 2:00 PM

HSINCHU, Taiwan - March 7, 2024 - MediaTek Research, a cutting-edge research arm of MediaTek, announced it is releasing a new open-source Large Language Model (LLM) called MediaTek Research Breeze-7B (MR Breeze-7B). This LLM is adept in Traditional Chinese and English, and comes on the heels of their pioneering traditional Chinese language model in early 2023. MR Breeze-7B, boasting 7 billion parameters, has been engineered on the widely acclaimed Mistral model.

MR Breeze-7B has absorbed a remarkable twentyfold additional knowledge compared to its predecessor, BLOOM-3B, enabling it to navigate the intricate linguistic and cultural nuances of the Traditional Chinese language with unprecedented precision. This advancement paves the way for more genuine and accurate bilingual interactions and content generation. Enhanced by MediaTek Research's optimizations, MR Breeze-7B outperforms its counterparts, including the Mistral and Llama models, in processing speed. It cuts the time and memory needed for complex Traditional Chinese inferences by half, providing users with a more seamless experience.

"With the rapid expansion of the AI industry, we're finding that there is a strong emphasis on English-based language models," said Dr. Da-shan Shiu, Managing Director of MediaTek Research. "As an open-source language model that's optimized for Traditional Chinese, MR Breeze-7B will significantly advance both academic and industrial AI technology, and this is just the beginning as we prepare to launch more open-source multimodal models to encourage additional AI collaborations."

Compared to other 7B Chinese-English language models, MR Breeze-7B delivers smoother, more accurate responses in both languages swiftly, with a keen ability to grasp context for relevant and coherent answers. This enhancement is crucial for scenarios demanding rapid bilingual interaction, such as live translation, business negotiations, and smart customer service. Furthermore, MR Breeze-7B's adeptness at parsing and producing tabular content is a game-changer for data-driven tasks like analytics, financial statements, and complex scheduling, proving indispensable for enterprises handling extensive structured data.

The release of MediaTek Research's open-source MR Breeze-7B model marks a significant step for researchers to further dissect and understand the intricacies of large language models, particularly in refining solutions for challenges such as hallucination and alignment in question-answering systems. MediaTek Research is preparing to unveil a new 47B parameter model, built upon the open-source Mixtral framework, for public testing in the near future.

###

About MediaTek Inc.

MediaTek Incorporated (TWSE: 2454) is a global fabless semiconductor company that enables nearly 2 billion connected devices a year. We are a market leader in developing innovative systems-on-chip (SoC) for mobile, home entertainment, connectivity and IoT products. Our dedication to innovation has positioned us as a driving market force in several key technology areas, including highly power-efficient mobile technologies, automotive solutions and a broad range of advanced multimedia products such as smartphones, tablets, digital televisions, 5G, Voice Assistant Devices (VAD) and wearables. MediaTek empowers and inspires people to expand their horizons and achieve their goals through smart technology, more easily and efficiently than ever before. We work with the brands you love to make great technology accessible to everyone, and it drives everything we do. Visit www.mediatek.com for more information.

Attachments

  • Original Link
  • Permalink

Disclaimer

Mediatek Inc. published this content on 07 March 2024 and is solely responsible for the information contained therein. Distributed by Public, unedited and unaltered, on 09 March 2024 02:39:08 UTC.