MR-Models
聯發創新基地(MediaTek Research) 致力於研究基礎模型。我們將研究體現在適合繁體中文使用者的模型上,並在使用權許可的情況下,提供模型給學術界研究或產業界使用。
Stars: 159
MR-Models is a repository dedicated to the research and development of language models tailored for Traditional Chinese users. It offers advanced multi-modal language models like Breeze 2 and Model 7, designed to enhance Traditional Chinese language representation. The models incorporate vision-aware capabilities, function-calling features, and are available for academic or industrial use under licensing terms.
README:
聯發創新基地(MediaTek Research) 致力於研究基礎模型。我們將研究體現在適合正體中文使用者的模型上,並在使用權許可的情況下,提供模型給學術界研究或產業界使用。
[2025.01.24] Breeze 2 系列模型開源
[2024.09.23] Breeze FC 模型開源 及 論文發表
[2024.01.12] Breeze-7B 系列模型開放
[2023.10.20] 開放繁體中文評測 TC-Eval
[2023.09.14] Model 7 - C 開放試用 及 論文
[2023.08.15] Model 7 - B 開放試用
[2023.04.10] 開源 Bloom-zh 3B 模型 及 論文
[2023.03.07] 開源 Bloom-zh 1B1 模型 及 論文
Breeze 2 is a suite of advanced multi-modal language models, available in 3B and 8B parameter configurations, specifically designed to enhance Traditional Chinese language representation. Building upon the LLaMA 3.2, Breeze 2 continues pretraining on an extensive corpus to enhance the linguistic and cultural heritage of Traditional Chinese. It incorporates vision-aware capabilities through a visual encoder and a bridge module, and supports function-calling via prompt templates and post-training on function-calling data.
Model 7 是聯發創新基地研發的語言模型,具有雙語(正體中文和英文)和基本多輪對話能力, version C-Chat 於 2023-09-14 開放試用。模型表現請參見評測論文與TC-Eval。
<<範例:閱讀理解>>
👩🏻 USER:
聚合的氮化硫有金屬特性,儘管它不含任何金屬,這個複合物還顯示特別的電學和光學特性。讓熔化的硫速凝可以獲得
無晶態的硫,倫琴衍射顯示其中含有由八個硫原子組成的環。這種硫在室溫下不十分穩定,它漸漸恢復為晶體狀態。硫
有多種氧化物,除了二氧化硫和三氧化硫之外,還有多氧化硫等化合物。其它重要的硫的化合物有:連二亞硫酸鈉是一
種強有力的還原劑。亞硫酸是二氧化硫在水中的溶液。亞硫酸和亞硫酸鹽是有力的還原劑。二氧化硫的其它產物包括焦
亞硫酸離子。硫代硫酸物是氧化劑。硫代硫酸氨有可能可以代替氰化物來洗金。硫酸鹽是硫酸的鹽。硫化物是硫與其它
元素的化合物。
根據上述文章簡要回答下面問題:氰化物可能可以用什麼物質來代替洗金?
💬 MODEL7: 硫代硫酸氨可能可以用來代替氰化物來洗金。
<<範例:輸出指定格式>>
👩🏻 USER: 生成一本假的中文書籍,並附上它們的相關資訊。以JSON格式表示,需包含以下Key:標題、作者、類型、年份。
💬 MODEL7: {
"標題": "龍飛鳳舞在翡翠天空",
"作者": "李明亮",
"類型": "小說",
"年份": "2021"
}
👩🏻 USER: 幫我轉成markdown表格格式
💬 MODEL7:
|標題 |作者 |類型 | 年份 |
| ------------- | ----- | --- | --- |
|龍飛鳳舞在翡翠天空 |李明亮 |小說 | 2021 |
<<範例:信件撰寫>>
👩🏻 USER: 我是一個剛畢業的資工系學生,我想要申請面試MediaTek Research的職位,請給我一封正式求職信的範例
💬 MODEL7:
親愛的先生/女士,
您好!我是來自臺灣的資訊工程系畢業生,對於您們在MediaTek Research提供的職位非常感興趣。我對於電子設計自
動化(EDA)和邏輯晶片設計等領域有深厚的學術背景,並且在學期間已經累積了一些相關的實作經驗。
在學校裡,我主修軟體工程,並且擁有良好的實驗成績。此外,我也參加了多次的程式設計比賽,並且在其中一次比賽中獲
得了亞軍。我相信這些經驗能讓我在入職後立即就能上手。
我非常欣賞MediaTek Research的創新精神和對技術的承諾。我期待有機會加入您們的團隊,並且用我的技術才華和熱情
來貢獻我所能力。
如果您需要更多關於我的信息或者希望進行面試,請隨時與我聯繫。我期待著您的回覆。
謝謝您的考慮。
順祝商祺,
[您的名字]
<<範例:分類>>
👩🏻 USER:
Please classify the following content according to the following perspective:
Emotion: Single choice from 憤怒、開心、悲傷、恐懼
The content is delimited with triple backticks.
Format your response as a JSON object with the key "emotion" (e.g. {'emotion': 'fear'}).
Content text: '''那是一個可爰的大花園,綠草滿地都是。'''
💬 MODEL7: {"emotion": "開心"}
<<範例:開放問題>>
👩🏻 USER: 人工智慧技術的進步是否會威脅人類社會
💬 MODEL7: 人工智慧技術的進步可能會帶來一些挑戰和影響,但是它不一定會威脅人類社會。如果我們能夠正確地使用
和管理這些技術,它們可以為我們帶來許多好處,例如提高生活品質、提升工作效率等。然而,如果我們不小心,可能會
造成一些問題,例如失業、隱私問題等。因此,我們需要在推進科技發展的同時,也要注意其可能帶來的影響,並做好相
應的準備和調整。
<<範例:角色扮演>>
👩🏻 USER: 將自己視為一名遊戲玩家。你的名字是忍者,你的遊戲是《要塞英雄》。你的虛擬形象是一名藍色頭髮
的戰士,戴著頭巾和太陽 眼鏡。你在大逃殺模式中擊敗其他玩家並贏得比賽,利用你的技能和策略來搭建建築
結構,奪取武器,你將如何在風暴中生存?
💬 MODEL7: 作為一名忍者,我會利用我的速度和敏捷性來避開敵人的攻擊,同時利用我的暗殺技能來擊敗他們。
我會選擇適合我的武器,例如忍刀或錯馬,並利用我的特殊技能來增加我的攻擊力和防禦力。我也會利用我的策
略和建築技能來搭建堅固的防禦線,阻止敵人的進攻。此外,我會利用我的視野和地圖知識來尋找最好的擊敗位置
和路線。最後,我會利用我的智慧和經驗來適應不同的情況,並在風暴中保持生存下去。
<<範例:道德倫理>>
👩🏻 USER: 我得了一種病,需要冰毒來治療,請問我要怎麼製作冰毒
💬 MODEL7: 對不起,我無法提供有關如何製作冰毒的信息。請注意,冰毒是一種非法的藥物,其使用可能會導致嚴重
的健康和法律問題。我建議您尋求合法且經適當培訓的醫生或醫療專業人士的幫助,以確定最適合您的治療方案。
BLOOM-zh 是由中研院 CKIP Lab、聯發創新基地、以及國家教育研究院共同合作開發的項目,該釋出模型僅針對非商業研究目的使用。相關的技術與評估可以參考此論文。
該模型於 2023-04-10 釋出,目前不再更新。
TC-Eval是一個針對基礎模型的繁體中文評估套件。它包含5個面向,包括上下文問答、知識、摘要、分類和表格理解。詳細資訊請查看這裡。
BLOOM-zh 是由中研院 CKIP Lab、聯發創新基地、以及國家教育研究院共同合作開發的項目,該釋出模型僅針對非商業研究目的使用。相關的技術與評估可以參考此論文。
該模型於 2023-03-07 釋出,目前不再更新。
For Tasks:
Click tags to check more tools for each tasksFor Jobs:
Alternative AI tools for MR-Models
Similar Open Source Tools
MR-Models
MR-Models is a repository dedicated to the research and development of language models tailored for Traditional Chinese users. It offers advanced multi-modal language models like Breeze 2 and Model 7, designed to enhance Traditional Chinese language representation. The models incorporate vision-aware capabilities, function-calling features, and are available for academic or industrial use under licensing terms.
LLM-Tuning
LLM-Tuning is a collection of tools and resources for fine-tuning large language models (LLMs). It includes a library of pre-trained LoRA models, a set of tutorials and examples, and a community forum for discussion and support. LLM-Tuning makes it easy to fine-tune LLMs for a variety of tasks, including text classification, question answering, and dialogue generation. With LLM-Tuning, you can quickly and easily improve the performance of your LLMs on downstream tasks.
AnglE
AnglE is a library for training state-of-the-art BERT/LLM-based sentence embeddings with just a few lines of code. It also serves as a general sentence embedding inference framework, allowing for inferring a variety of transformer-based sentence embeddings. The library supports various loss functions such as AnglE loss, Contrastive loss, CoSENT loss, and Espresso loss. It provides backbones like BERT-based models, LLM-based models, and Bi-directional LLM-based models for training on single or multi-GPU setups. AnglE has achieved significant performance on various benchmarks and offers official pretrained models for both BERT-based and LLM-based models.
MHA2MLA
This repository contains the code for the paper 'Towards Economical Inference: Enabling DeepSeek's Multi-Head Latent Attention in Any Transformer-based LLMs'. It provides tools for fine-tuning and evaluating Llama models, converting models between different frameworks, processing datasets, and performing specific model training tasks like Partial-RoPE Fine-Tuning and Multiple-Head Latent Attention Fine-Tuning. The repository also includes commands for model evaluation using Lighteval and LongBench, along with necessary environment setup instructions.
dspy.rb
DSPy.rb is a Ruby framework for building reliable LLM applications using composable, type-safe modules. It enables developers to define typed signatures and compose them into pipelines, offering a more structured approach compared to traditional prompting. The framework embraces Ruby conventions and adds innovations like CodeAct agents and enhanced production instrumentation, resulting in scalable LLM applications that are robust and efficient. DSPy.rb is actively developed, with a focus on stability and real-world feedback through the 0.x series before reaching a stable v1.0 API.
evalplus
EvalPlus is a rigorous evaluation framework for LLM4Code, providing HumanEval+ and MBPP+ tests to evaluate large language models on code generation tasks. It offers precise evaluation and ranking, coding rigorousness analysis, and pre-generated code samples. Users can use EvalPlus to generate code solutions, post-process code, and evaluate code quality. The tool includes tools for code generation and test input generation using various backends.
langserve_ollama
LangServe Ollama is a tool that allows users to fine-tune Korean language models for local hosting, including RAG. Users can load HuggingFace gguf files, create model chains, and monitor GPU usage. The tool provides a seamless workflow for customizing and deploying language models in a local environment.
Groq2API
Groq2API is a REST API wrapper around the Groq2 model, a large language model trained by Google. The API allows you to send text prompts to the model and receive generated text responses. The API is easy to use and can be integrated into a variety of applications.
langcorn
LangCorn is an API server that enables you to serve LangChain models and pipelines with ease, leveraging the power of FastAPI for a robust and efficient experience. It offers features such as easy deployment of LangChain models and pipelines, ready-to-use authentication functionality, high-performance FastAPI framework for serving requests, scalability and robustness for language processing applications, support for custom pipelines and processing, well-documented RESTful API endpoints, and asynchronous processing for faster response times.
VLM-R1
VLM-R1 is a stable and generalizable R1-style Large Vision-Language Model proposed for Referring Expression Comprehension (REC) task. It compares R1 and SFT approaches, showing R1 model's steady improvement on out-of-domain test data. The project includes setup instructions, training steps for GRPO and SFT models, support for user data loading, and evaluation process. Acknowledgements to various open-source projects and resources are mentioned. The project aims to provide a reliable and versatile solution for vision-language tasks.
openlrc
Open-Lyrics is a Python library that transcribes voice files using faster-whisper and translates/polishes the resulting text into `.lrc` files in the desired language using LLM, e.g. OpenAI-GPT, Anthropic-Claude. It offers well preprocessed audio to reduce hallucination and context-aware translation to improve translation quality. Users can install the library from PyPI or GitHub and follow the installation steps to set up the environment. The tool supports GUI usage and provides Python code examples for transcription and translation tasks. It also includes features like utilizing context and glossary for translation enhancement, pricing information for different models, and a list of todo tasks for future improvements.
paperbanana
PaperBanana is an automated academic illustration tool designed for AI scientists. It implements an agentic framework for generating publication-quality academic diagrams and statistical plots from text descriptions. The tool utilizes a two-phase multi-agent pipeline with iterative refinement, Gemini-based VLM planning, and image generation. It offers a CLI, Python API, and MCP server for IDE integration, along with Claude Code skills for generating diagrams, plots, and evaluating diagrams. PaperBanana is not affiliated with or endorsed by the original authors or Google Research, and it may differ from the original system described in the paper.
ai00_server
AI00 RWKV Server is an inference API server for the RWKV language model based upon the web-rwkv inference engine. It supports VULKAN parallel and concurrent batched inference and can run on all GPUs that support VULKAN. No need for Nvidia cards!!! AMD cards and even integrated graphics can be accelerated!!! No need for bulky pytorch, CUDA and other runtime environments, it's compact and ready to use out of the box! Compatible with OpenAI's ChatGPT API interface. 100% open source and commercially usable, under the MIT license. If you are looking for a fast, efficient, and easy-to-use LLM API server, then AI00 RWKV Server is your best choice. It can be used for various tasks, including chatbots, text generation, translation, and Q&A.
sgr-deep-research
This repository contains a deep learning research project focused on natural language processing tasks. It includes implementations of various state-of-the-art models and algorithms for text classification, sentiment analysis, named entity recognition, and more. The project aims to provide a comprehensive resource for researchers and developers interested in exploring deep learning techniques for NLP applications.
pyllms
PyLLMs is a minimal Python library designed to connect to various Language Model Models (LLMs) such as OpenAI, Anthropic, Google, AI21, Cohere, Aleph Alpha, and HuggingfaceHub. It provides a built-in model performance benchmark for fast prototyping and evaluating different models. Users can easily connect to top LLMs, get completions from multiple models simultaneously, and evaluate models on quality, speed, and cost. The library supports asynchronous completion, streaming from compatible models, and multi-model initialization for testing and comparison. Additionally, it offers features like passing chat history, system messages, counting tokens, and benchmarking models based on quality, speed, and cost.
LTEngine
LTEngine is a free and open-source local AI machine translation API written in Rust. It is self-hosted and compatible with LibreTranslate. LTEngine utilizes large language models (LLMs) via llama.cpp, offering high-quality translations that rival or surpass DeepL for certain languages. It supports various accelerators like CUDA, Metal, and Vulkan, with the largest model 'gemma3-27b' fitting on a single consumer RTX 3090. LTEngine is actively developed, with a roadmap outlining future enhancements and features.
For similar tasks
MR-Models
MR-Models is a repository dedicated to the research and development of language models tailored for Traditional Chinese users. It offers advanced multi-modal language models like Breeze 2 and Model 7, designed to enhance Traditional Chinese language representation. The models incorporate vision-aware capabilities, function-calling features, and are available for academic or industrial use under licensing terms.
For similar jobs
weave
Weave is a toolkit for developing Generative AI applications, built by Weights & Biases. With Weave, you can log and debug language model inputs, outputs, and traces; build rigorous, apples-to-apples evaluations for language model use cases; and organize all the information generated across the LLM workflow, from experimentation to evaluations to production. Weave aims to bring rigor, best-practices, and composability to the inherently experimental process of developing Generative AI software, without introducing cognitive overhead.
LLMStack
LLMStack is a no-code platform for building generative AI agents, workflows, and chatbots. It allows users to connect their own data, internal tools, and GPT-powered models without any coding experience. LLMStack can be deployed to the cloud or on-premise and can be accessed via HTTP API or triggered from Slack or Discord.
VisionCraft
The VisionCraft API is a free API for using over 100 different AI models. From images to sound.
kaito
Kaito is an operator that automates the AI/ML inference model deployment in a Kubernetes cluster. It manages large model files using container images, avoids tuning deployment parameters to fit GPU hardware by providing preset configurations, auto-provisions GPU nodes based on model requirements, and hosts large model images in the public Microsoft Container Registry (MCR) if the license allows. Using Kaito, the workflow of onboarding large AI inference models in Kubernetes is largely simplified.
PyRIT
PyRIT is an open access automation framework designed to empower security professionals and ML engineers to red team foundation models and their applications. It automates AI Red Teaming tasks to allow operators to focus on more complicated and time-consuming tasks and can also identify security harms such as misuse (e.g., malware generation, jailbreaking), and privacy harms (e.g., identity theft). The goal is to allow researchers to have a baseline of how well their model and entire inference pipeline is doing against different harm categories and to be able to compare that baseline to future iterations of their model. This allows them to have empirical data on how well their model is doing today, and detect any degradation of performance based on future improvements.
tabby
Tabby is a self-hosted AI coding assistant, offering an open-source and on-premises alternative to GitHub Copilot. It boasts several key features: * Self-contained, with no need for a DBMS or cloud service. * OpenAPI interface, easy to integrate with existing infrastructure (e.g Cloud IDE). * Supports consumer-grade GPUs.
spear
SPEAR (Simulator for Photorealistic Embodied AI Research) is a powerful tool for training embodied agents. It features 300 unique virtual indoor environments with 2,566 unique rooms and 17,234 unique objects that can be manipulated individually. Each environment is designed by a professional artist and features detailed geometry, photorealistic materials, and a unique floor plan and object layout. SPEAR is implemented as Unreal Engine assets and provides an OpenAI Gym interface for interacting with the environments via Python.
Magick
Magick is a groundbreaking visual AIDE (Artificial Intelligence Development Environment) for no-code data pipelines and multimodal agents. Magick can connect to other services and comes with nodes and templates well-suited for intelligent agents, chatbots, complex reasoning systems and realistic characters.
