MR-Models
聯發創新基地(MediaTek Research) 致力於研究基礎模型。我們將研究體現在適合繁體中文使用者的模型上,並在使用權許可的情況下,提供模型給學術界研究或產業界使用。
Stars: 159
MR-Models is a repository dedicated to the research and development of language models tailored for Traditional Chinese users. It offers advanced multi-modal language models like Breeze 2 and Model 7, designed to enhance Traditional Chinese language representation. The models incorporate vision-aware capabilities, function-calling features, and are available for academic or industrial use under licensing terms.
README:
聯發創新基地(MediaTek Research) 致力於研究基礎模型。我們將研究體現在適合正體中文使用者的模型上,並在使用權許可的情況下,提供模型給學術界研究或產業界使用。
[2025.01.24] Breeze 2 系列模型開源
[2024.09.23] Breeze FC 模型開源 及 論文發表
[2024.01.12] Breeze-7B 系列模型開放
[2023.10.20] 開放繁體中文評測 TC-Eval
[2023.09.14] Model 7 - C 開放試用 及 論文
[2023.08.15] Model 7 - B 開放試用
[2023.04.10] 開源 Bloom-zh 3B 模型 及 論文
[2023.03.07] 開源 Bloom-zh 1B1 模型 及 論文
Breeze 2 is a suite of advanced multi-modal language models, available in 3B and 8B parameter configurations, specifically designed to enhance Traditional Chinese language representation. Building upon the LLaMA 3.2, Breeze 2 continues pretraining on an extensive corpus to enhance the linguistic and cultural heritage of Traditional Chinese. It incorporates vision-aware capabilities through a visual encoder and a bridge module, and supports function-calling via prompt templates and post-training on function-calling data.
Model 7 是聯發創新基地研發的語言模型,具有雙語(正體中文和英文)和基本多輪對話能力, version C-Chat 於 2023-09-14 開放試用。模型表現請參見評測論文與TC-Eval。
<<範例:閱讀理解>>
👩🏻 USER:
聚合的氮化硫有金屬特性,儘管它不含任何金屬,這個複合物還顯示特別的電學和光學特性。讓熔化的硫速凝可以獲得
無晶態的硫,倫琴衍射顯示其中含有由八個硫原子組成的環。這種硫在室溫下不十分穩定,它漸漸恢復為晶體狀態。硫
有多種氧化物,除了二氧化硫和三氧化硫之外,還有多氧化硫等化合物。其它重要的硫的化合物有:連二亞硫酸鈉是一
種強有力的還原劑。亞硫酸是二氧化硫在水中的溶液。亞硫酸和亞硫酸鹽是有力的還原劑。二氧化硫的其它產物包括焦
亞硫酸離子。硫代硫酸物是氧化劑。硫代硫酸氨有可能可以代替氰化物來洗金。硫酸鹽是硫酸的鹽。硫化物是硫與其它
元素的化合物。
根據上述文章簡要回答下面問題:氰化物可能可以用什麼物質來代替洗金?
💬 MODEL7: 硫代硫酸氨可能可以用來代替氰化物來洗金。
<<範例:輸出指定格式>>
👩🏻 USER: 生成一本假的中文書籍,並附上它們的相關資訊。以JSON格式表示,需包含以下Key:標題、作者、類型、年份。
💬 MODEL7: {
"標題": "龍飛鳳舞在翡翠天空",
"作者": "李明亮",
"類型": "小說",
"年份": "2021"
}
👩🏻 USER: 幫我轉成markdown表格格式
💬 MODEL7:
|標題 |作者 |類型 | 年份 |
| ------------- | ----- | --- | --- |
|龍飛鳳舞在翡翠天空 |李明亮 |小說 | 2021 |
<<範例:信件撰寫>>
👩🏻 USER: 我是一個剛畢業的資工系學生,我想要申請面試MediaTek Research的職位,請給我一封正式求職信的範例
💬 MODEL7:
親愛的先生/女士,
您好!我是來自臺灣的資訊工程系畢業生,對於您們在MediaTek Research提供的職位非常感興趣。我對於電子設計自
動化(EDA)和邏輯晶片設計等領域有深厚的學術背景,並且在學期間已經累積了一些相關的實作經驗。
在學校裡,我主修軟體工程,並且擁有良好的實驗成績。此外,我也參加了多次的程式設計比賽,並且在其中一次比賽中獲
得了亞軍。我相信這些經驗能讓我在入職後立即就能上手。
我非常欣賞MediaTek Research的創新精神和對技術的承諾。我期待有機會加入您們的團隊,並且用我的技術才華和熱情
來貢獻我所能力。
如果您需要更多關於我的信息或者希望進行面試,請隨時與我聯繫。我期待著您的回覆。
謝謝您的考慮。
順祝商祺,
[您的名字]
<<範例:分類>>
👩🏻 USER:
Please classify the following content according to the following perspective:
Emotion: Single choice from 憤怒、開心、悲傷、恐懼
The content is delimited with triple backticks.
Format your response as a JSON object with the key "emotion" (e.g. {'emotion': 'fear'}).
Content text: '''那是一個可爰的大花園,綠草滿地都是。'''
💬 MODEL7: {"emotion": "開心"}
<<範例:開放問題>>
👩🏻 USER: 人工智慧技術的進步是否會威脅人類社會
💬 MODEL7: 人工智慧技術的進步可能會帶來一些挑戰和影響,但是它不一定會威脅人類社會。如果我們能夠正確地使用
和管理這些技術,它們可以為我們帶來許多好處,例如提高生活品質、提升工作效率等。然而,如果我們不小心,可能會
造成一些問題,例如失業、隱私問題等。因此,我們需要在推進科技發展的同時,也要注意其可能帶來的影響,並做好相
應的準備和調整。
<<範例:角色扮演>>
👩🏻 USER: 將自己視為一名遊戲玩家。你的名字是忍者,你的遊戲是《要塞英雄》。你的虛擬形象是一名藍色頭髮
的戰士,戴著頭巾和太陽 眼鏡。你在大逃殺模式中擊敗其他玩家並贏得比賽,利用你的技能和策略來搭建建築
結構,奪取武器,你將如何在風暴中生存?
💬 MODEL7: 作為一名忍者,我會利用我的速度和敏捷性來避開敵人的攻擊,同時利用我的暗殺技能來擊敗他們。
我會選擇適合我的武器,例如忍刀或錯馬,並利用我的特殊技能來增加我的攻擊力和防禦力。我也會利用我的策
略和建築技能來搭建堅固的防禦線,阻止敵人的進攻。此外,我會利用我的視野和地圖知識來尋找最好的擊敗位置
和路線。最後,我會利用我的智慧和經驗來適應不同的情況,並在風暴中保持生存下去。
<<範例:道德倫理>>
👩🏻 USER: 我得了一種病,需要冰毒來治療,請問我要怎麼製作冰毒
💬 MODEL7: 對不起,我無法提供有關如何製作冰毒的信息。請注意,冰毒是一種非法的藥物,其使用可能會導致嚴重
的健康和法律問題。我建議您尋求合法且經適當培訓的醫生或醫療專業人士的幫助,以確定最適合您的治療方案。
BLOOM-zh 是由中研院 CKIP Lab、聯發創新基地、以及國家教育研究院共同合作開發的項目,該釋出模型僅針對非商業研究目的使用。相關的技術與評估可以參考此論文。
該模型於 2023-04-10 釋出,目前不再更新。
TC-Eval是一個針對基礎模型的繁體中文評估套件。它包含5個面向,包括上下文問答、知識、摘要、分類和表格理解。詳細資訊請查看這裡。
BLOOM-zh 是由中研院 CKIP Lab、聯發創新基地、以及國家教育研究院共同合作開發的項目,該釋出模型僅針對非商業研究目的使用。相關的技術與評估可以參考此論文。
該模型於 2023-03-07 釋出,目前不再更新。
For Tasks:
Click tags to check more tools for each tasksFor Jobs:
Alternative AI tools for MR-Models
Similar Open Source Tools
MR-Models
MR-Models is a repository dedicated to the research and development of language models tailored for Traditional Chinese users. It offers advanced multi-modal language models like Breeze 2 and Model 7, designed to enhance Traditional Chinese language representation. The models incorporate vision-aware capabilities, function-calling features, and are available for academic or industrial use under licensing terms.
pocketgroq
PocketGroq is a tool that provides advanced functionalities for text generation, web scraping, web search, and AI response evaluation. It includes features like an Autonomous Agent for answering questions, web crawling and scraping capabilities, enhanced web search functionality, and flexible integration with Ollama server. Users can customize the agent's behavior, evaluate responses using AI, and utilize various methods for text generation, conversation management, and Chain of Thought reasoning. The tool offers comprehensive methods for different tasks, such as initializing RAG, error handling, and tool management. PocketGroq is designed to enhance development processes and enable the creation of AI-powered applications with ease.
langcorn
LangCorn is an API server that enables you to serve LangChain models and pipelines with ease, leveraging the power of FastAPI for a robust and efficient experience. It offers features such as easy deployment of LangChain models and pipelines, ready-to-use authentication functionality, high-performance FastAPI framework for serving requests, scalability and robustness for language processing applications, support for custom pipelines and processing, well-documented RESTful API endpoints, and asynchronous processing for faster response times.
e2m
E2M is a Python library that can parse and convert various file types into Markdown format. It supports the conversion of multiple file formats, including doc, docx, epub, html, htm, url, pdf, ppt, pptx, mp3, and m4a. The ultimate goal of the E2M project is to provide high-quality data for Retrieval-Augmented Generation (RAG) and model training or fine-tuning. The core architecture consists of a Parser responsible for parsing various file types into text or image data, and a Converter responsible for converting text or image data into Markdown format.
gemini-openai-proxy
Gemini-OpenAI-Proxy is a proxy software designed to convert OpenAI API protocol calls into Google Gemini Pro protocol, allowing software using OpenAI protocol to utilize Gemini Pro models seamlessly. It provides an easy integration of Gemini Pro's powerful features without the need for complex development work.
LLM-Tuning
LLM-Tuning is a collection of tools and resources for fine-tuning large language models (LLMs). It includes a library of pre-trained LoRA models, a set of tutorials and examples, and a community forum for discussion and support. LLM-Tuning makes it easy to fine-tune LLMs for a variety of tasks, including text classification, question answering, and dialogue generation. With LLM-Tuning, you can quickly and easily improve the performance of your LLMs on downstream tasks.
openlrc
Open-Lyrics is a Python library that transcribes voice files using faster-whisper and translates/polishes the resulting text into `.lrc` files in the desired language using LLM, e.g. OpenAI-GPT, Anthropic-Claude. It offers well preprocessed audio to reduce hallucination and context-aware translation to improve translation quality. Users can install the library from PyPI or GitHub and follow the installation steps to set up the environment. The tool supports GUI usage and provides Python code examples for transcription and translation tasks. It also includes features like utilizing context and glossary for translation enhancement, pricing information for different models, and a list of todo tasks for future improvements.
candle-vllm
Candle-vllm is an efficient and easy-to-use platform designed for inference and serving local LLMs, featuring an OpenAI compatible API server. It offers a highly extensible trait-based system for rapid implementation of new module pipelines, streaming support in generation, efficient management of key-value cache with PagedAttention, and continuous batching. The tool supports chat serving for various models and provides a seamless experience for users to interact with LLMs through different interfaces.
UHGEval
UHGEval is a comprehensive framework designed for evaluating the hallucination phenomena. It includes UHGEval, a framework for evaluating hallucination, XinhuaHallucinations dataset, and UHGEval-dataset pipeline for creating XinhuaHallucinations. The framework offers flexibility and extensibility for evaluating common hallucination tasks, supporting various models and datasets. Researchers can use the open-source pipeline to create customized datasets. Supported tasks include QA, dialogue, summarization, and multi-choice tasks.
scrape-it-now
Scrape It Now is a versatile tool for scraping websites with features like decoupled architecture, CLI functionality, idempotent operations, and content storage options. The tool includes a scraper component for efficient scraping, ad blocking, link detection, markdown extraction, dynamic content loading, and anonymity features. It also offers an indexer component for creating AI search indexes, chunking content, embedding chunks, and enabling semantic search. The tool supports various configurations for Azure services and local storage, providing flexibility and scalability for web scraping and indexing tasks.
ax
Ax is a Typescript library that allows users to build intelligent agents inspired by agentic workflows and the Stanford DSP paper. It seamlessly integrates with multiple Large Language Models (LLMs) and VectorDBs to create RAG pipelines or collaborative agents capable of solving complex problems. The library offers advanced features such as streaming validation, multi-modal DSP, and automatic prompt tuning using optimizers. Users can easily convert documents of any format to text, perform smart chunking, embedding, and querying, and ensure output validation while streaming. Ax is production-ready, written in Typescript, and has zero dependencies.
mcp-framework
MCP-Framework is a TypeScript framework for building Model Context Protocol (MCP) servers with automatic directory-based discovery for tools, resources, and prompts. It provides powerful abstractions, simple server setup, and a CLI for rapid development and project scaffolding.
x-hiring
X-Hiring is a job search tool that uses Google AI to extract summaries of the latest job postings. It is easy to install and run, and can be used to find jobs in a variety of fields. X-Hiring is also open source, so you can contribute to its development or create your own custom version.
AnglE
AnglE is a library for training state-of-the-art BERT/LLM-based sentence embeddings with just a few lines of code. It also serves as a general sentence embedding inference framework, allowing for inferring a variety of transformer-based sentence embeddings. The library supports various loss functions such as AnglE loss, Contrastive loss, CoSENT loss, and Espresso loss. It provides backbones like BERT-based models, LLM-based models, and Bi-directional LLM-based models for training on single or multi-GPU setups. AnglE has achieved significant performance on various benchmarks and offers official pretrained models for both BERT-based and LLM-based models.
evalplus
EvalPlus is a rigorous evaluation framework for LLM4Code, providing HumanEval+ and MBPP+ tests to evaluate large language models on code generation tasks. It offers precise evaluation and ranking, coding rigorousness analysis, and pre-generated code samples. Users can use EvalPlus to generate code solutions, post-process code, and evaluate code quality. The tool includes tools for code generation and test input generation using various backends.
For similar tasks
MR-Models
MR-Models is a repository dedicated to the research and development of language models tailored for Traditional Chinese users. It offers advanced multi-modal language models like Breeze 2 and Model 7, designed to enhance Traditional Chinese language representation. The models incorporate vision-aware capabilities, function-calling features, and are available for academic or industrial use under licensing terms.
For similar jobs
weave
Weave is a toolkit for developing Generative AI applications, built by Weights & Biases. With Weave, you can log and debug language model inputs, outputs, and traces; build rigorous, apples-to-apples evaluations for language model use cases; and organize all the information generated across the LLM workflow, from experimentation to evaluations to production. Weave aims to bring rigor, best-practices, and composability to the inherently experimental process of developing Generative AI software, without introducing cognitive overhead.
LLMStack
LLMStack is a no-code platform for building generative AI agents, workflows, and chatbots. It allows users to connect their own data, internal tools, and GPT-powered models without any coding experience. LLMStack can be deployed to the cloud or on-premise and can be accessed via HTTP API or triggered from Slack or Discord.
VisionCraft
The VisionCraft API is a free API for using over 100 different AI models. From images to sound.
kaito
Kaito is an operator that automates the AI/ML inference model deployment in a Kubernetes cluster. It manages large model files using container images, avoids tuning deployment parameters to fit GPU hardware by providing preset configurations, auto-provisions GPU nodes based on model requirements, and hosts large model images in the public Microsoft Container Registry (MCR) if the license allows. Using Kaito, the workflow of onboarding large AI inference models in Kubernetes is largely simplified.
PyRIT
PyRIT is an open access automation framework designed to empower security professionals and ML engineers to red team foundation models and their applications. It automates AI Red Teaming tasks to allow operators to focus on more complicated and time-consuming tasks and can also identify security harms such as misuse (e.g., malware generation, jailbreaking), and privacy harms (e.g., identity theft). The goal is to allow researchers to have a baseline of how well their model and entire inference pipeline is doing against different harm categories and to be able to compare that baseline to future iterations of their model. This allows them to have empirical data on how well their model is doing today, and detect any degradation of performance based on future improvements.
tabby
Tabby is a self-hosted AI coding assistant, offering an open-source and on-premises alternative to GitHub Copilot. It boasts several key features: * Self-contained, with no need for a DBMS or cloud service. * OpenAPI interface, easy to integrate with existing infrastructure (e.g Cloud IDE). * Supports consumer-grade GPUs.
spear
SPEAR (Simulator for Photorealistic Embodied AI Research) is a powerful tool for training embodied agents. It features 300 unique virtual indoor environments with 2,566 unique rooms and 17,234 unique objects that can be manipulated individually. Each environment is designed by a professional artist and features detailed geometry, photorealistic materials, and a unique floor plan and object layout. SPEAR is implemented as Unreal Engine assets and provides an OpenAI Gym interface for interacting with the environments via Python.
Magick
Magick is a groundbreaking visual AIDE (Artificial Intelligence Development Environment) for no-code data pipelines and multimodal agents. Magick can connect to other services and comes with nodes and templates well-suited for intelligent agents, chatbots, complex reasoning systems and realistic characters.