RAG_Hack
Hack Together: RAG Hack | Register, Learn, Hack
Stars: 310
RAGHack is a hackathon focused on building AI applications using the power of RAG (Retrieval Augmented Generation). RAG combines large language models with search engine knowledge to provide contextually relevant answers. Participants can learn to build RAG apps on Azure AI using various languages and retrievers, explore frameworks like LangChain and Semantic Kernel, and leverage technologies such as agents and vision models. The hackathon features live streams, hack submissions, and prizes for innovative projects.
README:
🛠️ Build, innovate, and #Hacktogether! 🛠️ It's time to start building AI applications using the power of RAG (Retrieval Augmented Generation). 🤖 + 📚 = 🔥
Large language models are powerful language generators, but they don't know everything about the world. RAG (Retrieval Augmented Generation) combines the power of large language models with the knowledge of a search engine. This allows you to ask questions of your own data, and get answers that are relevant to the context of your question.
RAGHack is your opportunity to get deep into RAG and start building RAG yourself. Across 25+ live streams, we'll show you how to build RAG apps on top of Azure AI in multiple languages (Python, Java, JS, C#) with multiple retrievers (AI Search, PostgreSQL, Azure SQL, Cosmos DB), with your own data sources! You'll learn about the most popular frameworks, like LangChain and Semantic Kernel, plus the latest technology, like agents and vision models. The possibilities are endless for what you can create... plus you can submit your hack for a chance to win exciting prizes! 🥳
The streams start September 3rd and end September 13th. Hack submissions are due September 16th, 11:59 PM PST. Join us!
Register for the hackathon using any of the sessions linked on the Reactor series home page. This will register you for both the selected session and the hackathon.
Introduce yourself and look for teammates here GitHub Discussions!
Read the official rules 📃
Day/Time | Topic | Resources |
---|---|---|
9/3, 04:30 PM UTC / 09:30 AM PT | RAG 101 | Link |
9/3, 06:00 PM UTC / 11:00 AM PT | RAG with .NET | Link |
9/3, 08:00 PM UTC / 01:00 PM PT | RAG with Azure AI Studio | Link |
9/3, 10:00 PM UTC / 03:00 PM PT | RAG with Python | Link |
9/4, 03:00 PM UTC / 08:00 AM PT | RAG with Langchain4J | Link |
9/4, 03:00 PM UTC / 08:00 AM PT | RAG with LangchainJS | Link |
9/4, 09:00 PM UTC / 02:00 PM PT | Scalable RAG with CosmosDB for NoSQL | Link |
9/5, 03:00 PM UTC / 08:00 AM PT | Responsible AI | Link |
9/5, 05:00 PM UTC / 10:00 AM PT | RAG on Cosmos DB MongoDB | Link |
9/5, 07:00 PM UTC / 12:00 PM PT | RAG with Azure AI Search | Link |
9/5, 09:00 PM UTC / 02:00 PM PT | RAG on PostgreSQL | Link |
9/5, 11:00 PM UTC / 04:00 PM PT | RAG on Azure SQL Server | Link |
9/6, 04:00 PM UTC / 09:00 AM PT | Intro to GraphRAG | Link |
9/6, 06:00 PM UTC / 11:00 AM PT | Add multi-channel communication in RAG apps | Link |
Day/Time | Topic | Resources |
---|---|---|
9/9, 03:00 PM UTC / 08:00 AM PT | RAG with Java + Semantic Kernel | Link |
9/9, 05:00 PM UTC / 10:00 AM PT | RAG with Java + Spring AI | Link |
9/9, 08:00 PM UTC / 01:00 PM PT | RAG with vision models | Link |
9/9, 11:00 PM UTC / 04:00 PM PT | Internationalization for RAG apps | Link |
9/10, 03:00 PM UTC / 08:00 AM PT | Use Phi-3 to create a VSCode chat agent extension | Link |
9/10, 05:00 PM UTC / 10:00 AM PT | Agentic RAG with Langchain | Link |
9/10, 10:00 PM UTC / 03:00 PM PT | Build an OpenAI code interpreter for Python | Link |
9/11, 03:00 PM UTC / 08:00 AM PT | Connections in Azure AI Studio | Link |
9/11, 05:00 PM UTC / 10:00 AM PT | Explore AutoGen concepts with AutoGen Studio | Link |
9/11, 08:00 PM UTC / 01:00 PM PT | RAG with Data Access Control | Link |
9/11, 10:00 PM UTC / 03:00 PM PT | RAFT: (RAG + Fine Tuning) in Azure AI Studio | Link |
9/12, 04:00 AM UTC / 09:00 AM PT | Pick the right model for the right job | Link |
9/12, 08:00 PM UTC / 01:00 PM PT | Evaluating your RAG Chat App | Link |
Day/Time | Topic | Resources |
---|---|---|
9/3, 03:00 PM UTC / 08:00 AM PT | RAG: Generación Aumentada de Recuperación | Link |
9/4, 03:00 PM UTC / 08:00 AM PT | RAG: Prácticas recomendadas de Azure AI Search | Link |
9/11, 03:00 PM UTC / 08:00 AM PT | AI Multi-Agentes: Patrones, Problemas y Soluciones | Link |
Day/Time | Topic | Resources |
---|---|---|
9/3, 03:00 PM UTC / 08:00 AM PT | RAG (Geração Aumentada de Busca) no Azure | Link |
9/12, 03:00 PM UTC / 08:00 AM PT | Construindo RAG com Azure AI Studio e Python | Link |
9/13, 03:00 PM UTC / 08:00 AM PT | Implantando RAG com .NET e Azure Developer CLI | Link |
Day/Time | Topic | Resources |
---|---|---|
9/3, 12:30 PM UTC / 05:30 AM PT | Global RAG Hack Together | Link |
9/10, 12:30 PM UTC / 05:30 AM PT | Create RAG apps with Azure AI SDK | Link |
9/12, 12:30 PM UTC / 05:30 AM PT | Create RAG applications with AI Toolkit VSCode Extension | Link |
9/14, 12:30 PM UTC / 05:30 AM PT | Intro to GraphRAG | Link |
For additional help with your hacks, you can drop by Office Hours in our AI Discord channel. Here are the Office Hours scheduled so far:
Day/Time | Topic/Hosts |
---|---|
9/4, 07:00 PM UTC / 12:00 PM PT | Python, AI Search, Postgres, with Pamela |
9/6, 07:00 PM UTC / 12:00 PM PT | .NET with Bruno |
Repository | Language/Retriever | Costs |
---|---|---|
azure-search-openai-demo | Python, Azure AI Search | Requires Azure deployment, follow guide for lower cost deployment |
azure-search-openai-demo-java | Java, Azure AI Search | Requires Azure deployment, see cost estimate for App Service deployment, Container Apps, Kubernetes |
serverless-chat-langchainjs | JavaScript, CosmosDB | Can be run locally for free with Ollama, see cost estimate for Azure deployment |
azure-search-openai-demo-csharp | C#, Azure AI Search | Requires Azure deployment, see cost estimate or follow guide for low cost deployment |
rag-postgres-openai-python | Python, PostgreSQL | Can be run locally for free with Ollama, see cost estimate for deployment for Azure deployment. |
Cosmic-Food-RAG-app | Python, Cosmos DB MongoDB | Requires Azure deployment, see cost estimate |
contoso-chat | Python, Azure AI Search, Azure AI Studio, PromptFlow | Requires Azure deployment, see cost estimate |
azure-sql-db-session-recommender-v2 | C#, Azure SQL | Can be run locally for free with Azure SQL Database free tier |
To find more samples, check out the following resources:
- Azure AI samples (Python)
- Azure AI samples (JavaScript)
- Azure AI samples (Java)
- Azure AI samples (C#)
- Azure AI samples (Go)
- Azure AI Studio Samples
- Cosmos DB AI Samples
- Azure SQL DB AI Samples
- AI learning and community hub
- Cloud skills challenge: Using Azure OpenAI Service
- Generative AI for Beginners
- Fundamentals of Generative AI
- Retrieval Augmented Generation in Azure AI Search
- Workshop - Create your own ChatGPT with Retrieval-Augmented-Generation
- OpenAI documentation
- Azure AI Search
- Azure OpenAI Service
- Comparing Azure OpenAI and OpenAI
- Azure Communication Services Chat SDK
- AI-in-a-Box
- Join the Azure AI Discord!
Hack submissions are due September 16th, 11:59 PM PST.
Submit your project here when it's ready: 🚀 Project Submission
Check out this video for step by step project submission guidance: Project Submission Video
Projects will be evaluated by a panel of judges, including Microsoft engineers, product managers, and developer advocates. Judging criteria will include innovation, impact, technical usability, and alignment with corresponding hackathon category.
Each winning team in the categories below will receive a cash prize of $500. 💸
- Best overall
- Best in JavaScript/TypeScript
- Best in Java
- Best in .NET
- Best in Python
- Best use of AI Studio
- Best use of AI Search
- Best use of PostgreSQL
- Best use of Cosmos DB
- Best use of Azure SQL
All hackathon participants who submit a project will receive a digital badge (sometime in October).
For Tasks:
Click tags to check more tools for each tasksFor Jobs:
Alternative AI tools for RAG_Hack
Similar Open Source Tools
RAG_Hack
RAGHack is a hackathon focused on building AI applications using the power of RAG (Retrieval Augmented Generation). RAG combines large language models with search engine knowledge to provide contextually relevant answers. Participants can learn to build RAG apps on Azure AI using various languages and retrievers, explore frameworks like LangChain and Semantic Kernel, and leverage technologies such as agents and vision models. The hackathon features live streams, hack submissions, and prizes for innovative projects.
Model-References
The 'Model-References' repository contains examples for training and inference using Intel Gaudi AI Accelerator. It includes models for computer vision, natural language processing, audio, generative models, MLPerf™ training, and MLPerf™ inference. The repository provides performance data and model validation information for various frameworks like PyTorch. Users can find examples of popular models like ResNet, BERT, and Stable Diffusion optimized for Intel Gaudi AI accelerator.
Awesome-LLM-Safety
Welcome to our Awesome-llm-safety repository! We've curated a collection of the latest, most comprehensive, and most valuable resources on large language model safety (llm-safety). But we don't stop there; included are also relevant talks, tutorials, conferences, news, and articles. Our repository is constantly updated to ensure you have the most current information at your fingertips.
ai-samples
AI Samples for .NET is a repository containing various samples demonstrating how to use AI in .NET applications. It provides quickstarts using Semantic Kernel and Azure OpenAI SDK, covers LLM Core Concepts, End to End Examples, Local Models, Local Embedding Models, Tokenizers, Vector Databases, and Reference Examples. The repository showcases different AI-related projects and tools for developers to explore and learn from.
auto-dev-vscode
AutoDev for VSCode is an AI-powered coding wizard with multilingual support, auto code generation, and a bug-slaying assistant. It offers customizable prompts and features like Auto Dev/Testing/Document/Agent. The tool aims to enhance coding productivity and efficiency by providing intelligent assistance and automation capabilities within the Visual Studio Code environment.
kangaroo
Kangaroo is an AI-powered SQL client and admin tool for popular databases like SQLite, MySQL, PostgreSQL, etc. It supports various functionalities such as table design, query, model, sync, export/import, and more. The tool is designed to be comfortable, fun, and developer-friendly, with features like code intellisense and autocomplete. Kangaroo aims to provide a seamless experience for database management across different operating systems.
nntrainer
NNtrainer is a software framework for training neural network models on devices with limited resources. It enables on-device fine-tuning of neural networks using user data for personalization. NNtrainer supports various machine learning algorithms and provides examples for tasks such as few-shot learning, ResNet, VGG, and product rating. It is optimized for embedded devices and utilizes CBLAS and CUBLAS for accelerated calculations. NNtrainer is open source and released under the Apache License version 2.0.
ml-road-map
The Machine Learning Road Map is a comprehensive guide designed to take individuals from various levels of machine learning knowledge to a basic understanding of machine learning principles using high-quality, free resources. It aims to simplify the complex and rapidly growing field of machine learning by providing a structured roadmap for learning. The guide emphasizes the importance of understanding AI for everyone, the need for patience in learning machine learning due to its complexity, and the value of learning from experts in the field. It covers five different paths to learning about machine learning, catering to consumers, aspiring AI researchers, ML engineers, developers interested in building ML applications, and companies looking to implement AI solutions.
are-copilots-local-yet
Current trends and state of the art for using open & local LLM models as copilots to complete code, generate projects, act as shell assistants, automatically fix bugs, and more. This document is a curated list of local Copilots, shell assistants, and related projects, intended to be a resource for those interested in a survey of the existing tools and to help developers discover the state of the art for projects like these.
CogVLM2
CogVLM2 is a new generation of open source models that offer significant improvements in benchmarks such as TextVQA and DocVQA. It supports 8K content length, image resolution up to 1344 * 1344, and both Chinese and English languages. The project provides basic calling methods, fine-tuning examples, and OpenAI API format calling examples to help developers quickly get started with the model.
EAGLE
Eagle is a family of Vision-Centric High-Resolution Multimodal LLMs that enhance multimodal LLM perception using a mix of vision encoders and various input resolutions. The model features a channel-concatenation-based fusion for vision experts with different architectures and knowledge, supporting up to over 1K input resolution. It excels in resolution-sensitive tasks like optical character recognition and document understanding.
RAGHub
RAGHub is a community-driven project focused on cataloging new and emerging frameworks, projects, and resources in the Retrieval-Augmented Generation (RAG) ecosystem. It aims to help users stay ahead of changes in the field by providing a platform for the latest innovations in RAG. The repository includes information on RAG frameworks, evaluation frameworks, optimization frameworks, citation frameworks, engines, search reranker frameworks, projects, resources, and real-world use cases across industries and professions.
oss-fuzz-gen
This framework generates fuzz targets for real-world `C`/`C++` projects with various Large Language Models (LLM) and benchmarks them via the `OSS-Fuzz` platform. It manages to successfully leverage LLMs to generate valid fuzz targets (which generate non-zero coverage increase) for 160 C/C++ projects. The maximum line coverage increase is 29% from the existing human-written targets.
awesome-mobile-llm
Awesome Mobile LLMs is a curated list of Large Language Models (LLMs) and related studies focused on mobile and embedded hardware. The repository includes information on various LLM models, deployment frameworks, benchmarking efforts, applications, multimodal LLMs, surveys on efficient LLMs, training LLMs on device, mobile-related use-cases, industry announcements, and related repositories. It aims to be a valuable resource for researchers, engineers, and practitioners interested in mobile LLMs.
sane-airscan
sane-airscan is a SANE backend that supports driverless scanning using Apple AirScan (eSCL) and Microsoft WSD protocols. It automatically chooses between the two protocols and has been tested with various devices from Brother, Canon, Dell, Kyocera, Lexmark, Epson, HP, OKI, Panasonic, Pantum, Ricoh, Samsung, and Xerox. The backend allows for automatic and manual device discovery and configuration, supports scanning from platen and ADF in color and grayscale modes, and works with both IPv4 and IPv6. It does not require installation and does not conflict with vendor-provided proprietary software.
llm-awq
AWQ (Activation-aware Weight Quantization) is a tool designed for efficient and accurate low-bit weight quantization (INT3/4) for Large Language Models (LLMs). It supports instruction-tuned models and multi-modal LMs, providing features such as AWQ search for accurate quantization, pre-computed AWQ model zoo for various LLMs, memory-efficient 4-bit linear in PyTorch, and efficient CUDA kernel implementation for fast inference. The tool enables users to run large models on resource-constrained edge platforms, delivering more efficient responses with LLM/VLM chatbots through 4-bit inference.
For similar tasks
dify
Dify is an open-source LLM app development platform that combines AI workflow, RAG pipeline, agent capabilities, model management, observability features, and more. It allows users to quickly go from prototype to production. Key features include: 1. Workflow: Build and test powerful AI workflows on a visual canvas. 2. Comprehensive model support: Seamless integration with hundreds of proprietary / open-source LLMs from dozens of inference providers and self-hosted solutions. 3. Prompt IDE: Intuitive interface for crafting prompts, comparing model performance, and adding additional features. 4. RAG Pipeline: Extensive RAG capabilities that cover everything from document ingestion to retrieval. 5. Agent capabilities: Define agents based on LLM Function Calling or ReAct, and add pre-built or custom tools. 6. LLMOps: Monitor and analyze application logs and performance over time. 7. Backend-as-a-Service: All of Dify's offerings come with corresponding APIs for easy integration into your own business logic.
intro-to-intelligent-apps
This repository introduces and helps organizations get started with building AI Apps and incorporating Large Language Models (LLMs) into them. The workshop covers topics such as prompt engineering, AI orchestration, and deploying AI apps. Participants will learn how to use Azure OpenAI, Langchain/ Semantic Kernel, Qdrant, and Azure AI Search to build intelligent applications.
runhouse
Runhouse is a tool that allows you to build, run, and deploy production-quality AI apps and workflows on your own compute. It provides simple, powerful APIs for the full lifecycle of AI development, from research to evaluation to production to updates to scaling to management, and across any infra. By automatically packaging your apps into scalable, secure, and observable services, Runhouse can also turn otherwise redundant AI activities into common reusable components across your team or company, which improves cost, velocity, and reproducibility.
Awesome-LLM-RAG-Application
Awesome-LLM-RAG-Application is a repository that provides resources and information about applications based on Large Language Models (LLM) with Retrieval-Augmented Generation (RAG) pattern. It includes a survey paper, GitHub repo, and guides on advanced RAG techniques. The repository covers various aspects of RAG, including academic papers, evaluation benchmarks, downstream tasks, tools, and technologies. It also explores different frameworks, preprocessing tools, routing mechanisms, evaluation frameworks, embeddings, security guardrails, prompting tools, SQL enhancements, LLM deployment, observability tools, and more. The repository aims to offer comprehensive knowledge on RAG for readers interested in exploring and implementing LLM-based systems and products.
sdfx
SDFX is the ultimate no-code platform for building and sharing AI apps with beautiful UI. It enables the creation of user-friendly interfaces for complex workflows by combining Comfy workflow with a UI. The tool is designed to merge the benefits of form-based UI and graph-node based UI, allowing users to create intricate graphs with a high-level UI overlay. SDFX is fully compatible with ComfyUI, abstracting the need for installing ComfyUI. It offers features like animated graph navigation, node bookmarks, UI debugger, custom nodes manager, app and template export, image and mask editor, and more. The tool compiles as a native app or web app, making it easy to maintain and add new features.
Build-Modern-AI-Apps
This repository serves as a hub for Microsoft Official Build & Modernize AI Applications reference solutions and content. It provides access to projects demonstrating how to build Generative AI applications using Azure services like Azure OpenAI, Azure Container Apps, Azure Kubernetes, and Azure Cosmos DB. The solutions include Vector Search & AI Assistant, Real-Time Payment and Transaction Processing, and Medical Claims Processing. Additionally, there are workshops like the Intelligent App Workshop for Microsoft Copilot Stack, focusing on infusing intelligence into traditional software systems using foundation models and design thinking.
RAG_Hack
RAGHack is a hackathon focused on building AI applications using the power of RAG (Retrieval Augmented Generation). RAG combines large language models with search engine knowledge to provide contextually relevant answers. Participants can learn to build RAG apps on Azure AI using various languages and retrievers, explore frameworks like LangChain and Semantic Kernel, and leverage technologies such as agents and vision models. The hackathon features live streams, hack submissions, and prizes for innovative projects.
For similar jobs
sweep
Sweep is an AI junior developer that turns bugs and feature requests into code changes. It automatically handles developer experience improvements like adding type hints and improving test coverage.
teams-ai
The Teams AI Library is a software development kit (SDK) that helps developers create bots that can interact with Teams and Microsoft 365 applications. It is built on top of the Bot Framework SDK and simplifies the process of developing bots that interact with Teams' artificial intelligence capabilities. The SDK is available for JavaScript/TypeScript, .NET, and Python.
ai-guide
This guide is dedicated to Large Language Models (LLMs) that you can run on your home computer. It assumes your PC is a lower-end, non-gaming setup.
classifai
Supercharge WordPress Content Workflows and Engagement with Artificial Intelligence. Tap into leading cloud-based services like OpenAI, Microsoft Azure AI, Google Gemini and IBM Watson to augment your WordPress-powered websites. Publish content faster while improving SEO performance and increasing audience engagement. ClassifAI integrates Artificial Intelligence and Machine Learning technologies to lighten your workload and eliminate tedious tasks, giving you more time to create original content that matters.
chatbot-ui
Chatbot UI is an open-source AI chat app that allows users to create and deploy their own AI chatbots. It is easy to use and can be customized to fit any need. Chatbot UI is perfect for businesses, developers, and anyone who wants to create a chatbot.
BricksLLM
BricksLLM is a cloud native AI gateway written in Go. Currently, it provides native support for OpenAI, Anthropic, Azure OpenAI and vLLM. BricksLLM aims to provide enterprise level infrastructure that can power any LLM production use cases. Here are some use cases for BricksLLM: * Set LLM usage limits for users on different pricing tiers * Track LLM usage on a per user and per organization basis * Block or redact requests containing PIIs * Improve LLM reliability with failovers, retries and caching * Distribute API keys with rate limits and cost limits for internal development/production use cases * Distribute API keys with rate limits and cost limits for students
uAgents
uAgents is a Python library developed by Fetch.ai that allows for the creation of autonomous AI agents. These agents can perform various tasks on a schedule or take action on various events. uAgents are easy to create and manage, and they are connected to a fast-growing network of other uAgents. They are also secure, with cryptographically secured messages and wallets.
griptape
Griptape is a modular Python framework for building AI-powered applications that securely connect to your enterprise data and APIs. It offers developers the ability to maintain control and flexibility at every step. Griptape's core components include Structures (Agents, Pipelines, and Workflows), Tasks, Tools, Memory (Conversation Memory, Task Memory, and Meta Memory), Drivers (Prompt and Embedding Drivers, Vector Store Drivers, Image Generation Drivers, Image Query Drivers, SQL Drivers, Web Scraper Drivers, and Conversation Memory Drivers), Engines (Query Engines, Extraction Engines, Summary Engines, Image Generation Engines, and Image Query Engines), and additional components (Rulesets, Loaders, Artifacts, Chunkers, and Tokenizers). Griptape enables developers to create AI-powered applications with ease and efficiency.