Best AI tools for< Ai Safety Researcher >
Infographic
20 - AI tool Sites

Center for AI Safety (CAIS)
The Center for AI Safety (CAIS) is a research and field-building nonprofit based in San Francisco. Their mission is to reduce societal-scale risks associated with artificial intelligence (AI) by conducting impactful research, building the field of AI safety researchers, and advocating for safety standards. They offer resources such as a compute cluster for AI/ML safety projects, a blog with in-depth examinations of AI safety topics, and a newsletter providing updates on AI safety developments. CAIS focuses on technical and conceptual research to address the risks posed by advanced AI systems.

Center for AI Safety (CAIS)
The Center for AI Safety (CAIS) is a research and field-building nonprofit organization based in San Francisco. They conduct impactful research, advocacy projects, and provide resources to reduce societal-scale risks associated with artificial intelligence (AI). CAIS focuses on technical AI safety research, field-building projects, and offers a compute cluster for AI/ML safety projects. They aim to develop and use AI safely to benefit society, addressing inherent risks and advocating for safety standards.

blog.biocomm.ai
blog.biocomm.ai is an AI safety blog that focuses on the existential threat posed by uncontrolled and uncontained AI technology. It curates and organizes information related to AI safety, including the risks and challenges associated with the proliferation of AI. The blog aims to educate and raise awareness about the importance of developing safe and regulated AI systems to ensure the survival of humanity.

Inspect
Inspect is an open-source framework for large language model evaluations created by the UK AI Safety Institute. It provides built-in components for prompt engineering, tool usage, multi-turn dialog, and model graded evaluations. Users can explore various solvers, tools, scorers, datasets, and models to create advanced evaluations. Inspect supports extensions for new elicitation and scoring techniques through Python packages.

Frontier Model Forum
The Frontier Model Forum (FMF) is a collaborative effort among leading AI companies to advance AI safety and responsibility. The FMF brings together technical and operational expertise to identify best practices, conduct research, and support the development of AI applications that meet society's most pressing needs. The FMF's core objectives include advancing AI safety research, identifying best practices, collaborating across sectors, and helping AI meet society's greatest challenges.

Anthropic
Anthropic is an AI safety and research company based in San Francisco. Our interdisciplinary team has experience across ML, physics, policy, and product. Together, we generate research and create reliable, beneficial AI systems.

Athina AI
Athina AI is a platform that provides research and guides for building safe and reliable AI products. It helps thousands of AI engineers in building safer products by offering tutorials, research papers, and evaluation techniques related to large language models. The platform focuses on safety, prompt engineering, hallucinations, and evaluation of AI models.

Granica AI
Granica AI is an AI Data Readiness Platform that helps users build and manage high-quality data for AI at scale. The platform uses AI to continuously improve the AI-readiness of data, making projects faster and more impactful over time. Granica offers solutions for data cost optimization, data privacy, data selection & curation, and research. The platform is trusted by category-defining companies and has been recognized in various industry awards and publications.

Faculty AI
Faculty AI is a leading applied AI consultancy and technology provider, specializing in helping customers transform their businesses through bespoke AI consultancy and Frontier, the world's first AI operating system. They offer services such as AI consultancy, generative AI solutions, and AI services tailored to various industries. Faculty AI is known for its expertise in AI governance and safety, as well as its partnerships with top AI platforms like OpenAI, AWS, and Microsoft.

OpenAI Strawberry Model
OpenAI Strawberry Model is a cutting-edge AI initiative that represents a significant leap in AI capabilities, focusing on enhancing reasoning, problem-solving, and complex task execution. It aims to improve AI's ability to handle mathematical problems, programming tasks, and deep research, including long-term planning and action. The project showcases advancements in AI safety and aims to reduce errors in AI responses by generating high-quality synthetic data for training future models. Strawberry is designed to achieve human-like reasoning and is expected to play a crucial role in the development of OpenAI's next major model, codenamed 'Orion.'

BoodleBox
BoodleBox is a platform for group collaboration with generative AI (GenAI) tools like ChatGPT, GPT-4, and hundreds of others. It allows teams to connect multiple bots, people, and sources of knowledge in one chat to keep discussions engaging, productive, and educational. BoodleBox also provides access to over 800 specialized AI bots and offers easy team management and billing to simplify access and usage across departments, teams, and organizations.

MIRI (Machine Intelligence Research Institute)
MIRI (Machine Intelligence Research Institute) is a non-profit research organization dedicated to ensuring that artificial intelligence has a positive impact on humanity. MIRI conducts foundational mathematical research on topics such as decision theory, game theory, and reinforcement learning, with the goal of developing new insights into how to build safe and beneficial AI systems.

Imbue
Imbue is a company focused on building AI systems that can reason and code, with the goal of rekindling the dream of the personal computer by creating practical AI agents that can accomplish larger goals and work safely in the real world. The company emphasizes innovation in AI technology and aims to push the boundaries of what AI can achieve in various fields.

O.XYZ
The O.XYZ website is a platform dedicated to decentralized super intelligence, governance, research, and innovation for the O Ecosystem. It focuses on cutting-edge AI research, community-driven allocation, and treasury management. The platform aims to empower the O Ecosystem through decentralization and responsible, safe, and ethical development of Super AI.

Aify.co
Aify.co is a website that covers all things artificial intelligence. It provides news, analysis, and opinion on the latest developments in AI, as well as resources for developers and users. The site is written by a team of experts in AI, and it is committed to providing accurate and up-to-date information on the field.

Lamini
Lamini is an enterprise-level LLM platform that offers precise recall with Memory Tuning, enabling teams to achieve over 95% accuracy even with large amounts of specific data. It guarantees JSON output and delivers massive throughput for inference. Lamini is designed to be deployed anywhere, including air-gapped environments, and supports training and inference on Nvidia or AMD GPUs. The platform is known for its factual LLMs and reengineered decoder that ensures 100% schema accuracy in the JSON output.

Texthub
Texthub is an AI-powered chatbot that allows users to engage in NSFW conversations and role-play scenarios. It is designed to provide a realistic and immersive experience, with the AI responding in a natural and engaging way. Texthub also offers a variety of features to enhance the user experience, such as the ability to customize the AI's appearance and personality.

Binah.ai
Binah.ai is an AI-powered Health Data Platform that offers a software solution for video-based vital signs monitoring. The platform enables users to measure various health and wellness indicators using a smartphone, tablet, or laptop. It provides support for continuous monitoring through a raw PPG signal from external sensors and offers a range of features such as blood pressure monitoring, heart rate variability, oxygen saturation, and more. Binah.ai aims to make health data more accessible for better care at lower costs by leveraging AI and deep learning algorithms.

Figure
Figure is a pioneering AI robotics company that is revolutionizing the industry by introducing a general-purpose humanoid robot to the global workforce. By combining cutting-edge AI technology with the dexterity of the human form, Figure aims to enhance human capabilities, address labor shortages, and improve workplace safety. The company's innovative approach is set to transform various sectors such as manufacturing, logistics, warehousing, and retail. With a team of experts boasting over 100 years of combined experience in AI and humanoid robotics, Figure is at the forefront of shaping the future of work.

DrugCard
DrugCard is an AI-enabled Data Intelligence platform designed to streamline drug safety routines by providing continuous and regular, traceable, transparent, accurate, complete, and scalable solutions for pharmacovigilance processes. The platform supports over 100 languages, monitors 1000+ local journals, and saves up to 70% of time compared to manual approaches. DrugCard caters to CROs, MAHs, and freelancers, offering efficient local literature screening services to ensure compliance with regulatory requirements and enhance productivity.
20 - Open Source Tools

alignment-handbook
The Alignment Handbook provides robust training recipes for continuing pretraining and aligning language models with human and AI preferences. It includes techniques such as continued pretraining, supervised fine-tuning, reward modeling, rejection sampling, and direct preference optimization (DPO). The handbook aims to fill the gap in public resources on training these models, collecting data, and measuring metrics for optimal downstream performance.

Awesome-System2-Reasoning-LLM
The Awesome-System2-Reasoning-LLM repository is dedicated to a survey paper titled 'From System 1 to System 2: A Survey of Reasoning Large Language Models'. It explores the development of reasoning Large Language Models (LLMs), their foundational technologies, benchmarks, and future directions. The repository provides resources and updates related to the research, tracking the latest developments in the field of reasoning LLMs.

Awesome-LLM-Safety
Welcome to our Awesome-llm-safety repository! We've curated a collection of the latest, most comprehensive, and most valuable resources on large language model safety (llm-safety). But we don't stop there; included are also relevant talks, tutorials, conferences, news, and articles. Our repository is constantly updated to ensure you have the most current information at your fingertips.

inspect_ai
Inspect AI is a framework developed by the UK AI Safety Institute for evaluating large language models. It offers various built-in components for prompt engineering, tool usage, multi-turn dialog, and model graded evaluations. Users can extend Inspect by adding new elicitation and scoring techniques through additional Python packages. The tool aims to provide a comprehensive solution for assessing the performance and safety of language models.

awesome-generative-ai
A curated list of Generative AI projects, tools, artworks, and models

awesome-generative-ai-apis
Awesome Generative AI & LLM APIs is a curated list of useful APIs that allow developers to integrate generative models into their applications without building the models from scratch. These APIs provide an interface for generating text, images, or other content, and include pre-trained language models for various tasks. The goal of this project is to create a hub for developers to create innovative applications, enhance user experiences, and drive progress in the AI field.

FigStep
FigStep is a black-box jailbreaking algorithm against large vision-language models (VLMs). It feeds harmful instructions through the image channel and uses benign text prompts to induce VLMs to output contents that violate common AI safety policies. The tool highlights the vulnerability of VLMs to jailbreaking attacks, emphasizing the need for safety alignments between visual and textual modalities.

awesome-open-ended
A curated list of open-ended learning AI resources focusing on algorithms that invent new and complex tasks endlessly, inspired by human advancements. The repository includes papers, safety considerations, surveys, perspectives, and blog posts related to open-ended AI research.

modelbench
ModelBench is a tool for running safety benchmarks against AI models and generating detailed reports. It is part of the MLCommons project and is designed as a proof of concept to aggregate measures, relate them to specific harms, create benchmarks, and produce reports. The tool requires LlamaGuard for evaluating responses and a TogetherAI account for running benchmarks. Users can install ModelBench from GitHub or PyPI, run tests using Poetry, and create benchmarks by providing necessary API keys. The tool generates static HTML pages displaying benchmark scores and allows users to dump raw scores and manage cache for faster runs. ModelBench is aimed at enabling users to test their own models and create tests and benchmarks.

awesome_LLM-harmful-fine-tuning-papers
This repository is a comprehensive survey of harmful fine-tuning attacks and defenses for large language models (LLMs). It provides a curated list of must-read papers on the topic, covering various aspects such as alignment stage defenses, fine-tuning stage defenses, post-fine-tuning stage defenses, mechanical studies, benchmarks, and attacks/defenses for federated fine-tuning. The repository aims to keep researchers updated on the latest developments in the field and offers insights into the vulnerabilities and safeguards related to fine-tuning LLMs.

driverlessai-recipes
This repository contains custom recipes for H2O Driverless AI, which is an Automatic Machine Learning platform for the Enterprise. Custom recipes are Python code snippets that can be uploaded into Driverless AI at runtime to automate feature engineering, model building, visualization, and interpretability. Users can gain control over the optimization choices made by Driverless AI by providing their own custom recipes. The repository includes recipes for various tasks such as data manipulation, data preprocessing, feature selection, data augmentation, model building, scoring, and more. Best practices for creating and using recipes are also provided, including security considerations, performance tips, and safety measures.

AI-Security-and-Privacy-Events
AI-Security-and-Privacy-Events is a curated list of academic events focusing on AI security and privacy. It includes seminars, conferences, workshops, tutorials, special sessions, and covers various topics such as NLP & LLM Security, Privacy and Security in ML, Machine Learning Security, AI System with Confidential Computing, Adversarial Machine Learning, and more.

HybridAGI
HybridAGI is the first Programmable LLM-based Autonomous Agent that lets you program its behavior using a **graph-based prompt programming** approach. This state-of-the-art feature allows the AGI to efficiently use any tool while controlling the long-term behavior of the agent. Become the _first Prompt Programmers in history_ ; be a part of the AI revolution one node at a time! **Disclaimer: We are currently in the process of upgrading the codebase to integrate DSPy**

awesome-ai-tools
Awesome AI Tools is a curated list of popular tools and resources for artificial intelligence enthusiasts. It includes a wide range of tools such as machine learning libraries, deep learning frameworks, data visualization tools, and natural language processing resources. Whether you are a beginner or an experienced AI practitioner, this repository aims to provide you with a comprehensive collection of tools to enhance your AI projects and research. Explore the list to discover new tools, stay updated with the latest advancements in AI technology, and find the right resources to support your AI endeavors.

100days_AI
The 100 Days in AI repository provides a comprehensive roadmap for individuals to learn Artificial Intelligence over a period of 100 days. It covers topics ranging from basic programming in Python to advanced concepts in AI, including machine learning, deep learning, and specialized AI topics. The repository includes daily tasks, resources, and exercises to ensure a structured learning experience. By following this roadmap, users can gain a solid understanding of AI and be prepared to work on real-world AI projects.

awesome-llm-security
Awesome LLM Security is a curated collection of tools, documents, and projects related to Large Language Model (LLM) security. It covers various aspects of LLM security including white-box, black-box, and backdoor attacks, defense mechanisms, platform security, and surveys. The repository provides resources for researchers and practitioners interested in understanding and safeguarding LLMs against adversarial attacks. It also includes a list of tools specifically designed for testing and enhancing LLM security.

Awesome-Attention-Heads
Awesome-Attention-Heads is a platform providing the latest research on Attention Heads, focusing on enhancing understanding of Transformer structure for model interpretability. It explores attention mechanisms for behavior, inference, and analysis, alongside feed-forward networks for knowledge storage. The repository aims to support researchers studying LLM interpretability and hallucination by offering cutting-edge information on Attention Head Mining.

Awesome-LLM-Post-training
The Awesome-LLM-Post-training repository is a curated collection of influential papers, code implementations, benchmarks, and resources related to Large Language Models (LLMs) Post-Training Methodologies. It covers various aspects of LLMs, including reasoning, decision-making, reinforcement learning, reward learning, policy optimization, explainability, multimodal agents, benchmarks, tutorials, libraries, and implementations. The repository aims to provide a comprehensive overview and resources for researchers and practitioners interested in advancing LLM technologies.
20 - OpenAI Gpts

GPT Safety Liaison
A liaison GPT for AI safety emergencies, connecting users to OpenAI experts.

AI Ethica Readify
Summarises AI ethics papers, provides context, and offers further assistance.

Alignment Navigator
AI Alignment guided by interdisciplinary wisdom and a future-focused vision.

Prompt Injection Detector
GPT used to classify prompts as valid inputs or injection attempts. Json output.

Individual Intelligence Oriented Alignment
Ask this AI anything about alignment and it will give the best scenario the superintelligence should do according to its Alignment Principals.

Chemistry Expert
Advanced AI for chemistry, offering innovative solutions, process optimizations, and safety assessments, powered by OpenAI.

Töökeskkonna spetialist Eestis 2024
Sa tead kõike töökeskkonnast ja seadustest mis sellega kaasnevad

香港地盤安全佬 HK Construction Site Safety Advisor
Upload a site photo to assess the potential hazard and seek advises from experience AI Safety Officer

PharmaFinder AI
Identifies medications and active ingredients from photos for user safety.

Buildwell AI - UK Construction Regs Assistant
Provides Construction Support relating to Planning Permission, Building Regulations, Party Wall Act and Fire Safety in the UK. Obtain instant Guidance for your Construction Project.

TrafficFlow
A specialized AI for optimizing traffic control, predicting bottlenecks, and improving road safety.

DateMate
Your friendly AI assistant for voice-based dating, offering personalized tips, safety advice, and fun interactions.

Hazard Analyst
Generates risk maps, emergency response plans and safety protocols for disaster management professionals.