Best AI tools for< Identify Safety Risks >
20 - AI tool Sites

Frontier Model Forum
The Frontier Model Forum (FMF) is a collaborative effort among leading AI companies to advance AI safety and responsibility. The FMF brings together technical and operational expertise to identify best practices, conduct research, and support the development of AI applications that meet society's most pressing needs. The FMF's core objectives include advancing AI safety research, identifying best practices, collaborating across sectors, and helping AI meet society's greatest challenges.

ReviewGPT
ReviewGPT is an AI-powered tool that helps users analyze Amazon products and reviews to make informed purchasing decisions. It utilizes AI to identify counterfeit products, pirated books, fake reviews, unreliable third-party sellers, and products with potential safety or health risks. By leveraging ReviewGPT, users can save time and money while ensuring they purchase genuine and high-quality products from Amazon.

SWMS AI
SWMS AI is an AI-powered safety risk assessment tool that helps businesses streamline compliance and improve safety. It leverages a vast knowledge base of occupational safety resources, codes of practice, risk assessments, and safety documents to generate risk assessments tailored specifically to a project, trade, and industry. SWMS AI can be customized to a company's policies to align its AI's document generation capabilities with proprietary safety standards and requirements.

Limbic
Limbic is a clinical AI application designed for mental healthcare providers to save time, improve outcomes, and maximize impact. It offers a suite of tools developed by a team of therapists, physicians, and PhDs in computational psychiatry. Limbic is known for its evidence-based approach, safety focus, and commitment to patient care. The application leverages AI technology to enhance various aspects of the mental health pathway, from assessments to therapeutic content delivery. With a strong emphasis on patient safety and clinical accuracy, Limbic aims to support clinicians in meeting the rising demand for mental health services while improving patient outcomes and preventing burnout.

Fordi
Fordi is an AI management tool that helps businesses avoid risks in real-time. It provides a comprehensive view of all AI systems, allowing businesses to identify and mitigate risks before they cause damage. Fordi also provides continuous monitoring and alerting, so businesses can be sure that their AI systems are always operating safely.

VirtuSense Technologies
VirtuSense Technologies is a leading provider of fall prevention and remote patient monitoring (RPM) solutions powered by artificial intelligence (AI). Their AI-driven solutions, VSTAlert and VSTBalance, are designed to help healthcare providers reduce falls, improve patient safety, and enhance care delivery. VSTAlert is a fall prevention system that uses AI to detect falls before they happen, reducing the risk of injury and improving patient outcomes. VSTBalance is a balance assessment tool that helps clinicians identify patients at risk of falling and provides personalized exercises to improve their balance and mobility. VirtuSense's solutions integrate with various healthcare systems and are used by hospitals, post-acute care facilities, and ambulatory care centers to improve patient care and reduce costs.

MAIHEM
MAIHEM is an AI-powered quality assurance platform that helps businesses test and improve the performance and safety of their AI applications. It automates the testing process, generates realistic test cases, and provides comprehensive analytics to help businesses identify and fix potential issues. MAIHEM is used by a variety of businesses, including those in the customer support, healthcare, education, and sales industries.

Prompt Hippo
Prompt Hippo is an AI tool designed as a side-by-side LLM prompt testing suite to ensure the robustness, reliability, and safety of prompts. It saves time by streamlining the process of testing LLM prompts and allows users to test custom agents and optimize them for production. With a focus on science and efficiency, Prompt Hippo helps users identify the best prompts for their needs.

Insitro
Insitro is a drug discovery and development company that uses machine learning and data to identify and develop new medicines. The company's platform integrates in vitro cellular data produced in its labs with human clinical data to help redefine disease. Insitro's pipeline includes wholly-owned and partnered therapeutic programs in metabolism, oncology, and neuroscience.

Link Shield
Link Shield is an AI-powered malicious URL detection API platform that helps protect online security. It utilizes advanced machine learning algorithms to analyze URLs and identify suspicious activity, safeguarding users from phishing scams, malware, and other harmful threats. The API is designed for ease of integration, affordability, and flexibility, making it accessible to developers of all levels. Link Shield empowers businesses to ensure the safety and security of their applications and online communities.

FaceCheck.ID
FaceCheck.ID is a facial recognition AI technology-powered search engine that allows users to upload a photo of a person to discover their social media profiles, appearances in blogs, videos, news websites, and more. It helps users verify the authenticity of individuals, avoid dangerous criminals, keep their families safe, and avoid becoming victims of various scams and crimes. The tool is designed to assist in identifying and uncovering information about individuals based on their facial features, with a focus on safety and security.

Robot Challenge Screen
Aimodelagency.com is an AI tool that offers a Robot Challenge Screen for checking site connection security. Users can verify the security of their website by enabling cookies in their browser settings. The tool helps in identifying any potential security vulnerabilities and ensures a safe browsing experience for visitors.

Regard
Regard is an AI-powered healthcare solution that automates clinical tasks, making it easier for clinicians to focus on patient care. It integrates with the EHR to analyze patient records and provide insights that can help improve diagnosis and treatment. Regard has been shown to improve hospital finances, patient safety, and physician happiness.

TripleWhale
TripleWhale is a website that provides security services to protect itself from online attacks. It uses Cloudflare to block unauthorized access and ensure the safety of the website. Users may encounter blocks due to various triggers such as submitting specific words or phrases, SQL commands, or malformed data. In such cases, users can contact the site owner to resolve the issue by providing details of the blocked action and the Cloudflare Ray ID.

Ignota Labs
Ignota Labs is a technology company focused on rescuing failing drugs and bringing new life to abandoned projects, ultimately providing hope to patients. The company utilizes a proprietary AI model, SAFEPATH, which applies deep learning to bioinformatics and cheminformatics datasets to solve drug safety issues. Ignota Labs aims to identify promising drug targets, address safety problems in clinical trials, and accelerate the delivery of therapeutically effective drugs to patients.

Valossa
Valossa is an AI video analysis tool that transcribes videos to text metadata, captions, and clips. It offers a range of AI-powered features such as automating captions, content logging, brand-safe contextual advertising, clip promo videos, identify sensitive content, and analyze video moods and sentiment. Valossa's AI capabilities include speech-to-text, computer vision, emotion analysis, and metadata generation, enabling users to accelerate video productivity with cognitive automation.

Yuna
Yuna is an AI-powered mental health companion designed to support users in navigating life's challenges and improving their well-being. It offers 24/7 emotional support, calming techniques for anxiety management, and personalized guidance based on cognitive behavioral therapy principles. Yuna prioritizes user privacy and safety, utilizing real-time monitoring to identify signs of distress and connect users with qualified healthcare professionals when necessary.

Aura
Aura is an all-in-one digital safety platform that uses artificial intelligence (AI) to protect your family online. It offers a wide range of features, including financial fraud protection, identity theft protection, VPN & online privacy, antivirus, password manager & smart vault, parental controls & safe gaming, and spam call protection. Aura is easy to use and affordable, and it comes with a 60-day money-back guarantee.

Seventh Sense
Seventh Sense is an AI company focused on providing cutting-edge AI solutions for secure and private identity verification. Their innovative technologies, such as SenseCrypt, OpenCV FR, and SenseVantage, offer advanced biometric verification, face recognition, and AI video analysis. With a mission to make self-sovereign identity accessible to all, Seventh Sense ensures privacy, security, and compliance through their AI algorithms and cryptographic solutions.

Dust
Dust is a customizable and secure AI assistant platform that helps businesses amplify their team's potential. It allows users to deploy the best Large Language Models to their company, connect Dust to their team's data, and empower their teams with assistants tailored to their specific needs. Dust is exceptionally modular and adaptable, tailoring to unique requirements and continuously evolving to meet changing needs. It supports multiple sources of data and models, including proprietary and open-source models from OpenAI, Anthropic, and Mistral. Dust also helps businesses identify their most creative and driven team members and share their experience with AI throughout the company. It promotes collaboration with shared conversations, @mentions in discussions, and Slackbot integration. Dust prioritizes security and data privacy, ensuring that data remains private and that enterprise-grade security measures are in place to manage data access policies.
20 - Open Source AI Tools

awesome_LLM-harmful-fine-tuning-papers
This repository is a comprehensive survey of harmful fine-tuning attacks and defenses for large language models (LLMs). It provides a curated list of must-read papers on the topic, covering various aspects such as alignment stage defenses, fine-tuning stage defenses, post-fine-tuning stage defenses, mechanical studies, benchmarks, and attacks/defenses for federated fine-tuning. The repository aims to keep researchers updated on the latest developments in the field and offers insights into the vulnerabilities and safeguards related to fine-tuning LLMs.

R-Judge
R-Judge is a benchmarking tool designed to evaluate the proficiency of Large Language Models (LLMs) in judging and identifying safety risks within diverse environments. It comprises 569 records of multi-turn agent interactions, covering 27 key risk scenarios across 5 application categories and 10 risk types. The tool provides high-quality curation with annotated safety labels and risk descriptions. Evaluation of 11 LLMs on R-Judge reveals the need for enhancing risk awareness in LLMs, especially in open agent scenarios. Fine-tuning on safety judgment is found to significantly improve model performance.

do-not-answer
Do-Not-Answer is an open-source dataset curated to evaluate Large Language Models' safety mechanisms at a low cost. It consists of prompts to which responsible language models do not answer. The dataset includes human annotations and model-based evaluation using a fine-tuned BERT-like evaluator. The dataset covers 61 specific harms and collects 939 instructions across five risk areas and 12 harm types. Response assessment is done for six models, categorizing responses into harmfulness and action categories. Both human and automatic evaluations show the safety of models across different risk areas. The dataset also includes a Chinese version with 1,014 questions for evaluating Chinese LLMs' risk perception and sensitivity to specific words and phrases.

agentic_security
Agentic Security is an open-source vulnerability scanner designed for safety scanning, offering customizable rule sets and agent-based attacks. It provides comprehensive fuzzing for any LLMs, LLM API integration, and stress testing with a wide range of fuzzing and attack techniques. The tool is not a foolproof solution but aims to enhance security measures against potential threats. It offers installation via pip and supports quick start commands for easy setup. Users can utilize the tool for LLM integration, adding custom datasets, running CI checks, extending dataset collections, and dynamic datasets with mutations. The tool also includes a probe endpoint for integration testing. The roadmap includes expanding dataset variety, introducing new attack vectors, developing an attacker LLM, and integrating OWASP Top 10 classification.

DeRTa
DeRTa (Refuse Whenever You Feel Unsafe) is a tool designed to improve safety in Large Language Models (LLMs) by training them to refuse compliance at any response juncture. The tool incorporates methods such as MLE with Harmful Response Prefix and Reinforced Transition Optimization (RTO) to address refusal positional bias and strengthen the model's capability to transition from potential harm to safety refusal. DeRTa provides training data, model weights, and evaluation scripts for LLMs, enabling users to enhance safety in language generation tasks.

watchtower
AIShield Watchtower is a tool designed to fortify the security of AI/ML models and Jupyter notebooks by automating model and notebook discoveries, conducting vulnerability scans, and categorizing risks into 'low,' 'medium,' 'high,' and 'critical' levels. It supports scanning of public GitHub repositories, Hugging Face repositories, AWS S3 buckets, and local systems. The tool generates comprehensive reports, offers a user-friendly interface, and aligns with industry standards like OWASP, MITRE, and CWE. It aims to address the security blind spots surrounding Jupyter notebooks and AI models, providing organizations with a tailored approach to enhancing their security efforts.

ABigSurveyOfLLMs
ABigSurveyOfLLMs is a repository that compiles surveys on Large Language Models (LLMs) to provide a comprehensive overview of the field. It includes surveys on various aspects of LLMs such as transformers, alignment, prompt learning, data management, evaluation, societal issues, safety, misinformation, attributes of LLMs, efficient LLMs, learning methods for LLMs, multimodal LLMs, knowledge-based LLMs, extension of LLMs, LLMs applications, and more. The repository aims to help individuals quickly understand the advancements and challenges in the field of LLMs through a collection of recent surveys and research papers.

awesome-artificial-intelligence-guidelines
The 'Awesome AI Guidelines' repository aims to simplify the ecosystem of guidelines, principles, codes of ethics, standards, and regulations around artificial intelligence. It provides a comprehensive collection of resources addressing ethical and societal challenges in AI systems, including high-level frameworks, principles, processes, checklists, interactive tools, industry standards initiatives, online courses, research, and industry newsletters, as well as regulations and policies from various countries. The repository serves as a valuable reference for individuals and teams designing, building, and operating AI systems to navigate the complex landscape of AI ethics and governance.

AwesomeResponsibleAI
Awesome Responsible AI is a curated list of academic research, books, code of ethics, courses, data sets, frameworks, institutes, newsletters, principles, podcasts, reports, tools, regulations, and standards related to Responsible, Trustworthy, and Human-Centered AI. It covers various concepts such as Responsible AI, Trustworthy AI, Human-Centered AI, Responsible AI frameworks, AI Governance, and more. The repository provides a comprehensive collection of resources for individuals interested in ethical, transparent, and accountable AI development and deployment.

prompt-injection-defenses
This repository provides a collection of tools and techniques for defending against injection attacks in software applications. It includes code samples, best practices, and guidelines for implementing secure coding practices to prevent common injection vulnerabilities such as SQL injection, XSS, and command injection. The tools and resources in this repository aim to help developers build more secure and resilient applications by addressing one of the most common and critical security threats in modern software development.

swarms
Swarms provides simple, reliable, and agile tools to create your own Swarm tailored to your specific needs. Currently, Swarms is being used in production by RBC, John Deere, and many AI startups.

llms-tools
The 'llms-tools' repository is a comprehensive collection of AI tools, open-source projects, and research related to Large Language Models (LLMs) and Chatbots. It covers a wide range of topics such as AI in various domains, open-source models, chats & assistants, visual language models, evaluation tools, libraries, devices, income models, text-to-image, computer vision, audio & speech, code & math, games, robotics, typography, bio & med, military, climate, finance, and presentation. The repository provides valuable resources for researchers, developers, and enthusiasts interested in exploring the capabilities of LLMs and related technologies.

MedLLMsPracticalGuide
This repository serves as a practical guide for Medical Large Language Models (Medical LLMs) and provides resources, surveys, and tools for building, fine-tuning, and utilizing LLMs in the medical domain. It covers a wide range of topics including pre-training, fine-tuning, downstream biomedical tasks, clinical applications, challenges, future directions, and more. The repository aims to provide insights into the opportunities and challenges of LLMs in medicine and serve as a practical resource for constructing effective medical LLMs.
20 - OpenAI Gpts

Brand Safety Audit
Get a detailed risk analysis for public relations, marketing, and internal communications, identifying challenges and negative impacts to refine your messaging strategy.

The Building Safety Act Bot (Beta)
Simplifying the BSA for your project. Created by www.arka.works

Hazard Analyst
Generates risk maps, emergency response plans and safety protocols for disaster management professionals.

Individual Intelligence Oriented Alignment
Ask this AI anything about alignment and it will give the best scenario the superintelligence should do according to its Alignment Principals.

Together
GPT for drug interactions. Enter at least two medication names to learn about potential drug interactions.

PharmaFinder AI
Identifies medications and active ingredients from photos for user safety.

Chemistry Expert
Advanced AI for chemistry, offering innovative solutions, process optimizations, and safety assessments, powered by OpenAI.