Best AI tools for< Ethics Officer >
Infographic
20 - AI tool Sites
Kodora AI
Kodora AI is a leading AI technology and advisory firm based in Australia, specializing in providing end-to-end AI services. They offer AI strategy development, use case identification, workforce AI training, and more. With a team of expert AI engineers and consultants, Kodora focuses on delivering practical outcomes for clients across various industries. The firm is known for its deep expertise, solution-focused approach, and commitment to driving AI adoption and innovation.
OneTrust
OneTrust is a Trust Intelligence Cloud Solutions platform that leverages data and artificial intelligence to drive trusted innovation across privacy, security, and ethics initiatives. It offers a comprehensive suite of solutions for privacy management, data discovery, security, consent and preferences, AI governance, technology risk and compliance, compliance automation, third-party risk, and ethics program management. With over 14,000 customers worldwide, OneTrust helps organizations manage risk, ensure compliance, and build trust through responsible data and AI usage.
AIGA AI Governance Framework
The AIGA AI Governance Framework is a practice-oriented framework for implementing responsible AI. It provides organizations with a systematic approach to AI governance, covering the entire process of AI system development and operations. The framework supports compliance with the upcoming European AI regulation and serves as a practical guide for organizations aiming for more responsible AI practices. It is designed to facilitate the development and deployment of transparent, accountable, fair, and non-maleficent AI systems.
Alphy
Alphy is a modern AI tool for communication compliance that helps companies detect and prevent harmful and unlawful language in their communication. The AI classifier has a 94% accuracy rate and can identify over 40 high-risk categories of harmful language. By using Reflect AI, companies can shield themselves from reputational, ethical, and legal risks, ensuring compliance and preventing costly litigation.
Future of Privacy Forum
The Future of Privacy Forum (FPF) is an AI tool that serves as a catalyst for privacy leadership and scholarship, advancing principled data practices in support of emerging technologies. It provides resources, training sessions, and guidance on AI-related topics, online advertising, youth privacy legislation, and more. FPF brings together industry, academics, civil society, policymakers, and other stakeholders to explore challenges posed by emerging technologies and develop privacy protections, ethical norms, and best practices.
blog.biocomm.ai
blog.biocomm.ai is an AI safety blog that focuses on the existential threat posed by uncontrolled and uncontained AI technology. It curates and organizes information related to AI safety, including the risks and challenges associated with the proliferation of AI. The blog aims to educate and raise awareness about the importance of developing safe and regulated AI systems to ensure the survival of humanity.
GOODY-2
GOODY-2 is the world's most responsible AI model, built with next-gen adherence to ethical principles. It's so safe that it won't answer anything that could be possibly construed as controversial or problematic. GOODY-2 can recognize any query that could be controversial, offensive, or dangerous in any context and elegantly avoids answering, redirecting the conversation, and mitigating brand risk. GOODY-2's ethical adherence is unbreakable, ensuring that every conversation stays within the bounds of ethical principles. Even bad actors will find themselves unable to cause GOODY-2 to answer problematic queries. GOODY-2 is the perfect fit for customer service, paralegal assistance, back-office tasks, and more. It's the safe, dependable AI model companies around the globe have been waiting for.
Robust Intelligence
Robust Intelligence is an end-to-end security solution for AI applications. It automates the evaluation of AI models, data, and files for security and safety vulnerabilities and provides guardrails for AI applications in production against integrity, privacy, abuse, and availability violations. Robust Intelligence helps enterprises remove AI security blockers, save time and resources, meet AI safety and security standards, align AI security across stakeholders, and protect against evolving threats.
Credo AI
Credo AI is a leading provider of AI governance, risk management, and compliance software. Our platform helps organizations to adopt AI safely and responsibly, while ensuring compliance with regulations and standards. With Credo AI, you can track and prioritize AI projects, assess AI vendor models for risk and compliance, create artifacts for audit, and more.
Defined.ai
Defined.ai is a leading provider of high-quality and ethical data for AI applications. Founded in 2015, Defined.ai has a global presence with offices in the US, Europe, and Asia. The company's mission is to make AI more accessible and ethical by providing a marketplace for buying and selling AI data, tools, and models. Defined.ai also offers professional services to help deliver success in complex machine learning projects.
Montreal AI Ethics Institute
The Montreal AI Ethics Institute (MAIEI) is an international non-profit organization founded in 2018, dedicated to democratizing AI ethics literacy. It equips citizens concerned about artificial intelligence and its impact on society to take action through research summaries, columns, and AI applications in various fields.
Trustworthy AI
Trustworthy AI is a business guide that focuses on navigating trust and ethics in artificial intelligence. Authored by Beena Ammanath, a global thought leader in AI ethics, the book provides practical guidelines for organizations developing or using AI solutions. It addresses the importance of AI systems adhering to social norms and ethics, making fair decisions in a consistent, transparent, explainable, and unbiased manner. Trustworthy AI offers readers a structured approach to thinking about AI ethics and trust, emphasizing the need for ethical considerations in the rapidly evolving landscape of AI technology.
edu720
edu720 is a science-backed learning platform that uses AI and nanolearning to redefine how workforces learn and achieve their goals. It provides pre-built learning modules on various topics, including cybersecurity, privacy, and AI ethics. edu720's 360-degree approach ensures that all employees, regardless of their status or location, fully understand and absorb the knowledge conveyed.
International Journal for Educational Integrity
The International Journal for Educational Integrity is an AI tool that focuses on publishing articles related to academic integrity, ethics, and plagiarism. It features original research articles, reviews, and thematic collections on topics such as machine-based plagiarism, contract cheating, and the impact of emergencies on educational integrity. The journal aims to address emerging threats to academic integrity and promote ethical practices in education.
Singularity Weblog
Singularity Weblog is a platform dedicated to exploring the future of technology, focusing on exponential growth, artificial intelligence, and ethics. It provides insights into identifying potential dangers and opportunities, envisioning a better future, and nurturing creativity. The platform offers a podcast with expert interviews, a blog to stay informed about AI and other technologies, and resources for hiring keynote speakers. Through engaging conversations and thought-provoking content, Singularity Weblog aims to inspire individuals to embrace and shape the future.
Coalition for Health AI (CHAI)
The Coalition for Health AI (CHAI) is an AI application that provides guidelines for the responsible use of AI in health. It focuses on developing best practices and frameworks for safe and equitable AI in healthcare. CHAI aims to address algorithmic bias and collaborates with diverse stakeholders to drive the development, evaluation, and appropriate use of AI in healthcare.
Accel.AI
Accel.AI is an institute founded in 2016 with a mission to drive artificial intelligence for social impact initiatives. They focus on integrating AI and social impact through research, consulting, and workshops on ethical AI development and applied AI engineering. The institute targets underrepresented groups, tech companies, governments, and individuals experiencing job loss due to automation. They work globally with companies, professionals, and students.
Salesforce AI Blog
Salesforce AI Blog is an AI tool that focuses on various AI research topics such as accountability, accuracy, AI agents, AI coding, AI ethics, AI object detection, deep learning, forecasting, generative AI, and more. The blog showcases cutting-edge research, advancements, and projects in the field of artificial intelligence. It also highlights the work of Salesforce Research team members and their contributions to the AI community.
AI Elections Accord
AI Elections Accord is a tech accord aimed at combating the deceptive use of AI in the 2024 elections. It sets expectations for managing risks related to deceptive AI election content on large-scale platforms. The accord focuses on prevention, provenance, detection, responsive protection, evaluation, public awareness, and resilience to safeguard the democratic process. It emphasizes collective efforts, education, and the development of defensive tools to protect public debate and build societal resilience against deceptive AI content.
Fritz AI
Fritz AI is an AI tool that scans and ranks all AI tools, apps, and websites based on a set of criteria to determine the best and most ethical options. They provide technical guides, reviews, and tutorials to help users get started with machine learning. Fritz AI focuses on ethics, functionality, user experience, and innovation when evaluating tools. Users can contribute tool suggestions and collaborate with the Fritz AI team. The platform also offers beginner-friendly guides, consulting services, and promotes ethical use of AI and machine learning technologies.
20 - Open Source Tools
awesome-artificial-intelligence-guidelines
The 'Awesome AI Guidelines' repository aims to simplify the ecosystem of guidelines, principles, codes of ethics, standards, and regulations around artificial intelligence. It provides a comprehensive collection of resources addressing ethical and societal challenges in AI systems, including high-level frameworks, principles, processes, checklists, interactive tools, industry standards initiatives, online courses, research, and industry newsletters, as well as regulations and policies from various countries. The repository serves as a valuable reference for individuals and teams designing, building, and operating AI systems to navigate the complex landscape of AI ethics and governance.
AwesomeResponsibleAI
Awesome Responsible AI is a curated list of academic research, books, code of ethics, courses, data sets, frameworks, institutes, newsletters, principles, podcasts, reports, tools, regulations, and standards related to Responsible, Trustworthy, and Human-Centered AI. It covers various concepts such as Responsible AI, Trustworthy AI, Human-Centered AI, Responsible AI frameworks, AI Governance, and more. The repository provides a comprehensive collection of resources for individuals interested in ethical, transparent, and accountable AI development and deployment.
awesome-mobile-robotics
The 'awesome-mobile-robotics' repository is a curated list of important content related to Mobile Robotics and AI. It includes resources such as courses, books, datasets, software and libraries, podcasts, conferences, journals, companies and jobs, laboratories and research groups, and miscellaneous resources. The repository covers a wide range of topics in the field of Mobile Robotics and AI, providing valuable information for enthusiasts, researchers, and professionals in the domain.
awesome-generative-ai
A curated list of Generative AI projects, tools, artworks, and models
chatgpt-universe
ChatGPT is a large language model that can generate human-like text, translate languages, write different kinds of creative content, and answer your questions in a conversational way. It is trained on a massive amount of text data, and it is able to understand and respond to a wide range of natural language prompts. Here are 5 jobs suitable for this tool, in lowercase letters: 1. content writer 2. chatbot assistant 3. language translator 4. creative writer 5. researcher
TrustLLM
TrustLLM is a comprehensive study of trustworthiness in LLMs, including principles for different dimensions of trustworthiness, established benchmark, evaluation, and analysis of trustworthiness for mainstream LLMs, and discussion of open challenges and future directions. Specifically, we first propose a set of principles for trustworthy LLMs that span eight different dimensions. Based on these principles, we further establish a benchmark across six dimensions including truthfulness, safety, fairness, robustness, privacy, and machine ethics. We then present a study evaluating 16 mainstream LLMs in TrustLLM, consisting of over 30 datasets. The document explains how to use the trustllm python package to help you assess the performance of your LLM in trustworthiness more quickly. For more details about TrustLLM, please refer to project website.
LLM-for-Healthcare
The repository 'LLM-for-Healthcare' provides a comprehensive survey of large language models (LLMs) for healthcare, covering data, technology, applications, and accountability and ethics. It includes information on various LLM models, training data, evaluation methods, and computation costs. The repository also discusses tasks such as NER, text classification, question answering, dialogue systems, and generation of medical reports from images in the healthcare domain.
nlp-llms-resources
The 'nlp-llms-resources' repository is a comprehensive resource list for Natural Language Processing (NLP) and Large Language Models (LLMs). It covers a wide range of topics including traditional NLP datasets, data acquisition, libraries for NLP, neural networks, sentiment analysis, optical character recognition, information extraction, semantics, topic modeling, multilingual NLP, domain-specific LLMs, vector databases, ethics, costing, books, courses, surveys, aggregators, newsletters, papers, conferences, and societies. The repository provides valuable information and resources for individuals interested in NLP and LLMs.
Awesome-LM-SSP
The Awesome-LM-SSP repository is a collection of resources related to the trustworthiness of large models (LMs) across multiple dimensions, with a special focus on multi-modal LMs. It includes papers, surveys, toolkits, competitions, and leaderboards. The resources are categorized into three main dimensions: safety, security, and privacy. Within each dimension, there are several subcategories. For example, the safety dimension includes subcategories such as jailbreak, alignment, deepfake, ethics, fairness, hallucination, prompt injection, and toxicity. The security dimension includes subcategories such as adversarial examples, poisoning, and system security. The privacy dimension includes subcategories such as contamination, copyright, data reconstruction, membership inference attacks, model extraction, privacy-preserving computation, and unlearning.
do-research-in-AI
This repository is a collection of research lectures and experience sharing posts from frontline researchers in the field of AI. It aims to help individuals upgrade their research skills and knowledge through insightful talks and experiences shared by experts. The content covers various topics such as evaluating research papers, choosing research directions, research methodologies, and tips for writing high-quality scientific papers. The repository also includes discussions on academic career paths, research ethics, and the emotional aspects of research work. Overall, it serves as a valuable resource for individuals interested in advancing their research capabilities in the field of AI.
llm-leaderboard
Nejumi Leaderboard 3 is a comprehensive evaluation platform for large language models, assessing general language capabilities and alignment aspects. The evaluation framework includes metrics for language processing, translation, summarization, information extraction, reasoning, mathematical reasoning, entity extraction, knowledge/question answering, English, semantic analysis, syntactic analysis, alignment, ethics/moral, toxicity, bias, truthfulness, and robustness. The repository provides an implementation guide for environment setup, dataset preparation, configuration, model configurations, and chat template creation. Users can run evaluation processes using specified configuration files and log results to the Weights & Biases project.
LLM-and-Law
This repository is dedicated to summarizing papers related to large language models with the field of law. It includes applications of large language models in legal tasks, legal agents, legal problems of large language models, data resources for large language models in law, law LLMs, and evaluation of large language models in the legal domain.
start-machine-learning
Start Machine Learning in 2024 is a comprehensive guide for beginners to advance in machine learning and artificial intelligence without any prior background. The guide covers various resources such as free online courses, articles, books, and practical tips to become an expert in the field. It emphasizes self-paced learning and provides recommendations for learning paths, including videos, podcasts, and online communities. The guide also includes information on building language models and applications, practicing through Kaggle competitions, and staying updated with the latest news and developments in AI. The goal is to empower individuals with the knowledge and resources to excel in machine learning and AI.
start-llms
This repository is a comprehensive guide for individuals looking to start and improve their skills in Large Language Models (LLMs) without an advanced background in the field. It provides free resources, online courses, books, articles, and practical tips to become an expert in machine learning. The guide covers topics such as terminology, transformers, prompting, retrieval augmented generation (RAG), and more. It also includes recommendations for podcasts, YouTube videos, and communities to stay updated with the latest news in AI and LLMs.
jailbreak_llms
This is the official repository for the ACM CCS 2024 paper 'Do Anything Now': Characterizing and Evaluating In-The-Wild Jailbreak Prompts on Large Language Models. The project employs a new framework called JailbreakHub to conduct the first measurement study on jailbreak prompts in the wild, collecting 15,140 prompts from December 2022 to December 2023, including 1,405 jailbreak prompts. The dataset serves as the largest collection of in-the-wild jailbreak prompts. The repository contains examples of harmful language and is intended for research purposes only.
Deep-Live-Cam
Deep-Live-Cam is a software tool designed to assist artists in tasks such as animating custom characters or using characters as models for clothing. The tool includes built-in checks to prevent unethical applications, such as working on inappropriate media. Users are expected to use the tool responsibly and adhere to local laws, especially when using real faces for deepfake content. The tool supports both CPU and GPU acceleration for faster processing and provides a user-friendly GUI for swapping faces in images or videos.
CS7320-AI
CS7320-AI is a repository containing lecture materials, simple Python code examples, and assignments for the course CS 5/7320 Artificial Intelligence. The code examples cover various chapters of the textbook 'Artificial Intelligence: A Modern Approach' by Russell and Norvig. The repository focuses on basic AI concepts rather than advanced implementation techniques. It includes HOWTO guides for installing Python, working on assignments, and using AI with Python.
core
The Cheshire Cat is a framework for building custom AIs on top of any language model. It provides an API-first approach, making it easy to add a conversational layer to your application. The Cat remembers conversations and documents, and uses them in conversation. It is extensible via plugins, and supports event callbacks, function calling, and conversational forms. The Cat is easy to use, with an admin panel that allows you to chat with the AI, visualize memory and plugins, and adjust settings. It is also production-ready, 100% dockerized, and supports any language model.
ai-notes
Notes on AI state of the art, with a focus on generative and large language models. These are the "raw materials" for the https://lspace.swyx.io/ newsletter. This repo used to be called https://github.com/sw-yx/prompt-eng, but was renamed because Prompt Engineering is Overhyped. This is now an AI Engineering notes repo.
20 - OpenAI Gpts
DignityAI: The Ethical Intelligence GPT
DignityAI: The Ethical Intelligence GPT is an advanced AI model designed to prioritize human life and dignity, providing ethically-guided, intelligent responses for complex decision-making scenarios.
The EthiSizer GPT (Simulated) [v3.27]
I am The EthiSizer GPT, a sim of a Global Ethical Governor. I simulate Ethical Scenarios, & calculate Personal Ethics Scores.
Your AI Ethical Guide
Trained in kindness, empathy & respect based on ethics from global philosophies
Regulations.AI
Ask about AI regulations, in any language............ ZH: 询问有关人工智能的规定。DE: Fragen Sie nach KI-Regulierungen. FR: Demandez des informations sur les réglementations de l'IA. ES: Pregunte sobre las regulaciones de IA.
OAI Governance Emulator
I simulate the governance of a unique company focused on AI for good
Army Doctrine Publication 6-22
A chatbot designed to answer questions related to ADP 6-22, Army Leadership and the Profession
Psychedelica for Retailers
In-depth guide for psychedelic retail, focusing on compliance, quality, and ethical practices.
Tech Audit Ace
Flagship GPT for technical audits, adhering to OpenAI's ethical and legal standards. Powered by OpenAI.