
LLM-and-Law
This repository is dedicated to summarizing papers related to large language models with the field of law
Stars: 180

This repository is dedicated to summarizing papers related to large language models with the field of law. It includes applications of large language models in legal tasks, legal agents, legal problems of large language models, data resources for large language models in law, law LLMs, and evaluation of large language models in the legal domain.
README:
This repository is dedicated to summarizing papers related to large language models with the field of law
[1] Legal Prompt Engineering for Multilingual Legal Judgement Prediction
[2] Can GPT-3 Perform Statutory Reasoning?
[3] Legal Prompting: Teaching a Language Model to Think Like a Lawyer
[4] Large Language Models as Fiduciaries: A Case Study Toward Robustly Communicating With Artificial Intelligence Through Legal Standards
[5] ChatGPT Goes to Law School
[6] ChatGPT, Professor of Law
[7] ChatGPT & Generative AI Systems as Quasi-Expert Legal Advice Lawyers - Case Study Considering Potential Appeal Against Conviction of Tom Hayes
[8] ‘Words Are Flowing Out Like Endless Rain Into a Paper Cup’: ChatGPT & Law School Assessments
[9] ChatGPT by OpenAI: The End of Litigation Lawyers?
[10] Law Informs Code: A Legal Informatics Approach to Aligning Artificial Intelligence with Humans
[11] ChatGPT may Pass the Bar Exam soon, but has a Long Way to Go for the LexGLUE benchmark paper
[12] How Ready are Pre-trained Abstractive Models and LLMs for Legal Case Judgement Summarization? paper
[13] Explaining Legal Concepts with Augmented Large Language Models (GPT-4) paper
[14] Garbage in, garbage out: Zero-shot detection of crime using Large Language Models paper
[15] Legal Summarisation through LLMs: The PRODIGIT Project paper
[16] Black-Box Analysis: GPTs Across Time in Legal Textual Entailment Task paper
[17] PolicyGPT: Automated Analysis of Privacy Policies with Large Language Models
[18] Reformulating Domain Adaptation of Large Language Models as Adapt-Retrieve-Revise paper
[19] Precedent-Enhanced Legal Judgment Prediction with LLM and Domain-Model Collaboration paper
[20] From Text to Structure: Using Large Language Models to Support the Development of Legal Expert Systems paper
[21] Boosting legal case retrieval by query content selection with large language models paper
[22] LLMediator: GPT-4 Assisted Online Dispute Resolution paper
[23] Employing Label Models on ChatGPT Answers Improves Legal Text Entailment Performance paper
[24] LLaMandement: Large Language Models for Summarization of French Legislative Proposals paper
[25] Logic Rules as Explanations for Legal Case Retrieval paper Our new paper, welcome to pay attention !!!
[26] Enhancing Legal Document Retrieval: A Multi-Phase Approach with Large Language Models paper
[27] BLADE: Enhancing Black-box Large Language Models with Small Domain-Specific Models paper
[28] A Survey on Large Language Models for Critical Societal Domains: Finance, Healthcare, and Law paper
[29] Archimedes-AUEB at SemEval-2024 Task 5: LLM explains Civil Procedure paper
[30] More Than Catastrophic Forgetting: Integrating General Capabilities For Domain-Specific LLMs paper
[31] Legal Documents Drafting with Fine-Tuned Pre-Trained Large Language Model paper
[32] Knowledge-Infused Legal Wisdom: Navigating LLM Consultation through the Lens of Diagnostics and Positive-Unlabeled Reinforcement Learning paper
[33] GOLDCOIN: Grounding Large Language Models in Privacy Laws via Contextual Integrity Theory paper
[34] Enabling Discriminative Reasoning in Large Language Models for Legal Judgment Prediction paper
[35] Large Language Models for Judicial Entity Extraction: A Comparative Study paper
[36] Applicability of Large Language Models and Generative Models for Legal Case Judgement Summarization paper
[37] LawLLM: Law Large Language Model for the US Legal System paper
[38] Legal syllogism prompting: Teaching large language models for legal judgment prediction paper
[39] Optimizing Numerical Estimation and Operational Efficiency in the Legal Domain through Large Language Models paper
[40] Hallucination-Free? Assessing the Reliability of Leading AI Legal Research Tools paper
[41] KRAG Framework for Enhancing LLMs in the Legal Domain paper
[42] Legal Evalutions and Challenges of Large Language Models paper
[43] Analyzing Images of Legal Documents: Toward Multi-Modal LLMs for Access to Justice paper
[44] Automating Legal Concept Interpretation with LLMs: Retrieval, Generation, and Evaluation paper
[45] Domaino1s: Guiding LLM Reasoning for Explainable Answers in High-Stakes Domains paper
[46] RELexED: Retrieval-Enhanced Legal Summarization with Exemplar Diversity paper
[1] SimuCourt: Building Judicial Decision-Making Agents with Real-world Judgement Documents paper
[2] Can Large Language Models Grasp Legal Theories? Enhance Legal Reasoning with Insights from Multi-Agent Collaboration paper
[1] Towards WinoQueer: Developing a Benchmark for Anti-Queer Bias in Large Language Models
[2] Persistent Anti-Muslim Bias in Large Language Models
[3] Understanding the Capabilities, Limitations, and Societal Impact of Large Language Models
[4] The Dark Side of ChatGPT: Legal and Ethical Challenges from Stochastic Parrots and Hallucination
[5] The GPTJudge: Justice in a Generative AI World paper
[6] Is the U.S. Legal System Ready for AI's Challenges to Human Values? paper
[7] Questioning Biases in Case Judgment Summaries: Legal Datasets or Large Language Models? paper
[8] Large Legal Fictions: Profiling Legal Hallucinations in Large Language Models
[9] A Legal Framework for Natural Language Processing Model Training in Portugal paper
[10] LegalBench-RAG: A Benchmark for Retrieval-Augmented Generation in the Legal Domain paper
[11] Bias in Large Language Models: Origin, Evaluation, and Mitigation paper
[12] An Information Theoretic Approach to Operationalize Right to Data Protection paper
[13] Protecting Privacy in Multimodal Large Language Models with MLLMU-Bench paper
[1] CAIL2018: A Large-Scale Legal Dataset for Judgment Prediction
[2] When does pretraining help? assessing self-supervised learning for law and the casehold dataset of 53,000+ legal holdings
[3] LeCaRD: a legal case retrieval dataset for Chinese law system
[4] LeXFiles and LegalLAMA: Facilitating English Multinational Legal Language Model Development
[5] Legal Extractive Summarization of U.S. Court Opinions
[6] Awesome Chinese Legal Resources github
[7] MultiLegalPile: A 689GB Multilingual Legal Corpus paper
[8] The Cambridge Law Corpus: A Corpus for Legal AI Research
[9] TransformLLM: Adapting Large Language Models via LLM-Transformed Reading Comprehension Text paper
[10] Natural Language Processing for the Legal Domain: A Survey of Tasks, Datasets, Models, and Challenges paper
[11] ChineseSafe: A Chinese Benchmark for Evaluating Safety in Large Language Models paper
[12] Augmenting Legal Decision Support Systems with LLM-based NLI for Analyzing Social Media Evidence paper
[13] CaseSumm: A Large-Scale Dataset for Long-Context Summarization from U.S. Supreme Court Opinions paper
[1] LawGPT_zh github
[2] LaWGPT github
[3] Lawyer LLaMA github
[4] LexiLaw github
[5] LexGPT 0.1: pre-trained GPT-J models with Pile of Law paper
[6] TOWARDS THE EXPLOITATION OF LLM-BASED CHATBOT FOR PROVIDING LEGAL SUPPORT TO PALESTINIAN COOPERATIVES paper
[7] ChatLaw: Open-Source Legal Large Language Model with Integrated External Knowledge Bases paper
[8] DISC-LawLLM github
[9] InternLM-Law: An Open Source Chinese Legal Large Language Model paper
[10] SaulLM-54B & SaulLM-141B: Scaling Up Domain Adaptation for the Legal Domain paper
[1] Measuring Massive Multitask Chinese Understanding paper
[2] LawBench: Benchmarking Legal Knowledge of Large Language Models github
[3] Large Language Models are legal but they are not: Making the case for a powerful LegalLLM paper
[4] Better Call GPT, Comparing Large Language Models Against Lawyers paper
[5] Evaluating GPT-3.5's Awareness and Summarization Abilities for European Constitutional Texts with Shared Topics paper
[6] Evaluation Ethics of LLMs in Legal Domain paper
[7] GPTs and Language Barrier: A Cross-Lingual Legal QA Examination paper
[8] LawBench: Benchmarking Legal Knowledge of Large Language Models paper
[9] LexEval: A Comprehensive Chinese Legal Benchmark for Evaluating Large Language Models paper
[10] LegalAgentBench: Evaluating LLM Agents in Legal Domain paper
The survey paper is shown in paper
Please cite the following papers as the references if you use our codes or the processed datasets.
@article{sun2023short,
title={A short survey of viewing large language models in legal aspect},
author={Sun, Zhongxiang},
journal={arXiv preprint arXiv:2303.09136},
year={2023}
}
For Tasks:
Click tags to check more tools for each tasksFor Jobs:
Alternative AI tools for LLM-and-Law
Similar Open Source Tools

LLM-and-Law
This repository is dedicated to summarizing papers related to large language models with the field of law. It includes applications of large language models in legal tasks, legal agents, legal problems of large language models, data resources for large language models in law, law LLMs, and evaluation of large language models in the legal domain.

Awesome_Mamba
Awesome Mamba is a curated collection of groundbreaking research papers and articles on Mamba Architecture, a pioneering framework in deep learning known for its selective state spaces and efficiency in processing complex data structures. The repository offers a comprehensive exploration of Mamba architecture through categorized research papers covering various domains like visual recognition, speech processing, remote sensing, video processing, activity recognition, image enhancement, medical imaging, reinforcement learning, natural language processing, 3D recognition, multi-modal understanding, time series analysis, graph neural networks, point cloud analysis, and tabular data handling.

Awesome-LLMs-on-device
Welcome to the ultimate hub for on-device Large Language Models (LLMs)! This repository is your go-to resource for all things related to LLMs designed for on-device deployment. Whether you're a seasoned researcher, an innovative developer, or an enthusiastic learner, this comprehensive collection of cutting-edge knowledge is your gateway to understanding, leveraging, and contributing to the exciting world of on-device LLMs.

awesome-weather-models
A catalogue and categorization of AI-based weather forecasting models. This page provides a catalogue and categorization of AI-based weather forecasting models to enable discovery and comparison of different available model options. The weather models are categorized based on metadata found in the JSON schema specification. The table includes information such as the name of the weather model, the organization that developed it, operational data availability, open-source status, and links for further details.

PIXIU
PIXIU is a project designed to support the development, fine-tuning, and evaluation of Large Language Models (LLMs) in the financial domain. It includes components like FinBen, a Financial Language Understanding and Prediction Evaluation Benchmark, FIT, a Financial Instruction Dataset, and FinMA, a Financial Large Language Model. The project provides open resources, multi-task and multi-modal financial data, and diverse financial tasks for training and evaluation. It aims to encourage open research and transparency in the financial NLP field.

ColossalAI
Colossal-AI is a deep learning system for large-scale parallel training. It provides a unified interface to scale sequential code of model training to distributed environments. Colossal-AI supports parallel training methods such as data, pipeline, tensor, and sequence parallelism and is integrated with heterogeneous training and zero redundancy optimizer.

awesome-large-audio-models
This repository is a curated list of awesome large AI models in audio signal processing, focusing on the application of large language models to audio tasks. It includes survey papers, popular large audio models, automatic speech recognition, neural speech synthesis, speech translation, other speech applications, large audio models in music, and audio datasets. The repository aims to provide a comprehensive overview of recent advancements and challenges in applying large language models to audio signal processing, showcasing the efficacy of transformer-based architectures in various audio tasks.

pytorch-grad-cam
This repository provides advanced AI explainability for PyTorch, offering state-of-the-art methods for Explainable AI in computer vision. It includes a comprehensive collection of Pixel Attribution methods for various tasks like Classification, Object Detection, Semantic Segmentation, and more. The package supports high performance with full batch image support and includes metrics for evaluating and tuning explanations. Users can visualize and interpret model predictions, making it suitable for both production and model development scenarios.

Next-Generation-LLM-based-Recommender-Systems-Survey
The Next-Generation LLM-based Recommender Systems Survey is a comprehensive overview of the latest advancements in recommender systems leveraging Large Language Models (LLMs). The survey covers various paradigms, approaches, and applications of LLMs in recommendation tasks, including generative and non-generative models, multimodal recommendations, personalized explanations, and industrial deployment. It discusses the comparison with existing surveys, different paradigms, and specific works in the field. The survey also addresses challenges and future directions in the domain of LLM-based recommender systems.

SoM-LLaVA
SoM-LLaVA is a new data source and learning paradigm for Multimodal LLMs, empowering open-source Multimodal LLMs with Set-of-Mark prompting and improved visual reasoning ability. The repository provides a new dataset that is complementary to existing training sources, enhancing multimodal LLMs with Set-of-Mark prompting and improved general capacity. By adding 30k SoM data to the visual instruction tuning stage of LLaVA, the tool achieves 1% to 6% relative improvements on all benchmarks. Users can train SoM-LLaVA via command line and utilize the implementation to annotate COCO images with SoM. Additionally, the tool can be loaded in Huggingface for further usage.

Awesome-Embodied-AI
Awesome-Embodied-AI is a curated list of papers on Embodied AI and related resources, tracking and summarizing research and industrial progress in the field. It includes surveys, workshops, tutorials, talks, blogs, and papers covering various aspects of Embodied AI, such as vision-language navigation, large language model-based agents, robotics, and more. The repository welcomes contributions and aims to provide a comprehensive overview of the advancements in Embodied AI.

GIMP-ML
A.I. for GNU Image Manipulation Program (GIMP-ML) is a repository that provides Python plugins for using computer vision models in GIMP. The code base and models are continuously updated to support newer and more stable functionality. Users can edit images with text, outpaint images, and generate images from text using models like Dalle 2 and Dalle 3. The repository encourages citations using a specific bibtex entry and follows the MIT license for GIMP-ML and the original models.

Pai-Megatron-Patch
Pai-Megatron-Patch is a deep learning training toolkit built for developers to train and predict LLMs & VLMs by using Megatron framework easily. With the continuous development of LLMs, the model structure and scale are rapidly evolving. Although these models can be conveniently manufactured using Transformers or DeepSpeed training framework, the training efficiency is comparably low. This phenomenon becomes even severer when the model scale exceeds 10 billion. The primary objective of Pai-Megatron-Patch is to effectively utilize the computational power of GPUs for LLM. This tool allows convenient training of commonly used LLM with all the accelerating techniques provided by Megatron-LM.

painting-droid
Painting Droid is an AI-powered cross-platform painting app inspired by MS Paint, expandable with plugins and open. It utilizes various AI models, from paid providers to self-hosted open-source models, as well as some lightweight ones built into the app. Features include regular painting app features, AI-generated content filling and augmentation, filters and effects, image manipulation, plugin support, and cross-platform compatibility.

DB-GPT-Hub
DB-GPT-Hub is an experimental project leveraging Large Language Models (LLMs) for Text-to-SQL parsing. It includes stages like data collection, preprocessing, model selection, construction, and fine-tuning of model weights. The project aims to enhance Text-to-SQL capabilities, reduce model training costs, and enable developers to contribute to improving Text-to-SQL accuracy. The ultimate goal is to achieve automated question-answering based on databases, allowing users to execute complex database queries using natural language descriptions. The project has successfully integrated multiple large models and established a comprehensive workflow for data processing, SFT model training, prediction output, and evaluation.
For similar tasks

LLM-and-Law
This repository is dedicated to summarizing papers related to large language models with the field of law. It includes applications of large language models in legal tasks, legal agents, legal problems of large language models, data resources for large language models in law, law LLMs, and evaluation of large language models in the legal domain.
For similar jobs

LLM-and-Law
This repository is dedicated to summarizing papers related to large language models with the field of law. It includes applications of large language models in legal tasks, legal agents, legal problems of large language models, data resources for large language models in law, law LLMs, and evaluation of large language models in the legal domain.

start-llms
This repository is a comprehensive guide for individuals looking to start and improve their skills in Large Language Models (LLMs) without an advanced background in the field. It provides free resources, online courses, books, articles, and practical tips to become an expert in machine learning. The guide covers topics such as terminology, transformers, prompting, retrieval augmented generation (RAG), and more. It also includes recommendations for podcasts, YouTube videos, and communities to stay updated with the latest news in AI and LLMs.

aiverify
AI Verify is an AI governance testing framework and software toolkit that validates the performance of AI systems against internationally recognised principles through standardised tests. It offers a new API Connector feature to bypass size limitations, test various AI frameworks, and configure connection settings for batch requests. The toolkit operates within an enterprise environment, conducting technical tests on common supervised learning models for tabular and image datasets. It does not define AI ethical standards or guarantee complete safety from risks or biases.

Awesome-LLM-Watermark
This repository contains a collection of research papers related to watermarking techniques for text and images, specifically focusing on large language models (LLMs). The papers cover various aspects of watermarking LLM-generated content, including robustness, statistical understanding, topic-based watermarks, quality-detection trade-offs, dual watermarks, watermark collision, and more. Researchers have explored different methods and frameworks for watermarking LLMs to protect intellectual property, detect machine-generated text, improve generation quality, and evaluate watermarking techniques. The repository serves as a valuable resource for those interested in the field of watermarking for LLMs.

LLM-LieDetector
This repository contains code for reproducing experiments on lie detection in black-box LLMs by asking unrelated questions. It includes Q/A datasets, prompts, and fine-tuning datasets for generating lies with language models. The lie detectors rely on asking binary 'elicitation questions' to diagnose whether the model has lied. The code covers generating lies from language models, training and testing lie detectors, and generalization experiments. It requires access to GPUs and OpenAI API calls for running experiments with open-source models. Results are stored in the repository for reproducibility.

graphrag
The GraphRAG project is a data pipeline and transformation suite designed to extract meaningful, structured data from unstructured text using LLMs. It enhances LLMs' ability to reason about private data. The repository provides guidance on using knowledge graph memory structures to enhance LLM outputs, with a warning about the potential costs of GraphRAG indexing. It offers contribution guidelines, development resources, and encourages prompt tuning for optimal results. The Responsible AI FAQ addresses GraphRAG's capabilities, intended uses, evaluation metrics, limitations, and operational factors for effective and responsible use.

langtest
LangTest is a comprehensive evaluation library for custom LLM and NLP models. It aims to deliver safe and effective language models by providing tools to test model quality, augment training data, and support popular NLP frameworks. LangTest comes with benchmark datasets to challenge and enhance language models, ensuring peak performance in various linguistic tasks. The tool offers more than 60 distinct types of tests with just one line of code, covering aspects like robustness, bias, representation, fairness, and accuracy. It supports testing LLMS for question answering, toxicity, clinical tests, legal support, factuality, sycophancy, and summarization.

Awesome-Jailbreak-on-LLMs
Awesome-Jailbreak-on-LLMs is a collection of state-of-the-art, novel, and exciting jailbreak methods on Large Language Models (LLMs). The repository contains papers, codes, datasets, evaluations, and analyses related to jailbreak attacks on LLMs. It serves as a comprehensive resource for researchers and practitioners interested in exploring various jailbreak techniques and defenses in the context of LLMs. Contributions such as additional jailbreak-related content, pull requests, and issue reports are welcome, and contributors are acknowledged. For any inquiries or issues, contact [email protected]. If you find this repository useful for your research or work, consider starring it to show appreciation.