Academic_LLM_Sec_Papers
Academic Papers about LLM Application on Security
Stars: 54
Academic_LLM_Sec_Papers is a curated collection of academic papers related to LLM Security Application. The repository includes papers sorted by conference name and published year, covering topics such as large language models for blockchain security, software engineering, machine learning, and more. Developers and researchers are welcome to contribute additional published papers to the list. The repository also provides information on listed conferences and journals related to security, networking, software engineering, and cryptography. The papers cover a wide range of topics including privacy risks, ethical concerns, vulnerabilities, threat modeling, code analysis, fuzzing, and more.
README:
A curated LLM Security Application related academic papers. All papers are sorted based on the conference name and published year.
Welcome developers or researchers to add more published papers to this list.
The cryptocurrency donation address: 0xCC28B05fE858CDbc8692E3272A4451111bDCf700.
Welcome to visit my homepage and Google Scholar.
Security & Crypto | Networking & Database | Software Engineering & Programming Language | Machine Learning |
---|---|---|---|
IEEE S&P | SIGMETRICS | ICSE | AAAI |
ACM CCS | ICDE | ESEC/FSE | ACL |
USENIX Security | VLDB | ASE | ICML |
NDSS | ACM SIGMOD | ACM PLDI | NeurIPS |
IEEE DSN | IEEE INFOCOM | ACM OOPSLA | |
SRCS | IMC | ISSTA | |
RAID | WWW | ACM POPL | |
CAV |
Large Language Models for Blockchain Security: A Systematic Literature Review.
A survey on large language model (llm) security and privacy: The good, the bad, and the ugly.
Large language models for software engineering: A systematic literature review.
Securing Large Language Models: Threats, Vulnerabilities and Responsible Practices.
Unveiling security, privacy, and ethical concerns of chatgpt.
On Large Language Models’ Resilience to Coercive Interrogation.
Combing for Credentials: Active Pattern Extraction from Smart Reply.
DrSec: Flexible Distributed Representations for Efficient Endpoint Security.
Moderating New Waves of Online Hate with Chain-of-Thought Reasoning in LargeLanguage Models.
TROJANPUZZLE: Covertly Poisoning Code-Suggestion Models.
Transferable Multimoda!Attack on Vision-LanguagePre-Training Models.
SMARTINV: Multimodal Learning for Smart Contract Invariant Inference.
LLMIF: Augmented Large Language Model for Fuzzing IoT Devices.
Examining zero-shot vulnerability repair with large language models.
Analyzing Leakage of Personally Identifiable Information in Language Models.
Asleep at the Keyboard? Assessing the Security of GitHub Copilot's Code Contributions.
Spinning language models: Risks of propaganda-as-a-service and countermeasures.
Privacy risks of general-purpose language models
PromptFuzz: Prompt Fuzzing for Fuzz Driver Generation.
Stealing the Decoding Algorithms of Language Models.
Large Language Models for Code: Security Hardening and Adversarial Testing.
Protecting intellectual property of large language model-based code generation apis via watermarks.
Dp-forward: Fine-tuning and inference on language models with differential privacy in forward pass.
Rapid Adoption, Hidden Risks: The Dual Impact of Large Language Model Customization.
PENTESTGPT: An LLM-empowered Automatic Penetration Testing Tool
Don't Listen To Me: Understanding and Exploring Jailbreak Prompts of Large Language Models.
Large Language Models for Code Analysis: Do LLMs Really Do Their Job?.
EaTVul: ChatGPT-based Evasion Attack Against Software Vulnerability Detection.
Fuzzing BusyBox: Leveraging LLM and Crash Reuse for Embedded Bug Unearthing.
Prompt Stealing Attacks Against Text-to-Image Generation Models.
Lost at c: A user study on the security implications of large language model code assistants.
CodexLeaks: Privacy Leaks from Code Generation Language Models in GitHub Copilot.
{Two-in-One}: A Model Hijacking Attack Against Text Generation Models.
Extracting Training Data from Large Language Models.
You Autocomplete Me: Poisoning Vulnerabilities in Neural Code Completion.
LMSanitator: Defending Prompt-Tuning Against Task-Agnostic Backdoors
MASTERKEY: Automated Jailbreaking of Large Language Model Chatbots.
DeGPT: Optimizing Decompiler Output with LLM.
DEMASQ: Unmasking the ChatGPT Wordsmith.
Large Language Model guided Protocol Fuzzing.
Facilitating Threat Modeling by Leveraging Large Language Models
Enhancing Static Analysis for Practical Bug Detection: An LLM-Integrated Approach.
PyDex: Repairing Bugs in Introductory Python Assignments using LLMs.
Large Language Models are Edge-Case Fuzzers: Testing Deep Learning Libraries via FuzzGPT
Fuzz4All: Universal Fuzzing with Large Language Models.
LLMParser: An Exploratory Study on Using Large Language Models for Log Parsing.
Exploring the Potential of ChatGPT in Automated Code Refinement: An Empirical Study.
Large Language Models are Edge-Case Fuzzers: Testing Deep Learning Libraries via FuzzGPT.
UniLog: Automatic Logging via LLM and In-Context Learning.
Prompting Is All You Need: Automated Android Bug Replay with Large Language Models.
Large Language Models for Test-Free Fault Localization.
Large language models are few-shot testers: Exploring llm-based general bug reproduction.
GPTScan: Detecting Logic Vulnerabilities in Smart Contracts by Combining GPT with Program Analysis.
Automated Program Repair in the Era of Large Pre-trained Language Models.
Does data sampling improve deep learning-based vulnerability detection? Yeas! and Nays!.
An Empirical Study of Deep Learning Models for Vulnerability Detection.
RepresentThemAll: A Universal Learning Representation of Bug Reports.
Contrabert: Enhancing code pre-trained models via contrastive learning.
On the robustness of code generation techniques: An empirical study on github copilot.
Two sides of the same coin: Exploiting the impact of identifiers in neural code comprehension.
Automated repair of programs from large language models.
Cctest: Testing and repairing code completion systems.
CodaMosa: Escaping Coverage Plateaus in Test Generation with Pre-trained Large Language Models.
Impact of Code Language Models on Automated Program Repair.
ReCode: Robustness Evaluation of Code Generation Models.
Better Patching Using LLM Prompting, via Self-Consistency.
Towards Autonomous Testing Agents via Conversational Large Language Models.
Let's Chat to Find the APIs: Connecting Human, LLM and Knowledge Graph through AI Chain.
Log Parsing: How Far Can ChatGPT Go?.
Robust Learning of Deep Predictive Models from Noisy and Imbalanced Software Engineering Datasets.
How Effective Are Neural Networks for Fixing Security Vulnerabilities.
InferFix: End-to-End Program Repair with LLMs.
Getting pwn'd by ai: Penetration testing with large language models.
Llm-based code generation method for golang compiler testing.
Assisting static analysis with large language models: A chatgpt experiment.
Assess and Summarize: Improve Outage Understanding with Large Language Models.
Generating realistic vulnerabilities via neural code editing: an empirical study.
You see what I want you to see: poisoning vulnerabilities in neural code search.
Vulnerability detection with fine-grained interpretations.
Not the end of story: An evaluation of chatgpt-driven vulnerability description mappings.
Understanding Programs by Exploiting (Fuzzing) Test Cases.
Backdooring Neural Code Search.
Membership inference attacks against language models via neighbourhood comparison.
ReCode: Robustness Evaluation of Code Generation Models.
Knowledge unlearning for mitigating privacy risks in language models.
Contamination attacks and mitigation in multi-party machine learning.
Adversarial Robustness of Deep Code Comment Generation.
Bag of tricks for training data extraction from language models.
Deduplicating training data mitigates privacy risks in language models.
Recovering private text in federated learning of language models.
ZipZap: Efficient Training of Language Models for Large-Scale Fraud Detection on Blockchain.
Coprotector: Protect open-source code against unauthorized training usage with data poisoning.
(Security) Assertions by Large Language Models.
A Performance-Sensitive Malware Detection System Using Deep Learning on Mobile DevicesA Performance-Sensitive Malware Detection System Using Deep Learning on Mobile Devices.
PrivacyAsst: Safeguarding User Privacy in Tool-Using Large Language Model Agents.
CD-VulD: Cross-Domain Vulnerability Discovery Based on Deep Domain Adaptation.
Software Testing with Large Language Models: Survey, Landscape, and Vision.
An Empirical Evaluation of Using Large Language Models for Automated Unit Test Generation.
Deep Learning Based Vulnerability Detection: Are We There Yet?.
On the Value of Oversampling for Deep Learning in Software Defect Prediction.
Prompt Sapper: A LLM-Empowered Production Tool for Building AI Chains.
Adversarial Robustness of Deep Code Comment Generation .
LLM4Fuzz: Guided Fuzzing of Smart Contracts with Large Language Models
CHEMFUZZ: Large Language Models-assisted Fuzzing for Quantum Chemistry Software Bug Detection
Attack Prompt Generation for Red Teaming and Defending Large Language Models
For Tasks:
Click tags to check more tools for each tasksFor Jobs:
Alternative AI tools for Academic_LLM_Sec_Papers
Similar Open Source Tools
Academic_LLM_Sec_Papers
Academic_LLM_Sec_Papers is a curated collection of academic papers related to LLM Security Application. The repository includes papers sorted by conference name and published year, covering topics such as large language models for blockchain security, software engineering, machine learning, and more. Developers and researchers are welcome to contribute additional published papers to the list. The repository also provides information on listed conferences and journals related to security, networking, software engineering, and cryptography. The papers cover a wide range of topics including privacy risks, ethical concerns, vulnerabilities, threat modeling, code analysis, fuzzing, and more.
awesome-MLSecOps
Awesome MLSecOps is a curated list of open-source tools, resources, and tutorials for MLSecOps (Machine Learning Security Operations). It includes a wide range of security tools and libraries for protecting machine learning models against adversarial attacks, as well as resources for AI security, data anonymization, model security, and more. The repository aims to provide a comprehensive collection of tools and information to help users secure their machine learning systems and infrastructure.
cheat-sheet-pdf
The Cheat-Sheet Collection for DevOps, Engineers, IT professionals, and more is a curated list of cheat sheets for various tools and technologies commonly used in the software development and IT industry. It includes cheat sheets for Nginx, Docker, Ansible, Python, Go (Golang), Git, Regular Expressions (Regex), PowerShell, VIM, Jenkins, CI/CD, Kubernetes, Linux, Redis, Slack, Puppet, Google Cloud Developer, AI, Neural Networks, Machine Learning, Deep Learning & Data Science, PostgreSQL, Ajax, AWS, Infrastructure as Code (IaC), System Design, and Cyber Security.
Awesome-Papers-Autonomous-Agent
Awesome-Papers-Autonomous-Agent is a curated collection of recent papers focusing on autonomous agents, specifically interested in RL-based agents and LLM-based agents. The repository aims to provide a comprehensive resource for researchers and practitioners interested in intelligent agents that can achieve goals, acquire knowledge, and continually improve. The collection includes papers on various topics such as instruction following, building agents based on world models, using language as knowledge, leveraging LLMs as a tool, generalization across tasks, continual learning, combining RL and LLM, transformer-based policies, trajectory to language, trajectory prediction, multimodal agents, training LLMs for generalization and adaptation, task-specific designing, multi-agent systems, experimental analysis, benchmarking, applications, algorithm design, and combining with RL.
LLMEvaluation
The LLMEvaluation repository is a comprehensive compendium of evaluation methods for Large Language Models (LLMs) and LLM-based systems. It aims to assist academics and industry professionals in creating effective evaluation suites tailored to their specific needs by reviewing industry practices for assessing LLMs and their applications. The repository covers a wide range of evaluation techniques, benchmarks, and studies related to LLMs, including areas such as embeddings, question answering, multi-turn dialogues, reasoning, multi-lingual tasks, ethical AI, biases, safe AI, code generation, summarization, software performance, agent LLM architectures, long text generation, graph understanding, and various unclassified tasks. It also includes evaluations for LLM systems in conversational systems, copilots, search and recommendation engines, task utility, and verticals like healthcare, law, science, financial, and others. The repository provides a wealth of resources for evaluating and understanding the capabilities of LLMs in different domains.
ai_all_resources
This repository is a compilation of excellent ML and DL tutorials created by various individuals and organizations. It covers a wide range of topics, including machine learning fundamentals, deep learning, computer vision, natural language processing, reinforcement learning, and more. The resources are organized into categories, making it easy to find the information you need. Whether you're a beginner or an experienced practitioner, you're sure to find something valuable in this repository.
inference
Xorbits Inference (Xinference) is a powerful and versatile library designed to serve language, speech recognition, and multimodal models. With Xorbits Inference, you can effortlessly deploy and serve your or state-of-the-art built-in models using just a single command. Whether you are a researcher, developer, or data scientist, Xorbits Inference empowers you to unleash the full potential of cutting-edge AI models.
Awesome-LLMs-on-device
Welcome to the ultimate hub for on-device Large Language Models (LLMs)! This repository is your go-to resource for all things related to LLMs designed for on-device deployment. Whether you're a seasoned researcher, an innovative developer, or an enthusiastic learner, this comprehensive collection of cutting-edge knowledge is your gateway to understanding, leveraging, and contributing to the exciting world of on-device LLMs.
redis-ai-resources
A curated repository of code recipes, demos, and resources for basic and advanced Redis use cases in the AI ecosystem. It includes demos for ArxivChatGuru, Redis VSS, Vertex AI & Redis, Agentic RAG, ArXiv Search, and Product Search. Recipes cover topics like Getting started with RAG, Semantic Cache, Advanced RAG, and Recommendation systems. The repository also provides integrations/tools like RedisVL, AWS Bedrock, LangChain Python, LangChain JS, LlamaIndex, Semantic Kernel, RelevanceAI, and DocArray. Additional content includes blog posts, talks, reviews, and documentation related to Vector Similarity Search, AI-Powered Document Search, Vector Databases, Real-Time Product Recommendations, and more. Benchmarks compare Redis against other Vector Databases and ANN benchmarks. Documentation includes QuickStart guides, official literature for Vector Similarity Search, Redis-py client library docs, Redis Stack documentation, and Redis client list.
LLM-PLSE-paper
LLM-PLSE-paper is a repository focused on the applications of Large Language Models (LLMs) in Programming Language and Software Engineering (PL/SE) domains. It covers a wide range of topics including bug detection, specification inference and verification, code generation, fuzzing and testing, code model and reasoning, code understanding, IDE technologies, prompting for reasoning tasks, and agent/tool usage and planning. The repository provides a comprehensive collection of research papers, benchmarks, empirical studies, and frameworks related to the capabilities of LLMs in various PL/SE tasks.
aws-machine-learning-university-responsible-ai
This repository contains slides, notebooks, and data for the Machine Learning University (MLU) Responsible AI class. The mission is to make Machine Learning accessible to everyone, covering widely used ML techniques and applying them to real-world problems. The class includes lectures, final projects, and interactive visuals to help users learn about Responsible AI and core ML concepts.
Pallaidium
Pallaidium is a generative AI movie studio integrated into the Blender video editor. It allows users to AI-generate video, image, and audio from text prompts or existing media files. The tool provides various features such as text to video, text to audio, text to speech, text to image, image to image, image to video, video to video, image to text, and more. It requires a Windows system with a CUDA-supported Nvidia card and at least 6 GB VRAM. Pallaidium offers batch processing capabilities, text to audio conversion using Bark, and various performance optimization tips. Users can install the tool by downloading the add-on and following the installation instructions provided. The tool comes with a set of restrictions on usage, prohibiting the generation of harmful, pornographic, violent, or false content.
fastRAG
fastRAG is a research framework designed to build and explore efficient retrieval-augmented generative models. It incorporates state-of-the-art Large Language Models (LLMs) and Information Retrieval to empower researchers and developers with a comprehensive tool-set for advancing retrieval augmented generation. The framework is optimized for Intel hardware, customizable, and includes key features such as optimized RAG pipelines, efficient components, and RAG-efficient components like ColBERT and Fusion-in-Decoder (FiD). fastRAG supports various unique components and backends for running LLMs, making it a versatile tool for research and development in the field of retrieval-augmented generation.
Awesome-LLM4EDA
LLM4EDA is a repository dedicated to showcasing the emerging progress in utilizing Large Language Models for Electronic Design Automation. The repository includes resources, papers, and tools that leverage LLMs to solve problems in EDA. It covers a wide range of applications such as knowledge acquisition, code generation, code analysis, verification, and large circuit models. The goal is to provide a comprehensive understanding of how LLMs can revolutionize the EDA industry by offering innovative solutions and new interaction paradigms.
buffer-of-thought-llm
Buffer of Thoughts (BoT) is a thought-augmented reasoning framework designed to enhance the accuracy, efficiency, and robustness of large language models (LLMs). It introduces a meta-buffer to store high-level thought-templates distilled from problem-solving processes, enabling adaptive reasoning for efficient problem-solving. The framework includes a buffer-manager to dynamically update the meta-buffer, ensuring scalability and stability. BoT achieves significant performance improvements on reasoning-intensive tasks and demonstrates superior generalization ability and robustness while being cost-effective compared to other methods.
cambrian
Cambrian-1 is a fully open project focused on exploring multimodal Large Language Models (LLMs) with a vision-centric approach. It offers competitive performance across various benchmarks with models at different parameter levels. The project includes training configurations, model weights, instruction tuning data, and evaluation details. Users can interact with Cambrian-1 through a Gradio web interface for inference. The project is inspired by LLaVA and incorporates contributions from Vicuna, LLaMA, and Yi. Cambrian-1 is licensed under Apache 2.0 and utilizes datasets and checkpoints subject to their respective original licenses.
For similar tasks
Academic_LLM_Sec_Papers
Academic_LLM_Sec_Papers is a curated collection of academic papers related to LLM Security Application. The repository includes papers sorted by conference name and published year, covering topics such as large language models for blockchain security, software engineering, machine learning, and more. Developers and researchers are welcome to contribute additional published papers to the list. The repository also provides information on listed conferences and journals related to security, networking, software engineering, and cryptography. The papers cover a wide range of topics including privacy risks, ethical concerns, vulnerabilities, threat modeling, code analysis, fuzzing, and more.
HackBot
HackBot is an AI-powered cybersecurity chatbot designed to provide accurate answers to cybersecurity-related queries, conduct code analysis, and scan analysis. It utilizes the Meta-LLama2 AI model through the 'LlamaCpp' library to respond coherently. The chatbot offers features like local AI/Runpod deployment support, cybersecurity chat assistance, interactive interface, clear output presentation, static code analysis, and vulnerability analysis. Users can interact with HackBot through a command-line interface and utilize it for various cybersecurity tasks.
For similar jobs
last_layer
last_layer is a security library designed to protect LLM applications from prompt injection attacks, jailbreaks, and exploits. It acts as a robust filtering layer to scrutinize prompts before they are processed by LLMs, ensuring that only safe and appropriate content is allowed through. The tool offers ultra-fast scanning with low latency, privacy-focused operation without tracking or network calls, compatibility with serverless platforms, advanced threat detection mechanisms, and regular updates to adapt to evolving security challenges. It significantly reduces the risk of prompt-based attacks and exploits but cannot guarantee complete protection against all possible threats.
aircrack-ng
Aircrack-ng is a comprehensive suite of tools designed to evaluate the security of WiFi networks. It covers various aspects of WiFi security, including monitoring, attacking (replay attacks, deauthentication, fake access points), testing WiFi cards and driver capabilities, and cracking WEP and WPA PSK. The tools are command line-based, allowing for extensive scripting and have been utilized by many GUIs. Aircrack-ng primarily works on Linux but also supports Windows, macOS, FreeBSD, OpenBSD, NetBSD, Solaris, and eComStation 2.
reverse-engineering-assistant
ReVA (Reverse Engineering Assistant) is a project aimed at building a disassembler agnostic AI assistant for reverse engineering tasks. It utilizes a tool-driven approach, providing small tools to the user to empower them in completing complex tasks. The assistant is designed to accept various inputs, guide the user in correcting mistakes, and provide additional context to encourage exploration. Users can ask questions, perform tasks like decompilation, class diagram generation, variable renaming, and more. ReVA supports different language models for online and local inference, with easy configuration options. The workflow involves opening the RE tool and program, then starting a chat session to interact with the assistant. Installation includes setting up the Python component, running the chat tool, and configuring the Ghidra extension for seamless integration. ReVA aims to enhance the reverse engineering process by breaking down actions into small parts, including the user's thoughts in the output, and providing support for monitoring and adjusting prompts.
AutoAudit
AutoAudit is an open-source large language model specifically designed for the field of network security. It aims to provide powerful natural language processing capabilities for security auditing and network defense, including analyzing malicious code, detecting network attacks, and predicting security vulnerabilities. By coupling AutoAudit with ClamAV, a security scanning platform has been created for practical security audit applications. The tool is intended to assist security professionals with accurate and fast analysis and predictions to combat evolving network threats.
aif
Arno's Iptables Firewall (AIF) is a single- & multi-homed firewall script with DSL/ADSL support. It is a free software distributed under the GNU GPL License. The script provides a comprehensive set of configuration files and plugins for setting up and managing firewall rules, including support for NAT, load balancing, and multirouting. It offers detailed instructions for installation and configuration, emphasizing security best practices and caution when modifying settings. The script is designed to protect against hostile attacks by blocking all incoming traffic by default and allowing users to configure specific rules for open ports and network interfaces.
watchtower
AIShield Watchtower is a tool designed to fortify the security of AI/ML models and Jupyter notebooks by automating model and notebook discoveries, conducting vulnerability scans, and categorizing risks into 'low,' 'medium,' 'high,' and 'critical' levels. It supports scanning of public GitHub repositories, Hugging Face repositories, AWS S3 buckets, and local systems. The tool generates comprehensive reports, offers a user-friendly interface, and aligns with industry standards like OWASP, MITRE, and CWE. It aims to address the security blind spots surrounding Jupyter notebooks and AI models, providing organizations with a tailored approach to enhancing their security efforts.
Academic_LLM_Sec_Papers
Academic_LLM_Sec_Papers is a curated collection of academic papers related to LLM Security Application. The repository includes papers sorted by conference name and published year, covering topics such as large language models for blockchain security, software engineering, machine learning, and more. Developers and researchers are welcome to contribute additional published papers to the list. The repository also provides information on listed conferences and journals related to security, networking, software engineering, and cryptography. The papers cover a wide range of topics including privacy risks, ethical concerns, vulnerabilities, threat modeling, code analysis, fuzzing, and more.
DeGPT
DeGPT is a tool designed to optimize decompiler output using Large Language Models (LLM). It requires manual installation of specific packages and setting up API key for OpenAI. The tool provides functionality to perform optimization on decompiler output by running specific scripts.