sunone_aimbot
Aim-bot based on AI for all FPS games
Stars: 294
Sunone Aimbot is an AI-powered aim bot for first-person shooter games. It leverages YOLOv8 and YOLOv10 models, PyTorch, and various tools to automatically target and aim at enemies within the game. The AI model has been trained on more than 30,000 images from popular first-person shooter games like Warface, Destiny 2, Battlefield 2042, CS:GO, Fortnite, The Finals, CS2, and more. The aimbot can be configured through the `config.ini` file to adjust various settings related to object search, capture methods, aiming behavior, hotkeys, mouse settings, shooting options, Arduino integration, AI model parameters, overlay display, debug window, and more. Users are advised to follow specific recommendations to optimize performance and avoid potential issues while using the aimbot.
README:
Sunone Aimbot is an AI-powered aim bot for first-person shooter games. It leverages the YOLOv8 and YOLOv10 models, PyTorch, and various other tools to automatically target and aim at enemies within the game. The AI model in repository has been trained on more than 30,000 images from popular first-person shooter games like Warface, Destiny 2, Battlefield (all series), Fortnite, The Finals, CS2 and more.
[!WARNING] Use it at your own risk, we do not guarantee that you may be blocked!
[!NOTE] This application only works on Nvidia graphics cards. AMD support is testing. See AI_enable_AMD option. The recommended graphics card for starting and more productive and stable operation starts with the rtx 20 series.
Before you get started, make sure you have the following prerequisites installed and pay attention to the versions in Tested Environment block, this may cause errors in launching the aimbot.
- Information in English:
- Информация на Русском языке:
- To launch the aimbot after all installations, type
py run.py
.
Windows | 10 and 11(priority) |
---|---|
Python: | 3.11.6 |
CUDA: | 12.4 |
TensorRT: | 10.0.1 |
Ultralytics: | 8.2.61 |
GitHub AI Model: | 0.4.1 (YOLOv8) |
Boosty AI Model: | 0.6.0 (YOLOv10) |
The behavior of the aimbot can be configured via the config.ini
file. Here are the available options:
- detection_window_width
int
: Horizontal resolution of the object search window. - detection_window_height
int
: Vertical resolution of the object search window.
- Bettercam_capture
bool
: Use Bettercam to capture images from the screen. - bettercam_capture_fps
int
: Specific fps value for screen capture. - bettercam_monitor_id
int
: Id of the monitor from which the images will be captured. - bettercam_gpu_id
int
: Id of the GPU to be used for image capture
- Obs_capture
bool
: Use Obs to capture images from the screen. - Obs_camera_id
str
orint
:auto
or number of Virtual Camera ID. - Obs_capture_fps
int
: Specific fps value for screen capture.
- body_y_offset
float
: Allows correction of y coordinates inside the body detected box if head is not detected. - hideout_targets
bool
: Allows shooting at targets on the range (for example in warface on the polygon or in aimlabs). - disable_headshot
bool
: Disable head targerting. - disable_prediction
bool
: Disable target position prediction. - prediction_interval
bool
: The time frame used to estimate an object's future position based on its current motion. - third_person
bool
: Turn on the third-person game mode. (sunxds_0.5.7+)
- The names of all the keys are here. Type
None
is empty button. - hotkey_targeting
str
: Aiming at the target. Supports multi keys, for examplehotkey_targeting = RightMouseButton,X2MouseButton
- hotkey_exit
str
: Exit. - hotkey_pause
str
: Pause AIM. - hotkey_reload_config
str
: Reload config.
- mouse_dpi
int
: Mouse DPI. - mouse_sensitivity
float
: Aim sensitivity. - mouse_fov_width
int
: The current horizontal value of the viewing angle in the game. - mouse_fov_height
int
: The current vertical value of the viewing angle in the game. - mouse_lock_target
bool
: True: Press once to permanently aim at the target, press again to turn off the aiming. False: Hold down the button to constantly aim at the target. - mouse_auto_aim
bool
: Automatic targeting. - mouse_ghub
bool
: Uses Logitech GHUB exploit for mouse movement. If the value is False, native win32 library is used for movement.
- auto_shoot
bool
: Automatic shooting. (For some games need arduino). - triggerbot
bool
: Automatic shooting at a target if it is in the scope, requires themouse_auto_shoot
option enabled, and aiming will also be automatically turned off. - force_click
bool
: Shooting will be performed even if the sight is not located within the object. - bScope_multiplier
float
: The multiplier of the target trigger size.
- arduino_move
bool
: Sends a command to the arduino to move the mouse. - arduino_shoot
bool
: Sends a command to the arduino to fire with the mouse. - arduino_port
str
: Arduino COM port. UseCOM1
orCOM2
... orauto
. - arduino_baudrate
int
: Custom Arduino baudrate. - arduino_16_bit_mouse
bool
: Send 16 bit data to the arduino port to move the mouse.
- AI_model_name
str
: AI model name. - AI_model_image_size
int
: AI model image size. - AI_conf
float
: How many percent is AI sure that this is the right goal. - AI_device
int
orstr
: Device to run on,0
,1
... orcpu
. - AI_enable_AMD
bool
: Enable support Amd GPUs. Install ROCm, Zluda and PyTorch. See AMD docs. - AI_mouse_net
bool
: Use a neural network to calculate mouse movements. See this repository.
- show_overlay
bool
: Enables the overlay. It is not recommended for gameplay, only for debugging. - overlay_show_borders
bool
: Displaying the borders of the overlay. - overlay_show_boxes
bool
: Display of boxes. - overlay_show_target_line
bool
: Displaying the line to the target. - overlay_show_target_prediction_line
bool
: Displaying the predictive line to the target. - overlay_show_labels
bool
: Displaying label names. - overlay_show_conf
bool
: Displaying the label names as well as the confidence level.
- show_window
bool
: Shows the OpenCV2 window for visual feedback. - show_detection_speed
bool
: Displays speed information inside the debug window. - show_window_fps
bool
: Displays FPS in the corner. - show_boxes
bool
: Displays detectable objects. - show_labels
bool
: Displays the name of the detected object. - show_conf
bool
: Displays object confidence threshold for detection. - show_target_line
bool
: Shows the mouse finishing line. - show_target_prediction_line
bool
: Show mouse prediction line. - show_bScope_box
bool
: Show the trigger box for auto shooting. - show_history_points
bool
: Show history points. - debug_window_always_on_top
bool
: The debug window will always be on top of other windows. - spawn_window_pos_x
int
: When the debugging window starts, it takes the x position. - spawn_window_pos_y
int
: When the debugging window starts, it takes the y position. - debug_window_scale_percent
int
: Adjusts the size of the debug window. - The names of the debugging window can be written in the file window_names.txt they will be randomly selected.
- *.pt: Default AI model.
- *.onnx: The model is optimized to run on processors.
- *.engine: Final exported model, which is faster than the previous two.
- All commands are executed in the console window:
- First, go to the aimbot directory using the command:
cd C:\Users\your_username\downloads\sunone_aimbot-main
- Then export the model from the .pt format in .engine format.
yolo export model="models/sunxds_0.4.1.pt" format=engine device=0 imgsz=480 half=True
-
model="model_path/model_name.pt"
: Path to model. -
format=engine
: TensorRT model format. -
half=true
: Use Half-precision floating-point format. -
device=0
: GPU id. -
workspace=8
: GPU max video memory. -
verbose=False
: Debug stuff. Convenient function, can show errors when exporting.
- Each model has its own image size with which it was trained, export only with the image size with which it was published.
- Limit the maximum value of frames per second in the game in which you will use it. And also do not set the screen resolution to high. Do not overload the graphics card.
- Do not set high graphics settings in games.
- Limit the browser (try not to watch YouTube while playing and working AI at the same time, for example (of course if you don't have a super duper graphics card)) and so on, which loads the video card.
- Try to use TensorRT for acceleration.
.pt
model is good, but does not have as much speed as.engine
. - Turn off the cv2 debug window, this saves system resources.
- Do not increase the object search window resolution, this may affect your search speed.
- If you have started the application and nothing happens, it may be working, close it with the F2 key and change the
show_window
option toTrue
in the file config.ini to make sure that the application is working.
I will post new models here.
This project is licensed under the MIT License. See LICENSE for details
For Tasks:
Click tags to check more tools for each tasksFor Jobs:
Alternative AI tools for sunone_aimbot
Similar Open Source Tools
sunone_aimbot
Sunone Aimbot is an AI-powered aim bot for first-person shooter games. It leverages YOLOv8 and YOLOv10 models, PyTorch, and various tools to automatically target and aim at enemies within the game. The AI model has been trained on more than 30,000 images from popular first-person shooter games like Warface, Destiny 2, Battlefield 2042, CS:GO, Fortnite, The Finals, CS2, and more. The aimbot can be configured through the `config.ini` file to adjust various settings related to object search, capture methods, aiming behavior, hotkeys, mouse settings, shooting options, Arduino integration, AI model parameters, overlay display, debug window, and more. Users are advised to follow specific recommendations to optimize performance and avoid potential issues while using the aimbot.
mods
AI for the command line, built for pipelines. LLM based AI is really good at interpreting the output of commands and returning the results in CLI friendly text formats like Markdown. Mods is a simple tool that makes it super easy to use AI on the command line and in your pipelines. Mods works with OpenAI, Groq, Azure OpenAI, and LocalAI To get started, install Mods and check out some of the examples below. Since Mods has built-in Markdown formatting, you may also want to grab Glow to give the output some _pizzazz_.
ML-Bench
ML-Bench is a tool designed to evaluate large language models and agents for machine learning tasks on repository-level code. It provides functionalities for data preparation, environment setup, usage, API calling, open source model fine-tuning, and inference. Users can clone the repository, load datasets, run ML-LLM-Bench, prepare data, fine-tune models, and perform inference tasks. The tool aims to facilitate the evaluation of language models and agents in the context of machine learning tasks on code repositories.
llm-vscode
llm-vscode is an extension designed for all things LLM, utilizing llm-ls as its backend. It offers features such as code completion with 'ghost-text' suggestions, the ability to choose models for code generation via HTTP requests, ensuring prompt size fits within the context window, and code attribution checks. Users can configure the backend, suggestion behavior, keybindings, llm-ls settings, and tokenization options. Additionally, the extension supports testing models like Code Llama 13B, Phind/Phind-CodeLlama-34B-v2, and WizardLM/WizardCoder-Python-34B-V1.0. Development involves cloning llm-ls, building it, and setting up the llm-vscode extension for use.
datadreamer
DataDreamer is an advanced toolkit designed to facilitate the development of edge AI models by enabling synthetic data generation, knowledge extraction from pre-trained models, and creation of efficient and potent models. It eliminates the need for extensive datasets by generating synthetic datasets, leverages latent knowledge from pre-trained models, and focuses on creating compact models suitable for integration into any device and performance for specialized tasks. The toolkit offers features like prompt generation, image generation, dataset annotation, and tools for training small-scale neural networks for edge deployment. It provides hardware requirements, usage instructions, available models, and limitations to consider while using the library.
agnai
Agnaistic is an AI roleplay chat tool that allows users to interact with personalized characters using their favorite AI services. It supports multiple AI services, persona schema formats, and features such as group conversations, user authentication, and memory/lore books. Agnaistic can be self-hosted or run using Docker, and it provides a range of customization options through its settings.json file. The tool is designed to be user-friendly and accessible, making it suitable for both casual users and developers.
codespin
CodeSpin.AI is a set of open-source code generation tools that leverage large language models (LLMs) to automate coding tasks. With CodeSpin, you can generate code in various programming languages, including Python, JavaScript, Java, and C++, by providing natural language prompts. CodeSpin offers a range of features to enhance code generation, such as custom templates, inline prompting, and the ability to use ChatGPT as an alternative to API keys. Additionally, CodeSpin provides options for regenerating code, executing code in prompt files, and piping data into the LLM for processing. By utilizing CodeSpin, developers can save time and effort in coding tasks, improve code quality, and explore new possibilities in code generation.
garak
Garak is a free tool that checks if a Large Language Model (LLM) can be made to fail in a way that is undesirable. It probes for hallucination, data leakage, prompt injection, misinformation, toxicity generation, jailbreaks, and many other weaknesses. Garak's a free tool. We love developing it and are always interested in adding functionality to support applications.
tiledesk-dashboard
Tiledesk is an open-source live chat platform with integrated chatbots written in Node.js and Express. It is designed to be a multi-channel platform for web, Android, and iOS, and it can be used to increase sales or provide post-sales customer service. Tiledesk's chatbot technology allows for automation of conversations, and it also provides APIs and webhooks for connecting external applications. Additionally, it offers a marketplace for apps and features such as CRM, ticketing, and data export.
holmesgpt
HolmesGPT is an open-source DevOps assistant powered by OpenAI or any tool-calling LLM of your choice. It helps in troubleshooting Kubernetes, incident response, ticket management, automated investigation, and runbook automation in plain English. The tool connects to existing observability data, is compliance-friendly, provides transparent results, supports extensible data sources, runbook automation, and integrates with existing workflows. Users can install HolmesGPT using Brew, prebuilt Docker container, Python Poetry, or Docker. The tool requires an API key for functioning and supports OpenAI, Azure AI, and self-hosted LLMs.
aiid
The Artificial Intelligence Incident Database (AIID) is a collection of incidents involving the development and use of artificial intelligence (AI). The database is designed to help researchers, policymakers, and the public understand the potential risks and benefits of AI, and to inform the development of policies and practices to mitigate the risks and promote the benefits of AI. The AIID is a collaborative project involving researchers from the University of California, Berkeley, the University of Washington, and the University of Toronto.
termax
Termax is an LLM agent in your terminal that converts natural language to commands. It is featured by: - Personalized Experience: Optimize the command generation with RAG. - Various LLMs Support: OpenAI GPT, Anthropic Claude, Google Gemini, Mistral AI, and more. - Shell Extensions: Plugin with popular shells like `zsh`, `bash` and `fish`. - Cross Platform: Able to run on Windows, macOS, and Linux.
chatgpt-cli
ChatGPT CLI provides a powerful command-line interface for seamless interaction with ChatGPT models via OpenAI and Azure. It features streaming capabilities, extensive configuration options, and supports various modes like streaming, query, and interactive mode. Users can manage thread-based context, sliding window history, and provide custom context from any source. The CLI also offers model and thread listing, advanced configuration options, and supports GPT-4, GPT-3.5-turbo, and Perplexity's models. Installation is available via Homebrew or direct download, and users can configure settings through default values, a config.yaml file, or environment variables.
cheating-based-prompt-engine
This is a vulnerability mining engine purely based on GPT, requiring no prior knowledge base, no fine-tuning, yet its effectiveness can overwhelmingly surpass most of the current related research. The core idea revolves around being task-driven, not question-driven, driven by prompts, not by code, and focused on prompt design, not model design. The essence is encapsulated in one word: deception. It is a type of code understanding logic vulnerability mining that fully stimulates the capabilities of GPT, suitable for real actual projects.
For similar tasks
sunone_aimbot
Sunone Aimbot is an AI-powered aim bot for first-person shooter games. It leverages YOLOv8 and YOLOv10 models, PyTorch, and various tools to automatically target and aim at enemies within the game. The AI model has been trained on more than 30,000 images from popular first-person shooter games like Warface, Destiny 2, Battlefield 2042, CS:GO, Fortnite, The Finals, CS2, and more. The aimbot can be configured through the `config.ini` file to adjust various settings related to object search, capture methods, aiming behavior, hotkeys, mouse settings, shooting options, Arduino integration, AI model parameters, overlay display, debug window, and more. Users are advised to follow specific recommendations to optimize performance and avoid potential issues while using the aimbot.
-Topaz-DeNoise-AI-Tool
Topaz DeNoise AI is a powerful tool designed for photographers and videographers to enhance image quality by reducing noise while preserving detail. It leverages advanced AI algorithms to clean up images, providing stunning results without sacrificing clarity. With features like AI-powered noise reduction, detail preservation, batch processing, and a user-friendly interface, users can easily improve the quality of their visuals. The tool offers a seamless workflow from downloading and installing the software to uploading images and applying noise reduction. Additionally, it provides documentation, contribution guidelines, and emphasizes security and responsible use.
tt-metal
TT-NN is a python & C++ Neural Network OP library. It provides a low-level programming model, TT-Metalium, enabling kernel development for Tenstorrent hardware.
mscclpp
MSCCL++ is a GPU-driven communication stack for scalable AI applications. It provides a highly efficient and customizable communication stack for distributed GPU applications. MSCCL++ redefines inter-GPU communication interfaces, delivering a highly efficient and customizable communication stack for distributed GPU applications. Its design is specifically tailored to accommodate diverse performance optimization scenarios often encountered in state-of-the-art AI applications. MSCCL++ provides communication abstractions at the lowest level close to hardware and at the highest level close to application API. The lowest level of abstraction is ultra light weight which enables a user to implement logics of data movement for a collective operation such as AllReduce inside a GPU kernel extremely efficiently without worrying about memory ordering of different ops. The modularity of MSCCL++ enables a user to construct the building blocks of MSCCL++ in a high level abstraction in Python and feed them to a CUDA kernel in order to facilitate the user's productivity. MSCCL++ provides fine-grained synchronous and asynchronous 0-copy 1-sided abstracts for communication primitives such as `put()`, `get()`, `signal()`, `flush()`, and `wait()`. The 1-sided abstractions allows a user to asynchronously `put()` their data on the remote GPU as soon as it is ready without requiring the remote side to issue any receive instruction. This enables users to easily implement flexible communication logics, such as overlapping communication with computation, or implementing customized collective communication algorithms without worrying about potential deadlocks. Additionally, the 0-copy capability enables MSCCL++ to directly transfer data between user's buffers without using intermediate internal buffers which saves GPU bandwidth and memory capacity. MSCCL++ provides consistent abstractions regardless of the location of the remote GPU (either on the local node or on a remote node) or the underlying link (either NVLink/xGMI or InfiniBand). This simplifies the code for inter-GPU communication, which is often complex due to memory ordering of GPU/CPU read/writes and therefore, is error-prone.
mlir-air
This repository contains tools and libraries for building AIR platforms, runtimes and compilers.
free-for-life
A massive list including a huge amount of products and services that are completely free! ⭐ Star on GitHub • 🤝 Contribute # Table of Contents * APIs, Data & ML * Artificial Intelligence * BaaS * Code Editors * Code Generation * DNS * Databases * Design & UI * Domains * Email * Font * For Students * Forms * Linux Distributions * Messaging & Streaming * PaaS * Payments & Billing * SSL
AIMr
AIMr is an AI aimbot tool written in Python that leverages modern technologies to achieve an undetected system with a pleasing appearance. It works on any game that uses human-shaped models. To optimize its performance, users should build OpenCV with CUDA. For Valorant, additional perks in the Discord and an Arduino Leonardo R3 are required.
aika
AIKA (Artificial Intelligence for Knowledge Acquisition) is a new type of artificial neural network designed to mimic the behavior of a biological brain more closely and bridge the gap to classical AI. The network conceptually separates activations from neurons, creating two separate graphs to represent acquired knowledge and inferred information. It uses different types of neurons and synapses to propagate activation values, binding signals, causal relations, and training gradients. The network structure allows for flexible topology and supports the gradual population of neurons and synapses during training.
For similar jobs
sweep
Sweep is an AI junior developer that turns bugs and feature requests into code changes. It automatically handles developer experience improvements like adding type hints and improving test coverage.
teams-ai
The Teams AI Library is a software development kit (SDK) that helps developers create bots that can interact with Teams and Microsoft 365 applications. It is built on top of the Bot Framework SDK and simplifies the process of developing bots that interact with Teams' artificial intelligence capabilities. The SDK is available for JavaScript/TypeScript, .NET, and Python.
ai-guide
This guide is dedicated to Large Language Models (LLMs) that you can run on your home computer. It assumes your PC is a lower-end, non-gaming setup.
classifai
Supercharge WordPress Content Workflows and Engagement with Artificial Intelligence. Tap into leading cloud-based services like OpenAI, Microsoft Azure AI, Google Gemini and IBM Watson to augment your WordPress-powered websites. Publish content faster while improving SEO performance and increasing audience engagement. ClassifAI integrates Artificial Intelligence and Machine Learning technologies to lighten your workload and eliminate tedious tasks, giving you more time to create original content that matters.
chatbot-ui
Chatbot UI is an open-source AI chat app that allows users to create and deploy their own AI chatbots. It is easy to use and can be customized to fit any need. Chatbot UI is perfect for businesses, developers, and anyone who wants to create a chatbot.
BricksLLM
BricksLLM is a cloud native AI gateway written in Go. Currently, it provides native support for OpenAI, Anthropic, Azure OpenAI and vLLM. BricksLLM aims to provide enterprise level infrastructure that can power any LLM production use cases. Here are some use cases for BricksLLM: * Set LLM usage limits for users on different pricing tiers * Track LLM usage on a per user and per organization basis * Block or redact requests containing PIIs * Improve LLM reliability with failovers, retries and caching * Distribute API keys with rate limits and cost limits for internal development/production use cases * Distribute API keys with rate limits and cost limits for students
uAgents
uAgents is a Python library developed by Fetch.ai that allows for the creation of autonomous AI agents. These agents can perform various tasks on a schedule or take action on various events. uAgents are easy to create and manage, and they are connected to a fast-growing network of other uAgents. They are also secure, with cryptographically secured messages and wallets.
griptape
Griptape is a modular Python framework for building AI-powered applications that securely connect to your enterprise data and APIs. It offers developers the ability to maintain control and flexibility at every step. Griptape's core components include Structures (Agents, Pipelines, and Workflows), Tasks, Tools, Memory (Conversation Memory, Task Memory, and Meta Memory), Drivers (Prompt and Embedding Drivers, Vector Store Drivers, Image Generation Drivers, Image Query Drivers, SQL Drivers, Web Scraper Drivers, and Conversation Memory Drivers), Engines (Query Engines, Extraction Engines, Summary Engines, Image Generation Engines, and Image Query Engines), and additional components (Rulesets, Loaders, Artifacts, Chunkers, and Tokenizers). Griptape enables developers to create AI-powered applications with ease and efficiency.