Chat-With-RTX-python-api

Chat-With-RTX-python-api

Chat With RTX Python API

Stars: 53

Visit
 screenshot

This repository contains a Python API for Chat With RTX, which allows users to interact with RTX models for natural language processing. The API provides functionality to send messages and receive responses from various LLM models. It also includes information on the speed of different models supported by Chat With RTX. The repository has a history of updates, including the removal of a feature and the addition of a new model for speech-to-text conversion. The repository is licensed under CC0.

README:

Python API for Chat With RTX

Usage

.\start_server.bat

import rtx_api_july_2024 as rtx_api

response = rtx_api.send_message("write fire emoji")
print(response)

Speed

Chat With RTX builds int4 (W4A16 AWQ) tensortRT engines for LLMs

Model On 4090
Mistral 457 char/sec
Llama2 315 char/sec
ChatGLM3 385 char/sec
Gemma 407 char/sec



Update History of Chat With RTX
3.2024  Removed youtube video transcript fetch
4.2024  Added Whisper Speech to text model
7.2024  Electron app ui

LICENSE: CC0

For Tasks:

Click tags to check more tools for each tasks

For Jobs:

Alternative AI tools for Chat-With-RTX-python-api

Similar Open Source Tools

For similar tasks

For similar jobs