Home

Awesome

<p align="center"> <a href="#"> <img src="https://github.com/user-attachments/assets/df1a3ec1-8c03-43c8-9432-c71358c35b9e" > </a> <br> <p align="center"> <a href="https://docs.gca.dev"> <img src="https://github.com/user-attachments/assets/c60562bf-540e-47d9-b578-994285071128" width="250"> </a> </p> <p align="center"> <a href="https://www.producthunt.com/posts/gpt-computer-assistant?embed=true&utm_source=badge-top-post-badge&utm_medium=badge&utm_souce=badge-gpt&#0045;computer&#0045;assistant" target="_blank"><img src="https://api.producthunt.com/widgets/embed-image/v1/top-post-badge.svg?post_id=465468&theme=dark&period=daily" alt="GPT&#0032;Computer&#0032;Assistant - Create&#0032;intelligence&#0032;for&#0032;your&#0032;products | Product Hunt" style="width: 250px; height: 54px;" width="250" height="54" /></a> . <a href="https://discord.gg/qApFmWMt8x"><img alt="Static Badge" src="https://img.shields.io/badge/Discord-Join?style=social&logo=discord" width=150></a> . <a href="https://x.com/GPTCompAsst"><img alt="Static Badge" src="https://img.shields.io/badge/X_App-Join?style=social&logo=x" width=150></a> </p> <p align="center"> <br /> Intelligence development framework <br /> </p> <br> <p align="center"> <a href="https://www.python.org/"> <img src="https://img.shields.io/badge/Made%20with-Python-1f425f.svg" alt="Made_with_python"> </a> . <img src="https://static.pepy.tech/personalized-badge/gpt-computer-assistant?period=total&units=international_system&left_color=grey&right_color=blue&left_text=PyPI%20Downloads" alt="pypi_downloads"> </p>

|ENGLISH|简体中文|正體中文|TÜRKÇE

GPT Computer Assistant(GCA)

Hi, this is an alternative work for providing ChatGPT MacOS app to Windows and Linux. In this way this is a fresh and stable work. You can easily install as Python library for this time but we will prepare a pipeline for providing native install scripts (.exe).

Powered by <a href="https://github.com/Upsonic/Tiger"><strong>Upsonic Tiger 🐅</strong></a> A function hub for llm agents.

1. Install and run

Python 3.10 or 3.11 is required

pip install 'gpt-computer-assistant[base]'
pip install 'gpt-computer-assistant[api]'

To run gpt-computer-assistant, simply type

computerassistant --api
<p align="center"> <a href="#"> <img src="https://github.com/user-attachments/assets/890b4e0a-4484-4870-a158-2d365b0d969e" > </a> </p> <p align="center"> <br> <br> <br> <br> <br> </p>

2. LLM Settings

from gpt_computer_assistant.remote import remote

remote.save_models("gpt-4o")
remote.save_openai_api_key("sk-**")
<p align="start"> <a href="https://docs.upsonic.co/gca/dev_guides/llm_settings"> <img src="https://github.com/user-attachments/assets/a75c8ddf-f9df-436b-9dc8-c5220211e15e" width="150"> </a> </p> <p align="center"> <br> <br> <br> </p>

3. Characteristic API

# Name of the assitant:
remote.change_name("X Intelligence")

#Developer personna of the assistant:
remote.change_developer("X Company")
<p align="start"> <a href="https://docs.upsonic.co/gca/dev_guides/characteristic"> <img src="https://github.com/user-attachments/assets/d7e02ac6-e40c-4b35-8e65-4621bf3fb9a1" width="150"> </a> </p> <p align="center"> <br> <br> <br> </p>

4. Connect Your Functions API

# Installing an library:
remote.install_library("numpy")



# Adding functianility as python functions:
@remote.custom_tool
def my_server_status() -> bool:
  """
  Check the server status.
  """
  return True
<p align="center"> <br> <br> <br> </p>

5. Interact with User API

remote.input

Talk with assistant, about user and computer. With this api you can create an consulting process.

`Hi, look to user window and return which app using now`

`Ask user to is user need any kind of supoprt`

`Extract the user redis config file.`

With this questions you will make a shortcut for your needs. You can collect informations from user computer or directly from user or user computer.

output = remote.input("Extract the user redis config file.", screen=False)
print(output)
<p align="start"> <a href="https://docs.upsonic.co/gca/dev_guides/interact"> <img src="https://github.com/user-attachments/assets/81614347-ab85-4965-9b77-225d0f2961e9" width="150"> </a> . <a href="https://docs.upsonic.co/gca/dev_guides/interact"> <img src="https://github.com/user-attachments/assets/ecaa7590-f4c5-4eda-9482-462cef54aeff" width="150"> </a> . <a href="https://docs.upsonic.co/gca/dev_guides/interact"> <img src="https://github.com/user-attachments/assets/0f35df10-b32e-4fa1-936e-b336be46b1bd" width="150"> </a> </p> <p align="start"> <a href="https://docs.upsonic.co/gca/dev_guides/interact"> <img src="https://github.com/user-attachments/assets/a893c50c-3ede-4b42-90ee-92e2fea82120" width="150"> </a> </p> <p align="center"> <br> <br> <br> </p>

Usage

options

Use cases

<img alt="Screenshot 2024-08-13 at 18 33 52" src="https://github.com/user-attachments/assets/8f994160-893a-4f56-bbf0-4a7aa87af650">

Roadmap

FeatureStatusTarget Release
Clear Chat HistoryCompletedQ2 2024
Long Audios Support (Split 20mb)CompletedQ2 2024
Text InputsCompletedQ2 2024
Just Text Mode (Mute Speech)CompletedQ2 2024
Added profiles (Different Chats)CompletedQ2 2024
More Feedback About Assistant StatusCompletedQ2 2024
Local Model Vision and Text (With Ollama, and vision models)CompletedQ2 2024
Our Customizable Agent InfrastructureCompletedQ2 2024
Supporting Groq ModelsCompletedQ2 2024
Adding Custom ToolsCompletedQ2 2024
Click on something on the screen (text and icon)CompletedQ2 2024
New UICompletedQ2 2024
Native Applications, exe, dmgCompletedQ3 2024
Collaborated Speaking Different Voice Models on long responses.CompletedQ2 2024
Auto Stop Recording, when you complate talkingCompletedQ2 2024
Wakeup WordCompletedQ2 2024
Continuously ConversationsCompletedQ2 2024
Adding more capability on deviceCompletedQ2 2024
Local TTSCompletedQ3 2024
Local STTCompletedQ3 2024
Tray MenuCompletedQ3 2024
Global HotkeyOn the wayQ3 2024
DeepFace Integration (Facial Recognition)PlannedQ3 2024

Capabilities

At this time we have many infrastructure elements. We just aim to provide whole things that already in ChatGPT app.

CapabilityStatus
Local LLM with Vision (Ollama)OK
Local text-to-speechOK
Local speech-to-textOK
Screen ReadOK
Click to and Text or Icon in the screenOK
Move to and Text or Icon in the screenOK
Typing SomethingOK
Pressing to Any KeyOK
ScrollingOK
MicrophoneOK
System AudioOK
MemoryOK
Open and Close AppOK
Open a URLOK
ClipboardOK
Search EnginesOK
Writing and running PythonOK
Writing and running SHOK
Using your Telegram AccountOK
Knowledge ManagementOK
Add more tool?

Predefined Agents

If you enable it your assistant will work with these teams:

Team NameStatus
search_on_internet_and_report_teamOK
generate_code_with_aim_team_OK
Add your own one?
<a href="#"> <img src="https://github.com/onuratakan/gpt-computer-assistant/assets/41792982/ba590bf8-6059-4cb6-8c4e-6d105ce4edd2" alt="Logo" > </a>

Contributors

<a href="https://github.com/onuratakan/gpt-computer-assistant/graphs/contributors"> <img src="https://contrib.rocks/image?repo=onuratakan/gpt-computer-assistant" /> </a>