Skip to content

beratcmn/local-intelligence

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

23 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Local Intelligence

Local Intelligence is a versatile desktop application designed to perform various text processing tasks using a sophisticated language model. It provides an intuitive GUI with buttons for summarizing text, composing emails, fixing grammar, extracting keywords, and explaining text. The application leverages advanced machine learning models to generate accurate and contextually appropriate outputs.

showcase_1.mp4

Table of Contents

Features

  • Summarize Text: Get concise summaries of lengthy text.
  • Compose Emails: Automatically draft emails based on input text.
  • Fix Grammar: Correct grammatical errors in your text.
  • Extract Keywords: Identify key terms and phrases.
  • Explain Text: Generate explanations for complex text.
  • Rephrase Generated Text: Rephrase the generated text based on the desired tone. Tones are categorized as Casual, Formal, Professional, Technical, Simple.

Todo

  • Add CPU only LLM inference
  • Add more features such as rephrasing based on the desired tone
  • Ensure the application is running smoothly on Mac
  • Ensure the application is running smoothly on Linux
  • Build the project to a standalone executable for Windows, macOS, and Linux
  • Create a user-friendly installer for the application
  • Better theme, icons, and overall design
  • Fix text-view spawning outside of the screen
  • Make text-view window draggable
  • Learn how to be a open source maintainer

Installation

NOTE: You have to install "llama-cpp-python" based on your operating system. You can find the installation instructions here.

Prerequisites

  • Python 3.8 or higher

Steps

  1. Clone the repository:

    git clone https://github.com/beratcmn/local-intelligence.git
    cd local-intelligence
  2. Create a virtual environment:

    python -m venv venv
    source venv/bin/activate  # On Windows use `venv\Scripts\activate.ps1 or venv\Scripts\activate.bat`
  3. Install the dependencies:

    pip install -r requirements.txt
  4. Download the pre-trained language model:

    • Don't worry, llama.cpp got you covered. It will automatically download the model for you to the ./models directory.

Usage

  1. Run the application:

    python src/app.py
  2. Use the GUI:

    • Press CTRL + SPACE to toggle the application window.
    • Copy the text you want to process to the clipboard. (For now, the application only supports clipboard input.)
    • Select the desired task by clicking the appropriate button.
    • The result will be displayed in a new window and copied to the clipboard.

Contributing

I welcome contributions from the community! Here's how you can help:

  1. Fork the repository.
  2. Create a new branch (git checkout -b feature-branch).
  3. Make your changes.
  4. Commit your changes (git commit -am 'Add new feature').
  5. Push to the branch (git push origin feature-branch).
  6. Open a Pull Request.

License

This project is licensed under the Apache 2.0 License - see the LICENSE file for details.

Contact

For any questions or issues, please open an issue in this repository or contact the project maintainers.

About

Something similar to Apple Intelligence?

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages