decompose_net
Decompose and Conquer: Introducing the First Open Source Large Language Model with Multi-Modal Task Decomposition Capabilities
Science Score: 44.0%
This score indicates how likely this project is to be science-related based on various indicators:
-
✓CITATION.cff file
Found CITATION.cff file -
✓codemeta.json file
Found codemeta.json file -
✓.zenodo.json file
Found .zenodo.json file -
○DOI references
-
○Academic publication links
-
○Academic email domains
-
○Institutional organization owner
-
○JOSS paper metadata
-
○Scientific vocabulary similarity
Low similarity (9.9%) to scientific vocabulary
Repository
Decompose and Conquer: Introducing the First Open Source Large Language Model with Multi-Modal Task Decomposition Capabilities
Basic Info
- Host: GitHub
- Owner: arash-shahmansoori
- License: mit
- Language: Jupyter Notebook
- Default Branch: main
- Size: 31.4 MB
Statistics
- Stars: 4
- Watchers: 1
- Forks: 0
- Open Issues: 0
- Releases: 0
Metadata Files
README.md
Decompose and Conquer: Introducing the First Open Source Large Language Model with Multi-Modal Task Decomposition Capabilities

Abstract
This repository introduces the first open-source Large Language Model (LLM) fine-tuned for task decomposition, utilizing the 7-billion-parameter Mistral LLM. Our model bridges a critical gap in designing controller agents for hierarchical swarm networks by efficiently decomposing composite tasks into actionable subtasks. This capability is underpinned by a specially curated multi-modal dataset following the alpaca format, designed for the decomposition of composite tasks into text, image, video, and audio subtasks, as well as combinations thereof. Leveraging Quantized Low Rank Adapters (QLORA) for fine-tuning, we push the boundaries of task-specific LLM adaptability and performance. The repository encompasses the training source code, adapters, the original dataset, and comprehensive evaluation results—marking significant strides in task decomposition and multi-modal AI research. https://medium.com/@arash.mansoori65/decompose-and-conquer-introducing-the-first-open-source-large-language-model-with-multi-modal-task-9d3683dd8bed
Main Contributions
- First Open Source LLM for Task Decomposition: We present the pioneering effort of finely tuning an LLM specifically aimed at decomposing composite tasks into manageable subtasks.
- Dataset for Multi-Modal Task Decomposition: Creation of an original dataset, designed to facilitate the analysis and generation of multi-modal composite tasks, setting a new benchmark for task decomposition studies.
- Design of Multi-Modal Dataset for Composite Task Decomposition: Our dataset is a testament to innovation in AI training materials, supporting a wide variety of complexity in multi-modal scenarios, including single-mode tasks.
Repository Contents
- Dataset: The custom-created dataset in alpaca format for multi-modal task decomposition.
- Fine-Tuning Source Code: Complete source code used for fine-tuning the Mistral LLM, featuring advanced techniques such as QLORA.
- Adapters and Tools: All adapters and tools developed and utilized during the fine-tuning process.
- Evaluation Results: Detailed results from the evaluation of the fine-tuned model on various task decomposition benchmarks.
- License: Open-source license detailing how the resources can be used.
Workflow Diagram and Explanation
mermaid
graph TD;
A[Dataset Creation] -->|Multi-Modal Task Decomposition| B(Fine-Tuning Mistral LLM with QLORA)
B --> C[Evaluate Fine-tuned Model]
C --> D[Task Decomposition Success]
C --> E[Adapt & Iterate]
D --> F[Open Source Code & Dataset]
B -.-> G[Hierarchical Swarm Controller Integration]
G --> H[Subtask Dispatching]
F -.-> G
E --> A
This diagram illustrates our iterative process, beginning with Dataset Creation specifically engineered for multi-modal task decomposition challenges. The dataset then drives the Fine-Tuning of the Mistral LLM utilizing QLORA, paving the way for comprehensive Evaluation. Upon successful evaluation, identified as Task Decomposition Success, the process leads to the release of our Open Source Code & Dataset. Parallelly, the fine-tuned model facilitates Hierarchical Swarm Controller Integration, crucial for Subtask Dispatching among agent networks. The cyclic arrow from Adapt & Iterate back to Dataset Creation signifies our commitment to continual improvement and adaptation of our tools and datasets based on evaluation feedback, ensuring that our resources remain at the vanguard of AI research and application.
How to Use
Installation
Install all the necessary requirements.
sh
pip install -r requirements.txt
Environment Variables
Create a .env file and set your openai API key and other required API keys of your choice, e.g., wandb and huggingface.
Dataset Generation
To create the dataset from scratch use the script provided in the dataset directory under create_dataset.py. We have already provided the 1k rows of the dataset for you to use in the data directory stored as 50 json files each of which containing 20 rows of the dataset.
Fine-Tuning
To run the fine-tuning, save the fine-tuned model and tokenizer, and the evaluation results use the following command.
sh
python main.py
Checkpoints
The final adapter checkpoints, tokenizer, and additional files are provided in the final_checkpoint directory.
Results
The evaluation results for decomposing multi-modal composite tasks are stored in the results directory. Other metrics including training and evaluation losses are provided in the assets directory.
Author
Arash Shahmansoori (arash.mansoori65@gmail.com)
License
This project is open-sourced under the MIT License, allowing for widespread use and contribution back to the community. Please refer to the MIT License file for detailed terms and conditions.
Acknowledgements
We extend our gratitude to the AI research community for the open-source tools and datasets that have made this work possible. Special thanks to the contributors of the Mistral LLM and the creators of the QLORA technique for their groundbreaking contributions to the field of AI.
We invite contributors, researchers, and AI enthusiasts to join us in advancing the field of task decomposition and multi-modal AI research. Together, we can build powerful AI systems capable of understanding and executing complex composite tasks.
Owner
- Login: arash-shahmansoori
- Kind: user
- Location: Ireland
- Repositories: 1
- Profile: https://github.com/arash-shahmansoori
Citation (CITATION.cff)
cff-version: 1.2.0
message: "If you use this software, please cite it as below."
authors:
- family-names: "Shahmansoori"
given-names: "Arash"
orcid: "https://orcid.org/0000-0001-5126-8005"
title: "Decompose and Conquer: Introducing the First Open Source Large Language Model with Multi-Modal Task Decomposition Capabilities"
version: "1.0.0"
date-released: "2024-02-24"
doi: ""
url: "https://github.com/arash-shahmansoori/decompose_net.git"
GitHub Events
Total
- Watch event: 1
Last Year
- Watch event: 1