Spaces:
Runtime error
Runtime error
title: GPT4Tools | |
emoji: π | |
colorFrom: indigo | |
colorTo: pink | |
sdk: gradio | |
sdk_version: 3.32.0 | |
app_file: app.py | |
pinned: false | |
license: apache-2.0 | |
# GPT4Tools: Teaching LLM to Use Tools via Self-instruction | |
[Lin Song](http://linsong.info/), [Yanwei Li](https://yanwei-li.com/), [Rui Yang](https://github.com/Yangr116), Sijie Zhao, [Yixiao Ge](https://geyixiao.com/), [Ying Shan](https://scholar.google.com/citations?user=4oXBp9UAAAAJ&hl=en) | |
GPT4Tools is a centralized system that can control multiple visual foundation models. | |
It is based on Vicuna (LLaMA), and 71K self-built instruction data. | |
By analyzing the language content, GPT4Tools is capable of automatically deciding, controlling, and utilizing different visual foundation models, allowing the user to interact with images during a conversation. | |
With this approach, GPT4Tools provides a seamless and efficient solution to fulfill various image-related requirements in a conversation. | |
Different from previous work, we support users teach their own LLM to use tools with simple refinement via self-instruction and LoRA. | |
<a href='https://github.com/StevenGrove/GPT4Tools'><img src='https://img.shields.io/badge/Project-Page-Green'></a> <a href='https://huggingface.co/stevengrove/gpt4tools-vicuna-13b-lora'><img src='https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-Model-blue'></a> [![YouTube](https://badges.aleen42.com/src/youtube.svg)](https://youtu.be/Qrj94ibQIT8) [![arXiv](https://img.shields.io/badge/arXiv-Paper-<COLOR>.svg)]() | |
ArXiv url: https://arxiv.org/abs//2305.18752 |