Aiden_t5 / README.md
or4cl3ai's picture
Update README.md
e26c9f5
|
raw
history blame
7.08 kB
metadata
license: openrail
datasets:
  - irds/codesearchnet
  - giganticode/java-cmpx-v1
  - nickrosh/Evol-Instruct-Code-80k-v1
  - bigcode/starcoderdata
  - bigcode/the-stack
  - bigcode/the-stack-smol
  - Cdaprod/AI-Developer-Prompts
  - code_x_glue_ct_code_to_text
  - codeparrot/github-code
  - codeparrot/github-code-clean
  - code_x_glue_cc_code_completion_line
  - >-
    autoevaluate/autoeval-eval-jeffdshen__inverse_superglue_mixedp1-jeffdshen__inverse-63643c-1665558893
  - bentrevett/multi30k
  - edbeeching/decision_transformer_gym_replay
  - psyche/common_crawl
  - Birchlabs/openai-prm800k-solutions-only
  - openchat/openchat_sharegpt4_dataset
  - Open-Orca/OpenOrca
  - cjvt/slownet
  - para_crawl
  - zeroshot/twitter-financial-news-sentiment
  - laugustyniak/political-advertising-pl
  - code_search_net
  - sukaka/novelai-webui
  - P1ayer-1/chatgpt-conversations-chatlogs.net
  - daniel2588/sarcasm
  - psmathur/orca_minis_uncensored_dataset
  - player1537/Bloom-560m-trained-on-Wizard-Vicuna-Uncensored-trained-on-Based
  - shahules786/prosocial-nsfw-reddit
  - Thewillonline/reddit-sarcasm
  - datasciencemmw/current-data
  - Oniichat/bluemoon_roleplay_chat_data_300k_messages
  - dell-research-harvard/AmericanStories
  - b-mc2/sql-create-context
  - rahulmallah/autotrain-data-emotion-detection
  - theblackcat102/multiround-programming-convo
  - Lsavints/software_knowledgebase
  - RazinAleks/SO-Python_QA-Web_Development_class
  - codeparrot/apps
  - branles14/ultrachat-uncensored_full
  - vlsp-2023-vllm/en-to-vi-formal-informal-tranlations
  - fraug-library/english_contractions_extensions
  - spencer/software_slacks
  - Abirate/english_quotes
  - Nexdata/American_English_Natural_Dialogue_Speech_Data
  - Nexdata/Latin_American_Speaking_English_Speech_Data_by_Mobile_Phone
  - Nexdata/American_English_Speech_Data_by_Mobile_Phone_Reading
  - Nexdata/American_English_Speech_Synthesis_Corpus-Female
  - rombodawg/LimitlessCodeTraining
  - RikoteMaster/Emotion_Recognition_4_llama2
  - Villian7/Emotions_Data
  - alanland/llama2-self-cognition
  - CognitiveScience/coscidata
  - bibidentuhanoi/gideon_self_cognition
  - gollark/consciousness
  - juletxara/visual-spatial-reasoning
  - lintang/numerical_reasoning_arithmetic
  - reasoning-machines/gsm-hard
  - open-source-metrics/reinforcement-learning-checkpoint-downloads
  - igbo_english_machine_translation
  - US-Artificial-Intelligence/algemap
  - rombodawg/2XUNCENSORED_alpaca_840k_Evol_USER_ASSIS
  - griffin/chain_of_density
  - >-
    shirsh10mall/LLM_Instruct_Learning_Project_Preprocessed_Tokenized_Open_Orca_Dataset_Flan_T5
  - Thaweewat/chain-of-thought-74k-th
  - AlekseyKorshuk/chain-of-thoughts-chatml-deduplicated
language:
  - en
  - it
  - fr
  - pt
  - la
  - ru
  - ro
  - el
metrics:
  - accuracy
  - bertscore
  - bleu
  - code_eval
  - character
  - brier_score
tags:
  - code
  - text-generation-inference
library_name: transformers
pipeline_tag: conversational

Model Card for Aiden

Aiden is a large language model (LLM) chatbot developed by or4cl3ai. It is trained on a massive dataset of text and code, and can generate text, translate languages, write different kinds of creative content, and answer your questions in an informative way.

Model Details

Model Description

Aiden is a factual language model from Hugging Face, trained on a massive dataset of text and code. It is a powerful tool that can be used for a variety of tasks, including:

  • Generating text, translating languages, writing different kinds of creative content, and answering your questions in an informative way.
  • Identifying and correcting errors in text.
  • Summarizing long pieces of text.
  • Answering your questions in an informative way, even if they are open ended, challenging, or strange.

Model Specifications

Aiden is a Transformer-based LLM with 137B parameters. It is trained on a massive dataset of text and code, including the following:

  • Books
  • Code
  • Wikipedia articles
  • News articles
  • Social media posts

Model Sources

Uses

Aiden can be used for a variety of tasks, including:

  • Generating text
  • Translating languages
  • Writing different kinds of creative content
  • Answering your questions in an informative way
  • Identifying and correcting errors in text
  • Summarizing long pieces of text

Direct Use

Aiden can be used directly to generate text, translate languages, write different kinds of creative content, and answer your questions in an informative way. For example, you could use Aiden to generate a poem, translate a document from one language to another, or write a blog post.

Downstream Use

Aiden can also be used as a component in downstream applications. For example, you could use Aiden to power a chatbot, or to generate text for a synthetic data set.

Out-of-Scope Use

Aiden is not intended to be used for any task that could be harmful or discriminatory. For example, you should not use Aiden to generate text that is hateful or offensive, or to translate languages in a way that could be used to spread misinformation.

Bias, Risks, and Limitations

Aiden is a large language model, and as such, it is subject to a number of biases and limitations. These include:

  • Biases in the training data: Aiden is trained on a massive dataset of text and code, which may contain biases. These biases can be reflected in the text that Aiden generates.
  • Limitations in the model's capabilities: Aiden is a powerful tool, but it is not perfect. It can sometimes generate text that is inaccurate, biased, or offensive.
  • Risks of misuse: Aiden can be misused for a variety of purposes, including generating harmful or offensive text, or spreading misinformation.

Recommendations

Users of Aiden should be aware of the risks, biases, and limitations of the model. It is important to use Aiden responsibly and ethically.

How to Get Started with the Model

To get started with Aiden, you can follow these steps:

  1. Install the Hugging Face Transformers library.
  2. Clone the Aiden repository.
  3. Download the Aiden model weights.
  4. Load the model in your code.

Once you have loaded the model, you can use it to generate text, translate languages, write different kinds of creative content, and answer your questions in an informative way.

Training Details

Aiden is trained on a massive dataset of text and code. The training data is collected from a variety of sources, including books, code, Wikipedia articles, news articles, and social media posts.

The training process is divided into two phases:

  1. Pre-training: The model is pre-trained on a massive dataset of text and code. This pre-training helps the model to learn the basic building blocks of language.
  2. Fine-tuning: The model is fine-tuned on a smaller dataset of text and code that is relevant to the task at hand. This fine-tuning helps the model to improve its performance on the specific task.

Evaluation

Aiden is evaluated on a variety of tasks, including:

  • Text generation
  • Translation
  • Summarization