danielhanchen commited on
Commit
06b68ef
1 Parent(s): cd3117a

Upload 9 files

Browse files
Alpaca_+_Codellama_34b_full_example.ipynb CHANGED
@@ -27,17 +27,10 @@
27
  "outputs": [],
28
  "source": [
29
  "%%capture\n",
30
- "import torch\n",
31
- "major_version, minor_version = torch.cuda.get_device_capability()\n",
32
- "# Must install separately since Colab has torch 2.2.1, which breaks packages\n",
33
- "!pip install \"unsloth[colab-new] @ git+https://github.com/unslothai/unsloth.git\"\n",
34
- "if major_version >= 8:\n",
35
- " # Use this for new GPUs like Ampere, Hopper GPUs (RTX 30xx, RTX 40xx, A100, H100, L40)\n",
36
- " !pip install --no-deps packaging ninja einops flash-attn xformers trl peft accelerate bitsandbytes\n",
37
- "else:\n",
38
- " # Use this for older GPUs (V100, Tesla T4, RTX 20xx)\n",
39
- " !pip install --no-deps xformers trl peft accelerate bitsandbytes\n",
40
- "pass"
41
  ]
42
  },
43
  {
 
27
  "outputs": [],
28
  "source": [
29
  "%%capture\n",
30
+ "# Installs Unsloth, Xformers (Flash Attention) and all other packages\n!",
31
+ "!pip install unsloth\n",
32
+ "# Get latest Unsloth\n",
33
+ "!pip install --upgrade --no-deps \"unsloth[colab-new] @ git+https://github.com/unslothai/unsloth.git\"\n"
 
 
 
 
 
 
 
34
  ]
35
  },
36
  {
Alpaca_+_Gemma_7b_full_example.ipynb CHANGED
@@ -29,17 +29,10 @@
29
  "outputs": [],
30
  "source": [
31
  "%%capture\n",
32
- "import torch\n",
33
- "major_version, minor_version = torch.cuda.get_device_capability()\n",
34
- "# Must install separately since Colab has torch 2.2.1, which breaks packages\n",
35
- "!pip install \"unsloth[colab-new] @ git+https://github.com/unslothai/unsloth.git\"\n",
36
- "if major_version >= 8:\n",
37
- " # Use this for new GPUs like Ampere, Hopper GPUs (RTX 30xx, RTX 40xx, A100, H100, L40)\n",
38
- " !pip install --no-deps packaging ninja einops flash-attn xformers trl peft accelerate bitsandbytes\n",
39
- "else:\n",
40
- " # Use this for older GPUs (V100, Tesla T4, RTX 20xx)\n",
41
- " !pip install --no-deps xformers trl peft accelerate bitsandbytes\n",
42
- "pass"
43
  ]
44
  },
45
  {
 
29
  "outputs": [],
30
  "source": [
31
  "%%capture\n",
32
+ "# Installs Unsloth, Xformers (Flash Attention) and all other packages\n!",
33
+ "!pip install unsloth\n",
34
+ "# Get latest Unsloth\n",
35
+ "!pip install --upgrade --no-deps \"unsloth[colab-new] @ git+https://github.com/unslothai/unsloth.git\"\n"
 
 
 
 
 
 
 
36
  ]
37
  },
38
  {
Alpaca_+_Llama_7b_full_example.ipynb CHANGED
@@ -27,17 +27,10 @@
27
  "outputs": [],
28
  "source": [
29
  "%%capture\n",
30
- "import torch\n",
31
- "major_version, minor_version = torch.cuda.get_device_capability()\n",
32
- "# Must install separately since Colab has torch 2.2.1, which breaks packages\n",
33
- "!pip install \"unsloth[colab-new] @ git+https://github.com/unslothai/unsloth.git\"\n",
34
- "if major_version >= 8:\n",
35
- " # Use this for new GPUs like Ampere, Hopper GPUs (RTX 30xx, RTX 40xx, A100, H100, L40)\n",
36
- " !pip install --no-deps packaging ninja einops flash-attn xformers trl peft accelerate bitsandbytes\n",
37
- "else:\n",
38
- " # Use this for older GPUs (V100, Tesla T4, RTX 20xx)\n",
39
- " !pip install --no-deps xformers trl peft accelerate bitsandbytes\n",
40
- "pass"
41
  ]
42
  },
43
  {
 
27
  "outputs": [],
28
  "source": [
29
  "%%capture\n",
30
+ "# Installs Unsloth, Xformers (Flash Attention) and all other packages\n!",
31
+ "!pip install unsloth\n",
32
+ "# Get latest Unsloth\n",
33
+ "!pip install --upgrade --no-deps \"unsloth[colab-new] @ git+https://github.com/unslothai/unsloth.git\"\n"
 
 
 
 
 
 
 
34
  ]
35
  },
36
  {
Alpaca_+_Mistral_7b_full_example.ipynb CHANGED
@@ -27,17 +27,10 @@
27
  "outputs": [],
28
  "source": [
29
  "%%capture\n",
30
- "import torch\n",
31
- "major_version, minor_version = torch.cuda.get_device_capability()\n",
32
- "# Must install separately since Colab has torch 2.2.1, which breaks packages\n",
33
- "!pip install \"unsloth[colab-new] @ git+https://github.com/unslothai/unsloth.git\"\n",
34
- "if major_version >= 8:\n",
35
- " # Use this for new GPUs like Ampere, Hopper GPUs (RTX 30xx, RTX 40xx, A100, H100, L40)\n",
36
- " !pip install --no-deps packaging ninja einops flash-attn xformers trl peft accelerate bitsandbytes\n",
37
- "else:\n",
38
- " # Use this for older GPUs (V100, Tesla T4, RTX 20xx)\n",
39
- " !pip install --no-deps xformers trl peft accelerate bitsandbytes\n",
40
- "pass"
41
  ]
42
  },
43
  {
 
27
  "outputs": [],
28
  "source": [
29
  "%%capture\n",
30
+ "# Installs Unsloth, Xformers (Flash Attention) and all other packages\n!",
31
+ "!pip install unsloth\n",
32
+ "# Get latest Unsloth\n",
33
+ "!pip install --upgrade --no-deps \"unsloth[colab-new] @ git+https://github.com/unslothai/unsloth.git\"\n"
 
 
 
 
 
 
 
34
  ]
35
  },
36
  {
Alpaca_+_TinyLlama_+_RoPE_Scaling_full_example.ipynb CHANGED
@@ -29,17 +29,10 @@
29
  "outputs": [],
30
  "source": [
31
  "%%capture\n",
32
- "import torch\n",
33
- "major_version, minor_version = torch.cuda.get_device_capability()\n",
34
- "# Must install separately since Colab has torch 2.2.1, which breaks packages\n",
35
- "!pip install \"unsloth[colab-new] @ git+https://github.com/unslothai/unsloth.git\"\n",
36
- "if major_version >= 8:\n",
37
- " # Use this for new GPUs like Ampere, Hopper GPUs (RTX 30xx, RTX 40xx, A100, H100, L40)\n",
38
- " !pip install --no-deps packaging ninja einops flash-attn xformers trl peft accelerate bitsandbytes\n",
39
- "else:\n",
40
- " # Use this for older GPUs (V100, Tesla T4, RTX 20xx)\n",
41
- " !pip install --no-deps xformers trl peft accelerate bitsandbytes\n",
42
- "pass"
43
  ]
44
  },
45
  {
 
29
  "outputs": [],
30
  "source": [
31
  "%%capture\n",
32
+ "# Installs Unsloth, Xformers (Flash Attention) and all other packages\n!",
33
+ "!pip install unsloth\n",
34
+ "# Get latest Unsloth\n",
35
+ "!pip install --upgrade --no-deps \"unsloth[colab-new] @ git+https://github.com/unslothai/unsloth.git\"\n"
 
 
 
 
 
 
 
36
  ]
37
  },
38
  {
ChatML_+_chat_templates_+_Mistral_7b_full_example.ipynb CHANGED
@@ -29,17 +29,10 @@
29
  "outputs": [],
30
  "source": [
31
  "%%capture\n",
32
- "import torch\n",
33
- "major_version, minor_version = torch.cuda.get_device_capability()\n",
34
- "# Must install separately since Colab has torch 2.2.1, which breaks packages\n",
35
- "!pip install \"unsloth[colab-new] @ git+https://github.com/unslothai/unsloth.git\"\n",
36
- "if major_version >= 8:\n",
37
- " # Use this for new GPUs like Ampere, Hopper GPUs (RTX 30xx, RTX 40xx, A100, H100, L40)\n",
38
- " !pip install --no-deps packaging ninja einops flash-attn xformers trl peft accelerate bitsandbytes\n",
39
- "else:\n",
40
- " # Use this for older GPUs (V100, Tesla T4, RTX 20xx)\n",
41
- " !pip install --no-deps xformers trl peft accelerate bitsandbytes\n",
42
- "pass"
43
  ]
44
  },
45
  {
 
29
  "outputs": [],
30
  "source": [
31
  "%%capture\n",
32
+ "# Installs Unsloth, Xformers (Flash Attention) and all other packages\n!",
33
+ "!pip install unsloth\n",
34
+ "# Get latest Unsloth\n",
35
+ "!pip install --upgrade --no-deps \"unsloth[colab-new] @ git+https://github.com/unslothai/unsloth.git\"\n"
 
 
 
 
 
 
 
36
  ]
37
  },
38
  {
DPO_Zephyr_Unsloth_Example.ipynb CHANGED
@@ -28,17 +28,10 @@
28
  "outputs": [],
29
  "source": [
30
  "%%capture\n",
31
- "import torch\n",
32
- "major_version, minor_version = torch.cuda.get_device_capability()\n",
33
- "# Must install separately since Colab has torch 2.2.1, which breaks packages\n",
34
- "!pip install \"unsloth[colab-new] @ git+https://github.com/unslothai/unsloth.git\"\n",
35
- "if major_version >= 8:\n",
36
- " # Use this for new GPUs like Ampere, Hopper GPUs (RTX 30xx, RTX 40xx, A100, H100, L40)\n",
37
- " !pip install --no-deps packaging ninja einops flash-attn xformers trl peft accelerate bitsandbytes\n",
38
- "else:\n",
39
- " # Use this for older GPUs (V100, Tesla T4, RTX 20xx)\n",
40
- " !pip install --no-deps xformers trl peft accelerate bitsandbytes\n",
41
- "pass"
42
  ]
43
  },
44
  {
 
28
  "outputs": [],
29
  "source": [
30
  "%%capture\n",
31
+ "# Installs Unsloth, Xformers (Flash Attention) and all other packages\n!",
32
+ "!pip install unsloth\n",
33
+ "# Get latest Unsloth\n",
34
+ "!pip install --upgrade --no-deps \"unsloth[colab-new] @ git+https://github.com/unslothai/unsloth.git\"\n"
 
 
 
 
 
 
 
35
  ]
36
  },
37
  {
Mistral_7b_Text_Completion_Raw_Text_training_full_example(1).ipynb ADDED
The diff for this file is too large to render. See raw diff
 
Mistral_7b_Text_Completion_Raw_Text_training_full_example.ipynb CHANGED
@@ -39,17 +39,10 @@
39
  "outputs": [],
40
  "source": [
41
  "%%capture\n",
42
- "import torch\n",
43
- "major_version, minor_version = torch.cuda.get_device_capability()\n",
44
- "# Must install separately since Colab has torch 2.2.1, which breaks packages\n",
45
- "!pip install \"unsloth[colab-new] @ git+https://github.com/unslothai/unsloth.git\"\n",
46
- "if major_version >= 8:\n",
47
- " # Use this for new GPUs like Ampere, Hopper GPUs (RTX 30xx, RTX 40xx, A100, H100, L40)\n",
48
- " !pip install --no-deps packaging ninja einops flash-attn xformers trl peft accelerate bitsandbytes\n",
49
- "else:\n",
50
- " # Use this for older GPUs (V100, Tesla T4, RTX 20xx)\n",
51
- " !pip install --no-deps xformers trl peft accelerate bitsandbytes\n",
52
- "pass"
53
  ]
54
  },
55
  {
 
39
  "outputs": [],
40
  "source": [
41
  "%%capture\n",
42
+ "# Installs Unsloth, Xformers (Flash Attention) and all other packages\n!",
43
+ "!pip install unsloth\n",
44
+ "# Get latest Unsloth\n",
45
+ "!pip install --upgrade --no-deps \"unsloth[colab-new] @ git+https://github.com/unslothai/unsloth.git\"\n"
 
 
 
 
 
 
 
46
  ]
47
  },
48
  {