Adds the missing layers from openai/clip-vit-large-patch14-336 to microsoft/LLM2CLIP-Openai-L-14-336 so that it can be loaded in ComfyUI.
import torch
from safetensors.torch import load_file, save_file
def newclip(original_path, new_path, combined_path):
original = load_file(original_path)
new = load_file(new_path)
combined = {}
for key in original.keys():
if key in new.keys():
combined[key] = new[key]
else:
combined[key] = original[key]
save_file(combined, combined_path)
newclip(
"./original-clip-l.safetensors",
"./new-llm2clip-clip-l.safetensors",
"./combined.safetensors"
)
Unable to determine this model's library. Check the
docs
.