--- language: - ja tags: - causal-lm - not-for-all-audiences - nsfw pipeline_tag: text-generation --- # Berghof Vanilla 7B drawing ## Model Description This is a 7B-parameter decoder-only Japanese language model fine-tuned on novel datasets, built on top of the base model Japanese Stable LM Base Gamma 7B. [Japanese Stable LM Instruct Gamma 7B](https://huggingface.co/stabilityai/japanese-stablelm-instruct-gamma-7b) ## Usage Ensure you are using Transformers 4.34.0 or newer. ```python import torch from transformers import AutoTokenizer, AutoModelForCausalLM tokenizer = AutoTokenizer.from_pretrained("Elizezen/Berghof-vanilla-7B") model = AutoModelForCausalLM.from_pretrained( "Elizezen/Berghof-vanilla-7B", torch_dtype="auto", ) model.eval() if torch.cuda.is_available(): model = model.to("cuda") input_ids = tokenizer.encode( "吾輩は猫である。名前はまだない",, add_special_tokens=True, return_tensors="pt" ) tokens = model.generate( input_ids.to(device=model.device), max_new_tokens=512, temperature=1, top_p=0.95, do_sample=True, ) out = tokenizer.decode(tokens[0][input_ids.shape[1]:], skip_special_tokens=True).strip() print(out) ``` ### Intended Use The model is mainly intended to be used for generating novels. It may not be so capable with instruction-based responses.