Issue in dataset / Escapes every special character

#2
by CyberTimon - opened

Hello
When using the model and it tries to write code etc it escapes every special character..
I think its an issue in the dataset.
Here is an example generation (It happens every time):

Sure, here's an example Python code that uses the Hugging Face Transformers library to train and evaluate a Google T5 model on some text data:

import transformers as tf
from transformers import GPT2LMHead, GPT2Tokenizer, GPT2ForSequenceClassification
# Load pre-trained tokenizer and classifier
tokenizer = GPT2Tokenizer.from\_pretrained('gpt2')
classifier = GPT2ForSequenceClassification.from\_pretrained('gpt2')
model = GPT2LMHead.from\_pretrained('gpt2')
# Define input/output sequences for training and evaluation
train\_data = ['This is my first sentence', 'This is another sentence']
valid\_data = [tf.constant(["I love this language!"]), tf.constant("It works well!")]
test\_data = ["This is yet another sentence"]
# Train the model using the train data
train\_dataset = tf.data.Dataset.from\_tensor\_slices((train\_data)).repeat().batch(1)
inputs = tokenizer(\_, padding=True, truncation=True)

It mostly happens before "_" characters

Cognitive Computations org

does this also happen in wizard-vicuna-13b? If so, this feedback should go to that repository.
I am pretty sure my cleanup script didn't introduce this issue, but if I made a regression I'm happy to fix it.

Hello
No it doesn't happen on wizard-vicuna-13b.
I tested many prompts and all the time your uncensored one escaped it and the normal one didn't.

Cognitive Computations org

ok thank you for this feedback i will look into it

Cognitive Computations org

I fixed the problem, and now I need to retrain the model. It will take about a week.

Oh okay, thank you for the effort!

Cognitive Computations org

The training is running now. It will be finished on May 17.

Great! Thank you so much

Thank you very much the issue is fixed! The model is great

CyberTimon changed discussion status to closed
Cognitive Computations org

Thank you for testing it

Sign up or log in to comment