# SVM Model with TF-IDF This repository provides a pre-trained Support Vector Machine (SVM) model for text classification using Term Frequency-Inverse Document Frequency (TF-IDF). The repository also includes utilities for data preprocessing and feature extraction: There are two ways to test our model: # 1.Colab (can see the test_example.py file for how the Colab looks like) ## Start
Download all the files.
Copy all the codes below into Colab
Before running the code, ensure you have all the required libraries installed: ```python pip install nltk beautifulsoup4 scikit-learn pandas datasets fsspec huggingface_hub ```
Download necessary NTLK resources for preprocessing. ```python import nltk nltk.download('stopwords') nltk.download('wordnet') nltk.download('omw-1.4') ```
Clean the Dataset ```python from data_cleaning import clean import pandas as pd import nltk nltk.download('stopwords') ```
You can replace with any datasets you want by changing the file name inside ```pd.read_csv()```. ```python df = pd.read_csv("hf://datasets/CIS5190abcd/headlines_test/test_cleaned_headlines.csv") cleaned_df = clean(df) ``` - Extract TF-IDF Features ```python from tfidf import tfidf X_new_tfidf = tfidf.transform(cleaned_df['title']) ``` - Make Predictions ```python from svm import svm_model ``` # 2. Termial ## Start:
Open your terminal.
Clone the repo by using the following command: ``` git clone https://huggingface.co/CIS5190abcd/svm ```
Go to the svm directory using following command: ``` cd svm ```
Run ```ls``` to check the files inside svm folder. Make sure ```tfidf.py```, ```svm.py``` and ```data_cleaning.py``` are existing in this directory. If not, run the folloing commands: ``` git checkout origin/main -- tfidf.py git checkout origin/main -- svm.py git checkout origin/main -- data_cleaning.py ```
Rerun ```ls```, double check all the required files(```tfidf.py```, ```svm.py``` and ```data_cleaning.py```) are existing. Should look like this: ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6755cffd784ff7ea9db10bd4/O9K5zYm7TKiIg9cYZpV1x.png)
keep inside the svm directory until ends. ## Installation
Before running the code, ensure you have all the required libraries installed: ```python pip install nltk beautifulsoup4 scikit-learn pandas datasets fsspec huggingface_hub ```
Go to Python which can be opened directly in terminal by typing the following command: ``` python ```
Download necessary NTLK resources for preprocessing. ```python import nltk nltk.download('stopwords') nltk.download('wordnet') nltk.download('omw-1.4') ```
After downloading all the required packages, **do not** exit. ## How to use: Training a new dataset with existing SVM model, follow the steps below: - Clean the Dataset ```python from data_cleaning import clean import pandas as pd import nltk nltk.download('stopwords') ```
You can replace with any datasets you want by changing the file name inside ```pd.read_csv()```. ```python df = pd.read_csv("hf://datasets/CIS5190abcd/headlines_test/test_cleaned_headlines.csv") cleaned_df = clean(df) ``` - Extract TF-IDF Features ```python from tfidf import tfidf X_new_tfidf = tfidf.transform(cleaned_df['title']) ``` - Make Predictions ```python from svm import svm_model ``` ```exit()``` if you want to leave python. ```cd ..```if you want to exit svm directory.