juncliu commited on
Commit
7c5c5e0
·
1 Parent(s): 35d42fa

update about

Browse files
Files changed (1) hide show
  1. src/about.py +6 -22
src/about.py CHANGED
@@ -43,41 +43,25 @@ points, spanning seven domains, 10 frequencies, multivariate inputs, and predict
43
  # Which evaluations are you running? how can people reproduce what you have?
44
  LLM_BENCHMARKS_TEXT = f"""
45
  How It Works
46
- To participate in the ContextualBench leaderboard, follow these steps to evaluate your Large Language Model (LLM) using the ContextualBench framework:
47
 
48
- Clone the Repository: Start by cloning the ContextualBench GitHub repository to your local machine using the following command:
 
 
49
 
50
  ```bash
51
- git clone https://github.com/SalesforceAIResearch/SFR-RAG
52
  ```
53
 
54
  Navigate to the Directory: Move into the cloned repository's directory:
55
 
56
 
57
- ``bash
58
- cd ContextualBench
59
- ```
60
-
61
- Install Dependencies: Install all necessary dependencies by executing:
62
-
63
  ```bash
64
- pip install -r requirements.txt
65
  ```
66
 
67
- Prepare Your Model and Dataset: Set up your model and dataset according to the guidelines provided in the repository's documentation.
68
- Run the Evaluation Script: Execute the evaluation script to generate outputs for your model on the specified dataset:
69
-
70
-
71
- ```bash
72
- python run.py [dataset_name]
73
- ```
74
 
75
- Collect and Format Outputs: Gather the outputs generated for each dataset and format them according to the leaderboard submission guidelines.
76
- Submit Your Results: Email the formatted outputs to the author's email address for evaluation. Our team will assess the performance and update the leaderboard accordingly.
77
 
78
- Reproducibility
79
- Ensuring reproducibility is a key aspect of the ContextualBench leaderboard.
80
- By following the standardized steps outlined above, participants can consistently reproduce evaluation results. This process not only facilitates fair comparisons across different models but also encourages transparency and reliability in model assessments. Participants are encouraged to adhere strictly to the submission guidelines to ensure their results are accurately reflected on the leaderboard.
81
  """
82
 
83
  EVALUATION_QUEUE_TEXT = """
 
43
  # Which evaluations are you running? how can people reproduce what you have?
44
  LLM_BENCHMARKS_TEXT = f"""
45
  How It Works
 
46
 
47
+ To participate in the GIFT-Eval leaderboard, follow these steps to evaluate your Time Series Model:
48
+
49
+ Clone the Repository: Start by cloning the GIFT-Eval GitHub repository to your local machine using the following command:
50
 
51
  ```bash
52
+ git clone https://github.com/SalesforceAIResearch/gift-eval
53
  ```
54
 
55
  Navigate to the Directory: Move into the cloned repository's directory:
56
 
57
 
 
 
 
 
 
 
58
  ```bash
59
+ cd gift-eval
60
  ```
61
 
62
+ Follow the instruction in the README.md file to install the required dependencies, set up your environment and obtain the evaluation results.
 
 
 
 
 
 
63
 
 
 
64
 
 
 
 
65
  """
66
 
67
  EVALUATION_QUEUE_TEXT = """