yentinglin commited on
Commit
6edca0c
1 Parent(s): 4483cea

Update src/about.py

Browse files
Files changed (1) hide show
  1. src/about.py +17 -5
src/about.py CHANGED
@@ -27,17 +27,29 @@ TITLE = """<h1 align="center" id="space-title">Demo leaderboard</h1>"""
27
 
28
  # What does your leaderboard evaluate?
29
  INTRODUCTION_TEXT = """
30
- Intro text
 
31
  """
32
 
33
  # Which evaluations are you running? how can people reproduce what you have?
34
  LLM_BENCHMARKS_TEXT = f"""
35
- ## How it works
36
 
37
- ## Reproducibility
38
- To reproduce our results, here is the commands you can run:
 
 
39
 
40
- please checkout this command https://github.com/adamlin120/lm-evaluation-harness/blob/main/run_all.sh
 
 
 
 
 
 
 
 
 
41
  """
42
 
43
  EVALUATION_QUEUE_TEXT = """
 
27
 
28
  # What does your leaderboard evaluate?
29
  INTRODUCTION_TEXT = """
30
+ This leaderboard showcases the performance of large language models (LLMs) on various Taiwanese Mandarin language understanding tasks. The models are evaluated on their accuracy across different benchmarks, providing insights into their strengths and weaknesses in comprehending and generating Taiwanese Mandarin text.
31
+ 這個排行榜展示了大型語言模型 (LLMs) 在各種臺灣繁體中文語言理解任務上的表現。
32
  """
33
 
34
  # Which evaluations are you running? how can people reproduce what you have?
35
  LLM_BENCHMARKS_TEXT = f"""
36
+ The leaderboard evaluates LLMs on the following benchmarks:
37
 
38
+ 1. TMLU (Taiwanese Mandarin Language Understanding): Measures the model's ability to understand Taiwanese Mandarin text across various domains.
39
+ 2. TW Truthful QA: Assesses the model's capability to provide truthful answers to questions in Taiwanese Mandarin, with a focus on Taiwan-specific context.
40
+ 3. TW Legal Eval: Evaluates the model's understanding of legal terminology and concepts in Taiwanese Mandarin, using questions from the Taiwanese bar exam for lawyers.
41
+ 4. MMLU (Massive Multitask Language Understanding): Tests the model's performance on a wide range of tasks in English.
42
 
43
+ To reproduce our results, please follow the instructions in the provided GitHub repository: https://github.com/adamlin120/lm-evaluation-harness/blob/main/run_all.sh
44
+
45
+ 該排行榜在以下考題上評估 LLMs:
46
+
47
+ 1. TMLU(臺灣國語語言理解):衡量模型理解各個領域臺灣國語文本的能力。
48
+ 2. TW Truthful QA:評估模型以臺灣國語提供真實答案的能力,重點關注臺灣特定的背景。
49
+ 3. TW Legal Eval:使用臺灣律師資格考試的問題,評估模型對臺灣國語法律術語和概念的理解。
50
+ 4. MMLU(大規模多任務語言理解):測試模型在英語中各種任務上的表現。
51
+
52
+ 要重現我們的結果,請按照:https://github.com/adamlin120/lm-evaluation-harness/blob/main/run_all.sh
53
  """
54
 
55
  EVALUATION_QUEUE_TEXT = """