Spaces:
Runtime error
Runtime error
Update src/assets/text_content.py
Browse files
src/assets/text_content.py
CHANGED
|
@@ -6,6 +6,10 @@ INTRODUCTION_TEXT = f"""
|
|
| 6 |
🐨 KG LLM Leaderboard aims to track, rank, and evaluate the performance of released Large Language Models on traditional KBQA/KGQA datasets.
|
| 7 |
|
| 8 |
The data on this page is sourced from a research paper. If you intend to use the data from this page, please remember to cite the following source: https://arxiv.org/abs/2303.07992
|
|
|
|
|
|
|
|
|
|
|
|
|
| 9 |
"""
|
| 10 |
|
| 11 |
LLM_BENCHMARKS_TEXT = f"""
|
|
|
|
| 6 |
🐨 KG LLM Leaderboard aims to track, rank, and evaluate the performance of released Large Language Models on traditional KBQA/KGQA datasets.
|
| 7 |
|
| 8 |
The data on this page is sourced from a research paper. If you intend to use the data from this page, please remember to cite the following source: https://arxiv.org/abs/2303.07992
|
| 9 |
+
|
| 10 |
+
We compare the current SOTA traditional KBQA models (fine-tuned (FT) and zero-shot (ZS)),
|
| 11 |
+
LLMs in the GPT family, and Other Non-GPT LLM. In QALD-9 and LC-quad2, the evaluation metric used is F1, while other datasets use Accuracy (Exact match).
|
| 12 |
+
|
| 13 |
"""
|
| 14 |
|
| 15 |
LLM_BENCHMARKS_TEXT = f"""
|