Search
Awesome LLM Human Preference Datasets
a collection of human preference datasets for LLM instruction tuning, RLHF and evaluation.
RWKV-howto
possibly useful materials and tutorial for learning RWKV.
ModelEditingPapers
A paper & resource list on model editing for large language models.
Awesome-Align-LLM-Human
A collection of papers and resources about aligning large language models (LLMs) with human.
Awesome-Code-LLM
An awesome and curated list of best code-LLM for research.
Awesome-LLM-Compression
Awesome LLM compression research papers and tools.
Awesome-LLM-Systems
Awesome LLM systems research papers.
awesome-llm-webapps
A collection of open source, actively maintained web apps for LLM applications.
awesome-japanese-llm
日本語LLMまとめ - Overview of Japanese LLMs.
Awesome-LLM-Healthcare
The paper list of the review on LLMs in medicine.
Awesome-LLM-Inference
A curated list of Awesome LLM Inference Paper with codes.
Awesome-LLM-3D
A curated list of Multi-modal Large Language Model in 3D world, including 3D understanding, reasoning, generation, and embodied agents.
...moreLLMDatahub
a curated collection of datasets specifically designed for chatbot training, including links, size, language, usage, and a brief description of each dataset
...moreAwesome-Chinese-LLM
整理开源的中文大语言模型,以规模较小、可私有化部署、训练成本较低的模型为主,包括底座模型,垂直领域微调及应用,数据集与教程等。
LLM4Opt
Applying Large language models (LLMs) for diverse optimization tasks (Opt) is an emerging research area. This is a collection of references and papers of LLM4Opt.
...moreawesome-language-model-analysis
This paper list focuses on the theoretical or empirical analysis of language models, e.g., the learning dynamics, expressive capacity, interpretability, generalization, and other interesting topics.
...moreChatbot Arena Leaderboard
a benchmark platform for large language models (LLMs) that features anonymous, randomized battles in a crowdsourced manner.
...moreOpen LLM Leaderboard
aims to track, rank, and evaluate LLMs and chatbots as they are released.
AlpacaEval
An Automatic Evaluator for Instruction-following Language Models using Nous benchmark suite.
ACLUE
an evaluation benchmark focused on ancient Chinese language comprehension.