Yuxuan Wang's Projects
Recent advancements propelled by large language models (LLMs), encompassing an array of domains including Vision, Audio, Agent, Robotics, Fundamental Sciences such as Mathematics, and Ominous.
Awesome-LLM: everything you need to know about Large Language Model
🔥🔥🔥Latest Papers, Codes and Datasets on Vid-LLMs.
CVPR and NeurIPS poster examples and templates. May we have in-person poster session soon!
[ACL2023] Shuo Wen Jie Zi is a new learning paradigm that enhances the semantics understanding ability of the Chinese PLMs with dictionary knowledge and structure of Chinese characters
A reading list of hallucination in Generative Models
贵校课程资料民间整理
Open Academic Research on Improving LLaMA to SOTA LLM
Language Modeling Research Hub, a comprehensive compendium for enthusiasts and scholars delving into the fascinating realm of language models (LMs), with a particular focus on large language models (LLMs)
Accelerating the development of large multimodal models (LMMs) with lmms-eval
[EMNLP2022] We propose a new collaborative reasoning method on mutli-modal graphs for multimodal dialogue
Multi-modal Dialogue Scene & Session Discrimination
multimodal-BART baseline for AVSD
Pressure Testing Large Video-Language Models (LVLM): Doing multimodal retrieval from LVLM at any video lengths to measure accuracy
Natural Language Processing Tutorial for Deep Learning Researchers
Multimodal Dialogue Understanding and Generation
Official repository for the paper PLLaVA
清华大学计算机系课程攻略 Guidance for courses in Department of Computer Science and Technology, Tsinghua University
VideoHallucer, The first comprehensive benchmark for hallucination detection in large video-language models (LVLMs)
[ACL2023] VSTAR is a multimodal dialogue dataset with scene and topic transition information