VideoPrism Collection VideoPrism is a foundational video encoder that enables state-of-the-art performance on a large variety of video understanding tasks. • 5 items • Updated Jul 16 • 12
LMLM-Models Collection LMLM-llama2 is a collection of LLaMA2-based models trained under the Large Memory Language Model (LMLM) framework • 5 items • Updated Jul 1 • 1
VideoPrism: A Foundational Visual Encoder for Video Understanding Paper • 2402.13217 • Published Feb 20, 2024 • 38
Pre-trained Large Language Models Learn Hidden Markov Models In-context Paper • 2506.07298 • Published Jun 8 • 26