"Vimo: Chat with Your Videos"
-
Updated
Aug 1, 2025 - Python
"Vimo: Chat with Your Videos"
This repository contains the training, inference, evaluation code for SpeechLLM models and details about the model releases on huggingface.
This repo contains a list of channels and sources from where LLMs should be learned
ExGra-Med: Medical Multi-Modal LLM with Extended Context Alignment
[NeurIPS 2024] Code, Dataset, Samples for the VATT paper “ Tell What You Hear From What You See - Video to Audio Generation Through Text”
FastLongSpeech is a novel framework designed to extend the capabilities of Large Speech-Language Models for efficient long-speech processing without necessitating dedicated long-speech training data.
Add a description, image, and links to the multi-modal-llms topic page so that developers can more easily learn about it.
To associate your repository with the multi-modal-llms topic, visit your repo's landing page and select "manage topics."