HyperAI

LLaVA-OneVision Multimodal All-round Vision Model Demo

One-click deployment of LLaVA-OneVision

Tutorial Introduction

LLaVA-OneVision is an open multimodal large model jointly developed in 2024 by researchers from ByteDance, Nanyang Technological University, the Chinese University of Hong Kong and the Hong Kong University of Science and Technology. It can process images, text, image-text interleaved inputs and videos. It is the first single model that can simultaneously break through the performance bottlenecks of open multimodal models in these three important computer vision scenarios.

It not only achieves strong transfer learning capabilities between different modalities and scenes, but also demonstrates its significant advantages in video understanding and cross-scene capabilities through task transfer. LLaVA-OneVision is characterized by its ability to handle a variety of visual tasks, whether it is the analysis of static images or the parsing of dynamic videos, it can provide high-quality output. In addition, the model is designed to focus on the consistency of the maximum number of visual tags, ensuring that the visual representation in different scenes can be balanced, thereby supporting cross-scene capability transfer.

Key Features:

  • Supports various input resolutions up to 2304*2304 pixels.
  • In anyres_max_9 mode, a single image input can be represented by at most 729*(9+1) tokens.
  • Supports multiple image and video input. Multiple image input is represented by 729 tokens per image, and video input is represented by 196 tokens per frame. Note: This tutorial requires a single card A6000 to start

How to run

1. 克隆并启动容器,待容器状态为「运行中」。由于模型较大,加载模型需要等待约 1 分钟,拷贝 API 地址到浏览器中打开即可
2. 可以看到如下界面
3. 点击下方上传单个/多个图片、文件或视频,并输入文本提示
4. 回车,生成回答

Discussion and Exchange

🖌️ 如果大家看到优质项目,欢迎后台留言推荐!另外,我们还建立了教程交流群,欢迎小伙伴们扫码备注【SD 教程】入群探讨各类技术问题、分享应用效果↓