Average Ratings 0 Ratings
Average Ratings 0 Ratings
Description
MonoQwen2-VL-v0.1 represents the inaugural visual document reranker aimed at improving the quality of visual documents retrieved within Retrieval-Augmented Generation (RAG) systems. Conventional RAG methodologies typically involve transforming documents into text through Optical Character Recognition (OCR), a process that can be labor-intensive and often leads to the omission of critical information, particularly for non-text elements such as graphs and tables. To combat these challenges, MonoQwen2-VL-v0.1 utilizes Visual Language Models (VLMs) that can directly interpret images, thus bypassing the need for OCR and maintaining the fidelity of visual information. The reranking process unfolds in two stages: it first employs distinct encoding to create a selection of potential documents, and subsequently applies a cross-encoding model to reorder these options based on their relevance to the given query. By implementing Low-Rank Adaptation (LoRA) atop the Qwen2-VL-2B-Instruct model, MonoQwen2-VL-v0.1 not only achieves impressive results but does so while keeping memory usage to a minimum. This innovative approach signifies a substantial advancement in the handling of visual data within RAG frameworks, paving the way for more effective information retrieval strategies.
Description
Nomic Embed is a comprehensive collection of open-source, high-performance embedding models tailored for a range of uses, such as multilingual text processing, multimodal content integration, and code analysis. Among its offerings, Nomic Embed Text v2 employs a Mixture-of-Experts (MoE) architecture that efficiently supports more than 100 languages with a remarkable 305 million active parameters, ensuring fast inference. Meanwhile, Nomic Embed Text v1.5 introduces flexible embedding dimensions ranging from 64 to 768 via Matryoshka Representation Learning, allowing developers to optimize for both performance and storage requirements. In the realm of multimodal applications, Nomic Embed Vision v1.5 works in conjunction with its text counterparts to create a cohesive latent space for both text and image data, enhancing the capability for seamless multimodal searches. Furthermore, Nomic Embed Code excels in embedding performance across various programming languages, making it an invaluable tool for developers. This versatile suite of models not only streamlines workflows but also empowers developers to tackle a diverse array of challenges in innovative ways.
API Access
Has API
API Access
Has API
Integrations
Baseten
Go
Java
JavaScript
PHP
Python
Ruby
Pricing Details
No price information available.
Free Trial
Free Version
Pricing Details
Free
Free Trial
Free Version
Deployment
Web-Based
On-Premises
iPhone App
iPad App
Android App
Windows
Mac
Linux
Chromebook
Deployment
Web-Based
On-Premises
iPhone App
iPad App
Android App
Windows
Mac
Linux
Chromebook
Customer Support
Business Hours
Live Rep (24/7)
Online Support
Customer Support
Business Hours
Live Rep (24/7)
Online Support
Types of Training
Training Docs
Webinars
Live Training (Online)
In Person
Types of Training
Training Docs
Webinars
Live Training (Online)
In Person
Vendor Details
Company Name
LightOn
Founded
2016
Country
France
Website
www.lighton.ai/lighton-blogs/monoqwen-vision
Vendor Details
Company Name
Nomic
Country
United States
Website
www.nomic.ai/embed