Vertex AI
Fully managed ML tools allow you to build, deploy and scale machine-learning (ML) models quickly, for any use case.
Vertex AI Workbench is natively integrated with BigQuery Dataproc and Spark. You can use BigQuery to create and execute machine-learning models in BigQuery by using standard SQL queries and spreadsheets or you can export datasets directly from BigQuery into Vertex AI Workbench to run your models there. Vertex Data Labeling can be used to create highly accurate labels for data collection.
Vertex AI Agent Builder empowers developers to design and deploy advanced generative AI applications for enterprise use. It supports both no-code and code-driven development, enabling users to create AI agents through natural language prompts or by integrating with frameworks like LangChain and LlamaIndex.
Learn more
RaimaDB
RaimaDB, an embedded time series database that can be used for Edge and IoT devices, can run in-memory. It is a lightweight, secure, and extremely powerful RDBMS. It has been field tested by more than 20 000 developers around the world and has been deployed in excess of 25 000 000 times.
RaimaDB is a high-performance, cross-platform embedded database optimized for mission-critical applications in industries such as IoT and edge computing. Its lightweight design makes it ideal for resource-constrained environments, supporting both in-memory and persistent storage options. RaimaDB offers flexible data modeling, including traditional relational models and direct relationships through network model sets. With ACID-compliant transactions and advanced indexing methods like B+Tree, Hash Table, R-Tree, and AVL-Tree, it ensures data reliability and efficiency. Built for real-time processing, it incorporates multi-version concurrency control (MVCC) and snapshot isolation, making it a robust solution for applications demanding speed and reliability.
Learn more
SmolVLM
SmolVLM-Instruct is a streamlined, AI-driven multimodal model that integrates vision and language processing capabilities, enabling it to perform functions such as image captioning, visual question answering, and multimodal storytelling. This model can process both text and image inputs efficiently, making it particularly suitable for smaller or resource-limited environments. Utilizing SmolLM2 as its text decoder alongside SigLIP as its image encoder, it enhances performance for tasks that necessitate the fusion of textual and visual data. Additionally, SmolVLM-Instruct can be fine-tuned for various specific applications, providing businesses and developers with a flexible tool that supports the creation of intelligent, interactive systems that leverage multimodal inputs. As a result, it opens up new possibilities for innovative application development across different industries.
Learn more
Qwen2.5-VL
Qwen2.5-VL marks the latest iteration in the Qwen vision-language model series, showcasing notable improvements compared to its predecessor, Qwen2-VL. This advanced model demonstrates exceptional capabilities in visual comprehension, adept at identifying a diverse range of objects such as text, charts, and various graphical elements within images. Functioning as an interactive visual agent, it can reason and effectively manipulate tools, making it suitable for applications involving both computer and mobile device interactions. Furthermore, Qwen2.5-VL is proficient in analyzing videos that are longer than one hour, enabling it to identify pertinent segments within those videos. The model also excels at accurately locating objects in images by creating bounding boxes or point annotations and supplies well-structured JSON outputs for coordinates and attributes. It provides structured data outputs for documents like scanned invoices, forms, and tables, which is particularly advantageous for industries such as finance and commerce. Offered in both base and instruct configurations across 3B, 7B, and 72B models, Qwen2.5-VL can be found on platforms like Hugging Face and ModelScope, further enhancing its accessibility for developers and researchers alike. This model not only elevates the capabilities of vision-language processing but also sets a new standard for future developments in the field.
Learn more