Viktor
Viktor is a fully autonomous AI coworker designed to operate directly inside your Slack workspace and execute real work across your organization. Rather than functioning as a simple chatbot, Viktor runs on its own cloud-based computer where it writes code, deploys applications, and performs complex multi-step tasks. It connects to more than 3,000 integrations through native APIs and browser automation, enabling it to manage advertising campaigns, analyze product metrics, update documents, and create tickets across tools like Linear and PostHog. Viktor proactively monitors systems and identifies anomalies, proposing concrete actions instead of merely sending alerts. It can run continuously for weeks while retaining context about team goals, project timelines, and previous decisions. Within Slack threads, team members can request data summaries, backend updates, marketing optimizations, or workflow automation and receive structured, actionable responses. Before executing changes, Viktor presents pending actions for approval, maintaining transparency and control. The platform supports scheduled tasks such as automated reports, audits, and recurring check-ins. Its persistent workspace context ensures continuity even as projects evolve over time. Available in Starter, Team, and Enterprise tiers, Viktor adapts to both small teams and large organizations. Built by experienced engineers and backed by leading investors, it positions itself as a productivity multiplier rather than a simple assistant. By embedding autonomous execution directly into Slack, Viktor transforms everyday collaboration into a coordinated, AI-powered operating system for modern teams.
Learn more
Vertex AI
Completely managed machine learning tools facilitate the rapid construction, deployment, and scaling of ML models tailored for various applications.
Vertex AI Workbench seamlessly integrates with BigQuery Dataproc and Spark, enabling users to create and execute ML models directly within BigQuery using standard SQL queries or spreadsheets; alternatively, datasets can be exported from BigQuery to Vertex AI Workbench for model execution. Additionally, Vertex Data Labeling offers a solution for generating precise labels that enhance data collection accuracy.
Furthermore, the Vertex AI Agent Builder allows developers to craft and launch sophisticated generative AI applications suitable for enterprise needs, supporting both no-code and code-based development. This versatility enables users to build AI agents by using natural language prompts or by connecting to frameworks like LangChain and LlamaIndex, thereby broadening the scope of AI application development.
Learn more
Qwen3.5
Qwen3.5 is an advanced open-weight multimodal AI system built to serve as the foundation for native digital agents capable of reasoning across text, images, and video. The primary release, Qwen3.5-397B-A17B, introduces a hybrid architecture that combines Gated DeltaNet linear attention with a sparse mixture-of-experts design, activating just 17 billion parameters per inference pass while maintaining a total parameter count of 397 billion. This selective activation dramatically improves decoding throughput and cost efficiency without sacrificing benchmark-level performance. Qwen3.5 demonstrates strong results across knowledge, multilingual reasoning, coding, STEM tasks, search agents, visual question answering, document understanding, and spatial intelligence benchmarks. The hosted Qwen3.5-Plus variant offers a default one-million-token context window and integrated tool usage such as web search and code interpretation for adaptive problem-solving. Expanded multilingual support now covers 201 languages and dialects, backed by a 250k vocabulary that enhances encoding and decoding efficiency across global use cases. The model is natively multimodal, using early fusion techniques and large-scale visual-text pretraining to outperform prior Qwen-VL systems in scientific reasoning and video analysis. Infrastructure innovations such as heterogeneous parallel training, FP8 precision pipelines, and disaggregated reinforcement learning frameworks enable near-text baseline throughput even with mixed multimodal inputs. Extensive reinforcement learning across diverse and generalized environments improves long-horizon planning, multi-turn interactions, and tool-augmented workflows. Designed for developers, researchers, and enterprises, Qwen3.5 supports scalable deployment through Alibaba Cloud Model Studio while paving the way toward persistent, economically aware, autonomous AI agents.
Learn more
GLM-5
GLM-5 is Z.ai’s most advanced open-source model to date, purpose-built for complex systems engineering, long-horizon planning, and autonomous agent workflows. Building on the foundation of GLM-4.5, it dramatically scales both total parameters and pre-training data while increasing active parameter efficiency. The integration of DeepSeek Sparse Attention allows GLM-5 to maintain strong long-context reasoning capabilities while reducing deployment costs. To improve post-training performance, Z.ai developed slime, an asynchronous reinforcement learning infrastructure that significantly boosts training throughput and iteration speed. As a result, GLM-5 achieves top-tier performance among open-source models across reasoning, coding, and general agent benchmarks. It demonstrates exceptional strength in long-term operational simulations, including leading results on Vending Bench 2, where it manages a year-long simulated business with strong financial outcomes. In coding evaluations such as SWE-bench and Terminal-Bench 2.0, GLM-5 delivers competitive results that narrow the gap with proprietary frontier systems. The model is fully open-sourced under the MIT License and available through Hugging Face, ModelScope, and Z.ai’s developer platforms. Developers can deploy GLM-5 locally using inference frameworks like vLLM and SGLang, including support for non-NVIDIA hardware through optimization and quantization techniques. Through Z.ai, users can access both Chat Mode for fast interactions and Agent Mode for tool-augmented, multi-step task execution. GLM-5 also enables structured document generation, producing ready-to-use .docx, .pdf, and .xlsx files for business and academic workflows. With compatibility across coding agents and cross-application automation frameworks, GLM-5 moves foundation models from conversational assistants toward full-scale work engines.
Learn more