Databricks has announced that Anthropic’s newest AI model, Claude Opus 4.6, is now available on the Data Intelligence Platform. This release significantly expands the company’s multi-model strategy, giving enterprise clients access to one of the most advanced systems for autonomous coding, specialized agents, and complex analytical workflows.
The integration, announced on February 5, allows Databricks users to deploy Claude Opus 4.6 directly on their managed enterprise data across AWS, Azure, and Google Cloud Platform. This milestone comes less than a year after Databricks and Anthropic signed a five-year, approximately $100 million strategic partnership aimed at bringing Claude models natively into the Databricks ecosystem.
Expanding the Multi-Model Ecosystem in Databricks
The addition of Claude Opus 4.6 continues the strategic roadmap established in March 2025, which first brought Claude 3.7 Sonnet to the platform. Since then, Databricks has rapidly integrated subsequent iterations, including Opus 4 and Sonnet 4, to offer customers choice and flexibility.
For many organizations, the challenge has been applying powerful reasoning models to proprietary data without exposing it to public API risks. This partnership addresses that by combining Anthropic’s “Constitutional AI” approach with Databricks’ Unity Catalog. This ensures that while the model is cutting-edge, the data remains governed, auditable, and secure.
Claude Opus 4.6 is now available on Databricks!@AnthropicAI's most advanced model yet is built to power agents that take on real-world work—from autonomous coding and cybersecurity to financial analysis and enterprise workflows. And now you can apply it directly to your… pic.twitter.com/H5VbJeHZYq
— Databricks (@databricks) February 5, 2026
Next-Generation Capabilities – Adaptive Thinking and Coding
Claude Opus 4.6, released by Anthropic on February 5, introduces specific features designed for enterprise-scale deployments. The model demonstrates significant improvements in software development tasks, including planning, code review, debugging, and maintaining stability within large codebases.
According to Anthropic, the model achieves a 65.4% score on Terminal-Bench 2.0, a benchmark for command-line programming tasks, and currently leads the Finance Agent benchmark for financial analysis operations.
Key technical features now available to Databricks customers include:
- Adaptive Thinking: The model automatically determines how much reasoning effort to apply to a specific task, optimizing performance and cost.
- Context Compression: Enables longer agent sessions without losing track of earlier instructions.
- Massive Context Window: Support for a 1 million token context window (in beta), with the ability to generate up to 128k output tokens.
Secure Deployment with Unity Catalog
Through this integration, Databricks customers can move beyond simple chat interfaces to build complex, automated workflows.
- Agent Bricks: Developers can build domain-specific agents that utilize Opus 4.6 for specialized tasks.
- AI Functions: Users can run complex analysis directly from SQL and Python, allowing the AI to query data and return structured insights.
- Workflow Automation: The model can orchestrate tasks across documents, spreadsheets, and internal databases.
Because the endpoint is hosted within the Databricks security perimeter, enterprises can fine-tune these models using their own data or use Retrieval Augmented Generation (RAG) while maintaining strict compliance controls.
Our Take – Moving from “Chat” to “Work”
At Multishoring, we view the integration of Claude Opus 4.6 not just as a model upgrade, but as a shift in how enterprises consume AI.
For the last two years, the conversation has been dominated by “Generative AI”—using models to write text or summarize documents. This release marks the maturity of “Agentic AI.” The specific capabilities of Opus 4.6—adaptive thinking and high-fidelity coding—mean we are no longer just asking the AI questions; we are assigning it jobs.
However, giving an AI agent permission to execute code or analyze financial data brings massive risk. This is why the Databricks Unity Catalog integration is the real story here.
- Data Gravity: By bringing the model to the data (instead of sending data to an API), we solve the latency and security issues that have stalled so many pilot projects.
- Governance as an Enabler:You cannot deploy autonomous agents without strict, row-level permissions. This integration allows CTOs to finally approve high-value use cases—like automated financial reconciliation or root-cause analysis on logs—because the governance layer remains intact.
We believe this will accelerate the move away from isolated “AI sandboxes” toward fully integrated Lakehouse workflows where AI is just another user with specific, managed permissions.

