- Cloud Database Insider
- Posts
- Databricks sued over patent dispute⚖️|Amazon S3 Files Now Generally Available📂|Master Agentic Snowpark with Top Best Practices❄️
Databricks sued over patent dispute⚖️|Amazon S3 Files Now Generally Available📂|Master Agentic Snowpark with Top Best Practices❄️
Deep Dive: Fundamental Database Concepts and Architectures

What’s in today’s newsletter:
Databricks sued for AI patent infringement dispute ⚖️
AWS releases S3 File for seamless file operations 📂
Snowpark enables autonomous data workflows with best practices ❄️
Also, check out the weekly Deep Dive - Fundamentals and Architectures of Databases
The IT strategy every team needs for 2026
2026 will redefine IT as a strategic driver of global growth. Automation, AI-driven support, unified platforms, and zero-trust security are becoming standard, especially for distributed teams. This toolkit helps IT and HR leaders assess readiness, define goals, and build a scalable, audit-ready IT strategy for the year ahead. Learn what’s changing and how to prepare.
DATABRICKS

TL;DR: Databricks faces a lawsuit alleging unauthorized use of patented AI model technologies, spotlighting growing intellectual property challenges that may set key legal precedents in AI innovation and patent enforcement.
Databricks is being sued for allegedly infringing patents related to AI model technology and data processing tools.
The lawsuit alleges Databricks used patented methodologies integral to AI model development without proper permission.
This legal challenge could clarify patent boundaries and influence enforcement in the rapidly evolving AI technology sector.
The case highlights the importance of navigating intellectual property rights to prevent costly disputes in AI innovation.
Why this matters: This lawsuit against Databricks underscores the escalating tensions around intellectual property in AI, a field where legal frameworks lag behind innovation. Its outcome could define patent boundaries, shaping the competitive landscape and compelling companies to adopt stringent IP strategies to avoid disruption and costly litigation.
AWS

TL;DR: AWS's new Amazon S3 File offers POSIX-style file operations on S3, enabling legacy apps seamless cloud storage access with enterprise-grade reliability, easing cloud migration and hybrid storage integration.
AWS launched the Amazon S3 File, enabling file operations like open, read, write, and close on S3 storage.
The API provides POSIX-style file access, allowing seamless use of S3 buckets as traditional file systems for apps.
It supports concurrent access, locking, and atomic writes, ensuring enterprise-grade file operation reliability on S3.
The S3 File API simplifies cloud migration for legacy workloads, reducing cost and risk while enhancing hybrid storage setups.
Why this matters: AWS’s new S3 File bridges legacy file access with cloud object storage, enabling enterprises to migrate workloads seamlessly without rewriting applications. This reduces costs and risks, supports hybrid architectures, and reinforces AWS’s leadership by combining file system familiarity with scalable, durable cloud storage benefits.
SNOWFLAKE

TL;DR: Snowpark enables autonomous agents in Snowflake for automating data workflows. Best practices emphasize modular design, clear behaviors, state management, error handling, and performance optimization for scalable, maintainable applications.
Agentic programming with Snowpark enables autonomous software agents to automate complex data workflows on Snowflake’s platform.
Best practices include modular code design, clear agent behavior definitions, effective state management, and robust error handling.
Optimizing performance involves leveraging lazy evaluation, caching, and balancing agent autonomy with control to avoid runaway processes.
These practices enhance automation scalability, improve decision speed, reduce manual work, and support maintainable, testable data applications.
Why this matters: Adopting agentic programming with Snowpark empowers organizations to automate and scale complex data workflows efficiently. Following best practices ensures robust, maintainable, and performant autonomous agents, accelerating decision-making while reducing risks and manual effort, paving the way for advanced cloud-native data applications and long-term success on Snowflake.

EVERYTHING ELSE IN CLOUD DATABASES
Databricks Co-Founder Claims AGI Is Here Now
LakeFlow slashes data ingestion costs by 98%
Dremio Boosts Apache Iceberg V3 Support in Cloud
DuckDB Public Beta Launches: One Binary Rules Data
KeeperDB Adds Zero Trust for Safer Database Access
Top 10 Data Visualization Tools for CIOs 2026
Supabase aims $10B value in funding talks
Iceberg V3 revolutionizes data management in 2026
Aurora PostgreSQL now supports multiple new versions
Snowflake boosts data governance with Iceberg v3 support
Neara boosts observability with ClickHouse ClickStack
Weaviate Agent Skills Boost AI Database Power
Snowflake Faces Class Action Over Investor Losses
Graph-Powered AI: Boost Insights with Neo4j
Dataveil v5 adds PostgreSQL masking support

DEEP DIVE
Fundamental Database Concepts and Architectures
I had a meeting about a week ago with a very talented and skilled Associate Director at work. We were discussing some migration plans. I brought up DTS packages and I had to explain that they were the progenitor of SSIS.
I have from that point, realized that I have been working with this data stuff for a very long time and also, I based my career on these fundamental database concepts. The following seem so far removed from the myriad database types and the ancillary platforms and software that surround them now in 2026.
I just want to share some of the concepts that serve as the grounding for my career and perhaps yours (You will notice a skew towards relational databases):
Declarative Referential Integrity
In the modern stack, DRI is the bedrock of automated data consistency, ensuring that relationships between tables remain intact through system-enforced primary and foreign key constraints. While some distributed systems trade this off for performance, maintaining these rules at the schema level prevents orphaned records and preserves the structural integrity required for complex data orchestration.
The Sybase/Microsoft split
The Sybase/Microsoft split in the early ‘90s is one of the most underrated fork-in-the-road moments in database history. What started as a joint effort ultimately led to two divergent paths: Sybase doubling down on enterprise data systems, while Microsoft evolved SQL Server into a dominant force in the Windows ecosystem. Fast forward to today, and you can trace a direct line from that split to Microsoft’s modern data stack—Azure SQL, Fabric, and beyond.
Third Normal Form
Third Normal Form is the backbone of relational data modeling, designed to eliminate redundancy and ensure data dependencies make logical sense. It forces you to think rigorously about how attributes relate to keys—removing transitive dependencies and tightening the structure of your data model. While modern analytics platforms often favor denormalization, 3NF remains critical for transactional systems where consistency and integrity are non-negotiable.
Star Schema and Snowflake Schema
These dimensional modeling patterns are the engines behind modern analytical processing, with the Star Schema prioritizing query simplicity and the Snowflake Schema focusing on normalized efficiency. Choosing between them is a strategic decision in lakehouse architecture, balancing the need for rapid join performance against the benefits of reduced data storage and structural clarity.
Transaction Isolation Levels
Isolation levels define the critical balance between data consistency and system concurrency, determining how changes made by one operation are visible to others. From "Read Uncommitted" to "Serializable," understanding these settings is vital for architects who must prevent anomalies like "dirty reads" or "phantoms" in high-throughput database environments.
E.F. Codd
E.F. Codd is the intellectual architect behind the relational model, and by extension, the entire modern database ecosystem. His work at IBM in the 1970s introduced the idea that data should be organized mathematically—through relations, tuples, and sets—rather than hierarchical or procedural structures. Decades later, every SQL query, every data warehouse, and every normalization rule traces back to his original vision.
Set Theory
At its core, every relational database is built on set theory—even if most practitioners never think about it that way. SQL operations like joins, unions, and intersections are direct implementations of set-based logic, enabling declarative data manipulation at scale. In an era of distributed query engines and vectorized execution, the principles of set theory are more relevant than ever—they’re just abstracted behind increasingly powerful engines.
These concepts might read like some dusty old tome found in some national archive but I think that they have shaped many of the incredible developments in the database world.
Gladstone Benjamin
🚀 Work With Cloud Database Insider
Looking to reach enterprise data engineers and architects?
Limited sponsorship slots available each month.

