Lentiq’s new EdgeLake architecture brings flexibility, analytical muscle, and opportunities for collaboration across organizations that have been limited by overly centralized data repositories.

Alex Woodie reports on EdgeLake in this story from Datanami:

Lentiq has a new take on data lakes. Instead of building a centralized data lake, the vendor’s EdgeLake offering provides users with a distributed architecture that’s composed of multiple interconnected “micro data lakes,” or data pools.

These data pools physically live across multiple clouds – AWS, Azure, and GCP – and store data in a publish/subscribe manner. EdgeLake is pre-loaded with “everything a data scientist or data engineer needs,” the company claims, including data management capabilities and notebook environments, as well as Apache Spark, Apache Kafka, and Streamsets software.

The data pools exist independently across different clouds, and governance rules are only enforced when the data moves. “This design allows data teams to leverage the best tools and skillsets available for the job, mitigate local infrastructure requirements and apply governance policies without hindering innovation and adaptability,” the vendor says.

The new service caught the eye of Kirk Borne, the principal data scientist and executive advisor for Booz Allen Hamilton, who previously taught astrophysics and computational science at George Mason University and has spoken at Tabor Communications‘ Leverage Big Data events, among others.

“Lentiq’s new Data Pools and EdgeLake architecture have changed the data analytics landscape in one comprehensive strike,” Borne wrote. “The data lake has come of age in Lentiq’s new Data Pool and EdgeLake by converging the best of data architecture, data curation, data sharing, data science modeling, analytics delivery, and production-level capabilities into a single user-driven, use case-configurable, flexible, agile platform for innovation.”