We recently put this together based on what we’re seeing across a number of AI projects.
A common pattern is that it’s not the model that’s holding things back, it’s the structure and quality of the data underneath. If the foundation isn’t consistent and well organised, AI outputs become hard to trust.
This piece goes into that in a bit more detail, including where approaches like Data Vault can help create a more reliable base for AI in a lakehouse setup.
Interested to hear if this aligns with what others are seeing or if you’ve tackled it differently.