What's it like to implement Ranger and Atlas on a data lake? As a data custodian I have implemented Ranger to protect sensitive data and to make sure our raw data is not altered. I've created all kinds of policies to do that. But I also gained a lot of experience using the Ranger REST API. In this session I will tell how to work with that.
Atlas is a nice product to store metadata. I call it the search engine for metadata in your data lake. But what kind of metadata do you store in it? What is useful? And who needs to access it? Atlas also comes with a REST API, which was a little harder to figure out, but in the end I managed to do a lot with it. In this session I'll explain how to use it as well.
I've used Ranger and Atlas for a year now, so expect a lot of insights for beginners as well as intermediate users.