Building DataLakes can become complicated when dealing with lots of data. Even more so, when working with heavily skewed data. In this post, we'll be looking at how we can improve your long running Spark jobs caused by skewed data.

Using AWS SSO (IAM Identity Center) to set up CLI and Athena JDBC access for the day

A brief article covering a method to dynamically calculate the ideal number of Spark partitions at runtime