Systems | Development | Analytics | API | Testing

Latest Videos

Managing Costs for Spark on Amazon EMR

Are you looking to optimize costs and resource usage for your Spark jobs on Amazon EMR? Then this is the webinar for you. Overallocating resources, such as memory, is a common fault when setting up Spark jobs. And for Spark jobs running on EMR, adding resources is a click away - but it’s an expensive click, so cost management is critical. Unravel Data is our AI-enabled observability platform for Spark jobs on Amazon EMR and other Big Data technologies. Unravel helps you right-size memory allocations, choose the right number of workers, and map your cluster needs to available instance types.

Managing Costs for Spark on Databricks Webinar

Are you looking to optimize costs and resource usage for your Spark jobs on Databricks? Then this is the webinar for you. Overallocating resources, such as memory, is a common fault when setting up Spark jobs. And for Spark jobs running on Databricks, adding resources is a click away - but it’s an expensive click, so cost management is critical.

Managing Cost & Resources Usage for Spark

Spark jobs require resources - and those resources? They can be pricey. If you're looking to speed up completion times, optimize costs, and reduce resource usage for your Spark jobs, this is the webinar for you.For Spark jobs running on-premises, optimizing resource usage is key. For Spark jobs running in the cloud, for example on Amazon EMR or Databricks, adding resources is a click away - but it’s an expensive click, so cost management is critical.

Troubleshooting Databricks

The popularity of Databricks is rocketing skyward, and it is now the leading multi-cloud platform for Spark and analytics workloads, offering fully managed Spark clusters in the cloud. Databricks is fast and organizations generally refactor their applications when moving them to Databricks. The result is strong performance. However, as usage of Databricks grows, so does the importance of reliability for Databricks jobs - especially big data jobs such as Spark workloads. But information you need for troubleshooting is scattered across multiple, voluminous log files.