Cloud-Fundis’ Post

View organization page for Cloud-Fundis, graphic

1,226 followers

We've talked about BrightSpark - our compute engine - for a few weeks. We're very proud of what we've built and of the plans we have for it going forward. You may be wondering what it has over simply running your data platform on EKS (as espoused here https://lnkd.in/ghPg4GpU). What is inside the box? * Your AWS Glue scripts can, with no, or only a small amount of modification, run on BrightSpark. * With BrightSpark, your engineers don't have to learn, manage or configure ECS or EKS. We handle all that. * It was designed with security at the top of the requirements. All your data remains in your account. No critical information leaves your organisation. * BrightSpark is a simple API call to create, submit, update or delete jobs * Switching the size of the compute from small -> extra-large is as simple as changing the t-shirt size on job run. * Monitoring and reporting are part of the service. How much a job costs? How long it ran? What resources did this job use? BrightSpark provides all that inside the box. * Tagging is a first-class citizen. Grouping of jobs in order to get reporting aggregates (How much did all these jobs cost over the last 7 days? Over the past 3 months, how has the runtime changed?). This makes organisation cross-billing easy. * BrightSpark is integrated with Azure EntraID (formerly Azure Active Directory) in both PySpark jobs as well as your Jupyter notebooks. BrightSpark (https://lnkd.in/dEBAz9Zm) has other applications too. Not only can we run #PySpark and Python jobs, but we can also run #ApacheRay jobs, shell scripts (think combining small files while simultaneously moving S3 data to Glacier), or run #Jupyter notebooks. #AWSPartner #Spark #BigData Talk to us about how BrightSpark can change your enterprise and reduce your #AWS bill.

To view or add a comment, sign in

Explore topics