AWS Batch vs Apache Spark comparison

Cancel
You must select at least 2 products to compare!
Apache Logo
3,093 views|2,345 comparisons
89% willing to recommend
Amazon Web Services (AWS) Logo
7,086 views|6,771 comparisons
100% willing to recommend
Comparison Buyer's Guide
Executive Summary

We performed a comparison between Apache Spark and AWS Batch based on real PeerSpot user reviews.

Find out in this report how the two Compute Service solutions compare in terms of features, pricing, service and support, easy of deployment, and ROI.
To learn more, read our detailed AWS Batch vs. Apache Spark Report (Updated: March 2024).
767,667 professionals have used our research since 2012.
Featured Review
Quotes From Members
We asked business professionals to review the solutions they use.
Here are some excerpts of what they said:
Pros
"I found the solution stable. We haven't had any problems with it.""The most valuable feature is the Fault Tolerance and easy binding with other processes like Machine Learning, graph analytics.""Apache Spark provides a very high-quality implementation of distributed data processing.""With Hadoop-related technologies, we can distribute the workload with multiple commodity hardware.""The solution has been very stable.""The data processing framework is good.""The product's deployment phase is easy.""AI libraries are the most valuable. They provide extensibility and usability. Spark has a lot of connectors, which is a very important and useful feature for AI. You need to connect a lot of points for AI, and you have to get data from those systems. Connectors are very wide in Spark. With a Spark cluster, you can get fast results, especially for AI."

More Apache Spark Pros →

"We can easily integrate AWS container images into the product.""AWS Batch manages the execution of computing workload, including job scheduling, provisioning, and scaling.""AWS Batch's deployment was easy.""There is one other feature in confirmation or call confirmation where you can have templates of what you want to do and just modify those to customize it to your needs. And these templates basically make it a lot easier for you to get started."

More AWS Batch Pros →

Cons
"The product could improve the user interface and make it easier for new users.""If you have a Spark session in the background, sometimes it's very hard to kill these sessions because of D allocation.""We've had problems using a Python process to try to access something in a large volume of data. It crashes if somebody gives me the wrong code because it cannot handle a large volume of data.""Dynamic DataFrame options are not yet available.""The solution needs to optimize shuffling between workers.""When you are working with large, complex tasks, the garbage collection process is slow and affects performance.""The setup I worked on was really complex.""The solution must improve its performance."

More Apache Spark Cons →

"When we run a lot of batch jobs, the UI must show the history.""AWS Batch needs to improve its documentation.""The main drawback to using AWS Batch would be the cost. It will be more expensive in some cases than using an HPC. It's more amenable to cases where you have spot requirements.""The solution should include better and seamless integration with other AWS services, like Amazon S3 data storage and EC2 compute resources."

More AWS Batch Cons →

Pricing and Cost Advice
  • "Since we are using the Apache Spark version, not the data bricks version, it is an Apache license version, the support and resolution of the bug are actually late or delayed. The Apache license is free."
  • "Apache Spark is open-source. You have to pay only when you use any bundled product, such as Cloudera."
  • "We are using the free version of the solution."
  • "Apache Spark is not too cheap. You have to pay for hardware and Cloudera licenses. Of course, there is a solution with open source without Cloudera."
  • "Apache Spark is an expensive solution."
  • "Spark is an open-source solution, so there are no licensing costs."
  • "On the cloud model can be expensive as it requires substantial resources for implementation, covering on-premises hardware, memory, and licensing."
  • "It is an open-source solution, it is free of charge."
  • More Apache Spark Pricing and Cost Advice →

  • "AWS Batch's pricing is good."
  • "The pricing is very fair."
  • "AWS Batch is a cheap solution."
  • More AWS Batch Pricing and Cost Advice →

    report
    Use our free recommendation engine to learn which Compute Service solutions are best for your needs.
    767,667 professionals have used our research since 2012.
    Questions from the Community
    Top Answer:We use Spark to process data from different data sources.
    Top Answer:In data analysis, you need to take real-time data from different data sources. You need to process this in a subsecond, and do the transformation in a subsecond
    Top Answer:AWS Lambda is a serverless solution. It doesn’t require any infrastructure, which allows for cost savings. There is no setup process to deal with, as the entire solution is in the cloud. If you use… more »
    Top Answer:AWS Batch manages the execution of computing workload, including job scheduling, provisioning, and scaling.
    Ranking
    5th
    out of 16 in Compute Service
    Views
    3,093
    Comparisons
    2,345
    Reviews
    25
    Average Words per Review
    432
    Rating
    8.7
    3rd
    out of 16 in Compute Service
    Views
    7,086
    Comparisons
    6,771
    Reviews
    4
    Average Words per Review
    973
    Rating
    9.0
    Comparisons
    Also Known As
    Amazon Batch
    Learn More
    Overview

    Spark provides programmers with an application programming interface centered on a data structure called the resilient distributed dataset (RDD), a read-only multiset of data items distributed over a cluster of machines, that is maintained in a fault-tolerant way. It was developed in response to limitations in the MapReduce cluster computing paradigm, which forces a particular linear dataflowstructure on distributed programs: MapReduce programs read input data from disk, map a function across the data, reduce the results of the map, and store reduction results on disk. Spark's RDDs function as a working set for distributed programs that offers a (deliberately) restricted form of distributed shared memory

    AWS Batch enables developers, scientists, and engineers to easily and efficiently run hundreds of thousands of batch computing jobs on AWS. AWS Batch dynamically provisions the optimal quantity and type of compute resources (e.g., CPU or memory optimized instances) based on the volume and specific resource requirements of the batch jobs submitted. With AWS Batch, there is no need to install and manage batch computing software or server clusters that you use to run your jobs, allowing you to focus on analyzing results and solving problems. AWS Batch plans, schedules, and executes your batch computing workloads across the full range of AWS compute services and features, such as Amazon EC2 and Spot Instances.

    Sample Customers
    NASA JPL, UC Berkeley AMPLab, Amazon, eBay, Yahoo!, UC Santa Cruz, TripAdvisor, Taboola, Agile Lab, Art.com, Baidu, Alibaba Taobao, EURECOM, Hitachi Solutions
    Hess, Expedia, Kelloggs, Philips, HyperTrack
    Top Industries
    REVIEWERS
    Computer Software Company30%
    Financial Services Firm15%
    University9%
    Marketing Services Firm6%
    VISITORS READING REVIEWS
    Financial Services Firm25%
    Computer Software Company13%
    Manufacturing Company7%
    Comms Service Provider6%
    VISITORS READING REVIEWS
    Financial Services Firm25%
    Computer Software Company13%
    Manufacturing Company6%
    Insurance Company5%
    Company Size
    REVIEWERS
    Small Business40%
    Midsize Enterprise19%
    Large Enterprise40%
    VISITORS READING REVIEWS
    Small Business17%
    Midsize Enterprise12%
    Large Enterprise71%
    VISITORS READING REVIEWS
    Small Business16%
    Midsize Enterprise12%
    Large Enterprise73%
    Buyer's Guide
    AWS Batch vs. Apache Spark
    March 2024
    Find out what your peers are saying about AWS Batch vs. Apache Spark and other solutions. Updated: March 2024.
    767,667 professionals have used our research since 2012.

    Apache Spark is ranked 5th in Compute Service with 60 reviews while AWS Batch is ranked 3rd in Compute Service with 4 reviews. Apache Spark is rated 8.4, while AWS Batch is rated 9.0. The top reviewer of Apache Spark writes "Reliable, able to expand, and handle large amounts of data well". On the other hand, the top reviewer of AWS Batch writes "User-friendly, good customization and offers exceptional scalability, allowing users to run jobs ranging from 32 cores to over 2,000 cores". Apache Spark is most compared with Spring Boot, Spark SQL, SAP HANA, Cloudera Distribution for Hadoop and AWS Lambda, whereas AWS Batch is most compared with AWS Lambda, AWS Fargate, Oracle Compute Cloud Service, Amazon EC2 Auto Scaling and Amazon EC2. See our AWS Batch vs. Apache Spark report.

    See our list of best Compute Service vendors.

    We monitor all Compute Service reviews to prevent fraudulent reviews and keep review quality high. We do not post reviews by company employees or direct competitors. We validate each review for authenticity via cross-reference with LinkedIn, and personal follow-up with the reviewer when necessary.