We asked business professionals to review the solutions they use. Here are some excerpts of what they said:
"The solution is very stable."
"I feel the streaming is its best feature."
"The features we find most valuable are the machine learning, data learning, and Spark Analytics."
"The main feature that we find valuable is that it is very fast."
"The processing time is very much improved over the data warehouse solution that we were using."
"The memory processing engine is the solution's most valuable aspect. It processes everything extremely fast, and it's in the cluster itself. It acts as a memory engine and is very effective in processing data correctly."
"AI libraries are the most valuable. They provide extensibility and usability. Spark has a lot of connectors, which is a very important and useful feature for AI. You need to connect a lot of points for AI, and you have to get data from those systems. Connectors are very wide in Spark. With a Spark cluster, you can get fast results, especially for AI."
"Its scalability and speed are very valuable. You can scale it a lot. It is a great technology for big data. It is definitely better than a lot of earlier warehouse or pipeline solutions, such as Informatica. Spark SQL is very compliant with normal SQL that we have been using over the years. This makes it easy to code in Spark. It is just like using normal SQL. You can use the APIs of Spark or you can directly write SQL code and run it. This is something that I feel is useful in Spark."
"Overall the solution is excellent."
"The speed of getting data."
"The performance is one of the most important features. It has an API to process the data in a functional manner."
"It is a stable solution."
"Data validation and ease of use are the most valuable features."
"The solution needs to optimize shuffling between workers."
"When you want to extract data from your HDFS and other sources then it is kind of tricky because you have to connect with those sources."
"We've had problems using a Python process to try to access something in a large volume of data. It crashes if somebody gives me the wrong code because it cannot handle a large volume of data."
"We use big data manager but we cannot use it as conditional data so whenever we're trying to fetch the data, it takes a bit of time."
"I would like to see integration with data science platforms to optimize the processing capability for these tasks."
"The graphical user interface (UI) could be a bit more clear. It's very hard to figure out the execution logs and understand how long it takes to send everything. If an execution is lost, it's not so easy to understand why or where it went. I have to manually drill down on the data processes which takes a lot of time. Maybe there could be like a metrics monitor, or maybe the whole log analysis could be improved to make it easier to understand and navigate."
"Stream processing needs to be developed more in Spark. I have used Flink previously. Flink is better than Spark at stream processing."
"Its UI can be better. Maintaining the history server is a little cumbersome, and it should be improved. I had issues while looking at the historical tags, which sometimes created problems. You have to separately create a history server and run it. Such things can be made easier. Instead of separately installing the history server, it can be made a part of the whole setup so that whenever you set it up, it becomes available."
"The solution needs to include graphing capabilities. Including financial charts would help improve everything overall."
"Anything to improve the GUI would be helpful."
"In the next update, we'd like to see better performance for small points of data. It is possible but there are better tools that are faster and cheaper."
"Being a new user, I am not able to find out how to partition it correctly. I probably need more information or knowledge. In other database solutions, you can easily optimize all partitions. I haven't found a quicker way to do that in Spark SQL. It would be good if you don't need a partition here, and the system automatically partitions in the best way. They can also provide more educational resources for new users."
"There should be better integration with other solutions."
Spark provides programmers with an application programming interface centered on a data structure called the resilient distributed dataset (RDD), a read-only multiset of data items distributed over a cluster of machines, that is maintained in a fault-tolerant way. It was developed in response to limitations in the MapReduce cluster computing paradigm, which forces a particular linear dataflowstructure on distributed programs: MapReduce programs read input data from disk, map a function across the data, reduce the results of the map, and store reduction results on disk. Spark's RDDs function as a working set for distributed programs that offers a (deliberately) restricted form of distributed shared memory
Apache Spark is ranked 1st in Hadoop with 10 reviews while Spark SQL is ranked 3rd in Hadoop with 5 reviews. Apache Spark is rated 8.6, while Spark SQL is rated 7.0. The top reviewer of Apache Spark writes "Good Streaming features enable to enter data and analysis within Spark Stream". On the other hand, the top reviewer of Spark SQL writes "GUI could be improved. Useful for speedily processing big data". Apache Spark is most compared with Spring Boot, Azure Stream Analytics, AWS Batch, AWS Lambda and SAP HANA, whereas Spark SQL is most compared with IBM Db2 Big SQL, Amazon EMR, Informatica Big Data Parser, AtScale Adaptive Analytics (A3) and Netezza Analytics. See our Apache Spark vs. Spark SQL report.
See our list of best Hadoop vendors.
We monitor all Hadoop reviews to prevent fraudulent reviews and keep review quality high. We do not post reviews by company employees or direct competitors. We validate each review for authenticity via cross-reference with LinkedIn, and personal follow-up with the reviewer when necessary.