We performed a comparison between Pentaho Data Integration and Analytics and Quest SharePlex based on real PeerSpot user reviews.
Find out in this report how the two Data Integration solutions compare in terms of features, pricing, service and support, easy of deployment, and ROI."It's my understanding that the product can scale."
"The fact that it's a low-code solution is valuable. It's good for more junior people who may not be as experienced with programming."
"Provides a good open source option."
"The way it has improved our product is by giving our users the ability to do ad hoc reports, which is very important to our users. We can do predictive analysis on trends coming in for contracts, which is what our product does. The product helps users decide which way to go based on the predictive analysis done by Pentaho. Pentaho is not doing predictions, but reporting on the predictions that our product is doing. This is a big part of our product."
"I absolutely love Hitachi. I'm one of the forefront supporters of Hitachi for my firm. It's so easy to integrate within our environments. In terms of being able to quickly build ETL jobs, transform, and then automate them, it's really easy to integrate throughout for data analytics."
"Sometimes, it took a whole team about two weeks to get all the data to prepare and present it. After the optimization of the data, it took about one to two hours to do the whole process. Therefore, it has helped a lot when you talk about money, because it doesn't take a whole team to do it, just one person to do one project at a time and run it when you want to run it. So, it has helped a lot on that side."
"Flexible deployment, in any environment, is very important to us. That is the key reason why we ended up with these tools. Because we have a very highly secure environment, we must be able to install it in multiple environments on multiple different servers. The fact that we could use the same tool in all our environments, on-prem and in the cloud, was very important to us."
"It makes it pretty simple to do some fairly complicated things. Both I and some of our other BI developers have made stabs at using, for example, SQL Server Integration Services, and we found them a little bit frustrating compared to Data Integration. So, its ease of use is right up there."
"The core replication and its performance. Performance is crucial, and SharePlex is by far the fastest. The way it handles replication to multiple targets along with basic filtering, as well as from multiple sources to a single target, is very efficient."
"There are some capabilities within SharePlex where you can see how the data is migrating and if it still maintains good data integrity. For example, if there are some tables that get out of sync, there are ways to find them and fix the problem on the spot. Since these are very common issues, we can easily fix these types of problems using utilities, like compare and repair. So, if you find something is out of sync, then you can just repair that table. It basically syncs that table from source to target to see if there are any differences. It will then replicate those differences to the target."
"Because of the volume of the transactions, we heavily use a feature that allows SharePlex to replicate thousands of transactions. It's called PEP, Post Enhancement Performance, and that has helped us scale tremendously."
"The core features of the solution we like are the reliability of the data transfer and the accuracy of data read and write. The stability of the solution is also excellent."
"I like SharePlex's Compare and Repair tool."
"As far as I remember, not all connectors worked very well. They can add more connectors and more drivers to the process to integrate with more flows."
"I'm still in the very recent stage concerning Pentaho Data Integration, but it can't really handle what I describe as "extreme data processing" i.e. when there is a huge amount of data to process. That is one area where Pentaho is still lacking."
"I work with different databases. I would like to work with more connectors to new databases, e.g., DynamoDB and MariaDB, and new cloud solutions, e.g., AWS, Azure, and GCP. If they had these connectors, that would be great. They could improve by building new connectors. If you have native connections to different databases, then you can make instructions more efficient and in a more natural way. You don't have to write any scripts to use that connector."
"If you're working with a larger data set, I'm not so sure it would be the best solution. The larger things got the slower it was."
"The reporting definitely needs improvement. There are a lot of general, basic features that it doesn't have. A simple feature you would expect a reporting tool to have is the ability to search the repository for a report. It doesn't even have that capability. That's been a feature that we've been asking for since the beginning and it hasn't been implemented yet."
"The performance could be improved. If they could have analytics perform well on large volumes, that would be a big deal for our products."
"The product needs more plugins."
"I could not connect to our Hadoop environment in an easy and flexible way, and it was important to scale our data warehouse."
"The reporting features need improvement. It would be very good for users to have a clear understanding of the status of replication."
"For its function in relation to replication (i.e. filtering), I'd give it a six or seven out of 10. GoldenGate has much more functionality by comparison."
"I don't know how easy it would be to change the architecture in an already implemented replication. For example, if we have a certain way of architecting for a particular database migration and want to change that during a period of time, is that an easy or difficult change? There was a need for us to change the architecture in-between the migration, but we didn't do it. We thought, "This is possibly complicated. Let's not change it in the middle because we were approaching our cutover date." That was one thing that we should have checked with support about for training."
"I would like the solution to have some kind of machine learning and AI capabilities. Often, if we want to improve the performance of posting, we have to bump up a parameter. That means we need to stop the process, come up with a figure that we want to bump the parameter up to, and then start SharePlex. Machine learning and AI capabilities for these kinds of improvement would tremendously help boost productivity for us."
"I would like more ability to automate installation and configuration in line with some of the DevOps processes that are more mature in the market. That would be a considerable improvement."
More Pentaho Data Integration and Analytics Pricing and Cost Advice →
Pentaho Data Integration and Analytics is ranked 15th in Data Integration with 48 reviews while Quest SharePlex is ranked 25th in Data Integration with 5 reviews. Pentaho Data Integration and Analytics is rated 8.0, while Quest SharePlex is rated 9.0. The top reviewer of Pentaho Data Integration and Analytics writes "It's flexible and can do almost anything I want it to do". On the other hand, the top reviewer of Quest SharePlex writes "It reduces the downtime and migration time exponentially". Pentaho Data Integration and Analytics is most compared with SSIS, Azure Data Factory, Talend Open Studio, Oracle Data Integrator (ODI) and AWS Glue, whereas Quest SharePlex is most compared with Oracle GoldenGate, AWS Database Migration Service, Qlik Replicate, Oracle Enterprise Manager and Fivetran. See our Pentaho Data Integration and Analytics vs. Quest SharePlex report.
See our list of best Data Integration vendors.
We monitor all Data Integration reviews to prevent fraudulent reviews and keep review quality high. We do not post reviews by company employees or direct competitors. We validate each review for authenticity via cross-reference with LinkedIn, and personal follow-up with the reviewer when necessary.