Try our new research platform with insights from 80,000+ expert users

Apache Hadoop vs Teradata vs VMware Tanzu Data Solutions comparison

 

Comparison Buyer's Guide

Executive Summary

Review summaries and opinions

We asked business professionals to review the solutions they use. Here are some excerpts of what they said:
 

ROI

Sentiment score
6.5
Apache Hadoop provides cost-effective data storage and processing, though ROI varies based on analytics use and sophistication.
Sentiment score
8.1
Teradata boosts analytics speed over 100%, enhancing customer service and satisfaction, with high ROI and user approval.
Sentiment score
7.0
VMware Tanzu Data Solutions is cost-efficient, stable, and valuable, offering high uptime without needing open-source alternatives.
 

Customer Service

Sentiment score
6.4
Customer service varies by Hadoop distributor, with Hortonworks rated highly; support depends on vendor, community resources, or external vendors.
Sentiment score
7.1
Teradata's customer service is praised for expertise but criticized for delays, with ratings ranging from 6 to 10 out of 10.
Sentiment score
6.7
VMware Tanzu Data Solutions' customer service is mixed, with varying experiences in response times and technical support effectiveness.
It's not structured support, which is why we don't use purely open-source projects without additional structured support.
The technical support from Teradata is quite advanced.
Customer support is very good, rated eight out of ten under our essential agreement.
 

Scalability Issues

Sentiment score
7.6
Apache Hadoop excels in scalability, allowing easy cluster expansion and efficient data handling, ideal for varied organizational needs.
Sentiment score
7.4
Teradata is praised for its scalability, speed, and flexibility, despite some complexity and cost challenges in cloud environments.
Sentiment score
7.3
VMware Tanzu Data Solutions offer scalability and adaptability but face challenges with complexity and resource limitations under high workloads.
It is a distributed file system and scales reasonably well as long as it is given sufficient resources.
This expansion can occur without incurring downtime or taking systems offline.
Scalability is complex as you need to purchase a license and coordinate with Teradata for additional disk space and CPU.
Most of our functions or jobs are queued due to that.
 

Stability Issues

Sentiment score
7.3
Apache Hadoop's stability, rated 8/10, improves with newer versions, though minor issues exist with memory and data processing.
Sentiment score
8.4
Teradata excels in stability with minimal downtime, robust architecture, 99.9% uptime, and reliable performance, despite minor large dataset issues.
Sentiment score
7.0
VMware Tanzu Data Solutions is stable for most users but experiences performance issues under heavy loads or specific configurations.
Continuous management in the way of upgrades and technical management is necessary to ensure that it remains effective.
I find the stability to be almost a ten out of ten.
The workload management and software maturity provide a reliable system.
I have faced stability issues, mainly due to the storage my organization has, though I am not sure if it's specifically due to the tool.
 

Room For Improvement

Apache Hadoop needs improved usability, integration, security, support, and performance for efficient high-volume data processing and better community resources.
Teradata users seek better transaction processing, enhanced scalability, modern interface, cloud focus, advanced analytics, and improved support and documentation.
VMware Tanzu Data Solutions need stability, scalability, and performance improvements, better integration, and alignment with PostgreSQL features.
The problem with Apache Hadoop arose when the guys that originally set it up left the firm, and the group that later owned it didn't have enough technical resources to properly maintain it.
Unlike SQL and Oracle, which have in-built replication capabilities, we don't have similar functionality with Teradata.
 

Setup Cost

Enterprise Hadoop offers cost benefits but varies with deployment type and distribution, impacting smaller organizations more heavily.
Teradata's high cost is justified by its superior performance, competitive total ownership costs, and flexible pricing models.
VMware Tanzu Data Solutions offers cost-effective open-source options with competitive pricing for enhanced features and optional support.
Initially, it may seem expensive compared to similar cloud databases, however, it offers significant value in performance, stability, and overall output once in use.
Teradata is much more expensive than SQL, which is well-performed and cheaper.
 

Valuable Features

Apache Hadoop excels with a scalable, cost-effective system handling diverse data types, integrating with tools, and supporting big data analytics.
Teradata offers efficient, scalable data management with fast query performance, robust security, automation, and cloud flexibility for businesses.
VMware Tanzu Data Solutions is praised for its scalability, integration, speed, security, and advanced analytics capabilities.
Hadoop is a distributed file system, and it scales reasonably well provided you give it sufficient resources.
The data mover is valuable over the last two years as it allows us to achieve data replication to our disaster recovery systems.
The product is not complex; I do not have to create stored procedures, functions, or views.
 

Mindshare comparison

As of May 2025, in the Data Warehouse category, the mindshare of Apache Hadoop is 5.1%, down from 5.6% compared to the previous year. The mindshare of Teradata is 15.9%, up from 15.5% compared to the previous year. The mindshare of VMware Tanzu Data Solutions is 3.6%, down from 5.3% compared to the previous year. It is calculated based on PeerSpot user engagement data.
Data Warehouse
 

Featured Reviews

Sushil Arya - PeerSpot reviewer
Provides ease of integration with the IT workflow of a business
When working with Kafka, I saw that the data came in an incremental order. The incremental data processing part is still not very effective in Apache Hadoop. If the data is already there, it can be processed very effectively, especially if the data is coming in every second. If you want to know the location of some data every second, then such data is not processed effectively in Apache Hadoop. I can say that one of the features where improvements are required revolves around the licensing cost of the tool. If the tool can build some licensing structures in a pay-per-use manner, organizations can get the look and feel of Apache Hadoop. Apache Hadoop can offer a licensing structure of the product that can be seen as similar to how AWS operates. Apache Hadoop can look into the capability of processing incremental data. The tool's setup process can be a scope of improvement. Also, it is not very simple because while doing the setup, we need to do all the server settings, including port listing and firewall configurations. If we look at other products on the market, then they can be made simpler. There are certain shortcomings when it comes to the product's technical support part, making it an area where improvements are required. The time frame for the resolution is an area that needs to be improved. The overall communication part of the technical support team also needs improvement.
SurjitChoudhury - PeerSpot reviewer
Offers seamless integration capabilities and performance optimization features, including extensive indexing and advanced tuning capabilities
We created and constructed the warehouse. We used multiple loading processes like MultiLoad, FastLoad, and Teradata Pump. But those are loading processes, and Teradata is a powerful tool because if we consider older technologies, its architecture with nodes, virtual processes, and nodes is a unique concept. Later, other technologies like Informatica also adopted the concept of nodes from Informatica PowerCenter version 7.x. Previously, it was a client-server architecture, but later, it changed to the nodes concept. Like, we can have the database available 24/7, 365 days. If one node fails, other nodes can take care of it. Informatica adopted all those concepts when it changed its architecture. Even Oracle databases have since adapted their architecture to them. However, this particular Teradata company initially started with its own different type of architecture, which major companies later adopted. It has grown now, but initially, whatever query we sent it would be mapped into a particular component. After that, it goes to the virtual processor and down to the disk, where the actual physical data is loaded. So, in between, there's a map, which acts like a data dictionary. It also holds information about each piece of data, where it's loaded, and on which particular virtual processor or node the data resides. Because Teradata comes with a four-node architecture, or however many nodes we choose, the cost is determined by that initially. So, what type of data does each and every node hold? It's a shared-no architecture. So, whatever task is given to a virtual processor it will be processed. If there's a failure, then it will be taken care of by another virtual processor. Moreover, this solution has impacted the query time and data performance. In Teradata, there's a lot of joining, partitioning, and indexing of records. There are primary and secondary indexes, hash indexing, and other indexing processes. To improve query performance, we first analyze the query and tune it. If a join needs a secondary index, which plays a major role in filtering records, we might reconstruct that particular table with the secondary index. This tuning involves partitioning and indexing. We use these tools and technologies to fine-tune performance. When it comes to integration, tools like Informatica seamlessly connect with Teradata. We ensure the Teradata database is configured correctly in Informatica, including the proper hostname and properties for the load process. We didn't find any major complexity or issues with integration. But, these technologies are quite old now. With newer big data technologies, we've worked with a four-layer architecture, pulling data from Hadoop Lake to Teradata. We configure Teradata with the appropriate hostname and credentials, and use BTEQ queries to load data. Previously, we converted the data warehouse to a CLD model as per Teradata's standardized procedures, moving from an ETL to an EMT process. This allowed us to perform gap analysis on missing entities based on the model and retrieve them from the source system again. We found Teradata integration straightforward and compatible with other tools.
Sushma Polavarapu - PeerSpot reviewer
Create essential reporting scripts with simplicity but has some operational challenges
We mainly create functions and scripts required for reporting The product is not complex; I do not have to create stored procedures, functions, or views. It's just simple to use. There is general room for improvement. I have been using it for five years. I have faced stability issues, mainly…
report
Use our free recommendation engine to learn which Data Warehouse solutions are best for your needs.
850,671 professionals have used our research since 2012.
 

Comparison Review

it_user232068 - PeerSpot reviewer
Aug 5, 2015
Netezza vs. Teradata
Original published at https://www.linkedin.com/pulse/should-i-choose-net Two leading Massively Parallel Processing (MPP) architectures for Data Warehousing (DW) are IBM PureData System for Analytics (formerly Netezza) and Teradata. I thought talking about the similarities and differences…
 

Top Industries

By visitors reading reviews
Financial Services Firm
33%
Computer Software Company
11%
University
6%
Energy/Utilities Company
6%
Financial Services Firm
26%
Computer Software Company
11%
Manufacturing Company
7%
Healthcare Company
7%
Financial Services Firm
28%
Computer Software Company
16%
Manufacturing Company
7%
Government
4%
 

Company Size

By reviewers
Large Enterprise
Midsize Enterprise
Small Business
 

Questions from the Community

What do you like most about Apache Hadoop?
It's primarily open source. You can handle huge data volumes and create your own views, workflows, and tables. I can ...
What is your experience regarding pricing and costs for Apache Hadoop?
The product is open-source, but some associated licensing fees depend on the subscription level. While it might be fr...
What needs improvement with Apache Hadoop?
The problem with Apache Hadoop arose when the guys that originally set it up left the firm, and the group that later ...
Comparing Teradata and Oracle Database, which product do you think is better and why?
I have spoken to my colleagues about this comparison and in our collective opinion, the reason why some people may d...
Which companies use Teradata and who is it most suitable for?
Before my organization implemented this solution, we researched which big brands were using Teradata, so we knew if ...
Is Teradata a difficult solution to work with?
Teradata is not a difficult product to work with, especially since they offer you technical support at all levels if ...
How does IBM MQ compare with VMware RabbitMQ?
IBM MQ has a great reputation behind it, and this solution is very robust with great stability. It is easy to use, si...
What is your experience regarding pricing and costs for VMware Tanzu Greenplum?
It’s an open-source solution. There are no expenses for using it.
 

Also Known As

No data available
IntelliFlex, Aster Data Map Reduce, , QueryGrid, Customer Interaction Manager, Digital Marketing Center, Data Mover, Data Stream Architecture
Greenplum, Pivotal Greenplum, VMware RabbitMQ, VMware Tanzu GemFire, VMware Postgres
 

Overview

 

Sample Customers

Amazon, Adobe, eBay, Facebook, Google, Hulu, IBM, LinkedIn, Microsoft, Spotify, AOL, Twitter, University of Maryland, Yahoo!, Cornell University Web Lab
Netflix
General Electric, Conversant, China CITIC Bank, Aridhia, Purdue University
Find out what your peers are saying about Snowflake Computing, Oracle, Teradata and others in Data Warehouse. Updated: April 2025.
850,671 professionals have used our research since 2012.