Which data warehouse tools are users most satisfied with?
What do BI analysts and big data professionals need to consider when choosing an ETL tool?
With all the software options on the cloud computing market, which data warehouse tools are users most satisfied with?
At IT Central Station, many users have written cloud data warehouse reviews that address these questions, as well as other prevalent topics among IT users currently engaged in BI analytics, data mining, and cloud-based data at large.
“Since we have lots of data sources and high volumes, we needed a unified and organized DB that can handle these amounts and will be our one single source of truth for the organization. Therefore, Redshift is the best solution.”
“One of our existing customers stores more than 500 terabytes of data in an AWS Redshift database and the warehouse performance was good.
We want to highlight that even if the warehouse size increases to petabytes, Redshift would still work fine, and there wouldn’t be any performance issues and would cost less also.”
“Query compilation time needs a lot of improvement for cases where you are generating queries dynamically. Also, it would help tremendously to have some more user-friendly, query optimization helper tools.”
JSON format support
Wasserman also discusses Amazon Redshift’s support for JSON format, explaining that “you can copy JSON to the column and have it analyzed using simple functions.” He also attributes the cloud computing tool’s value to “the parallel off/on where you can choose if you want it to unload to split files or into one file.”
How Redshift differs to other cloud computing tools
“I evaluated Hadoop and Spark, along with Redshift. I have no negative comments about those other products. Redshift is flexible in terms of configuration, maintenance, and security, especially VPC configuration, which secures our data a lot…
I have experience working in Hadoop as well. When I compare the two (Redshift & Hadoop), Redshift is more user-friendly in terms of configuration and maintenance.”
“Analyzing years of data requires high processing power and storage. IBM PDA has exactly that. Years of processed data (tables) can be queried and retrieved based on management requirements. This can be done in minutes for analysis.
This is extremely important in identifying trends for decision making in higher management to serve customers better in today’s business environment.”
“It is easy to use. Make sure you select the right ETL and reporting tool. Also, select the right tool for the organization to hold it in the long run.
It has a compression engine and FPGA on but you should still analyze your volume of data and decide on the right model and size.”
Regarding Netezza’s compression capabilities, Gunapalan adds that “I can't extend the storage, only up to 6x compress of data. You need to plan this when selecting the right product to buy.”
This user explains that Netezza “has been the primary driving technology behind the corporate-wide transition to Netezza as a standard data platform. A whole ecosystem is beginning to develop around the product.”
For valuable features, he lists several:
Ease of use
Lack of performance problems for analytics and massive data systems
Integration with Linux-based ETL and data streaming technologies
Integration with distributed computing platforms
Highly complicated architecture
Praveen also points out the challenge involved in using Netezza, which stems from it being “a highly complicated architecture and only IBM engineers/support, or someone who worked on the hardware side of the system can understand the system architecture completely.”
Hello peers,
I am a Senior Data Engineer at a large computer software company.
How can I convert SQL functions in Teradata to Redshift manually?
Thank you for your help.
Hello peers,
I am a Software Engineer Trainee for a large media company.
I am currently researching the migration process between Netezza and AWS Redshift. Can you please provide some information, links, or research papers about this migration?
Thank you for your help.
Here are some things to consider when migrating from Netezza to AWS Redshift:
A. Migrating your data from Netezza to Redshift may be done using methods such as:
o Use a third-party tool to export your data from Netezza and import it into Redshift.
o Use the AWS Data Migration Service to migrate your data between different data stores.
o Manually migrate your data by exporting it from Netezza and importing it into Redshift using SQL.
B. After you migrate your data, you need to model it in Redshift, which involves creating tables and schemas to store your data in a way that is efficient and easy to query.
C. Redshift is a columnar database, which means it is optimized for storing and querying large amounts of data, but you need to tune your Redshift cluster to get the best performance for your specific workload.
D. Redshift is a secure platform, but you need to take steps to secure your data, such as:
o Using encryption for protecting your data at rest and in transit
o Creating user accounts and permissions to control who has access to your data
o Monitoring your Redshift cluster for unauthorized access
The steps in migrating from Netezza to Redshift depend on your needs, but the ones mentioned above are general guidelines.