Lead Storage Engineer at a insurance company with 5,001-10,000 employees
Jul 29, 2019
Unified Manager, System Manager, and Cloud Manager are all GUI-based. It's easy for somebody who has not been exposed to this for years to pick it up and work with it.
NetApp's Cloud Manager automation capabilities are very good because it's REST-API-driven, so we can completely automate everything. It has a good overview if you want to just have a look into your environment as well.
The solution’s Snapshot copies and thin clones in terms of operational recovery are good. Snapshot copies are pretty much the write-in time data backups. Obviously, critical data is snapshotted a lot more frequently, and even clients and end users find it easier to restore whatever they need if it's file-based, statical, etc.
Storage Engineer at a media company with 5,001-10,000 employees
Jul 2, 2020
One of the most valuable features is its similarity to the physical app, which makes it familiar. It's almost identical to a real NetApp, which means you can run all of the associated NetApp processes and services with it. Otherwise, we would definitely have to deploy some hardware on a site somewhere, which could be a challenge in terms of CapEx.
The feature which I like the most is that it has the capabilities that the traditional storage system offers. It provides all the functionality. The deduplication and compression work exactly like ONTAP's traditional storage. So people who have experience with that find it very easy to manage.
Senior Analyst at a comms service provider with 5,001-10,000 employees
Nov 2, 2020
We're able to use the SnapMirror function and SnapMirror data from our on-prem environment into Azure. That is super-helpful. SnapMirror allows you to take data that exists on one NetApp, on a physical NetApp storage platform, and copy it over to another NetApp storage platform. It's a solid, proven technology, so we don't worry about whether data is getting lost or corrupted during the SnapMirror.
The storage tiering is definitely the most valuable feature... With respect to tiering, the inactive data is pushed to a lower tier where the storage cost is cheap, but the access cost is high.
Technology Consultant at a tech services company with 501-1,000 employees
Feb 1, 2022
The Cloud Manager application that's on the NetApp cloud site is easy to use. You can set up and schedule replications from there, so you don't have to go into the ONTAP system. Another feature we've recently started using is the scheduled power off. We started with one client and have been slowly implementing it with others. We can cut costs by not having the VM run all the time. It's only on when it's doing replication, but it powers off after.
Lead Storage Engineer at a insurance company with 5,001-10,000 employees
Jul 29, 2019
Some of the licensing is a little kludgy. We just created an HA environment in Azure and their licensing for SVMs per node is a little kludgy. They're working on it right now.
Scale-up and scale-out could be improved. It would be interesting to have multiple HA pairs on one cluster, for example, or to increase the single instances more, from a performance perspective. It would be good to get more performance out of a single HA pair.
How it handles erasure coding. I feel it the improvement should be there. Basically, it should be seamless. You don't want to have an underlying hardware issue or something, then suddenly there's no reads or writes. Luckily, it's at a replication site, so our main production site is still working and writing to it. But, the replication site has stopped right now while we try to bring that node back. Since we implemented in bare-metal, not in appliance, we had to go back to the original vendor. They didn't send it in time, and we had a hardware memory issue. Then, we had a hard disk issue, which brought the node down physically.
Storage Engineer at a media company with 5,001-10,000 employees
Jul 2, 2020
There is room for improvement with the capacity. There's a very hard limit to how many disks you can have and how much space you can have. That is something they should work to fix, because it's limiting. Right now, the limit is about 360 terabytes or 36 disks.
When it comes to a critical or a read-write-intensive application, it doesn't provide the performance that some applications require, especially for SAP. The SAP HANA database has a write-latency of less than 2 milliseconds and the CVO solution does not fit there. It could be used for other databases, where the requirements are not so demanding, especially when it comes to write-latency.
Senior Analyst at a comms service provider with 5,001-10,000 employees
Nov 2, 2020
When Azure does their maintenance, they do maintenance on one node at a time. With the two nodes of the CVO, it can automatically fail over from one node to the node that is staying up. And when the first node comes back online, it will fail back to the first node. We have had issues with everything failing back 100 percent correctly.
It definitely needs improvement with respect to clustering and with respect to more collaborative integrations with Azure. Right now, we have very limited functionalities with Azure, except for storage. If CVO could be integrated with Azure that would help. When there is any sort of maintenance happening in the cloud, it disrupts the service in Cloud Volumes ONTAP.
Technology Consultant at a tech services company with 501-1,000 employees
Feb 1, 2022
Cloud Volumes ONTAP's interface could use an overhaul. Sometimes you have to dig around in Cloud Manager a little bit to find certain things. The layout could be more intuitive.