Discover how NetApp’s AI Data Guardrails turn governance into a living system—enabling secure, compliant, and scalable AI platforms. From risk managem ...read more
By Mohammad Hossein Hajkazemi, Bhushan Jain, and Arpan Chowdhry
Introduction
Google Cloud NetApp Volumes is a fully managed, cloud-native storage s ...read more
NetApp Console delivers HIPAA (Health Insurance Portability and Accountability Act)- compliant data intelligence without storing ePHI
NetApp Console n ...read more
NetApp Console delivers simplicity with Console agent
NetApp® Console agent is the secure and trusted software from NetApp that enables the workflows ...read more
In today's fast-paced digital world, maintaining the continuity of your business operations during unforeseen disruptions is of paramount importance. Enter NetApp® Console, a robust interface that simplifies disaster recovery (DR) while reducing costs and complexity.
With traditional DR, the challenges can vary from having to manage VMware and on-premises storage independently, which can be daunting and resource intensive, to licensing requirements that lead to higher costs and inflexibility. In addition, many solutions depend on VMware vSphere APIs for data protection, which consume extra resources and fail to support on-premises native efficiency capabilities in the cloud.
... View more
GraphRAG has taken off fast, mostly because teams want AI systems that can explain themselves... But building and operating a full knowledge graph means managing schemas, ontologies, and graph infrastructure before you even know if the use case will pay off. This post walks through what sits in the middle; using a BM25-based retrieval mechanism. It looks at Hybrid RAG in practice and explains why treating retrieval as a first-class, controllable step matters more than throwing more embeddings at the problem. The focus is on how retrieval choices shape answers, trust, and long-term reliability without turning your stack into a research project.
... View more
If you’re looking to migrate or to analyze your data and free up storage, then you’re in the right place
In today’s data-driven world, it’s critical for you to effectively manage and protect your data. Manual data classification is not only impractical but also prone to human error. Enter the automated NetApp® Data Classification service, which is an absolute game-changer in managing your data. Let's delve a little deeper into the great benefits that you get with Data Classification!
... View more
Looking for a New Year’s resolution you can actually keep? Look no further! NetApp has been your go-to for certified Ansible modules, and now, with the release of the latest StorageGRID Ansible collection (version 21.16.0), automating your StorageGRID environment has never been easier. 🚀
Whether you're new to automation or a seasoned pro, this is the perfect time to dive into using these powerful modules. Our detailed guide walks you through onboarding a new tenant with a single Ansible playbook, making complex tasks a breeze. From creating tenants and buckets to generating access keys, we've got you covered.
Check out the full blog post for a step-by-step breakdown and start your automation journey today!
... View more
This blog provides a comprehensive guide to implementing data tiering in Hadoop environments using NetApp XCP, NFS, and S3 storage solutions. It covers setup, migration, verification, and automation strategies to optimize storage costs and performance.
• Benefits of Hadoop data tiering: Data tiering moves frequently accessed “hot” data to high-performance storage and infrequently accessed “cold” data to cost-effective object storage, optimizing storage costs and query performance while maintaining governance.
• Role of NetApp XCP: XCP facilitates high-throughput, scalable migrations from HDFS to NetApp NFS (hot storage) and S3 (cold storage), ensuring data integrity through verification features and supporting integration with Hadoop clusters.
• Architecture and process flow: The workflow involves classifying HDFS files by modification time into /hot and /cold directories, migrating these to NetApp NFS and S3 respectively using XCP, followed by verification of data integrity.
• Prerequisites and environment setup: The Hadoop cluster must be configured in HDFS mode with appropriate directories and storage policies (/hot as HOT, /cold as COLD). NetApp NFS and S3 targets must be configured and accessible from the XCP host, which requires specific environment variables for Java and Hadoop libraries.
• Data migration and verification examples: Sample commands demonstrate copying data from HDFS /hot to NetApp NFS and verifying the transfer using XCP. Migration to S3 requires professional support and proper configuration of AWS profiles and endpoints.
• Automated tiering script: A provided bash script classifies files by age, moves them to /hot or /cold, and runs XCP copy and verify commands for NFS and S3 targets. It supports dry-run mode and configurable parameters for flexible operation.
• Oozie workflow integration: The guide includes sample Oozie workflow and coordinator XML configurations to automate the tiering process on a scheduled basis, enabling repeatable and auditable execution within Cloudera Hadoop environments.
• Operational recommendations and outcomes: Running XCP as root with unique migration IDs and clean catalogs is advised. The process yields 40–60% storage capacity savings by reducing replicated data copies on enterprise storage, while maintaining high availability and data protection through NetApp features.
... View more