HDP is the industry's only true secure, enterprise-ready open source Apache™ Hadoop® distribution based on a centralized architecture (YARN). HDP addresses the complete needs of data-at-rest, powers real-time customer applications and delivers robust big data analytics that accelerate decision making and innovation.
YARN and Hadoop Distributed File System (HDFS) are the cornerstone components of Hortonworks Data Platform (HDP) for data-at-rest. While HDFS provides the scalable, fault-tolerant, cost-efficient storage for your big data lake, YARN provides the centralized architecture that enables you to process multiple workloads simultaneously. YARN provides the resource management and pluggable architecture for enabling a wide variety of data access methods.
Data streaming, processing and analytics engines for a variety of workloads
Hortonworks Data Platform includes a versatile range of processing engines that empower you to interact with the same data in multiple ways, at the same time. This means applications for big data analytics can interact with the data in the best way: from batch to interactive SQL or low latency access with NoSQL. Emerging use cases for data science, search and streaming are also supported with Apache Spark, Storm and Kafka.
HDP extends data access and management with powerful tools for data governance and integration. They provide a reliable, repeatable, and simple framework for managing the flow of data in and out of Hadoop. This control structure, along with a set of tooling to ease and automate the application of schema or metadata on sources is critical for successful integration of Hadoop into your modern data architecture.
Hortonworks has engineering relationships with many leading data management providers to enable their tools to work and integrate with HDP.
Authentication, authorization, and data protection
Security is woven and integrated into HDP in multiple layers. Critical features for authentication, authorization, accountability and data protection are in place to help secure HDP across these key requirements. Consistent with this approach throughout all of the enterprise Hadoop capabilities, HDP also ensures you can integrate and extend your current security solutions to provide a single, consistent, secure umbrella over your modern data architecture.
Operations teams deploy, monitor and manage a Hadoop cluster within their broader enterprise data ecosystem. Apache Ambari simplifies this experience. Ambari is an open source management platform for provisioning, managing, monitoring, and securing the Hortonworks Data Platform. It enables Hadoop to fit seamlessly into your enterprise environment.
Provision and manage Hadoop clusters in any cloud environment
Cloudbreak, as part of Hortonworks Data Platform and powered by Apache Ambari, allows simplified provisioning and Hadoop cluster management in any cloud environment including; Amazon Web Services, Microsoft Azure, Google Cloud Platform and OpenStack. It optimizes your use of cloud resources as workloads change.
Progressive Insurance es una de las compañías de seguros de vehículos más grandes de los EE.UU, El equipo recurrió a Hortonworks Data Platform para transformar su negocio con una ingesta masiva de nuevos tipos de datos. Progressive utiliza HDP para la colocación de anuncios y el almacenamiento de los datos de conducción para sus productos de seguros basados en el uso.
Download the Product Guide In this guide, discover: The technology components of HDP and its blueprint for Enterprise Hadoop. How HDP integrates and complements your existing data systems. The versatility...
You may be up all night wondering how enterprise organizations deal with large data volumes and data varieties without significantly increasing costs. And perhaps your existing data architectures are not...
In the financial services industry, fraud becomes an ever-increasing concern as the world becomes more connected through the web, electronic payments, mobility, and ever-changing technologies. The amount of data that...
Our customers increasingly leverage Data Science, and Machine Learning to solve complex predictive analytics problem. A few examples of these problems are churn prediction, predictive maintenance, image classification, and entity...
With the increases in data volumes, explosion in number of devices and sophistication of attackers, current cybersecurity tools are challenged processing the millions of events and providing insights fast enough....
Apache, Hadoop, Falcon, Atlas, Tez, Sqoop, Flume, Kafka, Pig, Hive, HBase, Accumulo, Storm, Solr, Spark, Ranger, Knox, Ambari, ZooKeeper, Oozie, Phoenix, NiFi, HAWQ, Zeppelin, Atlas, Slider, Mahout, MapReduce, HDFS, YARN, Metron and the Hadoop elephant and Apache project logos are either registered trademarks or trademarks of the Apache Software Foundation in the United States or other countries.