Apache Hadoop YARN is the modern distributed operating system for big data applications. It morphed the Hadoop compute layer to be a common resource management platform that can host a wide variety of applications. Many organizations leverage YARN in building their applications on top of Hadoop without themselves repeatedly worrying about resource management, isolation, multi-tenancy issues, etc.
In this talk, we’ll start with the current status of Apache Hadoop YARN—how it is used today in deployments large and small. We'll then move on to the exciting present and future of YARN—features that are further strengthening YARN as the first-class resource management platform for data centers running enterprise Hadoop.
We’ll discuss the current status as well as the future promise of features and initiatives like: powerful container placement, global scheduling, support for machine learning and deep learning workloads through GPU and FPGA support, extreme scale with YARN federation, containerized apps on YARN, support for long running services (alongside applications) natively without any changes, seamless application upgrades, powerful scheduling features like application priorities, intra-queue preemption across applications, and operational enhancements including insights through Timeline Service V2, a new web UI, and better queue management.
Speakers
Wangda Tan, Staff Software Engineer, Hortonworks
Billie Rinaldi, Principal Software Engineer I, Hortonworks
Many users are requesting most necessary features, so that’s why we release so fast.
Many of this featuress a necessary to run YARN cluster, such as RM fail-over/ HA, etc
2.6/2.7/2.8 are versions which most prod cluster are using, so many feature development remains in the background and many community effort are focusing on stablizing features.
Again, the most essential YARN doesn’t meet user’s requirement anymore, that’s why we released 3 minor releases in 6 months, and include features like:
YARN Federation (larger cluster)
GPU / FPGA, etc.
Even though TF provide options to use GPU memory less than whole device provided. But we cannot enforce this from external.
High level talk on ATSv2, it is scalable solution compared to 1.5.