As we begin to dive deeper into the connected world, there has been an explosion of structured and unstructured data. Additionally, advancements in Apache Hadoop and other Big Data technologies, cloud computing and machine learning tools all play into how this world will evolve. Over the last ten years, Apache Hadoop has proven to be a popular platform among seasoned developers who require a technology that can power large, complex applications. However, for customers, partners and application ISVs who write on-top of Hadoop, there is still one huge issue that remains; Interoperability. In this talk, john Mertic will take a closer look at how Apache Hadoop can become more interoperable to accelerate big data implementations.
5. @ODPiOrg
TOP 3 KEY CONSIDERATIONS FOR DATA
PROJECTS
What problem am I
trying to solve?
What data do I have and
where is it coming from?
What questions need
answered?
13. @ODPiOrg
WHY ADD A FEEDBACK LOOP
Modern data tools and
platforms are non-
trivial ( by design )
Enterprise deployment
pattern considerations
don’t make it upstream
Interoperability standards are
inconsistent, even though
platform vendors often derive
from the same codebase
16. @ODPiOrg
SUSTAINABLE DATA INVESTMENT
Successful Projects depend on
members, developers, infrastructure to
develop products that the market will
adopt.
Downstream constituents (ISVs/IHVs,
Solution Providers, End Users) as a
part of this helps them be an active
participant in their software supply
chain.
Google CES 2017/Mobile World Conference - pull out examples of craziest internet connected devices here.
Kolibree Ara Smart Toothbrush
HiMirror, A Smart Beauty Mirror
Smarter's FridgeCam
Zeeq Smart Pillow
Kerastase Hair Coach
Hadoop is used as a data strategy to solve #2 - has been implemented over the last 10 years. Even after 10 years it’s still growing and changing.
Quote out a few vendors doing this.
Progressive Insurance is quietly amassing a trove of driver data — 10 billion miles of driving data, to be exact — with Snapshot, its opt-in driving tracker device.
Hadoop is used as a data strategy to solve #2 - has been implemented over the last 10 years. Even after 10 years it’s still growing and changing.
Quote out a few vendors doing this.
Progressive Insurance is quietly amassing a trove of driver data — 10 billion miles of driving data, to be exact — with Snapshot, its opt-in driving tracker device.
But even after 10 years, adoption isn’t where people would expect it.
For Enterprise production of Hadoop, helping fill the gaps here would help the growth pattern
Talk to enterprises wanting to participate in the creation process – software supply chain.