There’s a lot of information about big data technologies, but splicing these technologies into an end-to-end enterprise data platform is a daunting task not widely covered. With this practical book, you’ll learn how to build big data infrastructure both on-premises and in the cloud and successfully architect a modern data platform. Ideal for enterprise architects, IT managers, application architects, and data engineers, this book shows you how to overcome the many challenges that emerge during Hadoop projects. You’ll explore the vast landscape of tools available in the Hadoop and big data realm in a thorough technical primer before diving into: Infrastructure: Look at all component layers in a modern data platform, from the server to the data center, to establish a solid foundation for data in your enterprise Platform: Understand aspects of deployment, operation, security, high availability, and disaster recovery, along with everything you need to know to integrate your platform with the rest of your enterprise IT Taking Hadoop to the cloud: Learn the important architectural aspects of running a big data platform in the cloud while maintaining enterprise security and high availability
Les mer
Ideal for enterprise architects, IT managers, application architects, and data engineers, this book shows you how to overcome the many challenges that emerge during Hadoop projects.

Produktdetaljer

ISBN
9781491969274
Publisert
2018-12-31
Utgiver
Vendor
O'Reilly Media
Vekt
666 gr
Høyde
233 mm
Bredde
178 mm
Dybde
15 mm
Aldersnivå
P, 06
Språk
Product language
Engelsk
Format
Product format
Heftet
Antall sider
636

Biographical note

Jan Kunigk has worked on enterprise Hadoop solutions since 2010. Before he joined Cloudera in 2014, Jan built optimized systems architectures for Hadoop at IBM and implemented a Hadoop-as-a-Service offering at T-Systems. In his current role as a Solutions Architect he makes Hadoop projects at Cloudera's enterprise customers successful, covering a wide spectrum of architectural decisions to the implementation of big data applications across all industry sectors on a day-to-day basis. Ian Buss began his journey into distributed computing with parallel computational electromagnetics whilst studying for a PhD in photonics at the University of Bristol. After simulating LEDs on supercomputers, he made the move from big compute in academia to big data in the public sector, first encountering Hadoop in 2012. After having fun building, deploying, managing and using Hadoop clusters, Ian joined Cloudera as a Solutions Architect in 2014. His day job now involves integrating Hadoop into enterprises and making stuff work in the real world. Paul Wilkinson has been wrestling with big data in the public sector since before Hadoop existed and was very glad when it arrived in his life in 2009. He became a Cloudera consultant in 2012, advising customers on all things hadoop: application design, information architecture, cluster management and infrastructure planning the FullStack. After a torrent of professional services work across financial services, cybersecurity, adtech, gaming and government, he's seen it all warts and all. Or at least, he hopes he has. Lars George has been involved with Hadoop and HBase since 2007, and became a full HBase committer in 2009. He has spoken at many Hadoop User Group meetings, and conferences such as Hadoop World and Hadoop Summit, ApacheCon, FOSDEM, QCon etc. He also started the Munich OpenHUG meetings. Lars worked for Cloudera for over five years, as the EMEA Chief Architect, acting as a liaison between the Cloudera professional services team and customers as well as partners in and around Europe, building the next data driven solutions. In 2016 he started with his own Hadoop advisory firm, extending on what he has learned and seen in the field for more than 8 years. He is also the author or O'Reilly's "HBase The Definitive Guide."