Skip to content

Pub Content

Santhosh-Devarajan edited this page May 22, 2018 · 5 revisions

Short Description (two sentences that describe the book):

The proposed book moves on to demonstrate the importance of a step by step learning of a workflow manager for Hadoop, Spark and associated software components in the Big-Data ecosystem. It introduces the native integration that simplifies and accelerates delivery of enterprise Big-data applications and its insights using the Azkaban workflow manager.

Long Description (300-350 words that describe the book and offer a benefit—think of it as back cover copy:

*Long Description of the book can be divided into 4 parts,

Part 1 - Focusing on what is the reason or a need to write a book on current technology. Briefly mentioning the shortcoming and how the same can be addressed OR how this technology would enhance process.

The insight provided by the applications in the Big Data ecosystem has become a prerequisite for companies to remain competitive in the emerging world. Creating an open source software project, Hadoop has becoming an integral part of a company’s IT ecosystem. Now the speed at which a company can generate business insights out of big data is vital to the data they have available when making important business decisions. The challenge is how to integrate and orchestrate the Big Data applications and processes into existing IT processes, enabling data scientists’ developers and business analysts, without causing major disruption and impacting businesses-usual operations. This book explores why organizations need a workflow manager to support their Big Data environment and it also highlights the main open source workflow solutions available today and exposes the limitations of such solutions.

Part 2 - Here we start the paragraph with “The Title” and how this new technology can bring about the needed change/enhancement in the current scenario.

“Beginners Guide to Big Data Workflow Manager using Azkaban“ will help the developers or business analyst to kick start their development process by integrating the various open source software components available in the big data ecosystem. The importance of a workflow manager is imperative for Big data components. The current enterprise landscape in the big data space is having an abundance of software technologies for making and bringing the solutions alive. This demands the introduction of workflow managers to enhance the speed at which developers can quickly build, operate and manage the life-cycle of the application development.

Part 3 – Here we dive deep into topics to be covered and highlight 4 to 5 points key points that would stand out.

Topics will be focusing more on a step by step approach of building ,integrating with various open source technology components available and typically combine Big Data technologies and legacy applications using pipeline building options available in Azkaban.

This book covers the following major titles for Azkaban

  • Introduction to workflow managers - 5 to 10 pages.
  • Azkaban Features - 5 to 10 pages.
  • Setting up your first Azkaban Instance - 5 to 10 pages.
  • Azkaban Configuration & User Management - 15 to 20 pages.
  • Creating the first Azkaban Job. - 5 to 10 pages.
  • Advanced Azkaban Job configuration- 30 to 40 pages
  • Azkaban Job Types - 20 to 30 pages.
  • Engineering Azkaban- 20 to 25 pages.
  • Securing Azkaban- 10 pages
  • Building an Enterprise use-case. - 20 pages.

Part 4 - Specifying addition of Practical example, Coding exercise or Use cases which may help the user and finally concluding as how using this technology will become the game changer in the complete eco system.

This book will be completely focusing on building and integrating the Azkaban workflow from the scratch. Approach will be a step-by-step learning mechanism with example and codes (shared via GIT). Initial Chapters will be focused individual components and its integration with the workflow, moving forward on the final chapters will be completely show casing a enterprise life-cycle orchestrating with workflow.

Readership (who’s the target audience?):

Developers, Big Data Programmers, Business Analysts, Software Architects , Consultants What you will learn: (Please give us a bulleted list of 5 or fewer items) We already have very good hands on these technologies.

USP of the book – 4 to 5 bullet points:

  1. No Similar books available in the market covering big-data workflow managers other than Apache Oozie.
  2. Very hot topic and every big data system needs a workflow manager and Azkaban comes latest in this category.
  3. Book takes us to the industries standard of deploying a workflow manager in the enterprise.
  4.  One stop solution for beginners to kick-start, build, operate and manage enterprise grade applications in a secure environment.
    

Channel Buyer: Why should a channel buyer or book shop exhibit the book – 3 to 4 line on topic and book should suffix. Book is taking us to the industries latest and hot selling workflow managers and similar books are also not available. So for a big data developer or an architect now it has become mandatory that they need to know at least one workflow in the big data industry.

Mini Outline (Cover each chapter. As an alternative, you can write a one- or two-sentence description of each chapter):

  • Introduction to workflow managers - This chapter covers the various aspects of different industry leading workflow managers in use and its comparison and the fitment.
  • Azkaban Features - Covering the aspects of Azkaban features that will distinguish or standout Azkaban among other workflow managers.
  • Setting up your first Azkaban Instance - Guided tour to build and deploy the first Azkaban instance with a minimal configuration.
  • Azkaban Configuration & User Management - Guided tour to each and every configuration aspect of Azkaban that would elevate the minimal installation to an enterprise grade deployment with access control features.
  • Creating the first Azkaban Job. - Step by Step instructions to build and run a simple Azkaban job using built-in job types.
  • Advanced Azkaban Job configuration- Complete life cycle management of creation, view and execution of flows using Azkaban Web-UI.
  • Azkaban Job Types - Azkaban plugin management and usage of various plugin provided by the community.
  • Engineering Azkaban- Interaction of Azkaban functionalities via RESTful services which will help the enterprise to seamlessly integrate Azkaban with their software frameworks.
  • Securing Azkaban- Securing Azkaban using Enterprise security standards.
  • Building an Enterprise use-case. - An industrial use-case which covers the end-to-end functionality with big-data components which leverages Azkaban as workflow manager.

Clone this wiki locally