If you want to discover one of the latest tools designed to produce stunning Big Data insights, this book features everything you need to get to grips with your data. Whether you are a data architect, developer, or a business strategist, HDInsight adds value in everything from development, administration, and reporting.
If you are a Big Data enthusiast and wish to use Hadoop v2 to solve your problems, then this book is for you. This book is for Java programmers with little to moderate knowledge of Hadoop MapReduce. This is also a one-stop reference for developers and system admins who want to quickly get up to speed with using Hadoop v2. It would be helpful to have a basic knowledge of software development using Java and a basic working knowledge of Linux.
Enter the world of Internet of Things with the power of data science with this highly practical, engaging book About This Book Explore real-world use cases from the Internet of Things (IoT) domain using decision science with this easy-to-follow, practical book Learn to make smarter decisions on top of your IoT solutions so that your IoT is smart in a real sense This highly practical, example-rich guide fills the gap between your knowledge of data science and IoT Who This Book Is For If you have a basic programming experience with R and want to solve business use cases in IoT using decision science then this book is for you. Even if your're a non-technical manager anchoring IoT projects, you can skip the code and still benefit from the book. What You Will Learn Explore decision science with respect to IoT Get to know the end to end analytics stack – Descriptive + Inquisitive + Predictive + Prescriptive Solve problems in IoT connected assets and connected operations Design and solve real-life IoT business use cases using cutting edge machine learning techniques Synthesize and assimilate results to form the perfect story for a business Master the art of problem solving when IoT meets decision science using a variety of statistical and machine learning techniques along with hands on tasks in R In Detail With an increasing number of devices getting connected to the Internet, massive amounts of data are being generated that can be used for analysis. This book helps you to understand Internet of Things in depth and decision science, and solve business use cases. With IoT, the frequency and impact of the problem is huge. Addressing a problem with such a huge impact requires a very structured approach. The entire journey of addressing the problem by defining it, designing the solution, and executing it using decision science is articulated in this book through engaging and easy-to-understand business use cases. You will get a detailed understanding of IoT, decision science, and the art of solving a business problem in IoT through decision science. By the end of this book, you'll have an understanding of the complex aspects of decision making in IoT and will be able to take that knowledge with you onto whatever project calls for it Style and approach This scenario-based tutorial approaches the topic systematically, allowing you to build upon what you learned in previous chapters.
Create outstanding enterprise solutions around DocumentDB using the latest technologies and programming tools with Azure About This Book Get to know the concepts of DocumentDB and learn to work your way around it Manipulate and query your documents using different modern technologies to access DocumentDB Build a real-life scenario using Microsoft Visual Studio and C# with this handy and practical guide Who This Book Is For This book is for novice developers and database architects who need a thorough knowledge of the features of DocumentDB and developing applications with it. Basic knowledge of SQL would be helpful. What You Will Learn Create, manage, and configure your DocumentDB environment Execute SQL queries from simple to complex and nested ones against your database Get to know about advanced DocumentDB techniques such as scopes, portioning, indexing, triggers, UDF's, and security Fine-tune your DocumentDB database to optimize performance and costs Interact with DocumentDB from different technologies and platforms Build a real-life scenario using C# and put DocumentDB at the heart of Azure solutions Understand how to migrate from your current datastore to DocumentDB In Detail Learning DocumentDB adopts a practical, step-by-step approach to help you learn the basics of DocumentDB and use your new-found abilities in real-life scenarios and enterprise solutions. We start with the absolute basics, such as setting up a DocumentDB environment, and guide you through managing your databases, and executing simple and complex queries. Next, we explain how to work with DocumentDB using the open REST protocol, and demonstrate how JavaScript works with DocumentDB. We'll also show you how to authenticate and execute queries. Moving on, you'll find out how to use DocumentDB from within Node.js to kick-start your Node.js projects. Next, you'll discover how to increase the performance of your DocumentDB database and fine-tune it. Finally, you'll get to grips with using DocumentDB in conjunction with other services offered from the Microsoft Azure platform. Style and approach This book can be used as a tutorial where you learn step by step, but also as a knowledge base to quickly look up recipes you can instantly utilize. Starting with the basics and moving on to advanced topics, every concept is explained in theory and demonstrated through easy-to-understand examples.
Explore architectural approaches to building Data Lakes that ingest, index, manage, and analyze massive amounts of data using Big Data technologies About This Book Comprehend the intricacies of architecting a Data Lake and build a data strategy around your current data architecture Efficiently manage vast amounts of data and deliver it to multiple applications and systems with a high degree of performance and scalability Packed with industry best practices and use-case scenarios to get you up-and-running Who This Book Is For This book is for architects and senior managers who are responsible for building a strategy around their current data architecture, helping them identify the need for a Data Lake implementation in an enterprise context. The reader will need a good knowledge of master data management and information lifecycle management, and experience of Big Data technologies. What You Will Learn Identify the need for a Data Lake in your enterprise context and learn to architect a Data Lake Learn to build various tiers of a Data Lake, such as data intake, management, consumption, and governance, with a focus on practical implementation scenarios Find out the key considerations to be taken into account while building each tier of the Data Lake Understand Hadoop-oriented data transfer mechanism to ingest data in batch, micro-batch, and real-time modes Explore various data integration needs and learn how to perform data enrichment and data transformations using Big Data technologies Enable data discovery on the Data Lake to allow users to discover the data Discover how data is packaged and provisioned for consumption Comprehend the importance of including data governance disciplines while building a Data Lake In Detail A Data Lake is a highly scalable platform for storing huge volumes of multistructured data from disparate sources with centralized data management services. This book explores the potential of Data Lakes and explores architectural approaches to building data lakes that ingest, index, manage, and analyze massive amounts of data using batch and real-time processing frameworks. It guides you on how to go about building a Data Lake that is managed by Hadoop and accessed as required by other Big Data applications. This book will guide readers (using best practices) in developing Data Lake's capabilities. It will focus on architect data governance, security, data quality, data lineage tracking, metadata management, and semantic data tagging. By the end of this book, you will have a good understanding of building a Data Lake for Big Data. Style and approach Data Lake Development with Big Data provides architectural approaches to building a Data Lake. It follows a use case-based approach where practical implementation scenarios of each key component are explained. It also helps you understand how these use cases are implemented in a Data Lake. The chapters are organized in a way that mimics the sequential data flow evidenced in a Data Lake.
Manage, analyze, and visualize data with Microsoft Excel 2013 to transform raw data into ready to use information About This Book Create formulas to help you analyze and explain findings Develop interactive spreadsheets that will impress your audience and give them the ability to slice and dice data A step-by-step guide to learn various ways to model data for businesses with the help of Excel 2013 Who This Book Is For If you want to start using Excel 2013 for data analysis and business modeling and enhance your skills in the data analysis life cycle then this book is for you, whether you're new to Excel or experienced. What You Will Learn Discover what Excel formulas are all about and how to use them in your spreadsheet development Identify bad data and learn cleaning strategies Create interactive spreadsheets that engage and appeal to your audience Leverage Excel's powerful built-in tools to get the median, maximum, and minimum values of your data Build impressive tables and combine datasets using Excel's built-in functionality Learn the powerful scripting language VBA, allowing you to implement your own custom solutions with ease In Detail Excel 2013 is one of the easiest to use data analysis tools you will ever come across. Its simplicity and powerful features has made it the go to tool for all your data needs. Complex operations with Excel, such as creating charts and graphs, visualization, and analyzing data make it a great tool for managers, data scientists, financial data analysts, and those who work closely with data. Learning data analysis and will help you bring your data skills to the next level. This book starts by walking you through creating your own data and bringing data into Excel from various sources. You'll learn the basics of SQL syntax and how to connect it to a Microsoft SQL Server Database using Excel's data connection tools. You will discover how to spot bad data and strategies to clean that data to make it useful to you. Next, you'll learn to create custom columns, identify key metrics, and make decisions based on business rules. You'll create macros using VBA and use Excel 2013's shiny new macros. Finally, at the end of the book, you'll be provided with useful shortcuts and tips, enabling you to do efficient data analysis and business modeling with Excel 2013. Style and approach This is a step-by-step guide to performing data analysis and business modelling with Excel 2013, complete with examples and tips.
Make life at the office easier for server administrators by helping them build resilient Ubuntu server systems About This Book Tackle the issues you come across in keeping your Ubuntu server up and running Build server machines and troubleshoot cloud computing related issues using Open Stack Discover tips and best practices to be followed for minimum maintenance of Ubuntu Server 3 Who This Book Is For This book is for a vast audience of Linux system administrators who primarily work on Debian-based systems and spend long hours trying fix issues with the enterprise server. Ubuntu is already one of the most popular OSes and this book targets the most common issues that most administrators have to deal with. With the right tools and definite solutions, you will be able to keep your Ubuntu servers in the pink of health. What You Will Learn Deploy packages and their dependencies with repositories Set up your own DNS and network for Ubuntu Server Authenticate and validate users and their access to various systems and services Maintain, monitor, and optimize your server resources and avoid tremendous load Get to know about processes, assigning and changing priorities, and running processes in background Optimize your shell with tools and provide users with an improved shell experience Set up separate environments for various services and run them safely in isolation Understand, build, and deploy OpenStack on your Ubuntu Server In Detail Ubuntu is becoming one of the favorite Linux flavors for many enterprises and is being adopted to a large extent. It supports a wide variety of common network systems and the use of standard Internet services including file serving, e-mail, Web, DNS, and database management. A large scale use and implementation of Ubuntu on servers has given rise to a vast army of Linux administrators who battle it out day in and day out to make sure the systems are in the right frame of operation and pre-empt any untoward incidents that may result in catastrophes for the businesses using it. Despite all these efforts, glitches and bugs occur that affect Ubuntu server's network, memory, application, and hardware and also generate cloud computing related issues using OpenStack. This book will help you end to end. Right from setting up your new Ubuntu Server to learning the best practices to host OpenStack without any hassles. You will be able to control the priority of jobs, restrict or allow access users to certain services, deploy packages, tackle issues related to server effectively, and reduce downtime. Also, you will learn to set up OpenStack, and manage and monitor its services while tuning the machine with best practices. You will also get to know about Virtualization to make services serve users better. Chapter by chapter, you will learn to add new features and functionalities and make your Ubuntu server a full-fledged, production-ready system. Style and approach This book contains topic-by-topic discussion in an easy-to-understand language with loads of examples to help you take care of Ubuntu Server. Plenty of screenshots will guide you through a step-by-step approach.
Get Started Fast with Apache Hadoop® 2, YARN, and Today’s Hadoop Ecosystem With Hadoop 2.x and YARN, Hadoop moves beyond MapReduce to become practical for virtually any type of data processing. Hadoop 2.x and the Data Lake concept represent a radical shift away from conventional approaches to data usage and storage. Hadoop 2.x installations offer unmatched scalability and breakthrough extensibility that supports new and existing Big Data analytics processing methods and models. Hadoop® 2 Quick-Start Guide is the first easy, accessible guide to Apache Hadoop 2.x, YARN, and the modern Hadoop ecosystem. Building on his unsurpassed experience teaching Hadoop and Big Data, author Douglas Eadline covers all the basics you need to know to install and use Hadoop 2 on personal computers or servers, and to navigate the powerful technologies that complement it. Eadline concisely introduces and explains every key Hadoop 2 concept, tool, and service, illustrating each with a simple “beginning-to-end” example and identifying trustworthy, up-to-date resources for learning more. This guide is ideal if you want to learn about Hadoop 2 without getting mired in technical details. Douglas Eadline will bring you up to speed quickly, whether you’re a user, admin, devops specialist, programmer, architect, analyst, or data scientist. Coverage Includes Understanding what Hadoop 2 and YARN do, and how they improve on Hadoop 1 with MapReduce Understanding Hadoop-based Data Lakes versus RDBMS Data Warehouses Installing Hadoop 2 and core services on Linux machines, virtualized sandboxes, or clusters Exploring the Hadoop Distributed File System (HDFS) Understanding the essentials of MapReduce and YARN application programming Simplifying programming and data movement with Apache Pig, Hive, Sqoop, Flume, Oozie, and HBase Observing application progress, controlling jobs, and managing workflows Managing Hadoop efficiently with Apache Ambari–including recipes for HDFS to NFSv3 gateway, HDFS snapshots, and YARN configuration Learning basic Hadoop 2 troubleshooting, and installing Apache Hue and Apache Spark
Get started fast with Apache Hadoop 2, with the first easy, accessible guide to this revolutionary Big Data technology. Building on his unsurpassed experience teaching Hadoop and Big Data, Dr. Douglas Eadline covers all the basics you need to know to install and use Hadoop 2 on both personal computers and servers, and navigate the entire Apache Hadoop ecosystem. Eadline demystifies Hadoop 2, explains the problems it solves, shows how it relates to Big Data, and demonstrates both administrators and users work with it. He explains the central role of MapReduce in Hadoop 1, and how (and why) YARN and Hadoop 2 move beyond MapReduce. You'll find essential information on: Planning and performing Hadoop 2 installations -- including decisions about hardware, software, clustering, and HDFS Using the Hadoop Distributed File System (HDFS) and working around its tradeoffs Running and benchmarking Hadoop 2 programs Working with MapReduce -- including basic programming examples Using higher-level tools, including Pig and Hive Getting started with Apache Hadoop YARN frameworks Administering Hadoop 2 with Ambari, rmadmin, and automated scripts From its Getting Started checklist/flowchart to its roadmap of additional resources, Hadoop 2 Quick-Start Guide is your perfect Hadoop 2 starting point -- and your fastest way to start mastering Big Data.
Unlock the power of your data with Hadoop 2.X ecosystem and its data warehousing techniques across large data sets About This Book Conquer the mountain of data using Hadoop 2.X tools The authors succeed in creating a context for Hadoop and its ecosystem Hands-on examples and recipes giving the bigger picture and helping you to master Hadoop 2.X data processing platforms Overcome the challenging data processing problems using this exhaustive course with Hadoop 2.X Who This Book Is For This course is for Java developers, who know scripting, wanting a career shift to Hadoop - Big Data segment of the IT industry. So if you are a novice in Hadoop or an expert, this book will make you reach the most advanced level in Hadoop 2.X. What You Will Learn Best practices for setup and configuration of Hadoop clusters, tailoring the system to the problem at hand Integration with relational databases, using Hive for SQL queries and Sqoop for data transfer Installing and maintaining Hadoop 2.X cluster and its ecosystem Advanced Data Analysis using the Hive, Pig, and Map Reduce programs Machine learning principles with libraries such as Mahout and Batch and Stream data processing using Apache Spark Understand the changes involved in the process in the move from Hadoop 1.0 to Hadoop 2.0 Dive into YARN and Storm and use YARN to integrate Storm with Hadoop Deploy Hadoop on Amazon Elastic MapReduce and Discover HDFS replacements and learn about HDFS Federation In Detail As Marc Andreessen has said “Data is eating the world,” which can be witnessed today being the age of Big Data, businesses are producing data in huge volumes every day and this rise in tide of data need to be organized and analyzed in a more secured way. With proper and effective use of Hadoop, you can build new-improved models, and based on that you will be able to make the right decisions. The first module, Hadoop beginners Guide will walk you through on understanding Hadoop with very detailed instructions and how to go about using it. Commands are explained using sections called “What just happened” for more clarity and understanding. The second module, Hadoop Real World Solutions Cookbook, 2nd edition, is an essential tutorial to effectively implement a big data warehouse in your business, where you get detailed practices on the latest technologies such as YARN and Spark. Big data has become a key basis of competition and the new waves of productivity growth. Hence, once you get familiar with the basics and implement the end-to-end big data use cases, you will start exploring the third module, Mastering Hadoop. So, now the question is if you need to broaden your Hadoop skill set to the next level after you nail the basics and the advance concepts, then this course is indispensable. When you finish this course, you will be able to tackle the real-world scenarios and become a big data expert using the tools and the knowledge based on the various step-by-step tutorials and recipes. Style and approach This course has covered everything right from the basic concepts of Hadoop till you master the advance mechanisms to become a big data expert. The goal here is to help you learn the basic essentials using the step-by-step tutorials and from there moving toward the recipes with various real-world solutions for you. It covers all the important aspects of Hadoop from system designing and configuring Hadoop, machine learning principles with various libraries with chapters illustrated with code fragments and schematic diagrams. This is a compendious course to explore Hadoop from the basics to the most advanced techniques available in Hadoop 2.X.
This book takes you on a fantastic journey to discover the attributes of big data using Apache Hive. Key Features Grasp the skills needed to write efficient Hive queries to analyze the Big Data Discover how Hive can coexist and work with other tools within the Hadoop ecosystem Uses practical, example-oriented scenarios to cover all the newly released features of Apache Hive 2.3.3 Book Description In this book, we prepare you for your journey into big data by frstly introducing you to backgrounds in the big data domain, alongwith the process of setting up and getting familiar with your Hive working environment. Next, the book guides you through discovering and transforming the values of big data with the help of examples. It also hones your skills in using the Hive language in an effcient manner. Toward the end, the book focuses on advanced topics, such as performance, security, and extensions in Hive, which will guide you on exciting adventures on this worthwhile big data journey. By the end of the book, you will be familiar with Hive and able to work effeciently to find solutions to big data problems What you will learn Create and set up the Hive environment Discover how to use Hive's definition language to describe data Discover interesting data by joining and filtering datasets in Hive Transform data by using Hive sorting, ordering, and functions Aggregate and sample data in different ways Boost Hive query performance and enhance data security in Hive Customize Hive to your needs by using user-defined functions and integrate it with other tools Who this book is for If you are a data analyst, developer, or simply someone who wants to quickly get started with Hive to explore and analyze Big Data in Hadoop, this is the book for you. Since Hive is an SQL-like language, some previous experience with SQL will be useful to get the most out of this book.
Demystify Big Data and discover how to bring operational intelligence to your data to revolutionize your work About This Book Get maximum use out of your data with Splunk's exceptional analysis and visualization capabilities Analyze and understand your operational data skillfully using this end-to-end course Full coverage of high-level Splunk techniques such as advanced searches, manipulations, and visualization Who This Book Is For This course is for software developers who wish to use Splunk for operational intelligence to make sense of their machine data. The content in this course will appeal to individuals from all facets of business, IT, security, product, marketing, and many more What You Will Learn Install and configure the latest version of Splunk. Use Splunk to gather, analyze, and report data Create Dashboards and Visualizations that make data meaningful Model and accelerate data and perform pivot-based reporting Integrate advanced JavaScript charts and leverage Splunk's APIs Develop and Manage apps in Splunk Integrate Splunk with R and Tableau using SDKs In Detail Splunk is an extremely powerful tool for searching, exploring, and visualizing data of all types. Splunk is becoming increasingly popular, as more and more businesses, both large and small, discover its ease and usefulness. Analysts, managers, students, and others can quickly learn how to use the data from their systems, networks, web traffic, and social media to make attractive and informative reports. This course will teach everything right from installing and configuring Splunk. The first module is for anyone who wants to manage data with Splunk. You'll start with very basics of Splunk— installing Splunk— before then moving on to searching machine data with Splunk. You will gather data from different sources, isolate them by indexes, classify them into source types, and tag them with the essential fields. With more than 70 recipes on hand in the second module that demonstrate all of Splunk's features, not only will you find quick solutions to common problems, but you'll also learn a wide range of strategies and uncover new ideas that will make you rethink what operational intelligence means to you and your organization. Dive deep into Splunk to find the most efficient solution to your data problems in the third module. Create the robust Splunk solutions you need to make informed decisions in big data machine analytics. From visualizations to enterprise integration, this well-organized high level guide has everything you need for Splunk mastery. This learning path combines some of the best that Packt has to offer into one complete, curated package. It includes content from the following Packt products: Splunk Essentials - Second Edition Splunk Operational Intelligence Cookbook - Second Edition Advanced Splunk Style and approach Packed with several step by step tutorials and a wide range of techniques to take advantage of Splunk and its wide range of capabilities to deliver operational intelligence within your enterpise