BIG DATA / ANALYTICS – Trainosys https://devtrainosys.slogninja.com The leader in training Fri, 14 May 2021 14:09:22 +0000 en-US hourly 1 https://wordpress.org/?v=6.6.2 https://devtrainosys.slogninja.com/wp-content/uploads/2020/09/logourl-100x100.png BIG DATA / ANALYTICS – Trainosys https://devtrainosys.slogninja.com 32 32 Cassandra Admin Advanced https://devtrainosys.slogninja.com/course/cassandra-admin-advanced/ Thu, 29 Apr 2021 06:07:20 +0000 https://trainosys.com/?post_type=product&p=1966 Pre-requisites:
  • The intended audience should have a background on:
    • Software Development
    • Unix and Hadoop
  • The lab setup for this course would be:
    • Windows 7 or Higher
    • 32GB RAM
    • VMWare Player

Target Audience:

  • Hadoop Developers
  • Data Engineers
  • Integration Engineers
  • Architects
  • Data Steward

Course Duration:

  • 21 hours – 3 days

Course Content:

  • Getting Started
  • Installation
    • Python
    • Choice of IDE
    • Required modules
  • Project Environment Setup
    • Creating a workspace folder
    • Testing if installation is correct
  • First Steps Towards Programming
  • Syntax
    • Expressions
    • Statements
    • Variables
  • Symbols Definition
  • Terminologies
  • Data Types
  • Primitives
    • Numbers
    • Strings
    • Booleans
      • Equivalence
    • Data Structures
      • Lists
      • Tuple
      • Dictionary
      • Set
    • Operators
      • Arithmetic Operators
      • Compound Operators
      • Comparison Operators
      • Membership Operators
      • Logical Operators
      • Identity Operators
    • Control Structures
      • Conditional
      • Loops
        • For loop
        • While loop
        • Loop with Else
      • The range() function
      • Break, Continue & Pass
    • Functions
      • Calling functions
      • Defining functions
      • Returning values
      • Accepting arguments
        • Default arguments
        • Keyword arguments
        • Arbitrary arguments list
  • Anonymous Functions
    • Map, Filter
  • Modules & Packages
    • Introduction to The Standard Library
      • Mathematics
      • Dates and Times
      • Random
      • Operating System Interface
    • Modules
    • Packages
    • Installing Third-Party Packages
    • Anaconda
  • Comprehension
    • General comprehension syntax
    • Types of comprehension
  • Generators
    • Generators vs Lists
    • General generator syntax
    • Functional generators
  • Object Oriented Programming
    • Introduction to Object Oriented Programming
    • Classes and Objects
      • Properties and Methods
      • Initializer & Destructor
    • Inheritance and Method Overwriting
  • Error Handling
    • When to use Exceptions
    • Handling Exceptions
    • Custom error handling routines
  • File and Directory Handling
    • Reading & Writing Text Files
    • Managing files and folders from Python
  • Databases
    • Introduction to databases
    • CRUD operations on SQLite3
    • Integrating with external databases
  • Virtual Environments
    • Creating virtual environments
  • Custom Content (depending on attendee’s requirements)
    • Exploration of relevant packages
    • Mini project

 

Course Customization Options

To request a customized training for this course, please contact us to arrange.

]]>
Basic SAP HANA https://devtrainosys.slogninja.com/course/basic-sap-hana/ Tue, 27 Apr 2021 05:09:08 +0000 https://trainosys.com/?post_type=product&p=1927 Course Overview:

Our Sap Hana Training course will help you understand and learn the fundamentals and will also felicitate on training hands-on for the better grasp on the course. Further, we have the highly qualified professionals who will train you about Sap Hana Studio, Modeling, Security features and its various other aspects. This course is designed in such a way that it will make you understand ‘why sap Hana is a fundamentally different database engine.

Course Objectives:

  • To help you learn about SAP HANA basics, features, architecture and terminology involved.
  • Train Hana Studio Modelers to help you gain expertise in modeling.
  • Help you master Sap Hana Interface, client, and connectivity.
  • Learn Auto-documentation process and SAP BO Analysis.
  • We will help you learn Sap Hana administration activities like managing users, storage and security.
  • The course will master you on System Replication, data provisioning, and ABAP Data flow.
  • You will learn all the process for configuring, reporting, and monitoring and dashboard creation.
  • To provide real-time and complex HANA projects for better understanding.
  • The course will help you writing Sap Hana queries and performance tuning.

Pre-requisites:

  • Basics of SQL, Database & Data Warehouse
  • Knowledge about SAP BW preferred
  • Basic understanding of any ERP business process would be an added advantage

Target Audience:

  • SAP Developer & BI consultants
  • Data warehouse professionals
  • Project Managers
  • Data Analytics Professionals along with Database Architectures

Course Duration:

  • 28 hours – 4 days

Course Content:

HANA Introduction

  • In-Memory Computing
  • Evolution of In-Memory computing at SAP
  • History of SAP HANA
  • HANA compare to BWA
  • In-Memory Basics
  • HANA Use cases

Architecture

  • Hana Engine Overview
  • Different HANA Engine Types
  • Row Store
  • Column Store
  • Persistency Layer
  • Business Impact of new architecture
  • Backup & Recovery

Modeling

  • Key Concepts in Data Modeling
  • Components of HANA data model & Views
  • Analytical Views
  • Attribute views
  • Calculation Views
  • Joins
  • Measures
  • Filters
  • Real Time Scenarios
  • HANA SQL Intro
  • Functions & Expressions
  • Procedures

Data Provisioning

  • Overview
  • Trigger Based Replication
  • ETL Based Replication
  • Log Based Replication
  • Intro to BODS 4
  • Basic Data service Connection types
  • Flat File upload in to HANA

Security

  • Creating Users
  • Creating Roles
  • Privileges
  • User Administration

Reporting

  • Connectivity options
  • Business Objects BI 4

 

Course Customization Options

To request a customized training for this course, please contact us to arrange.

]]>
Building Communication Services Using RabbitMQ and Node.JS https://devtrainosys.slogninja.com/course/building-communication-services-using-rabbitmq-and-node-js/ Mon, 26 Apr 2021 06:43:56 +0000 https://trainosys.com/?post_type=product&p=1920 Course Overview:

This training program is aimed at building high scalable server applications using Node.js and RabbitMQ. Participants will learn how to build server APIs using Node.js and understand and exploit the benefits of “Single Threaded Event Loop”. Participants will also learn to build scalable solutions using RabbitMQ as the messaging platform for communication between services and learn to exploit the various messaging strategies for better scalability and reliability.

Course Objectives:

  • This workshop is aimed at imparting the required know-how to build and manage applications using Node.js & RabbitMQ

Pre-requisites:

  • Experience with the Linux command line interface.
  • An understanding of enterprise application development concepts.

Target Audience:

  • IT specialists
  • Developers
  • System Architects

Course Duration:

  • 35 hours – 5 days

Course Content:

MESSAGING SOLUTIONS USING RABBITMQ

  1. Introduction to RabbitMQ
  1. Understanding Queueing Architecture
  • Message Queueing
  • Uses for Message Queueing
  1. Core Concepts
  • AMQP Messaging Standard
  • Exchanges
  • Queues, Bindings & Consumers
  • Standard Queue
  • Worker Queue
  • Publish & Subscribe
  • Direct Routing
  1. Exchanges & Bindings
  • Direct Exchange
  • Default Exchange
  • Topic Exchange
  • Fanout Exchange
  • Dead Letter Exchange
  1. Understanding Messaging
  • The role of a consumer
  • The role of a producer
  • Bindings consumers and producers
  • Messages and durability
  • How to verify delivery
  1. ES6
  1. New in ES6
  • Classes
  • let and const
  • Arrow functions
  • Enhanced Object Literals
  • Array & Object Destructuring
  • Spread Operator
  • Generators and Iterators
  • Symbols
  • Comprehensions
  • Async Programming using Promises
  • Async Await

BUILDING APPLICATIONS USING NODE.JS

  1. Introduction to Node.js
  • Installing Node.js
  • Node’s Event Loop
  • Alternatives to Node.js
  • Writing asynchronous code
  1. Modularizing code
  • Understanding built-in modules
  • Techniques for modularizing JavaScirpt code
  • Using require() to modularize application code
  • Using npm for third-party modules
  • Handling Exceptions
  1. Events and Streams
  • Understanding Events
  • EventEmitter class
  • Understanding Streams
  • Reading and writing streams
  • Using pipe()
  1. Accessing Local Resources
  • Process Object
  • Manipulating File System
  • Understanding Buffers
  1. js and the web
  • Handling web requests
  • Building a web server
  • Understanding the need for web sockets
  • Realtime interaction using socket.io
  1. Building web API using Express.js
  • Installing Express.js
  • Routing
  • Parameters and queries in routing
  • Working with cookies and sessions
  • Authentication and authorization
  • Error Handling
  1. Data Persistence in Node.js
  • Installing SQLite client module
  • Establishing connections
  • Executing Queries
  • Binding Parameters
  • Inserting, Updating and Deleting Data

 

  1. Building Network Servers & Clients
  • Creating a TCP Server
  • Creating a TCP Client
  • Creating a HTTP/HTTPS client
  • Using the Restler module
  1. Testing Node.js Applications
  • Introduction to TAP
  • Installing node-tap
  • Test Organization
  • Performing Assertions
  • Testing Async with Promises
  • Mocking dependencies using Nock
  • Test Coverage
  1. Scaling Node applications
  • The Child process model
  • exec, spawn, and fork functions
  • Using the Cluster module
  1. js Eco System
  • ONode Packages
  • Packages of interest
  • Deciding Factors
  1. Debugging Node Applications

BUILDING API FOR MESSAGING INTERFACES USING NODE.JS & RABBITMQ

  1. Building RabbitMQ service consumers
  • Installing RabbitMQ Client
  • Connecting to Message Broker
  • Sending Messages
  • Receiving Messages
  1. Implementing Queuing Code
  • Topic Based Publisher & Subscriber for Notifications
  • Simple Publish & Subscribe
  • Publisher Confirmation
  • Polymorphic Publish and Subscribe
  • Controlling Queues and Exchange Names
  • Remote Procedure Calls ?Synchronous RPC
  • Asynchronous RPC oTargeted Messaging
  • Using Send and Receive Pattern oFlexible Routing
  • Topic based publish & subscribe

RABBITMQ ADMINISTRATION

(Duration for this topic is reduced because of the usage of managed services)

  1. Administering RabbitMQ oClusters & Nodes
  • Creating and starting a managed instance
  1. Managing Connections & Channels
  • Starting and stopping nodes
  • RabbitMQ configuration files
  • How to manage privileges
  • Viewing statistics and analyzing logs
  • Sending alerts
  • How to set up parallel processing
  1. High availability with cluster
  • Architecture of a cluster
  • Queues in a cluster
  • Setting up a test cluster
  • Distributing the nodes to more machines
  • How to preserve messages: mirrored queues
  1. Implementing failover and replication
  • Setting up a load balancer-based master/slave
  • Installing the Shovel plugin
  • Configuring and running Shovel
  1. Web tools to administer RabbitMQ
  • The RabbitMQ Management plugin
  • Managing RabbitMQ from the web console
  • Administering users from the web console
  • Managing queue from the web console
  • Using the command line interface
  1. RabbitMQ and the REST API
  • REST API features
  • Accessing statistics
  • vhost and user provisioning
  1. Monitoring and securing RabbitMQ
  • Message durability and Message acknowledgement
  • Memory usage and process limits
  • Setting up SSL

 

 

Course Customization Options

To request a customized training for this course, please contact us to arrange.

]]>
Hadoop Development 0 to 100 https://devtrainosys.slogninja.com/course/hadoop-development-0-to-100/ Fri, 23 Apr 2021 03:08:04 +0000 https://trainosys.com/?post_type=product&p=1885 Course Overview:

Hadoop Development, Administration and BI Program is a one-stop course that introduces you to the domain of Hadoop Development as well as gives you technical know-how of the same.

Course Objectives:

At the end of this course you will be able to earn a credential of Hadoop professional and you will be capable of dealing with Terabyte scale of data and analyze it successfully using MapReduce

  • Learn the basics of Big Data and Hadoop
  • Play with Hadoop and Hadoop Ecosystem
  • Become a top-notch Hadoop Developer

Pre-requisites:

  • Typically, professionals with basic knowledge of software development, programming languages, and databases will find this course helpful. Basic knowledge should be enough to succeed at this course
  • Not For: Students who are absolute beginners at software development as a discipline will find it difficult to follow the course

Target Audience:

  • Developers

Course Duration:

  • 35 hours – 5days

Course Content:

Phase 1: Hadoop Fundamentals with multinode setup (Day 1,2)

Laying the foundation

Big Data

  • What is Big Data
  • Dimensions of Big Data
  • Big Data in Advertising
  • Big Data in Banking
  • Big Data in Telecom
  • Big Data in eCommerce
  • Big Data in Healthcare
  • Big Data in Defense
  • Processing options of Big Data
  • Hadoop as an option

Hadoop

  • What is Hadoop
  • How Hadoop Works
  • HDFS
  • Mapreduce Deep Dive
  • How Hadoop has an edge

Hadoop Ecosystem

  • Sqoop
  • Oozie
  • Pig
  • Hive
  • Flume

Hadoop Hands On

  • Running HDFS commands
  • Running your Mapreduce program
  • Running Sqoop Import and Sqoop Export
  • Creating Hive tables directly from Sqoop
  • Creating Hive tables
  • Querying Hive tables
  • Running an Oozie workflow

Phase 2: Hadoop Development (Day 3)

Become a Pro developer

Apache Spark

  • What is Spark?
  • Using Spark Shell
  • RDD Fundamentals
  • Functional Programming
  • Program

RDD in Depth

  • RDDs
  • Creating RDDs from files
  • Creating RDDs for another RDDs
  • RDD operations
  • Actions
  • Transformations
  • Pair RDDs
  • Joins using RDD

Spark platforms

  • Spark local mode
  • Spark standalone mode
  • Spark on YARN
  • Spark on Mesos

Spark Hands On

  • Python Spark Shell
  • Scala Spark Shell
  • Basic operations on RDDs
  • Pair RDD Hands On

Spark SQL & Dataframes

  • Spark SQL and the SQL Context
  • Creating Dataframes
  • Dataframe Queries and Transformations
  • Saving Dataframes
  • Dataframes and RDDs

Spark Dataframes Hands On

  • Dataframes on a JSON file
  • Dataframes on hive tables
  • Dataframes on JSON
  • Querying operations dataframes

Spark SQL & Dataframes

  • What is Spark Streaming
  • How it works
  • DStreams
  • Developing Spark Streaming Applications

Phase 3: Hadoop BI (Day 4) Analyze data using Pig and Hive

Hive

  • Introduction
  • Basic Data Analysis with Hive
  • Hive Data Management
  • Text Processing with Hive
  • Transformations in Hive
  • Optimizing Hive
  • Hive Hands On
  • Extending Hive

Impala

  • Introduction
  • Basic Data Analysis with Impala
  • Text Processing with Impala
  • Optimizing Impala
  • Impala Hands On

Pig

  • Introduction
  • How Pig works
  • Pig Hands On

Phase 4: NoSQL and Cluster Walkthrough (Day 5) Cluster Administration

NoSQL Databases

  • Why NoSQL
  • What is NoSQL Databases
  • Types of NoSQL Databases
  • Introduction to Cassandra and MongoDB and Hbase
  • Hbase Hands On

Cloudera Manager Setup

  • Why Cluster Manager software
  • Cloudera Manager
  • Using Cloudera Manager to setup cluster
  • Cluster walkthrough

Final Test

 

 

Course Customization Options

To request a customized training for this course, please contact us to arrange.

 

]]>
Hadoop with Spark https://devtrainosys.slogninja.com/course/hadoop-with-spark/ Fri, 23 Apr 2021 03:02:30 +0000 https://trainosys.com/?post_type=product&p=1884 Course Overview:

Hadoop Development, Administration and BI Program is a one-stop course that introduces you to the domain of Hadoop development as well as gives you technical knowhow of the same and it is the most popular Big Data processing framework. Spark is a general-purpose distributed data processing engine that is suitable for use in a wide range of circumstances. On top of the Spark core data processing engine, there are libraries for SQL, machine learning, graph computation, and stream processing, which can be used together in an application.

Course Objectives:

  • Learn the basics of Big Data and Spark
  • Play with Hadoop and Hadoop Ecosystem
  • Become a top-notch Spark Developer

Pre-requisites:

  • Typically, professionals with basic knowledge of software development, programming languages, and databases will find this course helpful. Basic knowledge should be enough to succeed at this course
  • Not For: Students who are absolute beginners at software development as a discipline will find it difficult to follow the course

Target Audience:

  • Developers
  • Data Analysts

Course Duration:

  • 35 hours – 3 days

Course Content:

Phase 1: Hadoop Fundamentals with single node setup (Day 1)

Laying the foundation

Introduction to Hadoop and Spark

  • Ecosystem
  • Big Data Overview
  • Key Roles in Big Data Project
  • Key Business Use cases
  • Hadoop and Spark Logical Architecture
  • Typical Big Data Project Pipeline

Basic Concepts of HDFS

  • HDFS Overview
  • Physical Architectures of HDFS
  • The Hadoop Distributed File System Hands-on.

Hadoop Ecosystem

  • Sqoop
  • Hive

Introduction to Sqoop

  • What is sqoop
  • Import / Import all tables Data
  • Sqoop Job/Eval and Sqoop Code-gen
  • List databases/tables

Hadoop Hands On

  • Running HDFS commands
  • Running Sqoop Import and Sqoop Export

Introduction to Spark

  • Spark Overview
  • Detailed discussion on “Why Spark”
  • Quick Recap of MapReduce
  • Spark vs MapReduce
  • Why Python for Spark
  • Just Enough Python for Spark
  • Understanding of CDH Spark and Apache Spark

Phase 2: Hadoop Development (Day 2)

Become a Pro developer with Spark and Hive Datawarehouse

Spark Core Framework and API

  • High level Spark Architecture
  • Role of Executor, Driver, SparkSession etc.
  • Resilient Distributed Datasets
  • Basic operations in Spark Core API i.e.
  • Actions and Transformations
  • Using the Spark REPL for performing interactive data analysis
  • Hands-on Exercises
  • Integrating with Hive

Delving Deeper into Spark API

  • Pair RDDs
  • Implementing Map Reduce Algorithms using Spark
  • Ways to create Pair RDDs JSON Processing Code Example on JSON Processing
  • XML Processing
  • Joins
  • Playing with Regular Expressions
  • Log File Processing using Regular Expressions
  • Hands-on Exercises

Executing a Spark Application

  • Writing Standalone Spark Application
  • Various commands to execute and configure
  • Spark Applications in various modes
  • Discussion on Application, Job, Stage,
  • Executor, Tasks
  • Interpreting RDD Metadata/Lineage/DAG
  • Controlling degree of Parallelism in Spark Job
  • Physical execution of a Spark application
  • Discussion on: How Spark is better than
  • MapReduce?
  • Hands-on Exercises

Phase 3: Hadoop with Spark Dataframes and park SQL (Day 3 and 4)

Spark SQL

  • Dataframes in Depth
  • Creating Dataframes
  • Discussion on Different file formatORC, Sequence, Avro, Parquet and sequence
  • Dataframe internals that makes it fast – Catalyst Optimizer and Tungsten
  • Load data into Spark from external data sources like relational databases
  • Saving dataframe to external sources like HDFS, RDBMS
  • SQL features of Data frame
  • Data formats – text format such csv, json, xml, binary formats such as parquet,orc
  • UDF in Spark Dataframe
  • When to use UDF of hive or not?
  • CDC use cases
  • Spark optimization techniques-joins?
  • Integration with Teradata- use case

Understanding of Hive

  • Hive as a Data Warehouse
  • Creating Tables for Analysis of data
  • Techniques of Loading Data into Tables
  • Difference between Internal and External Tables
  • Understanding Hive Data Types Joining,Union datasets
  • Join Optimizations
  • Partitions and Bucketing
  • Running a Spark SQL Application
  • Dataframes on a JSON file
  • Dataframes on hive tables
  • Dataframes on JSON
  • Querying operations dataframes
  • Hive Writing HSQL queries for data retrieval

Phase 4: NoSQL and Cluster Walkthrough (Day 5)

Know Kafka Tool and Spark Streaming

Introduction to Kafka

  • Kafka Overview
  • Salient Features of Kafka
  • Topics, Brokers and Partitions
  • Kafka Use cases

Kafka Connect and Spark Streaming

  • Kafka Connect
  • Hands-on Exercise

Structured Streaming

  • Structured Streaming Overview
  • How it is better than Kafka streaming?
  • Hands-on Exercises integrating with Kafka using Spark Streaming

 

 

Course Customization Options

To request a customized training for this course, please contact us to arrange.

]]>
Introduction to API Testing https://devtrainosys.slogninja.com/course/introduction-to-api-testing/ Fri, 23 Apr 2021 02:57:23 +0000 https://trainosys.com/?post_type=product&p=1883 Course Overview:

This course is designed to cover all aspects of API Testing & Postman tool and also to provide users with the knowledge required to work with Postman tool at a professional level. This course addresses planning and performing API tests, developing the pre-test & post-test scripts using Postman, executing the API tests and generating HTML reports using different tools like Newman.

The course covers real-world best practices for effective use of Postman tool for API testing with practical examples demonstrated. Postman is a popular automation tool for API testing. Postman offers features that simplify each step of testing an API.

Course Objectives:

  • Introduction to web application architecture
  • Advantages of API
  • What does API testing involve
  • Functional Testing of Web Services
  • Importing request from your browser

Pre-requisites:

  • An understanding of HTTP
  • NodeJS/NPM experience

Target Audience:

  • Software Testers

Course Duration:

  • 28 hours – 4 days

Course Content:

API Introduction

  • Introduction to web application architecture
  • Introduction to APIs
  • Introduction to Web-Services
  • How does an API works
  • What is API testing?

API TESTING USING POSTMAN– FOUNDATION COURSE

  • Advantages of API
  • API vs Web-Services
  • Introduction to API architecture
    • REST API
    • SOAP API

Understanding how REST API architecture works

  • Understanding how SOAP API architecture works
  • Understanding the HTTP methods
    • GET
    • POST
    • PUT
    • DELETE
    • PATCH
    • OPTIONS
    • HEAD

API TESTING

  • What does API testing involve
  • Validation techniques used in API testing
  • API testing steps
  • Understanding URI, end points, Resources, HTTP verbs
  • Understanding GET request
  • Understanding POST request
  • Understanding PUT request
  • Understanding DELETE request
  • GUI tools available for API testing
  • Command-line tools available for API testing
  • Best Practices for API testing

INTRODUCTION TO POSTMAN API TESTING TOOL

  • What is Postman tool
  • Installation of Native Postman tool
  • Installation of Postman tool as Chrome Add-on
  • Introduction to Postman landscape
  • Introduction to Postman Settings

API TESTING USING POSTMAN

  • Creating First API request using Postman
  • Functional Testing of Web Services
  • Understanding History
  • Setting up Test Project with REST APIs
  • Validating responses with Postman client
  • Validating the body of first JSON response
  • Examples on validating response headers and status codes
  • Understanding query parameters on sending request
  • Validating status code
  • Validating Response time
  • How to debug & troubleshoot
  • How to create and manage Workspaces

API TESTING USING POSTMAN – FOUNDATION COURSE

  • Importing request from your browser
  • Cookies
  • Saving Responses
  • Creating collections in Postman tool
  • Importing/Exportingcollections using Postman tool
  • Run Collections Remotely with URL
  • Creating variables at difference scope levels and how to refer them
  • How to set and get variables through scripting
  • Creating Environments
  • How to setup different URLs using environments
  • How to create quick scripts
  • Pre-request scripts
  • Test scripts/Assertions
  • Make the requests dynamic by taking advantage of variables
  • Generating code snippets in Postman tool

POSTMAN API AUTOMATION-PART 1

  • Collection Runner
  • Postman monitors
  • Automating with Newman tool
  • Basic introduction to NPM tool &NodeJS
  • Installing NodeJS
  • Installing Newman tool
  • Running a collection with Newman tool
  • Specifying environments in Newman tool
  • Creating an HTML report using Newman tool
  • Using Authentication/Authorization in Postman tool
  • Introducing delays for scheduled or continuous load API testing

POSTMAN API AUTOMATION – PART 2

  • Automating POST request with payload
  • Grabbing the response body into string for advance validations
  • Data-Driven testing using

CSV

JSON

  • Scheduling the API testing periodically
  • Load testing of API testing in Postman tool
  • Running and analyzing Load Test results
  • Shortcut methods to convert Raw data to XML/JSON
  • Logging feature to debug automation scripts
  • Using Postman to Create a Reusable API Framework
  • Testing Workflow with Twitter API

 

Course Customization Options

To request a customized training for this course, please contact us to arrange.

 

]]>
Kafka https://devtrainosys.slogninja.com/course/kafka/ Thu, 22 Apr 2021 04:10:48 +0000 https://trainosys.com/?post_type=product&p=1864 Course Overview:

Kafka provides high throughput and scalable messaging systems. Developed on LinkedIn, it can be used effectively in place of traditional messaging system like JMS, Apache MQ etc. Because Kafka can be integrated into systems like Spark, Storm, Hadoop etc, it is a highly recommended messaging system for solving big data problems involved in messaging system. Features of Kafka Cluster like scalability, Fault tolerance, High throughput, Durability etc. makes it a leader in today’s market in messaging systems. Due to extensive use of Kafka in market today, there is a steep increase in job opportunities. Kafka training courses cover integration of Kafka with other Big Data systems as well as how to set up real-time data pipelines for streaming applications.

Course Objectives:

  • Understand Apache Kafka Ecosystem, Architecture, Core Concepts and Operations
  • Master Concepts such as Topics, Partitions, Brokers, Producers, Consumers
  • Start a personal Kafka development environment
  • Learn major CLIs: kafka-topics, kafka-console-producer, kafka-console-consumer, kafka-consumergroups, kafka-configs
  • Create your Producers and Consumers in Java to interact with Kafka
  • Program a Real World Twitter Producer & ElasticSearch Consumer
  • Extended APIs Overview (Kafka Connect, Kafka Streams), Case Studies and Big Data Architecture
  • Practice and Understand Log Compaction

Pre-requisites:

To learn Kafka easily, step-by-step, you have come to the right place! No prior Kafka knowledge is required.

  • A recent Windows / Mac / Linux machine with minimum 16GB of RAM, 100 GB of disk space
  • Some understanding of Java Programming
  • Good to have knowledge about Linux command line
  • Desire to learn something awesome and new!

Target Audience:

  • Some understanding of Java Programming
  • Good to have knowledge about Linux command line

Course Duration:

  • 21 hours – 3 days

Course Content:

  • Introducing to Messaging Systems and their use cases, Problems and limitations (P2P, PubSub)
  • Understand publish-subscribe messaging and how it fits in the big data ecosystem.
  • JMS,RabbitMQ,ActiveMQ
  • Kafka for Big Data & Data Ingestion, Role in ETL
  • Why do we need Kafka? Components of Kafka
  • Kafka use cases
  • Setting Up a Kafka Cluster
  • Setting Up Zookeeper
  • A single node – a single broker cluster, Windows or Linux(Single VM)
  • A single node – multiple broker cluters, Windows or Linus (Single VM)
  • Multiples nodes – multiple broker clusters, EC2 Instances
  • The Kafka broker property list
  • Kafka Design, Leader, Follower, ISR, Offsets
  • Kafka design fundamentals – Broker, Producer, Consumer, Topic, Partition
  • Replication in Kafka
  • Writing Producers, The c#/java producer/consumer API
  • Avro Producer / Consumer c#/java
  • Creating a Java producer with custom partitioning
  • The Kafka producer property list acks, buffer.memory, compression.type, retries, batch.size, linger.ms client.id, max.in.flight.requests.per.connection, timeout.ms and metadata.fetch.timeout.ms
  • Kafka Streams (Java/Scala)
  • Kafka integration with Spark Java Streaming (Standalone VM)
  • Kafka administration tools
  • Kafka cluster mirroring and Monitoring using open source tools
  • Frequent Problems and Solutions
  • Kafka Manager Usage
  • Kafka configuration files details and troubleshooting
  • Partition management & config num. partitions, log. retention.ms, log.retention.bytes, log.segment.bytes log.segment.ms, message.max.bytes
  • Disaster Recovery
  • Low level clients
  • Lost Message detection and recovery
  • Use Kafka Connect to import/export data and Role of KSQL

 

Course Customization Options

To request a customized training for this course, please contact us to arrange.

]]>
Kafka Developer https://devtrainosys.slogninja.com/course/kafka-developer/ Thu, 22 Apr 2021 04:08:44 +0000 https://trainosys.com/?post_type=product&p=1862 Course Overview:

Apache Kafka Certification Training helps you in learning the concepts about Kafka Architecture, Configuring Kafka Cluster, Kafka Producer, Kafka Consumer, Kafka Monitoring.

Apache Kafka Certification Training is designed to provide insights into Integration of Kafka with Hadoop and Spark, understand Kafka Stream APIs, implement Twitter Streaming with Kafka, Cassandra through real life cases studies.

Course Objectives:

  • Traits
    • Traits in Scala
    • Mixing traits into Objects
    • Layered Traits
    • Traits for Rich Interfaces
    • Scala class hierarchy
  • Functions & Procedures
    • Local functions
    • First class functions
    • Partially applied functions
    • Currying
    • Default & Named Arguments
    • Varargs
    • Procedures
    • Closures
  • Classes in Scala
    • Simple Classes and Parameter less Methods
    • Access modifiers and rules
    • Type parameters
    • Object-Private Fields
    • Auxiliary Constructors
    • The Primary Constructor
    • Nested Classes
  • Higher Order functions in Collections
  • Anonymous Functions (aka Function Literals)
  • Pattern Matching in Depth
    • Using the match keyword to return a value
    • Using case classes for pattern matching
    • Adding pattern guards to match conditions
    • Partially specifying matches with wildcards
    • Deep matching using case constructors
    • Matching against collections of items
  • Using extractors instead of case classes
  • Exception Handling

Pre-requisites:

  • Software Development
  • Unix and Hadoop

Target Audience:

  • Hadoop Developers
  • Data Engineers
  • Integration Engineers
  • Architects
  • Data Steward

Course Duration:

  • 35 hours – 5 days

Course Content:

Phase 1: Understanding Big Data and Kafka (Day1)

  • What is Big Data & Dimensions of Big Data
  • Big Data in Advertising, Banking, Telecom, eCommerce, Healthcare, Defense
  • Input options of Big Data
  • Associated Challenges
  • Introduction to Kafka
  • Why Kafka
  • Kafka Use Cases
  • Introduction to Zookeeper
  • Zookeeper: Distributed Applications and Challenges
  • Zookeeper: Handling Partial Failures
  • Zookeeper: Leader Election
  • Aggregating User Activity Using Kafka
  • Zookeeper for Kafka
  • Kafka Architecture
  • What is Kafka Data Model?
  • What is a topic
  • What is a partition and Partition Distribution
  • What is a producer
  • What is a consumer

Phase 2: Kafka Deep Dive (Day 2 and 3)

  • Zookeeper Architecture
  • Zookeeper watcher
  • Kafka broker tuning parameters
  • Kafka Mirror Maker
  • Kafka Brokers
  • Kafka Replication
  • Persistence and Offsets
  • The Kafka broker property list
  • Customer Key and Value Serializer
  • Replication and Partition management in Kafka
  • Writing Producers– The Scala producer API
  • Creating a Scala producer with custom partitioning and Serializers The Kafka producer property list
  • Writing Consumers Kafka consumer APIs
  • Simple Scala consumers Reading messages from a topic and printing them
  • Publish-Subscribe System
  • Configuring Producers with Scala
  • Security options in Kafka

Phase 3: Understanding Scala Programming (Day 4 and 5)

  • Introduction to Functional Programming
  • Introduction to Scala
  • Scala Fundamentals
  • Functions & Procedures
  • Collections
    • Sets
    • Maps
    • Lists
    • Tuples
    • Arrays
  • Higher Order Functions on Collections
  • Objects in Scala oSingletons
    • Companion Objects
    • Objects Exten ding a Class or Trait
    • Application Objects
  • Inheritance
    • Extending a Class
    • Method Overriding
    • Protected Fields and Methods
    • Overriding
    • Anonymous Subclasses
    • Abstract Classes
    • Abstract Fields
    • Inheritance Hierarchy
    • Object Equality
    • Value Classes

 

 

Course Customization Options

To request a customized training for this course, please contact us to arrange.

]]>
MongoDB for Administrators https://devtrainosys.slogninja.com/course/mongodb-for-administrators/ Wed, 21 Apr 2021 06:29:40 +0000 https://trainosys.com/?post_type=product&p=1852 Course Overview:

This course covers everything a database administrator needs to know to successfully deploy and maintain MongoDB databases.

Course Objectives:

  • Understand MongoDB from a developer’s perspective, including its command shell, query API, and driver tools.
  • Deploy MongoDB in all its configurations – as a single server, with master/slave replication, as a replica set, and as a sharded cluster.
  • Evaluate applications and choose hardware appropriately.
  • Monitor MongoDB instances and integrate with standard monitoring software (Munin, Nagios, etc.)
  • Plan for backups and manage large data imports and exports.
  • Troubleshoot the most common developer issues and failure scenarios.
  • Each delegate will need to perform a series of practical exercises.

Pre-requisites:

  • Web and Java Application developer

Target Audience:

  • A Basic knowledge of core java
  • Experience with any other web framework will be helpful

Course Duration:

  • 21 hours – 3 days

Course Content:

MongoDB Architectural Overview

  • Origin, design goals, key features
  • Process structure (mongos, mongod, config servers)
  • Directory / file structure

Working with the MongoDB Shell

  • Documents and data types
  • CRUD (Inserts, queries, updates, deletes)
  • System commands

Single-server Configuration and Deployment

  • Configuration files
  • Data files and allocation
  • Log files
  • Hardware and file-system recommendations

Security

  • Built-in authentication
  • Recommendations for secure deployment
  • Monitoring MongoDB
  • mongostat
  • Analyzing memory and IO performance
  • Integration with monitoring tools: Munin / Cacti / Nagios
  • MongoDB’s web console

Indexing and Query Optimization

  • Managing indexes and MongoDB indexing internals
  • Single / Compound / Geo indexes
  • Identifying sub-optimal queries. Using the query profiler.

Introduction to drivers (Java/Python/Ruby/PHP/Perl)

  • How the drivers and shell communicate with MongoDB
  • BSON and the MongoDB Wire Protocol
  • Troubleshooting application connections

Intro to Read and Write scalability

Replication and Durability

  • Master-slave replication
  • Replica sets
  • Using write concern for durability
  • Handling replication failures

Auto-Sharding

  • How sharding works
  • Setting up a MongoDB shard cluster

Choosing a shard key

Sharding and indexes

  • Sharding and Replica Set Topologies
  • Administering a sharded cluster

Shard / Chunk Migration

Backup and Restore Plans

  • Filesystem-based strategies
  • mongodump / mongorestore
  • rsync
  • mongoimport / mongoexport
  • Summary and Closing Remarks

 

 

Course Customization Options

To request a customized training for this course, please contact us to arrange.

]]>
MySQL Basic https://devtrainosys.slogninja.com/course/mysql-basic/ Wed, 21 Apr 2021 06:25:03 +0000 https://trainosys.com/?post_type=product&p=1850 Course Overview:

The Introduction to MySQL course covers the basics of the SQL language as implemented by MySQL. The course is designed to give delegates practical experience in writing SQL statements using the MySQL client program and MySQL Workbench. The basic SQL statements, including the use of SQL functions and the basic table and view handling statements are introduced.

Course Objectives:

  • Create and alter a MySQL database and query
  • Modify data held within such a database

Pre-requisites:

  • A Basic knowledge of core java
  • Experience with any other web framework will be helpful.

Target Audience:

  • A Basic knowledge of core java
  • Experience with any other web framework will be helpful.

Course Duration:

  • 21 hours – 3 days

Course Content:

Day 1 Introduction to MySQL

  • Course Introduction
  • Administration and course materials
  • Course structure and agenda
  • Delegate and trainer introductions

Session 1: DATABASE CONCEPTS

  • What is a database?
  • Database management systems
  • Tables, rows and columns
  • Indexes, primary keys, unique constraints and foreign keys
  • Client-Server architecture
  • Supported data types
  • Storage engines and table types
  • The Information_Schema and MySQL Databases
  • Exercise: Using a MySQL Database

Session 2: USING THE MySQL CLIENT

  • What is the MySQL client?
  • Getting started and Logging in
  • Selecting a database
  • Client commands
  • Entering and executing SQL statements
  • Retrieving previous commands
  • Creating,editing and executing SQL files
  • Redirecting output into a file
  • Command line execution of MySQL Scripts
  • Exercise: Using the MySQL Client

Session 3: BASIC SELECT

  • The SQL SELECT statement
  • Case sensitivity
  • Quotes
  • Statement terminator
  • Syntax conventions
  • The select clause
  • The FROM clause
  • Conditions and the WHERE clause
  • Logical operators
  • The ORDER BY clause
  • Column aliases
  • Arithmetic expressions
  • Precedence of operators
  • Exercises: Querying the sample database

Session 4: LIMIT,UNION AND AGGREGATE FUNCTIONS

  • The LIMIT clause
  • UNION
  • UNION ALL
  • Aggregate functions
  • The GROUP BY clause
  • Using Rollup with GROUP BY
  • The HAVING clause
  • Exercises: Using Limit,Union and Aggregate Functions
  • Introduction to MySQL Training Course

Day 2 Session 5: SUBQUERIES AND JOINS

  • Subqueries
  • Cartesian products
  • Joins with original syntax
  • Table aliases
  • Natural joins
  • Join using
  • Join on
  • Multi-table joins
  • Exercises: Using Subqueries and Joins

Session 6: NUMERIC AND CHARACTER FUNCTIONS

  • Function types
  • Testing functions
  • Numeric functions
  • Character functions
  • Exercise: Using Numeric and Character Functions

Session 7: DATE,TIME AND OTHER FUNCTIONS

  • Date and time column types
  • Date and time formats
  • The date format function
  • Functions to return date time
  • Functions to extract components from date time
  • Date time arithmetic
  • Miscellaneous functions
  • Exercise: Using Date,Time and other functions
  • Introduction to MySQL Training Course

Day 3 Session 8: DATABASES AND TABLES

  • Creating a database
  • Selecting a database
  • Obtaining database metadata
  • Table properties
  • Creating tables
  • Auto_increment
  • Show create table
  • Column operations
  • Constraint operations
  • Copying tables
  • Renaming tables
  • Changing engine for tables
  • Obtaining table metadata
  • Altering tables
  • Dropping tables
  • Temporary tables
  • Memory tables
  • Exercises: Maintaining databases and tables

Session 9: INDEXES AND VIEWS

  • What is an index?
  • Structure of an index
  • Creating an index
  • Reviewing indexes
  • Composite indexes
  • Altering a table to add an index
  • Dropping indexes
  • What is a view?
  • Creating views
  • View restrictions
  • Dropping views
  • Exercise: Maintaining indexes and views

Session 10: MANAGING DATA

  • Inserting rows
  • Replacing rows
  • Updating rows
  • Deleting rows
  • The truncate statement
  • The COMMIT and ROLLBACK commands
  • Savepoints •Implicit commits
  • Exercise: Managing data

Session 11: ACCESS CONTROL

  • Creating users
  • Renaming users
  • Dropping users
  • Granting privileges
  • Revoking privileges
  • Exercise: Creating users, granting and revoking

Session 12: IMPORT AND EXPORT

  • Exporting using SQL
  • Importing using SQL
  • Exporting from the command line
  • Importing from the command line
  • Exercise: Export and import data

 

Course Customization Options

To request a customized training for this course, please contact us to arrange.

]]>