• Professional Development
  • Medicine & Nursing
  • Arts & Crafts
  • Health & Wellbeing
  • Personal Development

20 Data Architecture courses

🔥 Limited Time Offer 🔥

Get a 10% discount on your first order when you use this promo code at checkout: MAY24BAN3X

SQL NoSQL Big Data and Hadoop

4.7(160)

By Janets

Register on the SQL NoSQL Big Data and Hadoop today and build the experience, skills and knowledge you need to enhance your professional development and work towards your dream job. Study this course through online learning and take the first steps towards a long-term career. The course consists of a number of easy to digest, in-depth modules, designed to provide you with a detailed, expert level of knowledge. Learn through a mixture of instructional video lessons and online study materials. Receive online tutor support as you study the course, to ensure you are supported every step of the way. Get a digital certificate as a proof of your course completion. The SQL NoSQL Big Data and Hadoop is incredibly great value and allows you to study at your own pace. Access the course modules from any internet-enabled device, including computers, tablet, and smartphones. The course is designed to increase your employability and equip you with everything you need to be a success. Enrol on the now and start learning instantly! WHAT YOU GET WITH THE SQL NOSQL BIG DATA AND HADOOP * Receive a e-certificate upon successful completion of the course * Get taught by experienced, professional instructors * Study at a time and pace that suits your learning style * Get instant feedback on assessments  * 24/7 help and advice via email or live chat * Get full tutor support on weekdays (Monday to Friday) COURSE DESIGN The course is delivered through our online learning platform, accessible through any internet-connected device. There are no formal deadlines or teaching schedules, meaning you are free to study the course at your own pace. You are taught through a combination of * Video lessons * Online study materials CERTIFICATION Upon successful completion of the course, you will be able to obtain your course completion e-certificate free of cost. Print copy by post is also available at an additional cost of £9.99 and PDF Certificate at £4.99. WHO IS THIS COURSE FOR: The course is ideal for those who already work in this sector or are an aspiring professional. This course is designed to enhance your expertise and boost your CV. Learn key skills and gain a professional qualification to prove your newly-acquired knowledge. REQUIREMENTS: The online training is open to all students and has no formal entry requirements. To study the SQL NoSQL Big Data and Hadoop, all your need is a passion for learning, a good understanding of English, numeracy, and IT skills. You must also be over the age of 16. COURSE CONTENT Section 01: Introduction Introduction 00:07:00 Building a Data-driven Organization - Introduction 00:04:00 Data Engineering 00:06:00 Learning Environment & Course Material 00:04:00 Movielens Dataset 00:03:00 Section 02: Relational Database Systems Introduction to Relational Databases 00:09:00 SQL 00:05:00 Movielens Relational Model 00:15:00 Movielens Relational Model: Normalization vs Denormalization 00:16:00 MySQL 00:05:00 Movielens in MySQL: Database import 00:06:00 OLTP in RDBMS: CRUD Applications 00:17:00 Indexes 00:16:00 Data Warehousing 00:15:00 Analytical Processing 00:17:00 Transaction Logs 00:06:00 Relational Databases - Wrap Up 00:03:00 Section 03: Database Classification Distributed Databases 00:07:00 CAP Theorem 00:10:00 BASE 00:07:00 Other Classifications 00:07:00 Section 04: Key-Value Store Introduction to KV Stores 00:02:00 Redis 00:04:00 Install Redis 00:07:00 Time Complexity of Algorithm 00:05:00 Data Structures in Redis : Key & String 00:20:00 Data Structures in Redis II : Hash & List 00:18:00 Data structures in Redis III : Set & Sorted Set 00:21:00 Data structures in Redis IV : Geo & HyperLogLog 00:11:00 Data structures in Redis V : Pubsub & Transaction 00:08:00 Modelling Movielens in Redis 00:11:00 Redis Example in Application 00:29:00 KV Stores: Wrap Up 00:02:00 Section 05: Document-Oriented Databases Introduction to Document-Oriented Databases 00:05:00 MongoDB 00:04:00 MongoDB Installation 00:02:00 Movielens in MongoDB 00:13:00 Movielens in MongoDB: Normalization vs Denormalization 00:11:00 Movielens in MongoDB: Implementation 00:10:00 CRUD Operations in MongoDB 00:13:00 Indexes 00:16:00 MongoDB Aggregation Query - MapReduce function 00:09:00 MongoDB Aggregation Query - Aggregation Framework 00:16:00 Demo: MySQL vs MongoDB. Modeling with Spark 00:02:00 Document Stores: Wrap Up 00:03:00 Section 06: Search Engines Introduction to Search Engine Stores 00:05:00 Elasticsearch 00:09:00 Basic Terms Concepts and Description 00:13:00 Movielens in Elastisearch 00:12:00 CRUD in Elasticsearch 00:15:00 Search Queries in Elasticsearch 00:23:00 Aggregation Queries in Elasticsearch 00:23:00 The Elastic Stack (ELK) 00:12:00 Use case: UFO Sighting in ElasticSearch 00:29:00 Search Engines: Wrap Up 00:04:00 Section 07: Wide Column Store Introduction to Columnar databases 00:06:00 HBase 00:07:00 HBase Architecture 00:09:00 HBase Installation 00:09:00 Apache Zookeeper 00:06:00 Movielens Data in HBase 00:17:00 Performing CRUD in HBase 00:24:00 SQL on HBase - Apache Phoenix 00:14:00 SQL on HBase - Apache Phoenix - Movielens 00:10:00 Demo : GeoLife GPS Trajectories 00:02:00 Wide Column Store: Wrap Up 00:04:00 Section 08: Time Series Databases Introduction to Time Series 00:09:00 InfluxDB 00:03:00 InfluxDB Installation 00:07:00 InfluxDB Data Model 00:07:00 Data manipulation in InfluxDB 00:17:00 TICK Stack I 00:12:00 TICK Stack II 00:23:00 Time Series Databases: Wrap Up 00:04:00 Section 09: Graph Databases Introduction to Graph Databases 00:05:00 Modelling in Graph 00:14:00 Modelling Movielens as a Graph 00:10:00 Neo4J 00:04:00 Neo4J installation 00:08:00 Cypher 00:12:00 Cypher II 00:19:00 Movielens in Neo4J: Data Import 00:17:00 Movielens in Neo4J: Spring Application 00:12:00 Data Analysis in Graph Databases 00:05:00 Examples of Graph Algorithms in Neo4J 00:18:00 Graph Databases: Wrap Up 00:07:00 Section 10: Hadoop Platform Introduction to Big Data With Apache Hadoop 00:06:00 Big Data Storage in Hadoop (HDFS) 00:16:00 Big Data Processing : YARN 00:11:00 Installation 00:13:00 Data Processing in Hadoop (MapReduce) 00:14:00 Examples in MapReduce 00:25:00 Data Processing in Hadoop (Pig) 00:12:00 Examples in Pig 00:21:00 Data Processing in Hadoop (Spark) 00:23:00 Examples in Spark 00:23:00 Data Analytics with Apache Spark 00:09:00 Data Compression 00:06:00 Data serialization and storage formats 00:20:00 Hadoop: Wrap Up 00:07:00 Section 11: Big Data SQL Engines Introduction Big Data SQL Engines 00:03:00 Apache Hive 00:10:00 Apache Hive : Demonstration 00:20:00 MPP SQL-on-Hadoop: Introduction 00:03:00 Impala 00:06:00 Impala : Demonstration 00:18:00 PrestoDB 00:13:00 PrestoDB : Demonstration 00:14:00 SQL-on-Hadoop: Wrap Up 00:02:00 Section 12: Distributed Commit Log Data Architectures 00:05:00 Introduction to Distributed Commit Logs 00:07:00 Apache Kafka 00:03:00 Confluent Platform Installation 00:10:00 Data Modeling in Kafka I 00:13:00 Data Modeling in Kafka II 00:15:00 Data Generation for Testing 00:09:00 Use case: Toll fee Collection 00:04:00 Stream processing 00:11:00 Stream Processing II with Stream + Connect APIs 00:19:00 Example: Kafka Streams 00:15:00 KSQL : Streaming Processing in SQL 00:04:00 KSQL: Example 00:14:00 Demonstration: NYC Taxi and Fares 00:01:00 Streaming: Wrap Up 00:02:00 Section 13: Summary Database Polyglot 00:04:00 Extending your knowledge 00:08:00 Data Visualization 00:11:00 Building a Data-driven Organization - Conclusion 00:07:00 Conclusion 00:03:00 Resources Resources - SQL NoSQL Big Data And Hadoop 00:00:00

SQL NoSQL Big Data and Hadoop
Delivered Online On Demand
£25

SQL NoSQL Big Data and Hadoop

By Apex Learning

OVERVIEW This comprehensive course on SQL NoSQL Big Data and Hadoop will deepen your understanding on this topic. After successful completion of this course you can acquire the required skills in this sector. This SQL NoSQL Big Data and Hadoop comes with accredited certification from CPD, which will enhance your CV and make you worthy in the job market. So enrol in this course today to fast track your career ladder. HOW WILL I GET MY CERTIFICATE? At the end of the course there will be an online written test, which you can take either during or after the course. After successfully completing the test you will be able to order your certificate, these are included in the price. WHO IS THIS COURSE FOR? There is no experience or previous qualifications required for enrolment on this SQL NoSQL Big Data and Hadoop. It is available to all students, of all academic backgrounds. REQUIREMENTS Our SQL NoSQL Big Data and Hadoop is fully compatible with PC's, Mac's, Laptop, Tablet and Smartphone devices. This course has been designed to be fully compatible with tablets and smartphones so you can access your course on Wi-Fi, 3G or 4G. There is no time limit for completing this course, it can be studied in your own time at your own pace. CAREER PATH Learning this new skill will help you to advance in your career. It will diversify your job options and help you develop new techniques to keep up with the fast-changing world. This skillset will help you to- * Open doors of opportunities * Increase your adaptability * Keep you relevant * Boost confidence And much more! COURSE CURRICULUM 14 sections • 130 lectures • 22:34:00 total length •Introduction: 00:07:00 •Building a Data-driven Organization - Introduction: 00:04:00 •Data Engineering: 00:06:00 •Learning Environment & Course Material: 00:04:00 •Movielens Dataset: 00:03:00 •Introduction to Relational Databases: 00:09:00 •SQL: 00:05:00 •Movielens Relational Model: 00:15:00 •Movielens Relational Model: Normalization vs Denormalization: 00:16:00 •MySQL: 00:05:00 •Movielens in MySQL: Database import: 00:06:00 •OLTP in RDBMS: CRUD Applications: 00:17:00 •Indexes: 00:16:00 •Data Warehousing: 00:15:00 •Analytical Processing: 00:17:00 •Transaction Logs: 00:06:00 •Relational Databases - Wrap Up: 00:03:00 •Distributed Databases: 00:07:00 •CAP Theorem: 00:10:00 •BASE: 00:07:00 •Other Classifications: 00:07:00 •Introduction to KV Stores: 00:02:00 •Redis: 00:04:00 •Install Redis: 00:07:00 •Time Complexity of Algorithm: 00:05:00 •Data Structures in Redis : Key & String: 00:20:00 •Data Structures in Redis II : Hash & List: 00:18:00 •Data structures in Redis III : Set & Sorted Set: 00:21:00 •Data structures in Redis IV : Geo & HyperLogLog: 00:11:00 •Data structures in Redis V : Pubsub & Transaction: 00:08:00 •Modelling Movielens in Redis: 00:11:00 •Redis Example in Application: 00:29:00 •KV Stores: Wrap Up: 00:02:00 •Introduction to Document-Oriented Databases: 00:05:00 •MongoDB: 00:04:00 •MongoDB Installation: 00:02:00 •Movielens in MongoDB: 00:13:00 •Movielens in MongoDB: Normalization vs Denormalization: 00:11:00 •Movielens in MongoDB: Implementation: 00:10:00 •CRUD Operations in MongoDB: 00:13:00 •Indexes: 00:16:00 •MongoDB Aggregation Query - MapReduce function: 00:09:00 •MongoDB Aggregation Query - Aggregation Framework: 00:16:00 •Demo: MySQL vs MongoDB. Modeling with Spark: 00:02:00 •Document Stores: Wrap Up: 00:03:00 •Introduction to Search Engine Stores: 00:05:00 •Elasticsearch: 00:09:00 •Basic Terms Concepts and Description: 00:13:00 •Movielens in Elastisearch: 00:12:00 •CRUD in Elasticsearch: 00:15:00 •Search Queries in Elasticsearch: 00:23:00 •Aggregation Queries in Elasticsearch: 00:23:00 •The Elastic Stack (ELK): 00:12:00 •Use case: UFO Sighting in ElasticSearch: 00:29:00 •Search Engines: Wrap Up: 00:04:00 •Introduction to Columnar databases: 00:06:00 •HBase: 00:07:00 •HBase Architecture: 00:09:00 •HBase Installation: 00:09:00 •Apache Zookeeper: 00:06:00 •Movielens Data in HBase: 00:17:00 •Performing CRUD in HBase: 00:24:00 •SQL on HBase - Apache Phoenix: 00:14:00 •SQL on HBase - Apache Phoenix - Movielens: 00:10:00 •Demo : GeoLife GPS Trajectories: 00:02:00 •Wide Column Store: Wrap Up: 00:05:00 •Introduction to Time Series: 00:09:00 •InfluxDB: 00:03:00 •InfluxDB Installation: 00:07:00 •InfluxDB Data Model: 00:07:00 •Data manipulation in InfluxDB: 00:17:00 •TICK Stack I: 00:12:00 •TICK Stack II: 00:23:00 •Time Series Databases: Wrap Up: 00:04:00 •Introduction to Graph Databases: 00:05:00 •Modelling in Graph: 00:14:00 •Modelling Movielens as a Graph: 00:10:00 •Neo4J: 00:04:00 •Neo4J installation: 00:08:00 •Cypher: 00:12:00 •Cypher II: 00:19:00 •Movielens in Neo4J: Data Import: 00:17:00 •Movielens in Neo4J: Spring Application: 00:12:00 •Data Analysis in Graph Databases: 00:05:00 •Examples of Graph Algorithms in Neo4J: 00:18:00 •Graph Databases: Wrap Up: 00:07:00 •Introduction to Big Data With Apache Hadoop: 00:06:00 •Big Data Storage in Hadoop (HDFS): 00:16:00 •Big Data Processing : YARN: 00:11:00 •Installation: 00:13:00 •Data Processing in Hadoop (MapReduce): 00:14:00 •Examples in MapReduce: 00:25:00 •Data Processing in Hadoop (Pig): 00:12:00 •Examples in Pig: 00:21:00 •Data Processing in Hadoop (Spark): 00:23:00 •Examples in Spark: 00:23:00 •Data Analytics with Apache Spark: 00:09:00 •Data Compression: 00:06:00 •Data serialization and storage formats: 00:20:00 •Hadoop: Wrap Up: 00:07:00 •Introduction Big Data SQL Engines: 00:03:00 •Apache Hive: 00:10:00 •Apache Hive : Demonstration: 00:20:00 •MPP SQL-on-Hadoop: Introduction: 00:03:00 •Impala: 00:06:00 •Impala : Demonstration: 00:18:00 •PrestoDB: 00:13:00 •PrestoDB : Demonstration: 00:14:00 •SQL-on-Hadoop: Wrap Up: 00:02:00 •Data Architectures: 00:05:00 •Introduction to Distributed Commit Logs: 00:07:00 •Apache Kafka: 00:03:00 •Confluent Platform Installation: 00:10:00 •Data Modeling in Kafka I: 00:13:00 •Data Modeling in Kafka II: 00:15:00 •Data Generation for Testing: 00:09:00 •Use case: Toll fee Collection: 00:04:00 •Stream processing: 00:11:00 •Stream Processing II with Stream + Connect APIs: 00:19:00 •Example: Kafka Streams: 00:15:00 •KSQL : Streaming Processing in SQL: 00:04:00 •KSQL: Example: 00:14:00 •Demonstration: NYC Taxi and Fares: 00:01:00 •Streaming: Wrap Up: 00:02:00 •Database Polyglot: 00:04:00 •Extending your knowledge: 00:08:00 •Data Visualization: 00:11:00 •Building a Data-driven Organization - Conclusion: 00:07:00 •Conclusion: 00:03:00 •Assignment -SQL NoSQL Big Data and Hadoop: 00:00:00

SQL NoSQL Big Data and Hadoop
Delivered Online On Demand
£12

Introduction to Database Design

By iStudy UK

This Introduction to Database Design course teaches the basis of relational database design and explains how to make a good database design and become an expert on it. Designing a database is quite simple, but you've to understand a few rules before jump onto it. It is essential to know about these rules, otherwise, you will tend to make errors. If you find it hard to deal with database, scripts and all the technical parts, then this comprehensive course is just for you. The course includes the followings: * Creating database and database users * Introduction of data, different data types, foreign key constraints, and other relevant concepts used to create the best database * Importing database tables * Explore all types of relationships * Designing all types of relationships within a database. * Learn three common forms of database normalisation * Application of database for the business purpose * And much more... WHAT WILL I LEARN? * What is a database * Understand different types of databases * Understand normalization * Assign relationships * Eliminate repetition * Relate tables with IDs * Design rules REQUIREMENTS * Basic Microsoft Windows training or equivalent experience WHO IS THE TARGET AUDIENCE? * Students just getting started with designing databases and those who have been designing databases but looking for tips on more effective design Introduction Introduction FREE 00:03:00 Database Basics Overview 00:01:00 What is a Database? 00:03:00 Different Types of Databases 00:12:00 The Process of Database Design 00:08:00 Normalizing Overview - Normalizing 00:01:00 What is Normalization? 00:02:00 Basic Steps to Normalization 00:05:00 A. Brainstorm 00:01:00 B. Organize 00:03:00 C. Eliminate Repetition 00:12:00 D. Assign Relationships 00:01:00 D1. One to One Relationship 00:06:00 D2. One to Many Relationship 00:04:00 D3. Many to Many Relationship 00:06:00 Relating Tables with IDs 00:02:00 Examples of Bad Design 00:15:00 Examples of Good Design 00:06:00 Design Rules 00:05:00 Conclusion Conclusion 00:04:00 Course Certification

Introduction to Database Design
Delivered Online On Demand
£25

Building Data Analytics Solutions Using Amazon Redshift

By Nexus Human

Duration 1 Days 6 CPD hours This course is intended for This course is intended for data warehouse engineers, data platform engineers, and architects and operators who build and manage data analytics pipelines. Completed either AWS Technical Essentials or Architecting on AWS Completed Building Data Lakes on AWS Overview In this course, you will learn to: Compare the features and benefits of data warehouses, data lakes, and modern data architectures Design and implement a data warehouse analytics solution Identify and apply appropriate techniques, including compression, to optimize data storage Select and deploy appropriate options to ingest, transform, and store data Choose the appropriate instance and node types, clusters, auto scaling, and network topology for a particular business use case Understand how data storage and processing affect the analysis and visualization mechanisms needed to gain actionable business insights Secure data at rest and in transit Monitor analytics workloads to identify and remediate problems Apply cost management best practices In this course, you will build a data analytics solution using Amazon Redshift, a cloud data warehouse service. The course focuses on the data collection, ingestion, cataloging, storage, and processing components of the analytics pipeline. You will learn to integrate Amazon Redshift with a data lake to support both analytics and machine learning workloads. You will also learn to apply security, performance, and cost management best practices to the operation of Amazon Redshift. MODULE A: OVERVIEW OF DATA ANALYTICS AND THE DATA PIPELINE * Data analytics use cases * Using the data pipeline for analytics MODULE 1: USING AMAZON REDSHIFT IN THE DATA ANALYTICS PIPELINE * Why Amazon Redshift for data warehousing? * Overview of Amazon Redshift MODULE 2: INTRODUCTION TO AMAZON REDSHIFT * Amazon Redshift architecture Interactive Demo 1: Touring the Amazon Redshift console Amazon Redshift features Practice Lab 1: Load and query data in an Amazon Redshift cluster MODULE 3: INGESTION AND STORAGE * Ingestion * Interactive Demo 2: Connecting your Amazon Redshift cluster using a Jupyter notebook with Data API * Data distribution and storage * Interactive Demo 3: Analyzing semi-structured data using the SUPER data type * Querying data in Amazon Redshift * Practice Lab 2: Data analytics using Amazon Redshift Spectrum MODULE 4: PROCESSING AND OPTIMIZING DATA * Data transformation * Advanced querying * Practice Lab 3: Data transformation and querying in Amazon Redshift * Resource management * Interactive Demo 4: Applying mixed workload management on Amazon Redshift * Automation and optimization * Interactive demo 5: Amazon Redshift cluster resizing from the dc2.large to ra3.xlplus cluster MODULE 5: SECURITY AND MONITORING OF AMAZON REDSHIFT CLUSTERS * Securing the Amazon Redshift cluster * Monitoring and troubleshooting Amazon Redshift clusters MODULE 6: DESIGNING DATA WAREHOUSE ANALYTICS SOLUTIONS * Data warehouse use case review * Activity: Designing a data warehouse analytics workflow MODULE B: DEVELOPING MODERN DATA ARCHITECTURES ON AWS * Modern data architectures

Building Data Analytics Solutions Using Amazon Redshift
Delivered on-request, onlineDelivered Online
Price on Enquiry

Building Batch Data Analytics Solutions on AWS

By Nexus Human

Duration 1 Days 6 CPD hours This course is intended for This course is intended for: Data platform engineers Architects and operators who build and manage data analytics pipelines Overview In this course, you will learn to: Compare the features and benefits of data warehouses, data lakes, and modern data architectures Design and implement a batch data analytics solution Identify and apply appropriate techniques, including compression, to optimize data storage Select and deploy appropriate options to ingest, transform, and store data Choose the appropriate instance and node types, clusters, auto scaling, and network topology for a particular business use case Understand how data storage and processing affect the analysis and visualization mechanisms needed to gain actionable business insights Secure data at rest and in transit Monitor analytics workloads to identify and remediate problems Apply cost management best practices In this course, you will learn to build batch data analytics solutions using Amazon EMR, an enterprise-grade Apache Spark and Apache Hadoop managed service. You will learn how Amazon EMR integrates with open-source projects such as Apache Hive, Hue, and HBase, and with AWS services such as AWS Glue and AWS Lake Formation. The course addresses data collection, ingestion, cataloging, storage, and processing components in the context of Spark and Hadoop. You will learn to use EMR Notebooks to support both analytics and machine learning workloads. You will also learn to apply security, performance, and cost management best practices to the operation of Amazon EMR. MODULE A: OVERVIEW OF DATA ANALYTICS AND THE DATA PIPELINE * Data analytics use cases * Using the data pipeline for analytics MODULE 1: INTRODUCTION TO AMAZON EMR * Using Amazon EMR in analytics solutions * Amazon EMR cluster architecture * Interactive Demo 1: Launching an Amazon EMR cluster * Cost management strategies MODULE 2: DATA ANALYTICS PIPELINE USING AMAZON EMR: INGESTION AND STORAGE * Storage optimization with Amazon EMR * Data ingestion techniques MODULE 3: HIGH-PERFORMANCE BATCH DATA ANALYTICS USING APACHE SPARK ON AMAZON EMR * Apache Spark on Amazon EMR use cases * Why Apache Spark on Amazon EMR * Spark concepts * Interactive Demo 2: Connect to an EMR cluster and perform Scala commands using the Spark shell * Transformation, processing, and analytics * Using notebooks with Amazon EMR * Practice Lab 1: Low-latency data analytics using Apache Spark on Amazon EMR MODULE 4: PROCESSING AND ANALYZING BATCH DATA WITH AMAZON EMR AND APACHE HIVE * Using Amazon EMR with Hive to process batch data * Transformation, processing, and analytics * Practice Lab 2: Batch data processing using Amazon EMR with Hive * Introduction to Apache HBase on Amazon EMR MODULE 5: SERVERLESS DATA PROCESSING * Serverless data processing, transformation, and analytics * Using AWS Glue with Amazon EMR workloads * Practice Lab 3: Orchestrate data processing in Spark using AWS Step Functions MODULE 6: SECURITY AND MONITORING OF AMAZON EMR CLUSTERS * Securing EMR clusters * Interactive Demo 3: Client-side encryption with EMRFS * Monitoring and troubleshooting Amazon EMR clusters * Demo: Reviewing Apache Spark cluster history MODULE 7: DESIGNING BATCH DATA ANALYTICS SOLUTIONS * Batch data analytics use cases * Activity: Designing a batch data analytics workflow MODULE B: DEVELOPING MODERN DATA ARCHITECTURES ON AWS * Modern data architectures

Building Batch Data Analytics Solutions on AWS
Delivered on-request, onlineDelivered Online
Price on Enquiry

Working with Apache Kafka (for Developers) (TTDS6760)

By Nexus Human

Duration 2 Days 12 CPD hours This course is intended for This in an Introductory and beyond level course is geared for experienced Java developers seeking to be proficient in Apache Kafka. Attendees should be experienced developers who are comfortable with Java, and have reasonable experience working with databases. Overview Working in a hands-on learning environment, students will explore Overview of Streaming technologies Kafka concepts and architecture Programming using Kafka API Kafka Streams Monitoring Kafka Tuning / Troubleshooting Kafka Apache Kafka is a real-time data pipeline processor. It high-scalability, fault tolerance, execution speed, and fluid integrations are some of the key hallmarks that make it an integral part of many Enterprise Data architectures. In this lab intensive two day course, students will learn how to use Kafka to build streaming solutions. INTRODUCTION TO STREAMING SYSTEMS * Fast data * Streaming architecture * Lambda architecture * Message queues * Streaming processors INTRODUCTION TO KAFKA * Architecture * Comparing Kafka with other queue systems (JMS / MQ) * Kaka concepts : Messages, Topics, Partitions, Brokers, Producers, commit logs * Kafka & Zookeeper * Producing messages * Consuming messages (Consumers, Consumer Groups) * Message retention * Scaling Kafka PROGRAMMING WITH KAFKA * Configuration parameters * Producer API (Sending messages to Kafka) * Consumer API (consuming messages from Kafka) * Commits , Offsets, Seeking * Schema with Avro KAFKA STREAMS * Streams overview and architecture * Streams use cases and comparison with other platforms * Learning Kafka Streaming concepts (KStream, KTable, KStore) * KStreaming operations (transformations, filters, joins, aggregations) ADMINISTERING KAFKA * Hardware / Software requirements * Deploying Kafka * Configuration of brokers / topics / partitions / producers / consumers * Security: How secure Kafka cluster, and secure client communications (SASL, Kerberos) * Monitoring : monitoring tools * Capacity Planning : estimating usage and demand * Trouble shooting : failure scenarios and recovery MONITORING AND INSTRUMENTING KAFKA * Monitoring Kafka * Instrumenting with Metrics library * Instrument Kafka applications and monitor their performance

Working with Apache Kafka (for Developers) (TTDS6760)
Delivered on-request, onlineDelivered Online
Price on Enquiry

Enterprise Architecture (3- Days)

By Nexus Human

Duration 3 Days 18 CPD hours

Enterprise Architecture (3- Days)
Delivered on-request, onlineDelivered Online
Price on Enquiry

Enterprise Architecture (5 - Days)

By Nexus Human

Duration 5 Days 30 CPD hours

Enterprise Architecture (5 - Days)
Delivered on-request, onlineDelivered Online
Price on Enquiry

Enterprise Architecture

By Nexus Human

Duration 1 Days 6 CPD hours

Enterprise Architecture
Delivered on-request, onlineDelivered Online
Price on Enquiry

Designing and Building Big Data Applications

By Nexus Human

Duration 4 Days 24 CPD hours This course is intended for This course is best suited to developers, engineers, and architects who want to use use Hadoop and related tools to solve real-world problems. Overview Skills learned in this course include:Creating a data set with Kite SDKDeveloping custom Flume components for data ingestionManaging a multi-stage workflow with OozieAnalyzing data with CrunchWriting user-defined functions for Hive and ImpalaWriting user-defined functions for Hive and ImpalaIndexing data with Cloudera Search Cloudera University?s four-day course for designing and building Big Data applications prepares you to analyze and solve real-world problems using Apache Hadoop and associated tools in the enterprise data hub (EDH). INTRODUCTION APPLICATION ARCHITECTURE * Scenario Explanation * Understanding the Development Environment * Identifying and Collecting Input Data * Selecting Tools for Data Processing and Analysis * Presenting Results to the Use DEFINING & USING DATASETS * Metadata Management * What is Apache Avro? * Avro Schemas * Avro Schema Evolution * Selecting a File Format * Performance Considerations USING THE KITE SDK DATA MODULE * What is the Kite SDK? * Fundamental Data Module Concepts * Creating New Data Sets Using the Kite SDK * Loading, Accessing, and Deleting a Data Set IMPORTING RELATIONAL DATA WITH APACHE SQOOP * What is Apache Sqoop? * Basic Imports * Limiting Results * Improving Sqoop?s Performance * Sqoop 2 CAPTURING DATA WITH APACHE FLUME * What is Apache Flume? * Basic Flume Architecture * Flume Sources * Flume Sinks * Flume Configuration * Logging Application Events to Hadoop DEVELOPING CUSTOM FLUME COMPONENTS * Flume Data Flow and Common Extension Points * Custom Flume Sources * Developing a Flume Pollable Source * Developing a Flume Event-Driven Source * Custom Flume Interceptors * Developing a Header-Modifying Flume Interceptor * Developing a Filtering Flume Interceptor * Writing Avro Objects with a Custom Flume Interceptor MANAGING WORKFLOWS WITH APACHE OOZIE * The Need for Workflow Management * What is Apache Oozie? * Defining an Oozie Workflow * Validation, Packaging, and Deployment * Running and Tracking Workflows Using the CLI * Hue UI for Oozie PROCESSING DATA PIPELINES WITH APACHE CRUNCH * What is Apache Crunch? * Understanding the Crunch Pipeline * Comparing Crunch to Java MapReduce * Working with Crunch Projects * Reading and Writing Data in Crunch * Data Collection API Functions * Utility Classes in the Crunch API WORKING WITH TABLES IN APACHE HIVE * What is Apache Hive? * Accessing Hive * Basic Query Syntax * Creating and Populating Hive Tables * How Hive Reads Data * Using the RegexSerDe in Hive DEVELOPING USER-DEFINED FUNCTIONS * What are User-Defined Functions? * Implementing a User-Defined Function * Deploying Custom Libraries in Hive * Registering a User-Defined Function in Hive EXECUTING INTERACTIVE QUERIES WITH IMPALA * What is Impala? * Comparing Hive to Impala * Running Queries in Impala * Support for User-Defined Functions * Data and Metadata Management UNDERSTANDING CLOUDERA SEARCH * What is Cloudera Search? * Search Architecture * Supported Document Formats INDEXING DATA WITH CLOUDERA SEARCH * Collection and Schema Management * Morphlines * Indexing Data in Batch Mode * Indexing Data in Near Real Time PRESENTING RESULTS TO USERS * Solr Query Syntax * Building a Search UI with Hue * Accessing Impala through JDBC * Powering a Custom Web Application with Impala and Search

Designing and Building Big Data Applications
Delivered on-request, onlineDelivered Online
Price on Enquiry

Educators matching "Data Architecture"

Show all 4
Spyglass Education

spyglass education

London

As schools around the world respond during COVID-19, the need for remote learning tools has never been more urgent. To make the transition to distance learning easier, we've built solutions on Azure that innovate the way educators deliver coursework. With services like Teams, Azure WVD, Development & Test Labs and Lab Services, we can tailor a solution to fit your curriculum needs, work towards a connected remote learning experience and help alleviate some of the stress in a changing education environment. Interactive Learning Students today are intuitively digital. They have grown up with almost instant access to important information and a world of connected electronics at their fingertips. However, successfully executing Distance Learning is still proving to be a big challenge for many schools. We can help with our solutions that builds on the Microsoft Azure Bot Services, Conversational AI Services, and Augmented Platforms to design an interactive learning solution that aids students on their terms, at their convenience, and keeps them engaged. Education Quality Analytics Today’s complex world puts significant pressure on universities, from acquiring funding to attracting quality staff and students. People turn to data to tackle these demands, but millions of student records, disparate data sources, and lacking institutional standards hinder their efforts. To quickly execute, analyze, and react to these challenges, companies need a simplistic and intuitive view of their data. Spyglass MTG leverages an integrated data platform on Azure. We source data from everywhere whether data resides in extractable flat files, database silos, or anywhere else, with Power BI and the Azure Data Platform to build a customized set of performance metrics and a “visual storyboard” universities can leverage to take action.

iStudy UK

istudy uk

We understand more than anyone how important it is for you to get the right qualifications at the right time. We also understand that when you have a lot to do, you might not always have the time free to go to a place of study. That is why we created this site, so you can take all the time you need to learn more about your chosen topics as well as be able to do the whole thing from home. We believe in empowering people by offering them opportunities to expand and extend their knowledge and skill set as well as giving them the support they need to achieve great things. With thousands of courses available and a team who will do just about anything to help you, it is safe to say that you will not find a better course provider on the internet and so we would love to work with you to make sure that you get the best experience and best results out of your education. WHAT WE DO Here at iStudy we provide a range of online training courses for those who want to gain new skills, and qualifications and update their knowledge. iStudy training courses are delivered entirely online through our sophisticated student learning portal. The student learning portal is an online learning management system that allows students to study for their courses online. This innovative technology means there is no need to attend any classes or take time off work to study. Professionally accredited courses All our courses are delivered in partnership with nationally recognised awarding bodies so be assured that what you learn with us matters when it comes to achieving your career goals. Study that suits you You do not have to give up your job or your life to get a new qualification, you can learn anytime, anywhere.