• Professional Development
  • Medicine & Nursing
  • Arts & Crafts
  • Health & Wellbeing
  • Personal Development

36 Apache Spark courses

🔥 Limited Time Offer 🔥

Get a 10% discount on your first order when you use this promo code at checkout: MAY24BAN3X

SQL NoSQL Big Data and Hadoop

4.7(160)

By Janets

Register on the SQL NoSQL Big Data and Hadoop today and build the experience, skills and knowledge you need to enhance your professional development and work towards your dream job. Study this course through online learning and take the first steps towards a long-term career. The course consists of a number of easy to digest, in-depth modules, designed to provide you with a detailed, expert level of knowledge. Learn through a mixture of instructional video lessons and online study materials. Receive online tutor support as you study the course, to ensure you are supported every step of the way. Get a digital certificate as a proof of your course completion. The SQL NoSQL Big Data and Hadoop is incredibly great value and allows you to study at your own pace. Access the course modules from any internet-enabled device, including computers, tablet, and smartphones. The course is designed to increase your employability and equip you with everything you need to be a success. Enrol on the now and start learning instantly! WHAT YOU GET WITH THE SQL NOSQL BIG DATA AND HADOOP * Receive a e-certificate upon successful completion of the course * Get taught by experienced, professional instructors * Study at a time and pace that suits your learning style * Get instant feedback on assessments  * 24/7 help and advice via email or live chat * Get full tutor support on weekdays (Monday to Friday) COURSE DESIGN The course is delivered through our online learning platform, accessible through any internet-connected device. There are no formal deadlines or teaching schedules, meaning you are free to study the course at your own pace. You are taught through a combination of * Video lessons * Online study materials CERTIFICATION Upon successful completion of the course, you will be able to obtain your course completion e-certificate free of cost. Print copy by post is also available at an additional cost of £9.99 and PDF Certificate at £4.99. WHO IS THIS COURSE FOR: The course is ideal for those who already work in this sector or are an aspiring professional. This course is designed to enhance your expertise and boost your CV. Learn key skills and gain a professional qualification to prove your newly-acquired knowledge. REQUIREMENTS: The online training is open to all students and has no formal entry requirements. To study the SQL NoSQL Big Data and Hadoop, all your need is a passion for learning, a good understanding of English, numeracy, and IT skills. You must also be over the age of 16. COURSE CONTENT Section 01: Introduction Introduction 00:07:00 Building a Data-driven Organization - Introduction 00:04:00 Data Engineering 00:06:00 Learning Environment & Course Material 00:04:00 Movielens Dataset 00:03:00 Section 02: Relational Database Systems Introduction to Relational Databases 00:09:00 SQL 00:05:00 Movielens Relational Model 00:15:00 Movielens Relational Model: Normalization vs Denormalization 00:16:00 MySQL 00:05:00 Movielens in MySQL: Database import 00:06:00 OLTP in RDBMS: CRUD Applications 00:17:00 Indexes 00:16:00 Data Warehousing 00:15:00 Analytical Processing 00:17:00 Transaction Logs 00:06:00 Relational Databases - Wrap Up 00:03:00 Section 03: Database Classification Distributed Databases 00:07:00 CAP Theorem 00:10:00 BASE 00:07:00 Other Classifications 00:07:00 Section 04: Key-Value Store Introduction to KV Stores 00:02:00 Redis 00:04:00 Install Redis 00:07:00 Time Complexity of Algorithm 00:05:00 Data Structures in Redis : Key & String 00:20:00 Data Structures in Redis II : Hash & List 00:18:00 Data structures in Redis III : Set & Sorted Set 00:21:00 Data structures in Redis IV : Geo & HyperLogLog 00:11:00 Data structures in Redis V : Pubsub & Transaction 00:08:00 Modelling Movielens in Redis 00:11:00 Redis Example in Application 00:29:00 KV Stores: Wrap Up 00:02:00 Section 05: Document-Oriented Databases Introduction to Document-Oriented Databases 00:05:00 MongoDB 00:04:00 MongoDB Installation 00:02:00 Movielens in MongoDB 00:13:00 Movielens in MongoDB: Normalization vs Denormalization 00:11:00 Movielens in MongoDB: Implementation 00:10:00 CRUD Operations in MongoDB 00:13:00 Indexes 00:16:00 MongoDB Aggregation Query - MapReduce function 00:09:00 MongoDB Aggregation Query - Aggregation Framework 00:16:00 Demo: MySQL vs MongoDB. Modeling with Spark 00:02:00 Document Stores: Wrap Up 00:03:00 Section 06: Search Engines Introduction to Search Engine Stores 00:05:00 Elasticsearch 00:09:00 Basic Terms Concepts and Description 00:13:00 Movielens in Elastisearch 00:12:00 CRUD in Elasticsearch 00:15:00 Search Queries in Elasticsearch 00:23:00 Aggregation Queries in Elasticsearch 00:23:00 The Elastic Stack (ELK) 00:12:00 Use case: UFO Sighting in ElasticSearch 00:29:00 Search Engines: Wrap Up 00:04:00 Section 07: Wide Column Store Introduction to Columnar databases 00:06:00 HBase 00:07:00 HBase Architecture 00:09:00 HBase Installation 00:09:00 Apache Zookeeper 00:06:00 Movielens Data in HBase 00:17:00 Performing CRUD in HBase 00:24:00 SQL on HBase - Apache Phoenix 00:14:00 SQL on HBase - Apache Phoenix - Movielens 00:10:00 Demo : GeoLife GPS Trajectories 00:02:00 Wide Column Store: Wrap Up 00:04:00 Section 08: Time Series Databases Introduction to Time Series 00:09:00 InfluxDB 00:03:00 InfluxDB Installation 00:07:00 InfluxDB Data Model 00:07:00 Data manipulation in InfluxDB 00:17:00 TICK Stack I 00:12:00 TICK Stack II 00:23:00 Time Series Databases: Wrap Up 00:04:00 Section 09: Graph Databases Introduction to Graph Databases 00:05:00 Modelling in Graph 00:14:00 Modelling Movielens as a Graph 00:10:00 Neo4J 00:04:00 Neo4J installation 00:08:00 Cypher 00:12:00 Cypher II 00:19:00 Movielens in Neo4J: Data Import 00:17:00 Movielens in Neo4J: Spring Application 00:12:00 Data Analysis in Graph Databases 00:05:00 Examples of Graph Algorithms in Neo4J 00:18:00 Graph Databases: Wrap Up 00:07:00 Section 10: Hadoop Platform Introduction to Big Data With Apache Hadoop 00:06:00 Big Data Storage in Hadoop (HDFS) 00:16:00 Big Data Processing : YARN 00:11:00 Installation 00:13:00 Data Processing in Hadoop (MapReduce) 00:14:00 Examples in MapReduce 00:25:00 Data Processing in Hadoop (Pig) 00:12:00 Examples in Pig 00:21:00 Data Processing in Hadoop (Spark) 00:23:00 Examples in Spark 00:23:00 Data Analytics with Apache Spark 00:09:00 Data Compression 00:06:00 Data serialization and storage formats 00:20:00 Hadoop: Wrap Up 00:07:00 Section 11: Big Data SQL Engines Introduction Big Data SQL Engines 00:03:00 Apache Hive 00:10:00 Apache Hive : Demonstration 00:20:00 MPP SQL-on-Hadoop: Introduction 00:03:00 Impala 00:06:00 Impala : Demonstration 00:18:00 PrestoDB 00:13:00 PrestoDB : Demonstration 00:14:00 SQL-on-Hadoop: Wrap Up 00:02:00 Section 12: Distributed Commit Log Data Architectures 00:05:00 Introduction to Distributed Commit Logs 00:07:00 Apache Kafka 00:03:00 Confluent Platform Installation 00:10:00 Data Modeling in Kafka I 00:13:00 Data Modeling in Kafka II 00:15:00 Data Generation for Testing 00:09:00 Use case: Toll fee Collection 00:04:00 Stream processing 00:11:00 Stream Processing II with Stream + Connect APIs 00:19:00 Example: Kafka Streams 00:15:00 KSQL : Streaming Processing in SQL 00:04:00 KSQL: Example 00:14:00 Demonstration: NYC Taxi and Fares 00:01:00 Streaming: Wrap Up 00:02:00 Section 13: Summary Database Polyglot 00:04:00 Extending your knowledge 00:08:00 Data Visualization 00:11:00 Building a Data-driven Organization - Conclusion 00:07:00 Conclusion 00:03:00 Resources Resources - SQL NoSQL Big Data And Hadoop 00:00:00

SQL NoSQL Big Data and Hadoop
Delivered Online On Demand
£25

Apache Kafka Series - Learn Apache Kafka for Beginners v3

By Packt

A beginner-level course that follows a step-by-step approach to learning the fundamentals and core concepts of Apache Kafka 3.0. You will work through interesting activities such as programming a Twitter producer and Elasticsearch consumer to understand the various concepts.

Apache Kafka Series - Learn Apache Kafka for Beginners v3
Delivered Online On Demand
£35.99

Developer Training for Spark and Hadoop

By Nexus Human

Duration 4 Days 24 CPD hours This course is intended for Hadoop Developers Overview Through instructor-led discussion and interactive, hands-on exercises, participants will navigate the Hadoop ecosystem, learning topics such as:How data is distributed, stored, and processed in a Hadoop clusterHow to use Sqoop and Flume to ingest dataHow to process distributed data with Apache SparkHow to model structured data as tables in Impala and HiveHow to choose the best data storage format for different data usage patternsBest practices for data storage This training course is the best preparation for the challenges faced by Hadoop developers. Participants will learn to identify which tool is the right one to use in a given situation, and will gain hands-on experience in developing using those tools. COURSE OUTLINE * Introduction * Introduction to Hadoop and the Hadoop Ecosystem * Hadoop Architecture and HDFS * Importing Relational Data with Apache Sqoop * Introduction to Impala and Hive * Modeling and Managing Data with Impala and Hive * Data Formats * Data Partitioning * Capturing Data with Apache Flume * Spark Basics * Working with RDDs in Spark * Writing and Deploying Spark Applications * Parallel Programming with Spark * Spark Caching and Persistence * Common Patterns in Spark Data Processing * Spark SQL and DataFrames * Conclusion ADDITIONAL COURSE DETAILS: Nexus Humans Developer Training for Spark and Hadoop training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Developer Training for Spark and Hadoop course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.

Developer Training for Spark and Hadoop
Delivered on-request, onlineDelivered Online
Price on Enquiry

Cloudera Training for Apache HBase

By Nexus Human

Duration 4 Days 24 CPD hours This course is intended for This course is appropriate for developers and administrators who intend to use HBase. Overview Skills learned on the course include:The use cases and usage occasions for HBase, Hadoop, and RDBMSUsing the HBase shell to directly manipulate HBase tablesDesigning optimal HBase schemas for efficient data storage and recoveryHow to connect to HBase using the Java API, configure the HBase cluster, and administer an HBase clusterBest practices for identifying and resolving performance bottlenecks Cloudera University?s four-day training course for Apache HBase enables participants to store and access massive quantities of multi-structured data and perform hundreds of thousands of operations per second. INTRODUCTION TO HADOOP & HBASE * What Is Big Data? * Introducing Hadoop * Hadoop Components * What Is HBase? * Why Use HBase? * Strengths of HBase * HBase in Production * Weaknesses of HBase HBASE TABLES * HBase Concepts * HBase Table Fundamentals * Thinking About Table Design THE HBASE SHELL * Creating Tables with the HBase Shell * Working with Tables * Working with Table Data HBASE ARCHITECTURE FUNDAMENTALS * HBase Regions * HBase Cluster Architecture * HBase and HDFS Data Locality HBASE SCHEMA DESIGN * General Design Considerations * Application-Centric Design * Designing HBase Row Keys * Other HBase Table Features BASIC DATA ACCESS WITH THE HBASE API * Options to Access HBase Data * Creating and Deleting HBase Tables * Retrieving Data with Get * Retrieving Data with Scan * Inserting and Updating Data * Deleting Data MORE ADVANCED HBASE API FEATURES * Filtering Scans * Best Practices * HBase Coprocessors HBASE ON THE CLUSTER * How HBase Uses HDFS * Compactions and Splits HBASE READS & WRITES * How HBase Writes Data * How HBase Reads Data * Block Caches for Reading HBASE PERFORMANCE TUNING * Column Family Considerations * Schema Design Considerations * Configuring for Caching * Dealing with Time Series and Sequential Data * Pre-Splitting Regions HBASE ADMINISTRATION AND CLUSTER MANAGEMENT * HBase Daemons * ZooKeeper Considerations * HBase High Availability * Using the HBase Balancer * Fixing Tables with hbck * HBase Security HBASE REPLICATION & BACKUP * HBase Replication * HBase Backup * MapReduce and HBase Clusters USING HIVE & IMPALA WITH HBASE * Using Hive and Impala with HBase APPENDIX A: ACCESSING DATA WITH PYTHON AND THRIFT * Thrift Usage * Working with Tables * Getting and Putting Data * Scanning Data * Deleting Data * Counters * Filters APPENDIX B: OPENTSDB

Cloudera Training for Apache HBase
Delivered on-request, onlineDelivered Online
Price on Enquiry

Designing and Building Big Data Applications

By Nexus Human

Duration 4 Days 24 CPD hours This course is intended for This course is best suited to developers, engineers, and architects who want to use use Hadoop and related tools to solve real-world problems. Overview Skills learned in this course include:Creating a data set with Kite SDKDeveloping custom Flume components for data ingestionManaging a multi-stage workflow with OozieAnalyzing data with CrunchWriting user-defined functions for Hive and ImpalaWriting user-defined functions for Hive and ImpalaIndexing data with Cloudera Search Cloudera University?s four-day course for designing and building Big Data applications prepares you to analyze and solve real-world problems using Apache Hadoop and associated tools in the enterprise data hub (EDH). INTRODUCTION APPLICATION ARCHITECTURE * Scenario Explanation * Understanding the Development Environment * Identifying and Collecting Input Data * Selecting Tools for Data Processing and Analysis * Presenting Results to the Use DEFINING & USING DATASETS * Metadata Management * What is Apache Avro? * Avro Schemas * Avro Schema Evolution * Selecting a File Format * Performance Considerations USING THE KITE SDK DATA MODULE * What is the Kite SDK? * Fundamental Data Module Concepts * Creating New Data Sets Using the Kite SDK * Loading, Accessing, and Deleting a Data Set IMPORTING RELATIONAL DATA WITH APACHE SQOOP * What is Apache Sqoop? * Basic Imports * Limiting Results * Improving Sqoop?s Performance * Sqoop 2 CAPTURING DATA WITH APACHE FLUME * What is Apache Flume? * Basic Flume Architecture * Flume Sources * Flume Sinks * Flume Configuration * Logging Application Events to Hadoop DEVELOPING CUSTOM FLUME COMPONENTS * Flume Data Flow and Common Extension Points * Custom Flume Sources * Developing a Flume Pollable Source * Developing a Flume Event-Driven Source * Custom Flume Interceptors * Developing a Header-Modifying Flume Interceptor * Developing a Filtering Flume Interceptor * Writing Avro Objects with a Custom Flume Interceptor MANAGING WORKFLOWS WITH APACHE OOZIE * The Need for Workflow Management * What is Apache Oozie? * Defining an Oozie Workflow * Validation, Packaging, and Deployment * Running and Tracking Workflows Using the CLI * Hue UI for Oozie PROCESSING DATA PIPELINES WITH APACHE CRUNCH * What is Apache Crunch? * Understanding the Crunch Pipeline * Comparing Crunch to Java MapReduce * Working with Crunch Projects * Reading and Writing Data in Crunch * Data Collection API Functions * Utility Classes in the Crunch API WORKING WITH TABLES IN APACHE HIVE * What is Apache Hive? * Accessing Hive * Basic Query Syntax * Creating and Populating Hive Tables * How Hive Reads Data * Using the RegexSerDe in Hive DEVELOPING USER-DEFINED FUNCTIONS * What are User-Defined Functions? * Implementing a User-Defined Function * Deploying Custom Libraries in Hive * Registering a User-Defined Function in Hive EXECUTING INTERACTIVE QUERIES WITH IMPALA * What is Impala? * Comparing Hive to Impala * Running Queries in Impala * Support for User-Defined Functions * Data and Metadata Management UNDERSTANDING CLOUDERA SEARCH * What is Cloudera Search? * Search Architecture * Supported Document Formats INDEXING DATA WITH CLOUDERA SEARCH * Collection and Schema Management * Morphlines * Indexing Data in Batch Mode * Indexing Data in Near Real Time PRESENTING RESULTS TO USERS * Solr Query Syntax * Building a Search UI with Hue * Accessing Impala through JDBC * Powering a Custom Web Application with Impala and Search

Designing and Building Big Data Applications
Delivered on-request, onlineDelivered Online
Price on Enquiry

Cloudera Introduction to Machine Learning with Spark ML and MLlib

By Nexus Human

Duration 1 Days 6 CPD hours This course is intended for Software Engineers Overview The objective of this course is to learn the key language concepts to machine learning, Spark MLlib, and Spark ML. This course will teach you the key language concepts to machine learning, Spark MLlib, and Spark ML. The course includes coverage of collaborative filtering, clustering, classification, algorithms, and data volume. This course will teach you the key language concepts to machine learning, Spark MLlib, and Spark ML. The course includes coverage of collaborative filtering, clustering, classification, algorithms, and data volume.

Cloudera Introduction to Machine Learning with Spark ML and MLlib
Delivered on-request, onlineDelivered Online
Price on Enquiry

Educators matching "Apache Spark"

Show all 5
Nobleprog Pakistan

nobleprog pakistan

NobleProg is an international training and consultancy group, delivering high quality courses to every sector, covering: Artificial Intelligence, IT, Management, Applied Statistics. Over the last 17 years, we have trained more than 50,000 people from over 6000 companies and organisations. Our courses include classroom (both public and closed) and instructor-led online giving you choice and flexibility to suit your time, budget and level of expertise. We practice what we preach – we use a great deal of the technologies and methods that we teach, and continuously upgrade and improve our courses, keeping up to date with all the latest developments. Our trainers are hand picked and have been through rigorous checks and interviews, and all courses are evaluated by delegates ensuring continuous feedback and improvement. NobleProg in numbers 17 + years of experience 15 + offices all over the world 1000 + trainers cooperating with NobleProg 1400 + course outlines offered companies 6100 + companies that entrusted us satisfied participant 58 k. + satisfied participants NobleProg - The World’s Local Training Provider Our mission is to provide comprehensive training and consultancy solutions all over the world, in an effective and accessible way, tailored to consumers’ needs . We offer practical, real-world knowledge supported by a full understanding of the theory. Our expert trainers are skilled in the latest knowledge transfer techniques, blending presentation, demonstration and hands-on learning. We understand that our learners are excited to be gaining new skills and we thrive off that energy to deliver exceptional training events. Investing in upskilling or reskilling with NobleProg means you stay ahead. Our catalogue is constantly evolving and we offer the most in-demand courses, Java, JavaScript, SQL, Visual Basic for Applications (VBA), as well as Apache Spark, OpenStack, TensorFlow, Selenium, Artificial Intelligence, Data Analysis. Our offer consists of more than 1,400 training outlines covering more than 120 technologies. At NobleProg we emphasis a need of not only following the latest technological trends, but also anticipating changes. We focus on delivering professional skills and certifications that will have a real impact. See what sets us apart >> NobleProg's history NobleProg was established in 2005 in Krakow, Poland, and has gradually expanded its operations to other global markets since. In just two years the first international branch was opened in London. The overwhelming potential of NobleProg combined with the rising need for self-development programs, especially in the field of technological skills, prompted the company to change the business model into a franchise. By doing so, in a short period of time the company allowed a number of people passionate about education and new technologies to join the NobleProg Team. With each year the territorial reach of NobleProg was further expanding and we now have offices on every continent. NobleProg is the World's Local Training Provider.