• Professional Development
  • Medicine & Nursing
  • Arts & Crafts
  • Health & Wellbeing
  • Personal Development

93 Data Engineering courses

🔥 Limited Time Offer 🔥

Get a 10% discount on your first order when you use this promo code at checkout: MAY24BAN3X

SQL, NoSQL, Big Data and Hadoop Level 4

By Course Cloud

COURSE OVERVIEW The SQL, NoSQL, Big Data and Hadoop Level 4 course is designed to provide aspiring data engineers with the skills to fast track their career. It will introduce you to key database and data engineering concepts, taking you through the different classifications of databases and software. You will learn how to build a data-driven organisation step-by-step, best practices for data analysis, how to use Elasticsearch as a search engine, and much more. This course is open to everyone and can be studied on a part-time or full-time basis. It is ideal for anyone looking to work in this field or gain a better understanding of Hadoop from a database perspective. Fast track your career by enrolling today and learn tips and shortcuts from experienced industry professionals. This best selling SQL, NoSQL, Big Data and Hadoop Level 4 has been developed by industry professionals and has already been completed by hundreds of satisfied students. This in-depth SQL, NoSQL, Big Data and Hadoop Level 4 is suitable for anyone who wants to build their professional skill set and improve their expert knowledge. The SQL, NoSQL, Big Data and Hadoop Level 4 is CPD-accredited, so you can be confident you're completing a quality training course will boost your CV and enhance your career potential. The SQL, NoSQL, Big Data and Hadoop Level 4 is made up of several information-packed modules which break down each topic into bite-sized chunks to ensure you understand and retain everything you learn. After successfully completing the SQL, NoSQL, Big Data and Hadoop Level 4, you will be awarded a certificate of completion as proof of your new skills. If you are looking to pursue a new career and want to build your professional skills to excel in your chosen field, the certificate of completion from the SQL, NoSQL, Big Data and Hadoop Level 4 will help you stand out from the crowd. You can also validate your certification on our website. We know that you are busy and that time is precious, so we have designed the SQL, NoSQL, Big Data and Hadoop Level 4 to be completed at your own pace, whether that's part-time or full-time. Get full course access upon registration and access the course materials from anywhere in the world, at any time, from any internet-enabled device.  Our experienced tutors are here to support you through the entire learning process and answer any queries you may have via email.

SQL, NoSQL, Big Data and Hadoop Level 4
Delivered Online On Demand
£25

Data Science - Time Series Forecasting with Facebook Prophet in Python

By Packt

In this compact intermediate-level course, you will learn how to use Facebook Prophet to do time series analysis and forecasting. You will learn how Prophet works under the hood and the Prophet API. We will apply Prophet to a variety of datasets, including store sales and stock prices.

Data Science - Time Series Forecasting with Facebook Prophet in Python
Delivered Online On Demand
£82.99

MSc Advanced Motorsport Engineering

5.0(1)

By National Motorsport Academy

Study online for the Master’s Advanced Motorsport Engineering and boost your motorsport career. With the ability to fit your studies around your existing career and family, the MSc is flexible and affordable. Start on any date and study when and where suits you!

MSc Advanced Motorsport Engineering
Delivered on-request, onlineDelivered Online
£8950

DP-100T01 Designing and Implementing a Data Science Solution on Azure

By Nexus Human

Duration 4 Days 24 CPD hours This course is intended for This course is designed for data scientists with existing knowledge of Python and machine learning frameworks like Scikit-Learn, PyTorch, and Tensorflow, who want to build and operate machine learning solutions in the cloud. Overview Learn how to operate machine learning solutions at cloud scale using Azure Machine Learning. This course teaches you to leverage your existing knowledge of Python and machine learning to manage data ingestion and preparation, model training and deployment, and machine learning solution monitoring in Microsoft Azure. Learn how to operate machine learning solutions at cloud scale using Azure Machine Learning. This course teaches you to leverage your existing knowledge of Python and machine learning to manage data ingestion and preparation, model training and deployment, and machine learning solution monitoring with Azure Machine Learning and MLflow. Prerequisites Creating cloud resources in Microsoft Azure. Using Python to explore and visualize data. Training and validating machine learning models using common frameworks like Scikit-Learn, PyTorch, and TensorFlow. Working with containers AI-900T00: Microsoft Azure AI Fundamentals is recommended, or the equivalent experience. 1 - DESIGN A DATA INGESTION STRATEGY FOR MACHINE LEARNING PROJECTS * Identify your data source and format * Choose how to serve data to machine learning workflows * Design a data ingestion solution 2 - DESIGN A MACHINE LEARNING MODEL TRAINING SOLUTION * Identify machine learning tasks * Choose a service to train a machine learning model * Decide between compute options 3 - DESIGN A MODEL DEPLOYMENT SOLUTION * Understand how model will be consumed * Decide on real-time or batch deployment 4 - DESIGN A MACHINE LEARNING OPERATIONS SOLUTION * Explore an MLOps architecture * Design for monitoring * Design for retraining 5 - EXPLORE AZURE MACHINE LEARNING WORKSPACE RESOURCES AND ASSETS * Create an Azure Machine Learning workspace * Identify Azure Machine Learning resources * Identify Azure Machine Learning assets * Train models in the workspace 6 - EXPLORE DEVELOPER TOOLS FOR WORKSPACE INTERACTION * Explore the studio * Explore the Python SDK * Explore the CLI 7 - MAKE DATA AVAILABLE IN AZURE MACHINE LEARNING * Understand URIs * Create a datastore * Create a data asset 8 - WORK WITH COMPUTE TARGETS IN AZURE MACHINE LEARNING * Choose the appropriate compute target * Create and use a compute instance * Create and use a compute cluster 9 - WORK WITH ENVIRONMENTS IN AZURE MACHINE LEARNING * Understand environments * Explore and use curated environments * Create and use custom environments 10 - FIND THE BEST CLASSIFICATION MODEL WITH AUTOMATED MACHINE LEARNING * Preprocess data and configure featurization * Run an Automated Machine Learning experiment * Evaluate and compare models 11 - TRACK MODEL TRAINING IN JUPYTER NOTEBOOKS WITH MLFLOW * Configure MLflow for model tracking in notebooks * Train and track models in notebooks 12 - RUN A TRAINING SCRIPT AS A COMMAND JOB IN AZURE MACHINE LEARNING * Convert a notebook to a script * Run a script as a command job * Use parameters in a command job 13 - TRACK MODEL TRAINING WITH MLFLOW IN JOBS * Track metrics with MLflow * View metrics and evaluate models 14 - PERFORM HYPERPARAMETER TUNING WITH AZURE MACHINE LEARNING * Define a search space * Configure a sampling method * Configure early termination * Use a sweep job for hyperparameter tuning 15 - RUN PIPELINES IN AZURE MACHINE LEARNING * Create components * Create a pipeline * Run a pipeline job 16 - REGISTER AN MLFLOW MODEL IN AZURE MACHINE LEARNING * Log models with MLflow * Understand the MLflow model format * Register an MLflow model 17 - CREATE AND EXPLORE THE RESPONSIBLE AI DASHBOARD FOR A MODEL IN AZURE MACHINE LEARNING * Understand Responsible AI * Create the Responsible AI dashboard * Evaluate the Responsible AI dashboard 18 - DEPLOY A MODEL TO A MANAGED ONLINE ENDPOINT * Explore managed online endpoints * Deploy your MLflow model to a managed online endpoint * Deploy a model to a managed online endpoint * Test managed online endpoints 19 - DEPLOY A MODEL TO A BATCH ENDPOINT * Understand and create batch endpoints * Deploy your MLflow model to a batch endpoint * Deploy a custom model to a batch endpoint * Invoke and troubleshoot batch endpoints

DP-100T01 Designing and Implementing a Data Science Solution on Azure
Delivered Online5 days, Sept 30th, 13:00 + 1 more
£1785

Data Science Prerequisites - NumPy, Matplotlib, and Pandas in Python

By Packt

This course equips learners with a comprehensive understanding of the NumPy stack, including NumPy, Matplotlib, Pandas, and SciPy, to effectively tackle common challenges in deep learning and data science. Master the basics with this carefully structured course.

Data Science Prerequisites - NumPy, Matplotlib, and Pandas in Python
Delivered Online On Demand
£82.99

Preparing for the Professional Data Engineer Examination

By Nexus Human

Duration 1 Days 6 CPD hours This course is intended for This course is intended for the following participants:Cloud professionals interested in taking the Data Engineer certification exam.Data engineering professionals interested in taking the Data Engineer certification exam. Overview This course teaches participants the following skills: Position the Professional Data Engineer Certification Provide information, tips, and advice on taking the exam Review the sample case studies Review each section of the exam covering highest-level concepts sufficient to build confidence in what is known by the candidate and indicate skill gaps/areas of study if not known by the candidate Connect candidates to appropriate target learning This course will help prospective candidates plan their preparation for the Professional Data Engineer exam. The session will cover the structure and format of the examination, as well as its relationship to other Google Cloud certifications. Through lectures, quizzes, and discussions, candidates will familiarize themselves with the domain covered by the examination, to help them devise a preparation strategy. Rehearse useful skills including exam question reasoning and case comprehension. Tips and review of topics from the Data Engineering curriculum. UNDERSTANDING THE PROFESSIONAL DATA ENGINEER CERTIFICATION * Position the Professional Data Engineer certification among the offerings * Distinguish between Associate and Professional * Provide guidance between Professional Data Engineer and Associate Cloud Engineer * Describe how the exam is administered and the exam rules * Provide general advice about taking the exam SAMPLE CASE STUDIES FOR THE PROFESSIONAL DATA ENGINEER EXAM * Flowlogistic * MJTelco * DESIGNING AND BUILDING (REVIEW AND PREPARATION TIPS) * Designing data processing systems * Designing flexible data representations * Designing data pipelines * Designing data processing infrastructure * Build and maintain data structures and databases * Building and maintaining flexible data representations * Building and maintaining pipelines * Building and maintaining processing infrastructure * ANALYZING AND MODELING (REVIEW AND PREPARATION TIPS) * Analyze data and enable machine learning * Analyzing data * Machine learning * Machine learning model deployment * Model business processes for analysis and optimization * Mapping business requirements to data representations * Optimizing data representations, data infrastructure performance and cost * RELIABILITY, POLICY, AND SECURITY (REVIEW AND PREPARATION TIPS) * Design for reliability * Performing quality control * Assessing, troubleshooting, and improving data representation and data processing infrastructure * Recovering data * Visualize data and advocate policy * Building (or selecting) data visualization and reporting tools * Advocating policies and publishing data and reports * Design for security and compliance * Designing secure data infrastructure and processes * Designing for legal compliance * RESOURCES AND NEXT STEPS * Resources for learning more about designing data processing systems, data structures, and databases * Resources for learning more about data analysis, machine learning, business process analysis, and optimization * Resources for learning more about data visualization and policy Resources for learning more about reliability design * Resources for learning more about business process analysis and optimization * Resources for learning more about reliability, policies, security, and compliance ADDITIONAL COURSE DETAILS: Nexus Humans Preparing for the Professional Data Engineer Examination training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Preparing for the Professional Data Engineer Examination course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.

Preparing for the Professional Data Engineer Examination
Delivered on-request, onlineDelivered Online
Price on Enquiry

Advanced Data Analysis and Reconciliation

4.3(6)

By dbrownconsulting

Advanced Data Analysis and Reconciliation
Delivered Online3 weeks, Oct 22nd, 08:00
£900

SQL NoSQL Big Data and Hadoop

By Apex Learning

OVERVIEW This comprehensive course on SQL NoSQL Big Data and Hadoop will deepen your understanding on this topic. After successful completion of this course you can acquire the required skills in this sector. This SQL NoSQL Big Data and Hadoop comes with accredited certification from CPD, which will enhance your CV and make you worthy in the job market. So enrol in this course today to fast track your career ladder. HOW WILL I GET MY CERTIFICATE? At the end of the course there will be an online written test, which you can take either during or after the course. After successfully completing the test you will be able to order your certificate, these are included in the price. WHO IS THIS COURSE FOR? There is no experience or previous qualifications required for enrolment on this SQL NoSQL Big Data and Hadoop. It is available to all students, of all academic backgrounds. REQUIREMENTS Our SQL NoSQL Big Data and Hadoop is fully compatible with PC's, Mac's, Laptop, Tablet and Smartphone devices. This course has been designed to be fully compatible with tablets and smartphones so you can access your course on Wi-Fi, 3G or 4G. There is no time limit for completing this course, it can be studied in your own time at your own pace. CAREER PATH Learning this new skill will help you to advance in your career. It will diversify your job options and help you develop new techniques to keep up with the fast-changing world. This skillset will help you to- * Open doors of opportunities * Increase your adaptability * Keep you relevant * Boost confidence And much more! COURSE CURRICULUM 14 sections • 130 lectures • 22:34:00 total length •Introduction: 00:07:00 •Building a Data-driven Organization - Introduction: 00:04:00 •Data Engineering: 00:06:00 •Learning Environment & Course Material: 00:04:00 •Movielens Dataset: 00:03:00 •Introduction to Relational Databases: 00:09:00 •SQL: 00:05:00 •Movielens Relational Model: 00:15:00 •Movielens Relational Model: Normalization vs Denormalization: 00:16:00 •MySQL: 00:05:00 •Movielens in MySQL: Database import: 00:06:00 •OLTP in RDBMS: CRUD Applications: 00:17:00 •Indexes: 00:16:00 •Data Warehousing: 00:15:00 •Analytical Processing: 00:17:00 •Transaction Logs: 00:06:00 •Relational Databases - Wrap Up: 00:03:00 •Distributed Databases: 00:07:00 •CAP Theorem: 00:10:00 •BASE: 00:07:00 •Other Classifications: 00:07:00 •Introduction to KV Stores: 00:02:00 •Redis: 00:04:00 •Install Redis: 00:07:00 •Time Complexity of Algorithm: 00:05:00 •Data Structures in Redis : Key & String: 00:20:00 •Data Structures in Redis II : Hash & List: 00:18:00 •Data structures in Redis III : Set & Sorted Set: 00:21:00 •Data structures in Redis IV : Geo & HyperLogLog: 00:11:00 •Data structures in Redis V : Pubsub & Transaction: 00:08:00 •Modelling Movielens in Redis: 00:11:00 •Redis Example in Application: 00:29:00 •KV Stores: Wrap Up: 00:02:00 •Introduction to Document-Oriented Databases: 00:05:00 •MongoDB: 00:04:00 •MongoDB Installation: 00:02:00 •Movielens in MongoDB: 00:13:00 •Movielens in MongoDB: Normalization vs Denormalization: 00:11:00 •Movielens in MongoDB: Implementation: 00:10:00 •CRUD Operations in MongoDB: 00:13:00 •Indexes: 00:16:00 •MongoDB Aggregation Query - MapReduce function: 00:09:00 •MongoDB Aggregation Query - Aggregation Framework: 00:16:00 •Demo: MySQL vs MongoDB. Modeling with Spark: 00:02:00 •Document Stores: Wrap Up: 00:03:00 •Introduction to Search Engine Stores: 00:05:00 •Elasticsearch: 00:09:00 •Basic Terms Concepts and Description: 00:13:00 •Movielens in Elastisearch: 00:12:00 •CRUD in Elasticsearch: 00:15:00 •Search Queries in Elasticsearch: 00:23:00 •Aggregation Queries in Elasticsearch: 00:23:00 •The Elastic Stack (ELK): 00:12:00 •Use case: UFO Sighting in ElasticSearch: 00:29:00 •Search Engines: Wrap Up: 00:04:00 •Introduction to Columnar databases: 00:06:00 •HBase: 00:07:00 •HBase Architecture: 00:09:00 •HBase Installation: 00:09:00 •Apache Zookeeper: 00:06:00 •Movielens Data in HBase: 00:17:00 •Performing CRUD in HBase: 00:24:00 •SQL on HBase - Apache Phoenix: 00:14:00 •SQL on HBase - Apache Phoenix - Movielens: 00:10:00 •Demo : GeoLife GPS Trajectories: 00:02:00 •Wide Column Store: Wrap Up: 00:05:00 •Introduction to Time Series: 00:09:00 •InfluxDB: 00:03:00 •InfluxDB Installation: 00:07:00 •InfluxDB Data Model: 00:07:00 •Data manipulation in InfluxDB: 00:17:00 •TICK Stack I: 00:12:00 •TICK Stack II: 00:23:00 •Time Series Databases: Wrap Up: 00:04:00 •Introduction to Graph Databases: 00:05:00 •Modelling in Graph: 00:14:00 •Modelling Movielens as a Graph: 00:10:00 •Neo4J: 00:04:00 •Neo4J installation: 00:08:00 •Cypher: 00:12:00 •Cypher II: 00:19:00 •Movielens in Neo4J: Data Import: 00:17:00 •Movielens in Neo4J: Spring Application: 00:12:00 •Data Analysis in Graph Databases: 00:05:00 •Examples of Graph Algorithms in Neo4J: 00:18:00 •Graph Databases: Wrap Up: 00:07:00 •Introduction to Big Data With Apache Hadoop: 00:06:00 •Big Data Storage in Hadoop (HDFS): 00:16:00 •Big Data Processing : YARN: 00:11:00 •Installation: 00:13:00 •Data Processing in Hadoop (MapReduce): 00:14:00 •Examples in MapReduce: 00:25:00 •Data Processing in Hadoop (Pig): 00:12:00 •Examples in Pig: 00:21:00 •Data Processing in Hadoop (Spark): 00:23:00 •Examples in Spark: 00:23:00 •Data Analytics with Apache Spark: 00:09:00 •Data Compression: 00:06:00 •Data serialization and storage formats: 00:20:00 •Hadoop: Wrap Up: 00:07:00 •Introduction Big Data SQL Engines: 00:03:00 •Apache Hive: 00:10:00 •Apache Hive : Demonstration: 00:20:00 •MPP SQL-on-Hadoop: Introduction: 00:03:00 •Impala: 00:06:00 •Impala : Demonstration: 00:18:00 •PrestoDB: 00:13:00 •PrestoDB : Demonstration: 00:14:00 •SQL-on-Hadoop: Wrap Up: 00:02:00 •Data Architectures: 00:05:00 •Introduction to Distributed Commit Logs: 00:07:00 •Apache Kafka: 00:03:00 •Confluent Platform Installation: 00:10:00 •Data Modeling in Kafka I: 00:13:00 •Data Modeling in Kafka II: 00:15:00 •Data Generation for Testing: 00:09:00 •Use case: Toll fee Collection: 00:04:00 •Stream processing: 00:11:00 •Stream Processing II with Stream + Connect APIs: 00:19:00 •Example: Kafka Streams: 00:15:00 •KSQL : Streaming Processing in SQL: 00:04:00 •KSQL: Example: 00:14:00 •Demonstration: NYC Taxi and Fares: 00:01:00 •Streaming: Wrap Up: 00:02:00 •Database Polyglot: 00:04:00 •Extending your knowledge: 00:08:00 •Data Visualization: 00:11:00 •Building a Data-driven Organization - Conclusion: 00:07:00 •Conclusion: 00:03:00 •Assignment -SQL NoSQL Big Data and Hadoop: 00:00:00

SQL NoSQL Big Data and Hadoop
Delivered Online On Demand
£12

Snowflake - Build and Architect Data Pipelines Using AWS

By Packt

The course helps you learn Snowflake from scratch and explore a few of its important features. You will build automated pipelines with Snowflake and use the AWS cloud with Snowflake as a data warehouse. You will also explore Snowpark to be worked on the data pipelines.

Snowflake - Build and Architect Data Pipelines Using AWS
Delivered Online On Demand
£52.99

Data Science Model Deployments and Cloud Computing on GCP

By Packt

Are you interested in learning and deploying applications at scale using Google Cloud platform? Do you lack hands-on exposure when it comes to deploying applications and seeing them in action? Then this course is for you. You will also learn microservices and event-driven architectures with real-world use case implementations.

Data Science Model Deployments and Cloud Computing on GCP
Delivered Online On Demand
£56.99