WHAT YOU WILL LEARN
Work with NoSQL databases to insert, replace, delete, question, index, mixture, and shard/partition information.
Develop hands-on NoSQL expertise working with MongoDB, Apache Cassandra, and IBM Cloudant.
Develop foundational data of Large Knowledge and achieve hands-on lab expertise utilizing Apache Hadoop, MapReduce, Apache Spark, Spark SQL, and Kubernetes.
Carry out Extract, Remodel and Load (ETL) processing and Machine Studying mannequin coaching and deployment with Apache Spark.
SKILLS YOU WILL GAIN
- Apache Hadoop
- Mongodb
- Large Knowledge
- Apache Spark
- NoSQL Databases
- NoSQL
- Cloud Database
- Cloudant
- Cassandra
- SparkSQL
- SparkML
About this Specialization
Large Knowledge Engineers and professionals with NoSQL abilities are extremely wanted within the information administration business. This Specialization is designed for these looking for to develop basic abilities for working with Large Knowledge, Apache Spark, and NoSQL databases. Three information-packed programs cowl fashionable NoSQL databases like MongoDB and Apache Cassandra, the extensively used Apache Hadoop ecosystem of Large Knowledge instruments, in addition to Apache Spark analytics engine for large-scale information processing.
You begin with an outline of assorted classes of NoSQL (Not solely SQL) information repositories, after which work hands-on with a number of of them together with IBM Cloudant, MonogoDB and Cassandra. You’ll carry out numerous information administration duties, similar to creating & replicating databases, inserting, updating, deleting, querying, indexing, aggregating & sharding information. Subsequent, you’ll achieve basic data of Large Knowledge applied sciences similar to Hadoop, MapReduce, HDFS, Hive, and HBase, adopted by a extra in depth working data of Apache Spark, Spark Dataframes, Spark SQL, PySpark, the Spark Utility UI, and scaling Spark with Kubernetes. Within the remaining course, you’ll study to work with Spark Structured Streaming Spark ML – for performing Extract, Remodel and Load processing (ETL) and machine studying duties.
This specialization is appropriate for freshmen within the fields of NoSQL and Large Knowledge – whether or not you’re or making ready to be a Knowledge Engineer, Software program Developer, IT Architect, Knowledge Scientist, or IT Supervisor.
Utilized Studying Undertaking
The emphasis on this specialization is on studying by doing. As such, every course consists of hands-on labs to apply & apply the NoSQL and Large Knowledge abilities you study throughout lectures.
Within the first course, you’ll work hands-on with a number of NoSQL databases- MongoDB, Apache Cassandra, and IBM Cloudant to carry out a wide range of duties: creating the database, including paperwork, querying information, using the HTTP API, performing Create, Learn, Replace & Delete (CRUD) operations, limiting & sorting data, indexing, aggregation, replication, utilizing CQL shell, keyspace operations, & different desk operations.
Within the subsequent course, you’ll launch a Hadoop cluster utilizing Docker and run Map Cut back jobs. You’ll discover working with Spark utilizing Jupyter notebooks on a Python kernel. You’ll construct your Spark abilities utilizing DataFrames, Spark SQL, and scale your jobs utilizing Kubernetes.
Within the remaining course you’ll use Spark for ETL processing, and Machine Studying mannequin coaching and deployment utilizing IBM Watson.
0 Comments