What Is Spatial Computing?
November 25, 2024
Article · 8 min read
Cultivate your career with expert-led programs, job-ready certificates, and 10,000 ways to grow. All for $25/month, billed annually. Save now
Springboard your Big Data career. Master fundamentals of NoSQL, Big Data, and Apache Spark with hands-on job-ready skills in machine learning and data engineering.
Instructors: IBM Skills Network Team
11,680 already enrolled
Included with
(191 reviews)
Recommended experience
Beginner level
The courses in the specialization require that you have basic computer and data literacy skills, as well as some programming background with languages such as with Python and SQL. No prior knowledge or experience of Big Data and NoSQL is required.
(191 reviews)
Recommended experience
Beginner level
The courses in the specialization require that you have basic computer and data literacy skills, as well as some programming background with languages such as with Python and SQL. No prior knowledge or experience of Big Data and NoSQL is required.
Work with NoSQL databases to insert, update, delete, query, index, aggregate, and shard/partition data.
Develop hands-on NoSQL experience working with MongoDB, Apache Cassandra, and IBM Cloudant.
Develop foundational knowledge of Big Data and gain hands-on lab experience using Apache Hadoop, MapReduce, Apache Spark, Spark SQL, and Kubernetes.
Perform Extract, Transform and Load (ETL) processing and Machine Learning model training and deployment with Apache Spark.
Add to your LinkedIn profile
Add this credential to your LinkedIn profile, resume, or CV
Share it on social media and in your performance review
Big Data Engineers and professionals with NoSQL skills are highly sought after in the data management industry. This Specialization is designed for those seeking to develop fundamental skills for working with Big Data, Apache Spark, and NoSQL databases. Three information-packed courses cover popular NoSQL databases like MongoDB and Apache Cassandra, the widely used Apache Hadoop ecosystem of Big Data tools, as well as Apache Spark analytics engine for large-scale data processing.
You start with an overview of various categories of NoSQL (Not only SQL) data repositories, and then work hands-on with several of them including IBM Cloudant, MonogoDB and Cassandra. You’ll perform various data management tasks, such as creating & replicating databases, inserting, updating, deleting, querying, indexing, aggregating & sharding data. Next, you’ll gain fundamental knowledge of Big Data technologies such as Hadoop, MapReduce, HDFS, Hive, and HBase, followed by a more in depth working knowledge of Apache Spark, Spark Dataframes, Spark SQL, PySpark, the Spark Application UI, and scaling Spark with Kubernetes. In the final course, you will learn to work with Spark Structured Streaming Spark ML - for performing Extract, Transform and Load processing (ETL) and machine learning tasks.
This specialization is suitable for beginners in the fields of NoSQL and Big Data – whether you are or preparing to be a Data Engineer, Software Developer, IT Architect, Data Scientist, or IT Manager.
Applied Learning Project
The emphasis in this specialization is on learning by doing. As such, each course includes hands-on labs to practice & apply the NoSQL and Big Data skills you learn during lectures.
In the first course, you will work hands-on with several NoSQL databases- MongoDB, Apache Cassandra, and IBM Cloudant to perform a variety of tasks: creating the database, adding documents, querying data, utilizing the HTTP API, performing Create, Read, Update & Delete (CRUD) operations, limiting & sorting records, indexing, aggregation, replication, using CQL shell, keyspace operations, & other table operations.
In the next course, you’ll launch a Hadoop cluster using Docker and run Map Reduce jobs. You’ll explore working with Spark using Jupyter notebooks on a Python kernel. You’ll build your Spark skills using DataFrames, Spark SQL, and scale your jobs using Kubernetes.
In the final course you will use Spark for ETL processing, and Machine Learning model training and deployment using IBM Watson.
Differentiate among the four main categories of NoSQL repositories.
Describe the characteristics, features, benefits, limitations, and applications of the more popular Big Data processing tools.
Perform common tasks using MongoDB tasks including create, read, update, and delete (CRUD) operations.
Execute keyspace, table, and CRUD operations in Cassandra.
Explain the impact of big data, including use cases, tools, and processing methods.
Describe Apache Hadoop architecture, ecosystem, practices, and user-related applications, including Hive, HDFS, HBase, Spark, and MapReduce.
Apply Spark programming basics, including parallel programming basics for DataFrames, data sets, and Spark SQL.
Use Spark’s RDDs and data sets, optimize Spark SQL using Catalyst and Tungsten, and use Spark’s development and runtime environment options.
Describe ML, explain its role in data engineering, summarize generative AI, discuss Spark's uses, and analyze ML pipelines and model persistence.
Evaluate ML models, distinguish between regression, classification, and clustering models, and compare data engineering pipelines with ML pipelines.
Construct the data analysis processes using Spark SQL, and perform regression, classification, and clustering using SparkML.
Demonstrate connecting to Spark clusters, build ML pipelines, perform feature extraction and transformation, and model persistence.
At IBM, we know how rapidly tech evolves and recognize the crucial need for businesses and professionals to build job-ready, hands-on skills quickly. As a market-leading tech innovator, we’re committed to helping you thrive in this dynamic landscape. Through IBM Skills Network, our expertly designed training programs in AI, software development, cybersecurity, data science, business management, and more, provide the essential skills you need to secure your first job, advance your career, or drive business success. Whether you’re upskilling yourself or your team, our courses, Specializations, and Professional Certificates build the technical expertise that ensures you, and your organization, excel in a competitive world.
Unlimited access to 10,000+ world-class courses, hands-on projects, and job-ready certificate programs - all included in your subscription
Earn a degree from world-class universities - 100% online
Upskill your employees to excel in the digital economy
The specialization requires 36-42 hours of effort to complete. Working 6-8 hours a week, it can be completed within 1-2 months. Working 3-4 hours a week, it can be completed in 4-6 months.
Basic computer literacy, a grounding in IT systems, working experience with one or more Operating Systems, and programming languages such as Python, data literacy skills, some knowledge of SQL, and a willingness to self-learn online. No prior knowledge of Big Data or NoSQL is required.
It is recommended that you complete the courses in the order in which they occur in the Specialization. Course 2 is a pre-requisite for Course 3.
University credit is currently not available for this Specialization.
Upon successful completion of the Specialization, you will have the practical knowledge and experience to start tackling Data Engineering tasks involving NoSQL Databases, Big Data and Apache Spark.
This course is completely online, so there’s no need to show up to a classroom in person. You can access your lectures, readings and assignments anytime and anywhere via the web or your mobile device.
If you subscribed, you get a 7-day free trial during which you can cancel at no penalty. After that, we don’t give refunds, but you can cancel your subscription at any time. See our full refund policy.
Yes! To get started, click the course card that interests you and enroll. You can enroll and complete the course to earn a shareable certificate, or you can audit it to view the course materials for free. When you subscribe to a course that is part of a Specialization, you’re automatically subscribed to the full Specialization. Visit your learner dashboard to track your progress.
Yes. In select learning programs, you can apply for financial aid or a scholarship if you can’t afford the enrollment fee. If fin aid or scholarship is available for your learning program selection, you’ll find a link to apply on the description page.
When you enroll in the course, you get access to all of the courses in the Specialization, and you earn a certificate when you complete the work. If you only want to read and view the course content, you can audit the course for free. If you cannot afford the fee, you can apply for financial aid.
Financial aid available,