This comprehensive course explores the essential
technologies, architectures, and methodologies for
efficiently processing, storing, and analyzing massive
datasets—from petabytes to exabytes—that challenge
traditional systems. Students will gain hands-on
expertise in designing and implementing robust,
scalable data processing pipelines and advanced
analytics workflows, utilizing industry-leading
distributed computing frameworks like Apache Spark,
Hadoop, and Kafka.
Want to receive push notifications for all major on-site activities?
Want to receive push notifications for all major on-site activities?