Manipulating big data distributed over a cluster using functional concepts is rampant in industry, and is arguably one of the first widespread industrial uses of functional ideas. This is evidenced by the popularity of MapReduce and Hadoop, and most recently Apache Spark, a fast, in-memory distributed collections framework written in Scala. In this course, we'll see how the data parallel paradigm can be extended to the distributed case, using Spark throughout. We'll cover Spark's programming model in detail, being careful to understand how and when it differs from familiar programming models, like shared-memory parallel collections or sequential Scala collections. Through hands-on examples in Spark and Scala, we'll learn when important issues related to distribution like latency and network communication should be considered and how they can be addressed effectively for improved performance.
이 강좌에 대하여
귀하가 습득할 기술
- 5 stars73.02%
- 4 stars21.13%
- 3 stars4.36%
- 2 stars0.66%
- 1 star0.81%
BIG DATA ANALYSIS WITH SCALA AND SPARK의 최상위 리뷰
Awesome course and awesome teacher! Nevertheless, to grasp the most of this course, you should do the previous 3 courses of the "Functional Programming in Scala" specialization.
some of the questions are unnecessarily specific (i.e. needs to be rounded to 1 decimal and sorted exactly for it to work)
but otherwise, great lecturer and great content
Good overview of the subject, covering all important aspects. Assignments were well prepared, with a couple of unclear points that were quickly discovered and explained on the forums.
goot as introduction about spark and big data.
Small notice: it is incorrect to compare performance hadoop and spark. As I understand, spark was expected to be compacred with MapReduce.
Functional Programming in Scala 특화 과정 정보
자주 묻는 질문
강의 및 과제를 언제 이용할 수 있게 되나요?
이 전문 분야를 구독하면 무엇을 이용할 수 있나요?
재정 지원을 받을 수 있나요?
궁금한 점이 더 있으신가요? 학습자 도움말 센터를 방문해 보세요.