申请试用
HOT
登录
注册
 
Apache Spark Core—Deep Dive—Proper Optimization (continues)

Apache Spark Core—Deep Dive—Proper Optimization (continues)

Spark开源社区
/
发布于
/
7752
人观看
Optimizing spark jobs through a true understanding of spark core. Learn: What is a partition? What is the difference between read/shuffle/write partitions? How to increase parallelism and decrease output files? Where does shuffle data go between stages? What is the “right” size for your spark partitions and files? Why does a job slow down with only a few tasks left and never finish? Why doesn’t adding nodes decrease my compute time?
0 点赞
2 收藏
40下载
确认
3秒后跳转登录页面
去登陆