For some reason, I want to create as many partitions as the number of executors * number of tasks. This is what I'll pass to the second parameter of sc.parallelize. Now my question is, can I programmatically get the number of executors and number of tasks per executor in Spark.
            Asked
            
        
        
            Active
            
        
            Viewed 131 times
        
    1
            
            
        
        Cœur
        
- 37,241
 - 25
 - 195
 - 267
 
        MetallicPriest
        
- 29,191
 - 52
 - 200
 - 356