[ https://issues.apache.org/jira/browse/SPARK-5137?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14269263#comment-14269263
]
Al M edited comment on SPARK-5137 at 1/8/15 12:30 PM:
------------------------------------------------------
That's right. _emphasis_a_emphasis_ has 11 partitions and _emphasis_b_emphasis_ has a lot
more. I can see why you wouldn't want to force a shuffle on _emphasis_a_emphasis_ since that's
unnecessary processing.
Thanks for your detailed explanation and quick response. I'll close this since I agree that
it behaves correctly.
was (Author: alrocks47):
That's right. {code}a{code} has 11 partitions and {code}b{code} has a lot more. I can see
why you wouldn't want to force a shuffle on {code}a{code} since that's unnecessary processing.
Thanks for your detailed explanation and quick response. I'll close this since I agree that
it behaves correctly.
> subtract does not take the spark.default.parallelism into account
> -----------------------------------------------------------------
>
> Key: SPARK-5137
> URL: https://issues.apache.org/jira/browse/SPARK-5137
> Project: Spark
> Issue Type: Bug
> Affects Versions: 1.2.0
> Environment: CENTOS 6; scala
> Reporter: Al M
> Priority: Trivial
>
> The 'subtract' function (PairRDDFunctions.scala) in scala does not use the default parallelism
value set in the config (spark.default.parallelism). This is easy enough to work around.
I can just load the property and pass it in as an argument.
> It would be great if subtract used the default value, just like all the other PairRDDFunctions.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org
|