spark-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Som Lima <somplastic...@gmail.com>
Subject Re: Scala version compatibility
Date Mon, 06 Apr 2020 20:10:53 GMT
Those who followed best practices in software development
would start with a clean environment I.e.  installation of
operating system. Then install development tools keeping a  record of
version numbers. So that  at the time of deployment unforeseen errors are
avoided by duplicating  development environment , test environment  , final
client runtime environment.

That is now the purpose of docker.

So if you want run same source code or byte code in different environments
then you should run it. If you can programme and ask this question then you
should know how to.








On Mon, 6 Apr 2020, 20:50 Andrew Melo, <andrew.melo@gmail.com> wrote:

> Hello all,
>
> I'm aware that Scala is not binary compatible between revisions. I have
> some Java code whose only Scala dependency is the transitive dependency
> through Spark. This code calls a Spark API which returns a Seq<Path>, which
> I then convert into a List<Path> with
> JavaConverters.seqAsJavaListConverter. Will this usage cause binary
> incompatibility if the jar is compiled in one Scala version and executed in
> another?
>
> I tried grokking
> https://docs.scala-lang.org/overviews/core/binary-compatibility-of-scala-releases.html,
> and wasn't quite able to make heads or tails of this particular case.
>
> Thanks!
> Andrew
>
>
>

Mime
View raw message