Hi All,
there is information in 1.0.0 Spark's documentation that 
there is an option "--cores" that one can use to set the number of cores
that spark-shell uses on the cluster:

You can also pass an option --cores <numCores> to control the number of cores that spark-shell uses on the cluster.

This option does not seem to work for me.
If run the following command:
./spark-shell --cores 12
I'm keep on getting an error:
bad option: '--cores'

Is there any other way of controlling the total number of cores used by sparkshell?

Thanks,
Marek