Mohit, This doesn't seems to be working can you please provide more details? when I use "from pyspark import SparkContext" it is disabled in pycharm. I use pycharm community edition. Where should I set the environment variables in same python script or different python script?Also, should I run any Spark local cluster so Spark program runs on top of that?Appreciate your help-SathishOn Wed, Aug 6, 2014 at 6:22 PM, Mohit Singh <email@example.com> wrote:My naive set up..Addingos.environ['SPARK_HOME'] = "/path/to/spark"sys.path.append("/path/to/spark/python")on top of my script.from pyspark import SparkContextfrom pyspark import SparkConfExecution works from within pycharm...Though my next step is to figure out autocompletion and I bet there are better ways to develop apps for spark..--On Wed, Aug 6, 2014 at 4:16 PM, Sathish Kumaran Vairavelu <firstname.lastname@example.org> wrote:
Hello,I am trying to use the python IDE PyCharm for Spark application development. How can I use pyspark with Python IDE? Can anyone help me with this?ThanksSathish
"When you want success as badly as you want the air, then you will get it. There is no other secret of success."