>>> 16/06/30 10:44:34 ERROR util.Utils: Uncaught exception in thread stdout writer for python

java.lang.AbstractMethodError: pyspark_cassandra.DeferringRowReader.read(Lcom/datastax/driver/core/Row;Lcom/datastax/spark/connector/CassandraRowMetadata;)Ljava/lang/Object;

>> You are trying to call an abstract method.  Please check the method DeferringRowReader.read

 

Do not know how to fix this issue.

Have seen in many tutorials around the net and those ones made the same calling I am currently doing

 

from pyspark_cassandra import CassandraSparkContext, Row

from pyspark import SparkContext, SparkConf

from pyspark.sql import SQLContext

conf = SparkConf().setAppName("test").setMaster("spark://192.168.23.31:7077").set("spark.cassandra.connection.host", "192.168.23.31")

sc = CassandraSparkContext(conf=conf)

table = sc.cassandraTable("lebara_diameter_codes","nl_lebara_diameter_codes")

food_count = table.select("errorcode2001").groupBy("errorcode2001").count()

food_count.collect()

 

I am really new to this psark thing. Was able to configure it correctly nd now learning the API.

This email is confidential and may be subject to privilege. If you are not the intended recipient, please do not copy or disclose its content but contact the sender immediately upon receipt.