I seem to be having issues compiling it though...
 def readAsRdd[T: ClassManifest](sc: org.apache.spark.SparkContext, uri:String, clazz:java.lang.Class[_]) = {
    val rdd = sc.sequenceFile[org.apache.hadoop.io.Text, org.apache.hadoop.io.BytesWritable](uri)
      val sz = l._2.getLength
      val b = l._2.getBytes.slice(0,sz)
      val parseFrom = clazz.getMethod("parseFrom",Class.forName("[B"))

[ERROR] /Users/shay/sb/experiment/sps-emr/ue/src/main/scala/ue/proto.scala:57: error: could not find implicit value for parameter kcf: () => org.apache.spark.WritableConverter[org.apache.hadoop.io.Text]
[ERROR] Error occurred in an application involving default arguments.
[INFO]     val rdd = sc.sequenceFile[org.apache.hadoop.io.Text, org.apache.hadoop.io.BytesWritable](uri)

On Fri, Oct 18, 2013 at 9:37 AM, Matei Zaharia <matei.zaharia@gmail.com> wrote:
Don't worry about the implicit params, those are filled in by the compiler. All you need to do is provide a key and value type, and a path. Look at how sequenceFile gets used in this test:


In particular, the K and V in Spark can be any Writable class, *or* primitive types like Int, Double, etc, or String. For the latter ones, Spark automatically uses the correct Hadoop Writable (e.g. IntWritable, DoubleWritable, Text).


On Oct 17, 2013, at 5:35 PM, Shay Seng <shay@1618labs.com> wrote:

Hey gurus,

I'm having a little trouble deciphering the docs for 

sequenceFile[KV](path: StringminSplits: Int = defaultMinSplits)(implicit km: ClassManifest[K]vm: ClassManifest[V]kcf: () ⇒WritableConverter[K]vcf: () ⇒ WritableConverter[V])RDD[(K, V)]

Does anyone have a short example snippet?