spark-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Renjith (JIRA)" <j...@apache.org>
Subject [jira] [Updated] (SPARK-23865) Not able to load file from Spark Dataframes
Date Wed, 04 Apr 2018 12:50:00 GMT

     [ https://issues.apache.org/jira/browse/SPARK-23865?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]

Renjith updated SPARK-23865:
----------------------------
    Description: 
Hello,

I am in the phase of learning Spark as part of it trying examples. I am using the following
lines of code as below for my file named df.scala:

import org.apache.spark.sql.SparkSession
 val spark = SparkSession.builder().getOrCreate()
 val df = spark.read.csv("CitiGroup2006_2008")
 df.Head(5)

In my Scala Terminal:

scala> :load df.scala

Loading df.scala...
 import org.apache.spark.sql.SparkSession
 spark: org.apache.spark.sql.SparkSession = org.apache.spark.sql.SparkSession@4756e5cc
 org.apache.spark.sql.AnalysisException: Path does not exist: [file:/C:/Spark/MyPrograms/Scala_and_Spark_Bootcamp_master/SparkD|file:///C:/Spark/MyPrograms/Scala_and_Spark_Bootcamp_master/SparkD]
 ataFrames/CitiGroup2006_2008;
 at org.apache.spark.sql.execution.datasources.DataSource$.org$apache$spark$sql$execution$datasources$DataSource$$checkAndGl
 obPathIfNecessary(DataSource.scala:715)
 at org.apache.spark.sql.execution.datasources.DataSource$$anonfun$15.apply(DataSource.scala:389)
 at org.apache.spark.sql.execution.datasources.DataSource$$anonfun$15.apply(DataSource.scala:389)
 at scala.collection.TraversableLike$$anonfun$flatMap$1.apply(TraversableLike.scala:241)
 at scala.collection.TraversableLike$$anonfun$flatMap$1.apply(TraversableLike.scala:241)
 at scala.collection.immutable.List.foreach(List.scala:381)
 at scala.collection.TraversableLike$class.flatMap(TraversableLike.scala:241)
 at scala.collection.immutable.List.flatMap(List.scala:344)

at org.apache.spark.sql.execution.datasources.DataSource.resolveRelation(DataSource.scala:388)
 at org.apache.spark.sql.DataFrameReader.loadV1Source(DataFrameReader.scala:239)
 at org.apache.spark.sql.DataFrameReader.load(DataFrameReader.scala:227)
 at org.apache.spark.sql.DataFrameReader.csv(DataFrameReader.scala:594)
 at org.apache.spark.sql.DataFrameReader.csv(DataFrameReader.scala:473)
 ... 72 elided
 <console>:25: error: not found: value df
 df.Head(5)
 ^

all environment variables are set and pointed. Is this a version issue of Spark 2.3.0 or should
i degrade the version if so please let me know which version is stable to do my practicals

  was:
Hello,

I am in the phase of learning Spark as part of it trying examples. I am using the following
lines of code as below for my file named df.scala:

import org.apache.spark.sql.SparkSession
val spark = SparkSession.builder().getOrCreate()
val df = spark.read.csv("CitiGroup2006_2008")
df.Head(5)

In my Scala Terminal:

scala> :load df.scala

Loading df.scala...
import org.apache.spark.sql.SparkSession
spark: org.apache.spark.sql.SparkSession = org.apache.spark.sql.SparkSession@4756e5cc
org.apache.spark.sql.AnalysisException: Path does not exist: file:/C:/Spark/MyPrograms/Scala_and_Spark_Bootcamp_master/SparkD
ataFrames/CitiGroup2006_2008;
 at org.apache.spark.sql.execution.datasources.DataSource$.org$apache$spark$sql$execution$datasources$DataSource$$checkAndGl
obPathIfNecessary(DataSource.scala:715)
 at org.apache.spark.sql.execution.datasources.DataSource$$anonfun$15.apply(DataSource.scala:389)
 at org.apache.spark.sql.execution.datasources.DataSource$$anonfun$15.apply(DataSource.scala:389)
 at scala.collection.TraversableLike$$anonfun$flatMap$1.apply(TraversableLike.scala:241)
 at scala.collection.TraversableLike$$anonfun$flatMap$1.apply(TraversableLike.scala:241)
 at scala.collection.immutable.List.foreach(List.scala:381)
 at scala.collection.TraversableLike$class.flatMap(TraversableLike.scala:241)
 at scala.collection.immutable.List.flatMap(List.scala:344)

at org.apache.spark.sql.execution.datasources.DataSource.resolveRelation(DataSource.scala:388)
 at org.apache.spark.sql.DataFrameReader.loadV1Source(DataFrameReader.scala:239)
 at org.apache.spark.sql.DataFrameReader.load(DataFrameReader.scala:227)
 at org.apache.spark.sql.DataFrameReader.csv(DataFrameReader.scala:594)
 at org.apache.spark.sql.DataFrameReader.csv(DataFrameReader.scala:473)
 ... 72 elided
<console>:25: error: not found: value df
 df.Head(5)
 ^

all environment variables are set and pointed. Is this a version of Spark 2.3.0 or should
i dgrade the version if so please let me know which version is stable to do my practicals


> Not able to load file from Spark Dataframes
> -------------------------------------------
>
>                 Key: SPARK-23865
>                 URL: https://issues.apache.org/jira/browse/SPARK-23865
>             Project: Spark
>          Issue Type: Bug
>          Components: Examples
>    Affects Versions: 2.3.0
>         Environment: Executed in Atom Editor.
>            Reporter: Renjith
>            Priority: Major
>              Labels: newbie
>
> Hello,
> I am in the phase of learning Spark as part of it trying examples. I am using the following
lines of code as below for my file named df.scala:
> import org.apache.spark.sql.SparkSession
>  val spark = SparkSession.builder().getOrCreate()
>  val df = spark.read.csv("CitiGroup2006_2008")
>  df.Head(5)
> In my Scala Terminal:
> scala> :load df.scala
> Loading df.scala...
>  import org.apache.spark.sql.SparkSession
>  spark: org.apache.spark.sql.SparkSession = org.apache.spark.sql.SparkSession@4756e5cc
>  org.apache.spark.sql.AnalysisException: Path does not exist: [file:/C:/Spark/MyPrograms/Scala_and_Spark_Bootcamp_master/SparkD|file:///C:/Spark/MyPrograms/Scala_and_Spark_Bootcamp_master/SparkD]
>  ataFrames/CitiGroup2006_2008;
>  at org.apache.spark.sql.execution.datasources.DataSource$.org$apache$spark$sql$execution$datasources$DataSource$$checkAndGl
>  obPathIfNecessary(DataSource.scala:715)
>  at org.apache.spark.sql.execution.datasources.DataSource$$anonfun$15.apply(DataSource.scala:389)
>  at org.apache.spark.sql.execution.datasources.DataSource$$anonfun$15.apply(DataSource.scala:389)
>  at scala.collection.TraversableLike$$anonfun$flatMap$1.apply(TraversableLike.scala:241)
>  at scala.collection.TraversableLike$$anonfun$flatMap$1.apply(TraversableLike.scala:241)
>  at scala.collection.immutable.List.foreach(List.scala:381)
>  at scala.collection.TraversableLike$class.flatMap(TraversableLike.scala:241)
>  at scala.collection.immutable.List.flatMap(List.scala:344)
> at org.apache.spark.sql.execution.datasources.DataSource.resolveRelation(DataSource.scala:388)
>  at org.apache.spark.sql.DataFrameReader.loadV1Source(DataFrameReader.scala:239)
>  at org.apache.spark.sql.DataFrameReader.load(DataFrameReader.scala:227)
>  at org.apache.spark.sql.DataFrameReader.csv(DataFrameReader.scala:594)
>  at org.apache.spark.sql.DataFrameReader.csv(DataFrameReader.scala:473)
>  ... 72 elided
>  <console>:25: error: not found: value df
>  df.Head(5)
>  ^
> all environment variables are set and pointed. Is this a version issue of Spark 2.3.0
or should i degrade the version if so please let me know which version is stable to do my
practicals



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org


Mime
View raw message