mine is version 1.5.2 Ted

Thanks


On 25 March 2016 at 22:45, Ted Yu <yuzhihong@gmail.com> wrote:
Strange: the JIRAs below were marked Fixed in 1.5.0

On Fri, Mar 25, 2016 at 3:43 PM, Mich Talebzadeh <mich.talebzadeh@gmail.com> wrote:
Is this 1.6 Ted?


On 25 March 2016 at 22:40, Ted Yu <yuzhihong@gmail.com> wrote:
Looks like database support was fixed by:

[SPARK-7943] [SPARK-8105] [SPARK-8435] [SPARK-8714] [SPARK-8561] Fixes multi-database support

On Fri, Mar 25, 2016 at 3:35 PM, Ashok Kumar <ashok34668@yahoo.com> wrote:
1.5.2 Ted.

Those two lines I don't know where they come. It finds and gets the table info OK

HTH


On Friday, 25 March 2016, 22:32, Ted Yu <yuzhihong@gmail.com> wrote:


Which release of Spark do you use, Mich ?

In master branch, the message is more accurate (sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/NoSuchItemException.scala):

  override def getMessage: String = s"Table $table not found in database $db"


On Fri, Mar 25, 2016 at 3:21 PM, Mich Talebzadeh <mich.talebzadeh@gmail.com> wrote:
You can use DESCRIBE FORMATTED <DATABASE>.<TABLE_NAME> to get that info.

This is based on the same command in Hive however, it throws two erroneous error lines as shown below (don't see them in Hive DESCRIBE ...)

Example

scala> sql("describe formatted test.t14").collect.foreach(println)
16/03/25 22:32:38 ERROR Hive: Table test not found: test.test table not found
16/03/25 22:32:38 ERROR Hive: Table test not found: test.test table not found
[# col_name             data_type               comment             ]
[                ]
[invoicenumber          int                                         ]
[paymentdate            date                                        ]
[net                    decimal(20,2)                               ]
[vat                    decimal(20,2)                               ]
[total                  decimal(20,2)                               ]
[                ]
[# Detailed Table Information            ]
[Database:              test                     ]
[Owner:                 hduser                   ]
[CreateTime:            Fri Mar 25 22:13:44 GMT 2016     ]
[LastAccessTime:        UNKNOWN                  ]
[Protect Mode:          None                     ]
[Retention:             0                        ]
[Location:              hdfs://rhes564:9000/user/hive/warehouse/test.db/t14      ]
[Table Type:            MANAGED_TABLE            ]
[Table Parameters:               ]
[       COLUMN_STATS_ACCURATE   {\"BASIC_STATS\":\"true\"}]
[       comment                 from csv file from excel sheet]
[       numFiles                2                   ]
[       orc.compress            ZLIB                ]
[       totalSize               1090                ]
[       transient_lastDdlTime   1458944025          ]
[                ]
[# Storage Information           ]
[SerDe Library:         org.apache.hadoop.hive.ql.io.orc.OrcSerde        ]
[InputFormat:           org.apache.hadoop.hive.ql.io.orc.OrcInputFormat  ]
[OutputFormat:          org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat         ]
[Compressed:            No                       ]
[Num Buckets:           -1                       ]
[Bucket Columns:        []                       ]
[Sort Columns:          []                       ]
[Storage Desc Params:            ]
[       serialization.format    1                   ]

HTH


On 25 March 2016 at 22:12, Ashok Kumar <ashok34668@yahoo.com.invalid> wrote:
Experts,

I would like to know when a table was created in Hive database using Spark shell?

Thanks