Hi Divya In general you will get better performance if you can minimize your use of UDFs. Spark 2.0/ tungsten does a lot of code generation. It will have to treat your UDF as a block box. Andy From: Rishabh Bhardwaj Date: Wednesday, July 20, 2016 at 4:22 AM To: Rabin Banerjee Cc: Divya Gehlot , "user @spark" Subject: Re: write and call UDF in spark dataframe > Hi Divya, > > There is already "from_unixtime" exists in org.apache.spark.sql.frunctions, > Rabin has used that in the sql query,if you want to use it in dataframe DSL > you can try like this, > >> val new_df = df.select(from_unixtime($"time").as("newtime")) > > Thanks, > Rishabh. > > On Wed, Jul 20, 2016 at 4:21 PM, Rabin Banerjee > wrote: >> Hi Divya , >> >> Try, >> >> val df = sqlContext.sql("select from_unixtime(ts,'YYYY-MM-dd') as `ts` from >> mr") >> Regards, >> Rabin >> >> On Wed, Jul 20, 2016 at 12:44 PM, Divya Gehlot >> wrote: >>> Hi, >>> Could somebody share example of writing and calling udf which converts unix >>> tme stamp to date tiime . >>> >>> >>> Thanks, >>> Divya >> >