drill-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Herman Tan <her...@redcubesg.com>
Subject Re: ERROR is reading parquet data after create table
Date Tue, 02 Oct 2018 08:44:05 GMT
Hi Divya and everyone,

The problem has disappeared.
Drill was not restarted.
This appears to be intermittent.
Before I submitted the error report, I ran the script several times and it
failed all the time.
Today I ran it again and it succeeded.
I will restart and test again.

Regards,
Herman



On Thu, Sep 27, 2018 at 11:50 AM Divya Gehlot <divya.htconex@gmail.com>
wrote:

> Hi Herman,
> Just to ensure that  your parquet file format is not corrupted , Can you
> please query a folder like just 2001 or some of the files underneath
> .Instead of querying the whole data set at once .
>
> Thanks,
> Divya
>
> On Wed, 26 Sep 2018 at 15:35, Herman Tan <herman@redcubesg.com> wrote:
>
> > Hi Kunal,
> >
> > ----
> > That said, could you provide some details about the parquet data you've
> > created, like the schema, parquet version and the tool used to generate.
> > Usually, the schema (and meta) provides most of these details for any
> > parquet file.
> > ----
> >
> > 1. The schema is under dfs.tmp, the queries to generate are all
> documented
> > below.
> > 2. I don't know how to find the parquet version of the data file
> > 3. The tool used to generate the parquest is apache drill.  The CTAS is
> > detailed below.
> >
> > Regards,
> > Herman
> > ____________
> >
> > *This is the Text data*
> >
> > This is the folders of the files
> > Total # of lines about 50 million rows
> > ----------
> > show files from dfs.`D:\retail_sandbox\pos\sales_pos_detail\pos_details_
> > 20180825`
> > ;
> > show files from dfs.`D:\retail_sandbox\pos\sales_pos_detail\pos_details_
> > 20180825\2011`
> > ;
> > -----
> > sales_pos_detail
> >   \pos_details_20180825
> >     \2007
> >     \2008
> >     \2009
> >     \2010
> >     \2011
> >   \pos_details_0.csv
> >   \pos_details_1.csv
> >   \pos_details_2.csv
> >   \pos_details_3.csv
> >   \pos_details_4.csv
> >   \pos_details_5.csv
> >   \pos_details_6.csv
> >   \pos_details_7.csv
> >   \pos_details_8.csv
> >     \2012
> >     \2013
> >     \2014
> >     \2015
> >     \2016
> >     \2017
> >     \2018
> >     \others
> > -----
> >
> > *This is the view with the metadata defined:*
> >
> > create or replace view dfs.tmp.load_pos_sales_detail as
> > SELECT
> > -- dimension keys
> >  cast(dim_date_key as int) dim_date_key
> > ,cast(dim_site_key as int) dim_site_key
> > ,cast(dim_pos_header_key as bigint) dim_pos_header_key
> > ,cast(dim_pos_cashier_key as int) dim_pos_cashier_key
> > ,cast(dim_card_number_key as int) dim_card_number_key
> > ,cast(dim_hour_minute_key as int) dim_hour_minute_key
> > ,cast(dim_pos_clerk_key as int) dim_pos_clerk_key
> > ,cast(dim_product_key as int) dim_product_key
> > ,cast(dim_pos_employee_purchase_key as int) dim_pos_employee_purchase_key
> > ,cast(dim_pos_terminal_key as int) dim_pos_terminal_key
> > ,cast(dim_campaign_key as int) dim_campaign_key
> > ,cast(dim_promo_key as int) dim_promo_key
> > ,cast( case when dim_site_lfl_key = '' then 0 else dim_site_lfl_key end
> as
> > int) dim_site_lfl_key
> > -- derived from keys
> > ,dim_date_str
> > ,`year` as `trx_year`
> > -- Measures
> > ,Product_Sales_Qty
> > ,Product_Sales_Price
> > ,Product_Cost_Price
> > ,Product_Cost_Amt
> > ,Product_Sales_Gross_Amt
> > ,Product_Sales_Promo_Disc_Amt
> > ,Product_Sales_Add_Promo_Disc_Amt
> > ,Product_Sales_Total_Promo_Disc_Amt
> > ,Product_Sales_Retail_Promo_Amt
> > ,Product_Sales_Retail_Amt
> > ,Product_Sales_VAT_Amt
> > ,Product_Sales_Product_Margin_Amt
> > ,Product_Sales_Initial_Margin_Amt
> > from dfs.`D:\retail_sandbox\pos\sales_pos_detail\pos_details_20180825`
> > ;
> >
> >
> > *This is the CTAS that generates the parquet from the view above:*
> >
> > drop table if exists dfs.tmp.load_pos_sales_detail_tbl
> > ;
> >
> > create table dfs.tmp.load_pos_sales_detail_tbl AS
> > SELECT
> > -- dimension keys
> >  dim_date_key
> > ,dim_site_key
> > ,dim_pos_header_key
> > ,dim_pos_cashier_key
> > ,dim_card_number_key
> > ,dim_hour_minute_key
> > ,dim_pos_clerk_key
> > ,dim_product_key
> > ,dim_pos_employee_purchase_key
> > ,dim_pos_terminal_key
> > ,dim_campaign_key
> > ,dim_promo_key
> > ,dim_site_lfl_key
> > -- derived from keys
> > ,dim_date_str
> > ,`trx_year`
> > -- Measures
> > ,Product_Sales_Qty Sales_Qty
> > ,Product_Sales_Price Sales_Price
> > ,Product_Cost_Price Cost_Price
> > ,Product_Cost_Amt Cost_Amt
> > ,Product_Sales_Gross_Amt Sales_Gross_Amt
> > ,Product_Sales_Promo_Disc_Amt Sales_Promo_Disc_Amt
> > ,Product_Sales_Add_Promo_Disc_Amt Add_Promo_Disc_Amt
> > ,Product_Sales_Total_Promo_Disc_Amt Total_Promo_Disc_Amt
> > ,Product_Sales_Retail_Promo_Amt Retail_Promo_Amt
> > ,Product_Sales_Retail_Amt Retail_Amt
> > ,Product_Sales_VAT_Amt VAT_Amt
> > ,Product_Sales_Product_Margin_Amt Product_Margin_Amt
> > ,Product_Sales_Initial_Margin_Amt Initial_Margin_Amt
> > from dfs.tmp.load_pos_sales_detail
> > ;
> >
> >
> > *This is the select query that generated the error:*
> >
> > select *
> > from dfs.tmp.load_pos_sales_detail_tbl
> > ;
> >
> > ----- ERROR ----------------------------
> >
> > SQL Error: RESOURCE ERROR: Waited for 30000 ms, but only 10 tasks for
> > 'Fetch parquet metadata' are complete. Total number of tasks 29,
> > parallelism 16.
> >
> >
> > On Mon, Sep 24, 2018 at 9:08 AM, Kunal Khatua <kunal@apache.org> wrote:
> >
> > > Hi Herman
> > >
> > > Assuming that you're doing analytics on your data. If that's the case,
> > > parquet format is the way to go.
> > >
> > > That said, could you provide some details about the parquet data you've
> > > created, like the schema, parquet version and the tool used to
> generate.
> > > Usually, the schema (and meta) provides most of these details for any
> > > parquet file.
> > >
> > > It'll be useful to know if there is a pattern in the failure because of
> > > which there might be corruption occurring.
> > >
> > > Kunal
> > >
> > >
> > > On 9/22/2018 11:49:36 PM, Herman Tan <herman@redcubesg.com> wrote:
> > > Hi Karthik,
> > >
> > > Thank you for pointing me to the mail archive in May 2018.
> > > That is exactly the same problem I am facing.
> > >
> > > I thought of using Drill as an ETL where I load the warehouse parquet
> > > tables from text source files.
> > > Then I query the parquet tables.
> > > It works on some parquet tables but am having problems with large ones
> > that
> > > consist of several files. (I think)
> > > Still investigating.
> > > Anyone in the community have other experience?
> > > Should I work with all text files instead of parquet?
> > >
> > >
> > > Herman
> > >
> > >
> > > On Fri, Sep 21, 2018 at 2:15 AM, Karthikeyan Manivannan
> > > kmanivannan@mapr.com> wrote:
> > >
> > > > Hi Herman,
> > > >
> > > > I am not sure what the exact problem here is but can you check to see
> > if
> > > > you are not hitting the problem described here:
> > > >
> > > > http://mail-archives.apache.org/mod_mbox/drill-user/201805.mbox/%
> > > > 3CCACwRgneXLXoP2vCYuGsA4Gwd1jGS8F+rcpzQ8rHuatFW5fmRaQ@mail.gmail.com
> > %3E
> > > >
> > > > Thanks
> > > >
> > > > Karthik
> > > >
> > > > On Wed, Sep 19, 2018 at 7:02 PM Herman Tan wrote:
> > > >
> > > > > Hi,
> > > > >
> > > > > I encountered the following error.
> > > > > The Steps I did are as follows:
> > > > > 1. Create a view to fix the data type of fields with cast
> > > > > 2. Create table (parquet) using the view
> > > > > 3. Query select * from table (query a field also does not work)
> > > > >
> > > > > The error:
> > > > > SQL Error: RESOURCE ERROR: Waited for 30000 ms, but only 10 tasks
> for
> > > > > 'Fetch parquet metadata' are complete. Total number of tasks 29,
> > > > > parallelism 16.
> > > > >
> > > > > When I re-run this, the number of tasks will vary.
> > > > >
> > > > > What could be the problem?
> > > > >
> > > > > Regards,
> > > > > Herman Tan
> > > > >
> > > > > More info below:
> > > > >
> > > > > This is the folders of the files
> > > > > Total # of lines, 50 million
> > > > > ----------
> > > > > show files from
> > > > > dfs.`D:\retail_sandbox\pos\sales_pos_detail\pos_details_20180825`
> > > > > ;
> > > > > show files from
> > > > >
> > dfs.`D:\retail_sandbox\pos\sales_pos_detail\pos_details_20180825\2011`
> > > > > ;
> > > > > -----
> > > > > sales_pos_detail
> > > > > \pos_details_20180825
> > > > > \2007
> > > > > \2008
> > > > > \2009
> > > > > \2010
> > > > > \2011
> > > > > \pos_details_0.csv
> > > > > \pos_details_1.csv
> > > > > \pos_details_2.csv
> > > > > \pos_details_3.csv
> > > > > \pos_details_4.csv
> > > > > \pos_details_5.csv
> > > > > \pos_details_6.csv
> > > > > \pos_details_7.csv
> > > > > \pos_details_8.csv
> > > > > \2012
> > > > > \2013
> > > > > \2014
> > > > > \2015
> > > > > \2016
> > > > > \2017
> > > > > \2018
> > > > > \others
> > > > > -----
> > > > >
> > > > > create or replace view dfs.tmp.load_pos_sales_detail as
> > > > > SELECT
> > > > > -- dimension keys
> > > > > cast(dim_date_key as int) dim_date_key
> > > > > ,cast(dim_site_key as int) dim_site_key
> > > > > ,cast(dim_pos_header_key as bigint) dim_pos_header_key
> > > > > ,cast(dim_pos_cashier_key as int) dim_pos_cashier_key
> > > > > ,cast(dim_card_number_key as int) dim_card_number_key
> > > > > ,cast(dim_hour_minute_key as int) dim_hour_minute_key
> > > > > ,cast(dim_pos_clerk_key as int) dim_pos_clerk_key
> > > > > ,cast(dim_product_key as int) dim_product_key
> > > > > ,cast(dim_pos_employee_purchase_key as int)
> > > > dim_pos_employee_purchase_key
> > > > > ,cast(dim_pos_terminal_key as int) dim_pos_terminal_key
> > > > > ,cast(dim_campaign_key as int) dim_campaign_key
> > > > > ,cast(dim_promo_key as int) dim_promo_key
> > > > > ,cast( case when dim_site_lfl_key = '' then 0 else dim_site_lfl_key
> > end
> > > > as
> > > > > int) dim_site_lfl_key
> > > > > -- derived from keys
> > > > > ,dim_date_str
> > > > > ,`year` as `trx_year`
> > > > > -- Measures
> > > > > ,Product_Sales_Qty
> > > > > ,Product_Sales_Price
> > > > > ,Product_Cost_Price
> > > > > ,Product_Cost_Amt
> > > > > ,Product_Sales_Gross_Amt
> > > > > ,Product_Sales_Promo_Disc_Amt
> > > > > ,Product_Sales_Add_Promo_Disc_Amt
> > > > > ,Product_Sales_Total_Promo_Disc_Amt
> > > > > ,Product_Sales_Retail_Promo_Amt
> > > > > ,Product_Sales_Retail_Amt
> > > > > ,Product_Sales_VAT_Amt
> > > > > ,Product_Sales_Product_Margin_Amt
> > > > > ,Product_Sales_Initial_Margin_Amt
> > > > > from
> > dfs.`D:\retail_sandbox\pos\sales_pos_detail\pos_details_20180825`
> > > > > ;
> > > > >
> > > > > drop table if exists dfs.tmp.load_pos_sales_detail_tbl
> > > > > ;
> > > > >
> > > > > create table dfs.tmp.load_pos_sales_detail_tbl AS
> > > > > SELECT
> > > > > -- dimension keys
> > > > > dim_date_key
> > > > > ,dim_site_key
> > > > > ,dim_pos_header_key
> > > > > ,dim_pos_cashier_key
> > > > > ,dim_card_number_key
> > > > > ,dim_hour_minute_key
> > > > > ,dim_pos_clerk_key
> > > > > ,dim_product_key
> > > > > ,dim_pos_employee_purchase_key
> > > > > ,dim_pos_terminal_key
> > > > > ,dim_campaign_key
> > > > > ,dim_promo_key
> > > > > ,dim_site_lfl_key
> > > > > -- derived from keys
> > > > > ,dim_date_str
> > > > > ,`trx_year`
> > > > > -- Measures
> > > > > ,Product_Sales_Qty Sales_Qty
> > > > > ,Product_Sales_Price Sales_Price
> > > > > ,Product_Cost_Price Cost_Price
> > > > > ,Product_Cost_Amt Cost_Amt
> > > > > ,Product_Sales_Gross_Amt Sales_Gross_Amt
> > > > > ,Product_Sales_Promo_Disc_Amt Sales_Promo_Disc_Amt
> > > > > ,Product_Sales_Add_Promo_Disc_Amt Add_Promo_Disc_Amt
> > > > > ,Product_Sales_Total_Promo_Disc_Amt Total_Promo_Disc_Amt
> > > > > ,Product_Sales_Retail_Promo_Amt Retail_Promo_Amt
> > > > > ,Product_Sales_Retail_Amt Retail_Amt
> > > > > ,Product_Sales_VAT_Amt VAT_Amt
> > > > > ,Product_Sales_Product_Margin_Amt Product_Margin_Amt
> > > > > ,Product_Sales_Initial_Margin_Amt Initial_Margin_Amt
> > > > > from dfs.tmp.load_pos_sales_detail
> > > > > ;
> > > > >
> > > > > select *
> > > > > from dfs.tmp.load_pos_sales_detail_tbl
> > > > > ;
> > > > >
> > > > > ----- ERROR ----------------------------
> > > > >
> > > > > SQL Error: RESOURCE ERROR: Waited for 30000 ms, but only 10 tasks
> for
> > > > > 'Fetch parquet metadata' are complete. Total number of tasks 29,
> > > > > parallelism 16.
> > > > >
> > > > >
> > > > > [Error Id: 3b079174-f5d0-4313-8097-25a0b3070854 on
> > > > > IORA-G9KY9P2.stf.nus.edu.sg:31010]
> > > > > RESOURCE ERROR: Waited for 30000 ms, but only 10 tasks for 'Fetch
> > > > parquet
> > > > > metadata' are complete. Total number of tasks 29, parallelism 16.
> > > > >
> > > > >
> > > > > [Error Id: 3b079174-f5d0-4313-8097-25a0b3070854 on
> > > > > IORA-G9KY9P2.stf.nus.edu.sg:31010]
> > > > > RESOURCE ERROR: Waited for 30000 ms, but only 10 tasks for 'Fetch
> > > > > parquet metadata' are complete. Total number of tasks 29,
> parallelism
> > > 16.
> > > > >
> > > > >
> > > > > [Error Id: 3b079174-f5d0-4313-8097-25a0b3070854 on
> > > > > IORA-G9KY9P2.stf.nus.edu.sg:31010]
> > > > > RESOURCE ERROR: Waited for 30000 ms, but only 10 tasks for 'Fetch
> > > > > parquet metadata' are complete. Total number of tasks 29,
> parallelism
> > > 16.
> > > > >
> > > > >
> > > > > [Error Id: 3b079174-f5d0-4313-8097-25a0b3070854 on
> > > > > IORA-G9KY9P2.stf.nus.edu.sg:31010]
> > > > >
> > > > > ----------------------------------------
> > > > > From Drill log:
> > > > >
> > > > > 2018-09-20 08:58:12,035
> > [245d0f5a-ae5f-bfa2-ff04-40f7bdd1c2bf:foreman]
> > > > > INFO o.a.drill.exec.work.foreman.Foreman - Query text for query id
> > > > > 245d0f5a-ae5f-bfa2-ff04-40f7bdd1c2bf: select *
> > > > > from dfs.tmp.load_pos_sales_detail_tbl
> > > > >
> > > > > 2018-09-20 08:58:53,068
> > [245d0f5a-ae5f-bfa2-ff04-40f7bdd1c2bf:foreman]
> > > > > ERROR o.a.d.e.s.parquet.metadata.Metadata - Waited for 30000 ms,
> but
> > > > only
> > > > > 10 tasks for 'Fetch parquet metadata' are complete. Total number of
> > > tasks
> > > > > 29, parallelism 16.
> > > > > java.util.concurrent.CancellationException: null
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.TimedCallable$FutureMapper.
> > > > apply(TimedCallable.java:86)
> > > > > ~[drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.TimedCallable$FutureMapper.
> > > > apply(TimedCallable.java:57)
> > > > > ~[drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.common.collections.Collectors.lambda$
> > > > toList$2(Collectors.java:97)
> > > > > ~[drill-common-1.14.0.jar:1.14.0]
> > > > > at java.util.ArrayList.forEach(ArrayList.java:1257) ~[na:1.8.0_172]
> > > > > at
> > > > > org.apache.drill.common.collections.Collectors.toList(
> > > > Collectors.java:97)
> > > > > ~[drill-common-1.14.0.jar:1.14.0]
> > > > > at org.apache.drill.exec.store.TimedCallable.run(
> > > TimedCallable.java:214)
> > > > > ~[drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.parquet.metadata.Metadata.
> > > > getParquetFileMetadata_v3(Metadata.java:340)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.parquet.metadata.Metadata.
> > > > getParquetTableMetadata(Metadata.java:324)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.parquet.metadata.Metadata.
> > > > getParquetTableMetadata(Metadata.java:305)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.parquet.metadata.Metadata.
> > > > getParquetTableMetadata(Metadata.java:124)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.parquet.ParquetGroupScan.
> > > > initInternal(ParquetGroupScan.java:254)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.parquet.AbstractParquetGroupScan.init(
> > > > AbstractParquetGroupScan.java:380)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.parquet.ParquetGroupScan.
> > > > init>(ParquetGroupScan.java:132)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.parquet.ParquetGroupScan.
> > > > init>(ParquetGroupScan.java:102)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > >
> org.apache.drill.exec.store.parquet.ParquetFormatPlugin.getGroupScan(
> > > > ParquetFormatPlugin.java:180)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > >
> org.apache.drill.exec.store.parquet.ParquetFormatPlugin.getGroupScan(
> > > > ParquetFormatPlugin.java:70)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.dfs.FileSystemPlugin.getPhysicalScan(
> > > > FileSystemPlugin.java:136)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.AbstractStoragePlugin.getPhysicalScan(
> > > > AbstractStoragePlugin.java:116)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.AbstractStoragePlugin.getPhysicalScan(
> > > > AbstractStoragePlugin.java:111)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.logical.DrillTable.
> > > > getGroupScan(DrillTable.java:99)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.logical.DrillScanRel.(
> > > > DrillScanRel.java:89)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.logical.DrillScanRel.(
> > > > DrillScanRel.java:69)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.logical.DrillScanRel.(
> > > > DrillScanRel.java:62)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.logical.DrillScanRule.onMatch(
> > > > DrillScanRule.java:38)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.calcite.plan.volcano.VolcanoRuleCall.
> > > > onMatch(VolcanoRuleCall.java:212)
> > > > > [calcite-core-1.16.0-drill-r6.jar:1.16.0-drill-r6]
> > > > > at
> > > > >
> > > > > org.apache.calcite.plan.volcano.VolcanoPlanner.
> > > > findBestExp(VolcanoPlanner.java:652)
> > > > > [calcite-core-1.16.0-drill-r6.jar:1.16.0-drill-r6]
> > > > > at org.apache.calcite.tools.Programs$RuleSetProgram.run(
> > > > Programs.java:368)
> > > > > [calcite-core-1.16.0-drill-r6.jar:1.16.0-drill-r6]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.sql.handlers.
> > > DefaultSqlHandler.transform(
> > > > DefaultSqlHandler.java:429)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.sql.handlers.
> > > DefaultSqlHandler.transform(
> > > > DefaultSqlHandler.java:369)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.
> > > > convertToRawDrel(DefaultSqlHandler.java:255)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.
> > > > convertToDrel(DefaultSqlHandler.java:318)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > >
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.getPlan(
> > > > DefaultSqlHandler.java:180)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.sql.DrillSqlWorker.
> > > > getQueryPlan(DrillSqlWorker.java:145)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan(
> > > > DrillSqlWorker.java:83)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at org.apache.drill.exec.work
> > .foreman.Foreman.runSQL(Foreman.java:567)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at org.apache.drill.exec.work
> .foreman.Foreman.run(Foreman.java:266)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > java.util.concurrent.ThreadPoolExecutor.runWorker(
> > > > ThreadPoolExecutor.java:1149)
> > > > > [na:1.8.0_172]
> > > > > at
> > > > >
> > > > > java.util.concurrent.ThreadPoolExecutor$Worker.run(
> > > > ThreadPoolExecutor.java:624)
> > > > > [na:1.8.0_172]
> > > > > at java.lang.Thread.run(Thread.java:748) [na:1.8.0_172]
> > > > > 2018-09-20 08:58:53,080
> > [245d0f5a-ae5f-bfa2-ff04-40f7bdd1c2bf:foreman]
> > > > > INFO o.a.d.e.s.parquet.metadata.Metadata - User Error Occurred:
> > Waited
> > > > for
> > > > > 30000 ms, but only 10 tasks for 'Fetch parquet metadata' are
> > complete.
> > > > > Total number of tasks 29, parallelism 16. (null)
> > > > > org.apache.drill.common.exceptions.UserException: RESOURCE ERROR:
> > > Waited
> > > > > for 30000 ms, but only 10 tasks for 'Fetch parquet metadata' are
> > > > complete.
> > > > > Total number of tasks 29, parallelism 16.
> > > > >
> > > > >
> > > > > [Error Id: f887dcae-9f55-469c-be52-b6ce2a37eeb0 ]
> > > > > at
> > > > >
> > > > > org.apache.drill.common.exceptions.UserException$
> > > > Builder.build(UserException.java:633)
> > > > > ~[drill-common-1.14.0.jar:1.14.0]
> > > > > at org.apache.drill.exec.store.TimedCallable.run(
> > > TimedCallable.java:253)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.parquet.metadata.Metadata.
> > > > getParquetFileMetadata_v3(Metadata.java:340)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.parquet.metadata.Metadata.
> > > > getParquetTableMetadata(Metadata.java:324)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.parquet.metadata.Metadata.
> > > > getParquetTableMetadata(Metadata.java:305)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.parquet.metadata.Metadata.
> > > > getParquetTableMetadata(Metadata.java:124)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.parquet.ParquetGroupScan.
> > > > initInternal(ParquetGroupScan.java:254)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.parquet.AbstractParquetGroupScan.init(
> > > > AbstractParquetGroupScan.java:380)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.parquet.ParquetGroupScan.
> > > > init>(ParquetGroupScan.java:132)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.parquet.ParquetGroupScan.
> > > > init>(ParquetGroupScan.java:102)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > >
> org.apache.drill.exec.store.parquet.ParquetFormatPlugin.getGroupScan(
> > > > ParquetFormatPlugin.java:180)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > >
> org.apache.drill.exec.store.parquet.ParquetFormatPlugin.getGroupScan(
> > > > ParquetFormatPlugin.java:70)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.dfs.FileSystemPlugin.getPhysicalScan(
> > > > FileSystemPlugin.java:136)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.AbstractStoragePlugin.getPhysicalScan(
> > > > AbstractStoragePlugin.java:116)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.AbstractStoragePlugin.getPhysicalScan(
> > > > AbstractStoragePlugin.java:111)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.logical.DrillTable.
> > > > getGroupScan(DrillTable.java:99)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.logical.DrillScanRel.(
> > > > DrillScanRel.java:89)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.logical.DrillScanRel.(
> > > > DrillScanRel.java:69)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.logical.DrillScanRel.(
> > > > DrillScanRel.java:62)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.logical.DrillScanRule.onMatch(
> > > > DrillScanRule.java:38)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.calcite.plan.volcano.VolcanoRuleCall.
> > > > onMatch(VolcanoRuleCall.java:212)
> > > > > [calcite-core-1.16.0-drill-r6.jar:1.16.0-drill-r6]
> > > > > at
> > > > >
> > > > > org.apache.calcite.plan.volcano.VolcanoPlanner.
> > > > findBestExp(VolcanoPlanner.java:652)
> > > > > [calcite-core-1.16.0-drill-r6.jar:1.16.0-drill-r6]
> > > > > at org.apache.calcite.tools.Programs$RuleSetProgram.run(
> > > > Programs.java:368)
> > > > > [calcite-core-1.16.0-drill-r6.jar:1.16.0-drill-r6]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.sql.handlers.
> > > DefaultSqlHandler.transform(
> > > > DefaultSqlHandler.java:429)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.sql.handlers.
> > > DefaultSqlHandler.transform(
> > > > DefaultSqlHandler.java:369)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.
> > > > convertToRawDrel(DefaultSqlHandler.java:255)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.
> > > > convertToDrel(DefaultSqlHandler.java:318)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > >
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.getPlan(
> > > > DefaultSqlHandler.java:180)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.sql.DrillSqlWorker.
> > > > getQueryPlan(DrillSqlWorker.java:145)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan(
> > > > DrillSqlWorker.java:83)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at org.apache.drill.exec.work
> > .foreman.Foreman.runSQL(Foreman.java:567)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at org.apache.drill.exec.work
> .foreman.Foreman.run(Foreman.java:266)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > java.util.concurrent.ThreadPoolExecutor.runWorker(
> > > > ThreadPoolExecutor.java:1149)
> > > > > [na:1.8.0_172]
> > > > > at
> > > > >
> > > > > java.util.concurrent.ThreadPoolExecutor$Worker.run(
> > > > ThreadPoolExecutor.java:624)
> > > > > [na:1.8.0_172]
> > > > > at java.lang.Thread.run(Thread.java:748) [na:1.8.0_172]
> > > > > Caused by: java.util.concurrent.CancellationException: null
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.TimedCallable$FutureMapper.
> > > > apply(TimedCallable.java:86)
> > > > > ~[drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.TimedCallable$FutureMapper.
> > > > apply(TimedCallable.java:57)
> > > > > ~[drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.common.collections.Collectors.lambda$
> > > > toList$2(Collectors.java:97)
> > > > > ~[drill-common-1.14.0.jar:1.14.0]
> > > > > at java.util.ArrayList.forEach(ArrayList.java:1257) ~[na:1.8.0_172]
> > > > > at
> > > > > org.apache.drill.common.collections.Collectors.toList(
> > > > Collectors.java:97)
> > > > > ~[drill-common-1.14.0.jar:1.14.0]
> > > > > at org.apache.drill.exec.store.TimedCallable.run(
> > > TimedCallable.java:214)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > ... 33 common frames omitted
> > > > > 2018-09-20 09:02:10,608 [UserServer-1] WARN
> > > > > o.a.drill.exec.rpc.user.UserServer - Message of mode REQUEST of rpc
> > > > type 3
> > > > > took longer than 500ms. Actual duration was 2042ms.
> > > > > 2018-09-20 09:02:10,608
> > [245d0e6f-0dc1-2a4b-12a4-b9aaad4182fc:foreman]
> > > > > INFO o.a.drill.exec.work.foreman.Foreman - Query text for query id
> > > > > 245d0e6f-0dc1-2a4b-12a4-b9aaad4182fc: select *
> > > > > from dfs.tmp.load_pos_sales_detail_tbl
> > > > >
> > > > > 2018-09-20 09:02:42,615
> > [245d0e6f-0dc1-2a4b-12a4-b9aaad4182fc:foreman]
> > > > > ERROR o.a.d.e.s.parquet.metadata.Metadata - Waited for 30000 ms,
> but
> > > > only
> > > > > 10 tasks for 'Fetch parquet metadata' are complete. Total number of
> > > tasks
> > > > > 29, parallelism 16.
> > > > > java.util.concurrent.CancellationException: null
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.TimedCallable$FutureMapper.
> > > > apply(TimedCallable.java:86)
> > > > > ~[drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.TimedCallable$FutureMapper.
> > > > apply(TimedCallable.java:57)
> > > > > ~[drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.common.collections.Collectors.lambda$
> > > > toList$2(Collectors.java:97)
> > > > > ~[drill-common-1.14.0.jar:1.14.0]
> > > > > at java.util.ArrayList.forEach(ArrayList.java:1257) ~[na:1.8.0_172]
> > > > > at
> > > > > org.apache.drill.common.collections.Collectors.toList(
> > > > Collectors.java:97)
> > > > > ~[drill-common-1.14.0.jar:1.14.0]
> > > > > at org.apache.drill.exec.store.TimedCallable.run(
> > > TimedCallable.java:214)
> > > > > ~[drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.parquet.metadata.Metadata.
> > > > getParquetFileMetadata_v3(Metadata.java:340)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.parquet.metadata.Metadata.
> > > > getParquetTableMetadata(Metadata.java:324)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.parquet.metadata.Metadata.
> > > > getParquetTableMetadata(Metadata.java:305)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.parquet.metadata.Metadata.
> > > > getParquetTableMetadata(Metadata.java:124)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.parquet.ParquetGroupScan.
> > > > initInternal(ParquetGroupScan.java:254)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.parquet.AbstractParquetGroupScan.init(
> > > > AbstractParquetGroupScan.java:380)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.parquet.ParquetGroupScan.
> > > > init>(ParquetGroupScan.java:132)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.parquet.ParquetGroupScan.
> > > > init>(ParquetGroupScan.java:102)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > >
> org.apache.drill.exec.store.parquet.ParquetFormatPlugin.getGroupScan(
> > > > ParquetFormatPlugin.java:180)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > >
> org.apache.drill.exec.store.parquet.ParquetFormatPlugin.getGroupScan(
> > > > ParquetFormatPlugin.java:70)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.dfs.FileSystemPlugin.getPhysicalScan(
> > > > FileSystemPlugin.java:136)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.AbstractStoragePlugin.getPhysicalScan(
> > > > AbstractStoragePlugin.java:116)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.AbstractStoragePlugin.getPhysicalScan(
> > > > AbstractStoragePlugin.java:111)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.logical.DrillTable.
> > > > getGroupScan(DrillTable.java:99)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.logical.DrillScanRel.(
> > > > DrillScanRel.java:89)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.logical.DrillScanRel.(
> > > > DrillScanRel.java:69)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.logical.DrillScanRel.(
> > > > DrillScanRel.java:62)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.logical.DrillScanRule.onMatch(
> > > > DrillScanRule.java:38)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.calcite.plan.volcano.VolcanoRuleCall.
> > > > onMatch(VolcanoRuleCall.java:212)
> > > > > [calcite-core-1.16.0-drill-r6.jar:1.16.0-drill-r6]
> > > > > at
> > > > >
> > > > > org.apache.calcite.plan.volcano.VolcanoPlanner.
> > > > findBestExp(VolcanoPlanner.java:652)
> > > > > [calcite-core-1.16.0-drill-r6.jar:1.16.0-drill-r6]
> > > > > at org.apache.calcite.tools.Programs$RuleSetProgram.run(
> > > > Programs.java:368)
> > > > > [calcite-core-1.16.0-drill-r6.jar:1.16.0-drill-r6]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.sql.handlers.
> > > DefaultSqlHandler.transform(
> > > > DefaultSqlHandler.java:429)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.sql.handlers.
> > > DefaultSqlHandler.transform(
> > > > DefaultSqlHandler.java:369)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.
> > > > convertToRawDrel(DefaultSqlHandler.java:255)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.
> > > > convertToDrel(DefaultSqlHandler.java:318)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > >
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.getPlan(
> > > > DefaultSqlHandler.java:180)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.sql.DrillSqlWorker.
> > > > getQueryPlan(DrillSqlWorker.java:145)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan(
> > > > DrillSqlWorker.java:83)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at org.apache.drill.exec.work
> > .foreman.Foreman.runSQL(Foreman.java:567)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at org.apache.drill.exec.work
> .foreman.Foreman.run(Foreman.java:266)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > java.util.concurrent.ThreadPoolExecutor.runWorker(
> > > > ThreadPoolExecutor.java:1149)
> > > > > [na:1.8.0_172]
> > > > > at
> > > > >
> > > > > java.util.concurrent.ThreadPoolExecutor$Worker.run(
> > > > ThreadPoolExecutor.java:624)
> > > > > [na:1.8.0_172]
> > > > > at java.lang.Thread.run(Thread.java:748) [na:1.8.0_172]
> > > > > 2018-09-20 09:02:42,625
> > [245d0e6f-0dc1-2a4b-12a4-b9aaad4182fc:foreman]
> > > > > INFO o.a.d.e.s.parquet.metadata.Metadata - User Error Occurred:
> > Waited
> > > > for
> > > > > 30000 ms, but only 10 tasks for 'Fetch parquet metadata' are
> > complete.
> > > > > Total number of tasks 29, parallelism 16. (null)
> > > > > org.apache.drill.common.exceptions.UserException: RESOURCE ERROR:
> > > Waited
> > > > > for 30000 ms, but only 10 tasks for 'Fetch parquet metadata' are
> > > > complete.
> > > > > Total number of tasks 29, parallelism 16.
> > > > >
> > > > >
> > > > > [Error Id: 3b079174-f5d0-4313-8097-25a0b3070854 ]
> > > > > at
> > > > >
> > > > > org.apache.drill.common.exceptions.UserException$
> > > > Builder.build(UserException.java:633)
> > > > > ~[drill-common-1.14.0.jar:1.14.0]
> > > > > at org.apache.drill.exec.store.TimedCallable.run(
> > > TimedCallable.java:253)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.parquet.metadata.Metadata.
> > > > getParquetFileMetadata_v3(Metadata.java:340)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.parquet.metadata.Metadata.
> > > > getParquetTableMetadata(Metadata.java:324)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.parquet.metadata.Metadata.
> > > > getParquetTableMetadata(Metadata.java:305)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.parquet.metadata.Metadata.
> > > > getParquetTableMetadata(Metadata.java:124)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.parquet.ParquetGroupScan.
> > > > initInternal(ParquetGroupScan.java:254)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.parquet.AbstractParquetGroupScan.init(
> > > > AbstractParquetGroupScan.java:380)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.parquet.ParquetGroupScan.
> > > > init>(ParquetGroupScan.java:132)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.parquet.ParquetGroupScan.
> > > > init>(ParquetGroupScan.java:102)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > >
> org.apache.drill.exec.store.parquet.ParquetFormatPlugin.getGroupScan(
> > > > ParquetFormatPlugin.java:180)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > >
> org.apache.drill.exec.store.parquet.ParquetFormatPlugin.getGroupScan(
> > > > ParquetFormatPlugin.java:70)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.dfs.FileSystemPlugin.getPhysicalScan(
> > > > FileSystemPlugin.java:136)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.AbstractStoragePlugin.getPhysicalScan(
> > > > AbstractStoragePlugin.java:116)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.AbstractStoragePlugin.getPhysicalScan(
> > > > AbstractStoragePlugin.java:111)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.logical.DrillTable.
> > > > getGroupScan(DrillTable.java:99)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.logical.DrillScanRel.(
> > > > DrillScanRel.java:89)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.logical.DrillScanRel.(
> > > > DrillScanRel.java:69)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.logical.DrillScanRel.(
> > > > DrillScanRel.java:62)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.logical.DrillScanRule.onMatch(
> > > > DrillScanRule.java:38)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.calcite.plan.volcano.VolcanoRuleCall.
> > > > onMatch(VolcanoRuleCall.java:212)
> > > > > [calcite-core-1.16.0-drill-r6.jar:1.16.0-drill-r6]
> > > > > at
> > > > >
> > > > > org.apache.calcite.plan.volcano.VolcanoPlanner.
> > > > findBestExp(VolcanoPlanner.java:652)
> > > > > [calcite-core-1.16.0-drill-r6.jar:1.16.0-drill-r6]
> > > > > at org.apache.calcite.tools.Programs$RuleSetProgram.run(
> > > > Programs.java:368)
> > > > > [calcite-core-1.16.0-drill-r6.jar:1.16.0-drill-r6]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.sql.handlers.
> > > DefaultSqlHandler.transform(
> > > > DefaultSqlHandler.java:429)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.sql.handlers.
> > > DefaultSqlHandler.transform(
> > > > DefaultSqlHandler.java:369)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.
> > > > convertToRawDrel(DefaultSqlHandler.java:255)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.
> > > > convertToDrel(DefaultSqlHandler.java:318)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > >
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.getPlan(
> > > > DefaultSqlHandler.java:180)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.sql.DrillSqlWorker.
> > > > getQueryPlan(DrillSqlWorker.java:145)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan(
> > > > DrillSqlWorker.java:83)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at org.apache.drill.exec.work
> > .foreman.Foreman.runSQL(Foreman.java:567)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at org.apache.drill.exec.work
> .foreman.Foreman.run(Foreman.java:266)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > java.util.concurrent.ThreadPoolExecutor.runWorker(
> > > > ThreadPoolExecutor.java:1149)
> > > > > [na:1.8.0_172]
> > > > > at
> > > > >
> > > > > java.util.concurrent.ThreadPoolExecutor$Worker.run(
> > > > ThreadPoolExecutor.java:624)
> > > > > [na:1.8.0_172]
> > > > > at java.lang.Thread.run(Thread.java:748) [na:1.8.0_172]
> > > > > Caused by: java.util.concurrent.CancellationException: null
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.TimedCallable$FutureMapper.
> > > > apply(TimedCallable.java:86)
> > > > > ~[drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.exec.store.TimedCallable$FutureMapper.
> > > > apply(TimedCallable.java:57)
> > > > > ~[drill-java-exec-1.14.0.jar:1.14.0]
> > > > > at
> > > > >
> > > > > org.apache.drill.common.collections.Collectors.lambda$
> > > > toList$2(Collectors.java:97)
> > > > > ~[drill-common-1.14.0.jar:1.14.0]
> > > > > at java.util.ArrayList.forEach(ArrayList.java:1257) ~[na:1.8.0_172]
> > > > > at
> > > > > org.apache.drill.common.collections.Collectors.toList(
> > > > Collectors.java:97)
> > > > > ~[drill-common-1.14.0.jar:1.14.0]
> > > > > at org.apache.drill.exec.store.TimedCallable.run(
> > > TimedCallable.java:214)
> > > > > [drill-java-exec-1.14.0.jar:1.14.0]
> > > > > ... 33 common frames omitted
> > > > >
> > > > >
> > > > > ----------------------------------------
> > > > > ----------
> > > > >
> > > >
> > >
> >
>

Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message