beam-builds mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Apache Jenkins Server <jenk...@builds.apache.org>
Subject Build failed in Jenkins: beam_PostCommit_Java_PortabilityApi #575
Date Wed, 02 Jan 2019 23:08:18 GMT
See <https://builds.apache.org/job/beam_PostCommit_Java_PortabilityApi/575/display/redirect?page=changes>

Changes:

[aaltay] [BEAM-5315] Python 3 port io.filebased_* modules (#7386)

------------------------------------------
[...truncated 1.53 MB...]
    INFO: Adding BigQueryIO.Write/BatchLoads/View.AsSingleton/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/WithKeys/AddKeys/Map
as step s45
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/View.AsSingleton/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey
as step s46
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/View.AsSingleton/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/Combine.GroupedValues
as step s47
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/View.AsSingleton/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Values/Values/Map
as step s48
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/View.AsSingleton/Combine.GloballyAsSingletonView/View.VoidKeyToMultimapMaterialization/ParDo(VoidKeyToMultimapMaterialization)
as step s49
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/View.AsSingleton/Combine.GloballyAsSingletonView/View.CreatePCollectionView
as step s50
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/Create.Values/Read(CreateSource)/Impulse as step
s51
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/Create.Values/Read(CreateSource)/ParDo(SplitBoundedSource)
as step s52
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/Create.Values/Read(CreateSource)/Reshuffle.ViaRandomKey/Pair
with random key as step s53
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/Create.Values/Read(CreateSource)/Reshuffle.ViaRandomKey/Reshuffle/Window.Into()/Window.Assign
as step s54
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/Create.Values/Read(CreateSource)/Reshuffle.ViaRandomKey/Reshuffle/GroupByKey
as step s55
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/Create.Values/Read(CreateSource)/Reshuffle.ViaRandomKey/Reshuffle/ExpandIterable
as step s56
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/Create.Values/Read(CreateSource)/Reshuffle.ViaRandomKey/Values/Values/Map
as step s57
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/Create.Values/Read(CreateSource)/ParDo(ReadFromBoundedSource)
as step s58
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/GetTempFilePrefix as step s59
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/TempFilePrefixView/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/WithKeys/AddKeys/Map
as step s60
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/TempFilePrefixView/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey
as step s61
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/TempFilePrefixView/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/Combine.GroupedValues
as step s62
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/TempFilePrefixView/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Values/Values/Map
as step s63
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/TempFilePrefixView/Combine.GloballyAsSingletonView/View.VoidKeyToMultimapMaterialization/ParDo(VoidKeyToMultimapMaterialization)
as step s64
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/TempFilePrefixView/Combine.GloballyAsSingletonView/View.CreatePCollectionView
as step s65
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/rewindowIntoGlobal/Window.Assign as step s66
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/WriteBundlesToFiles as step s67
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/GroupByDestination as step s68
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/WriteGroupedRecords as step s69
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/FlattenFiles as step s70
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyResults/View.AsIterable/View.VoidKeyToMultimapMaterialization/ParDo(VoidKeyToMultimapMaterialization)
as step s71
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyResults/View.AsIterable/View.CreatePCollectionView
as step s72
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyResults/Create.Values/Read(CreateSource)/Impulse
as step s73
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyResults/Create.Values/Read(CreateSource)/ParDo(SplitBoundedSource)
as step s74
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyResults/Create.Values/Read(CreateSource)/Reshuffle.ViaRandomKey/Pair
with random key as step s75
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyResults/Create.Values/Read(CreateSource)/Reshuffle.ViaRandomKey/Reshuffle/Window.Into()/Window.Assign
as step s76
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyResults/Create.Values/Read(CreateSource)/Reshuffle.ViaRandomKey/Reshuffle/GroupByKey
as step s77
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyResults/Create.Values/Read(CreateSource)/Reshuffle.ViaRandomKey/Reshuffle/ExpandIterable
as step s78
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyResults/Create.Values/Read(CreateSource)/Reshuffle.ViaRandomKey/Values/Values/Map
as step s79
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyResults/Create.Values/Read(CreateSource)/ParDo(ReadFromBoundedSource)
as step s80
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyResults/ParDo(Anonymous) as step s81
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/WritePartitionUntriggered as step s82
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/MultiPartitionsReshuffle/Window.Into()/Window.Assign
as step s83
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/MultiPartitionsReshuffle/GroupByKey as step s84
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/MultiPartitionsReshuffle/ExpandIterable as step
s85
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/MultiPartitionsWriteTables/ParMultiDo(WriteTables)
as step s86
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/MultiPartitionsWriteTables/WithKeys/AddKeys/Map
as step s87
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/MultiPartitionsWriteTables/Window.Into()/Window.Assign
as step s88
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/MultiPartitionsWriteTables/GroupByKey as step
s89
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/MultiPartitionsWriteTables/Values/Values/Map
as step s90
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/MultiPartitionsWriteTables/ParDo(GarbageCollectTemporaryFiles)
as step s91
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyRenameInput/View.AsIterable/View.VoidKeyToMultimapMaterialization/ParDo(VoidKeyToMultimapMaterialization)
as step s92
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyRenameInput/View.AsIterable/View.CreatePCollectionView
as step s93
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyRenameInput/Create.Values/Read(CreateSource)/Impulse
as step s94
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyRenameInput/Create.Values/Read(CreateSource)/ParDo(SplitBoundedSource)
as step s95
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyRenameInput/Create.Values/Read(CreateSource)/Reshuffle.ViaRandomKey/Pair
with random key as step s96
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyRenameInput/Create.Values/Read(CreateSource)/Reshuffle.ViaRandomKey/Reshuffle/Window.Into()/Window.Assign
as step s97
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyRenameInput/Create.Values/Read(CreateSource)/Reshuffle.ViaRandomKey/Reshuffle/GroupByKey
as step s98
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyRenameInput/Create.Values/Read(CreateSource)/Reshuffle.ViaRandomKey/Reshuffle/ExpandIterable
as step s99
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyRenameInput/Create.Values/Read(CreateSource)/Reshuffle.ViaRandomKey/Values/Values/Map
as step s100
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyRenameInput/Create.Values/Read(CreateSource)/ParDo(ReadFromBoundedSource)
as step s101
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/ReifyRenameInput/ParDo(Anonymous) as step s102
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/WriteRenameUntriggered as step s103
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/SinglePartitionsReshuffle/Window.Into()/Window.Assign
as step s104
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/SinglePartitionsReshuffle/GroupByKey as step
s105
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/SinglePartitionsReshuffle/ExpandIterable as step
s106
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/SinglePartitionWriteTables/ParMultiDo(WriteTables)
as step s107
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/SinglePartitionWriteTables/WithKeys/AddKeys/Map
as step s108
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/SinglePartitionWriteTables/Window.Into()/Window.Assign
as step s109
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/SinglePartitionWriteTables/GroupByKey as step
s110
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/SinglePartitionWriteTables/Values/Values/Map
as step s111
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/SinglePartitionWriteTables/ParDo(GarbageCollectTemporaryFiles)
as step s112
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/CreateEmptyFailedInserts/Read(CreateSource)/Impulse
as step s113
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/CreateEmptyFailedInserts/Read(CreateSource)/ParDo(SplitBoundedSource)
as step s114
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/CreateEmptyFailedInserts/Read(CreateSource)/Reshuffle.ViaRandomKey/Pair
with random key as step s115
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/CreateEmptyFailedInserts/Read(CreateSource)/Reshuffle.ViaRandomKey/Reshuffle/Window.Into()/Window.Assign
as step s116
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/CreateEmptyFailedInserts/Read(CreateSource)/Reshuffle.ViaRandomKey/Reshuffle/GroupByKey
as step s117
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/CreateEmptyFailedInserts/Read(CreateSource)/Reshuffle.ViaRandomKey/Reshuffle/ExpandIterable
as step s118
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/CreateEmptyFailedInserts/Read(CreateSource)/Reshuffle.ViaRandomKey/Values/Values/Map
as step s119
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding BigQueryIO.Write/BatchLoads/CreateEmptyFailedInserts/Read(CreateSource)/ParDo(ReadFromBoundedSource)
as step s120
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator
addStep
    INFO: Adding DropInputs as step s121
    Jan 02, 2019 11:01:54 PM org.apache.beam.runners.dataflow.DataflowRunner run
    INFO: Staging pipeline description to gs://temp-storage-for-end-to-end-tests/testpipeline-jenkins-0102230143-5edf358/output/results/staging/
    Jan 02, 2019 11:01:55 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
    INFO: Uploading <349675 bytes, hash fvCimWd-KUxS-2QE52OvqQ> to gs://temp-storage-for-end-to-end-tests/testpipeline-jenkins-0102230143-5edf358/output/results/staging/pipeline-fvCimWd-KUxS-2QE52OvqQ.pb

org.apache.beam.sdk.io.gcp.bigquery.BigQueryToTableIT > testStandardQueryWithoutCustom
STANDARD_OUT
    Dataflow SDK version: 2.10.0-SNAPSHOT

org.apache.beam.sdk.io.gcp.bigquery.BigQueryToTableIT > testStandardQueryWithoutCustom
STANDARD_ERROR
    Jan 02, 2019 11:02:03 PM org.apache.beam.runners.dataflow.DataflowRunner run
    INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-01-02_15_01_55-7855797073510943290?project=apache-beam-testing

org.apache.beam.sdk.io.gcp.bigquery.BigQueryToTableIT > testStandardQueryWithoutCustom
STANDARD_OUT
    Submitted job: 2019-01-02_15_01_55-7855797073510943290

org.apache.beam.sdk.io.gcp.bigquery.BigQueryToTableIT > testStandardQueryWithoutCustom
STANDARD_ERROR
    Jan 02, 2019 11:02:03 PM org.apache.beam.runners.dataflow.DataflowRunner run
    INFO: To cancel the job using the 'gcloud' tool, run:
    > gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2019-01-02_15_01_55-7855797073510943290
    Jan 02, 2019 11:02:03 PM org.apache.beam.runners.dataflow.TestDataflowRunner run
    INFO: Running Dataflow job 2019-01-02_15_01_55-7855797073510943290 with 0 expected assertions.
    Jan 02, 2019 11:08:14 PM org.apache.beam.runners.dataflow.DataflowPipelineJob waitUntilFinish
    WARNING: No terminal state was returned. State value UNKNOWN
    Jan 02, 2019 11:08:14 PM org.apache.beam.runners.dataflow.TestDataflowRunner checkForPAssertSuccess
    WARNING: Metrics not present for Dataflow job 2019-01-02_15_01_55-7855797073510943290.
    Jan 02, 2019 11:08:14 PM org.apache.beam.runners.dataflow.TestDataflowRunner run
    WARNING: Dataflow job 2019-01-02_15_01_55-7855797073510943290 did not output a success
or failure metric.
    Jan 02, 2019 11:08:14 PM org.apache.beam.sdk.io.gcp.bigquery.BigQueryToTableIT cleanBqEnvironment
    INFO: Start to clean up tables and datasets.
    Jan 02, 2019 11:08:14 PM org.apache.beam.sdk.io.gcp.testing.BigqueryClient deleteDataset
    INFO: Successfully deleted dataset: bq_query_to_table_1546470102747_787

Gradle Test Executor 3 finished executing tests.

> Task :beam-runners-google-cloud-dataflow-java:googleCloudPlatformFnApiWorkerIntegrationTest
FAILED

org.apache.beam.sdk.io.gcp.bigquery.BigQueryToTableIT > testStandardQueryWithoutCustom
FAILED
    java.lang.RuntimeException: Dataflow job 2019-01-02_15_01_55-7855797073510943290 terminated
in state UNKNOWN but did not return a failure reason.
        at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:134)
        at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:90)
        at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:55)
        at org.apache.beam.sdk.Pipeline.run(Pipeline.java:313)
        at org.apache.beam.sdk.Pipeline.run(Pipeline.java:299)
        at org.apache.beam.sdk.io.gcp.bigquery.BigQueryToTableIT.runBigQueryToTablePipeline(BigQueryToTableIT.java:113)
        at org.apache.beam.sdk.io.gcp.bigquery.BigQueryToTableIT.testStandardQueryWithoutCustom(BigQueryToTableIT.java:297)

7 tests completed, 1 failed
Finished generating test XML results (0.029 secs) into: <https://builds.apache.org/job/beam_PostCommit_Java_PortabilityApi/ws/src/runners/google-cloud-dataflow-java/build/test-results/googleCloudPlatformFnApiWorkerIntegrationTest>
Generating HTML test report...
Finished generating test html results (0.019 secs) into: <https://builds.apache.org/job/beam_PostCommit_Java_PortabilityApi/ws/src/runners/google-cloud-dataflow-java/build/reports/tests/googleCloudPlatformFnApiWorkerIntegrationTest>
:beam-runners-google-cloud-dataflow-java:googleCloudPlatformFnApiWorkerIntegrationTest (Thread[Task
worker for ':',5,main]) completed. Took 33 mins 28.688 secs.
:beam-runners-google-cloud-dataflow-java:cleanUpDockerImages (Thread[Task worker for ':',5,main])
started.

> Task :beam-runners-google-cloud-dataflow-java:cleanUpDockerImages
Caching disabled for task ':beam-runners-google-cloud-dataflow-java:cleanUpDockerImages':
Caching has not been enabled for the task
Task ':beam-runners-google-cloud-dataflow-java:cleanUpDockerImages' is not up-to-date because:
  Task has not declared any outputs despite executing actions.
Starting process 'command 'docker''. Working directory: <https://builds.apache.org/job/beam_PostCommit_Java_PortabilityApi/ws/src/runners/google-cloud-dataflow-java>
Command: docker rmi us.gcr.io/apache-beam-testing/java-postcommit-it/java:20190102221941
Successfully started process 'command 'docker''
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20190102221941
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:621cbb2d68eda4bf5cc119d194de616ea57b3bedfebda204b7dc34e1b1bbd81d
Starting process 'command 'gcloud''. Working directory: <https://builds.apache.org/job/beam_PostCommit_Java_PortabilityApi/ws/src/runners/google-cloud-dataflow-java>
Command: gcloud --quiet container images delete --force-delete-tags us.gcr.io/apache-beam-testing/java-postcommit-it/java:20190102221941
Successfully started process 'command 'gcloud''
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:621cbb2d68eda4bf5cc119d194de616ea57b3bedfebda204b7dc34e1b1bbd81d
  Associated tags:
 - 20190102221941
Tags:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java:20190102221941
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20190102221941].
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:621cbb2d68eda4bf5cc119d194de616ea57b3bedfebda204b7dc34e1b1bbd81d].
:beam-runners-google-cloud-dataflow-java:cleanUpDockerImages (Thread[Task worker for ':',5,main])
completed. Took 1.872 secs.

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':beam-runners-google-cloud-dataflow-java:googleCloudPlatformFnApiWorkerIntegrationTest'.
> There were failing tests. See the report at: file://<https://builds.apache.org/job/beam_PostCommit_Java_PortabilityApi/ws/src/runners/google-cloud-dataflow-java/build/reports/tests/googleCloudPlatformFnApiWorkerIntegrationTest/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to get more log
output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 5.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/4.10.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 48m 46s
90 actionable tasks: 84 executed, 5 from cache, 1 up-to-date

Publishing build scan...
https://gradle.com/s/in5x5772z4y4a

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Mime
View raw message