spark empty dataframe with column not able to write in big query #6
Description
Writing empty data frame with column name show following error ;
Data frame as below
+------+---------+-------+
|id|start|end|
+------+---------+-------+
+------+---------+-------+
root
|-- id: string (nullable = true)
|-- start: long (nullable = false)
|-- end: long (nullable = false)
-ERROR-:com.miraisolutions.spark.bigquery.client.BigQueryClient.waitForJob(BigQueryClient.scala:375), com.miraisolutions.spark.bigquery.client.BigQueryClient.importTable(BigQueryClient.scala:356), com.miraisolutions.spark.bigquery.DefaultSource$$anonfun$createRelation$2$$anonfun$apply$4.apply(DefaultSource.scala:74), com.miraisolutions.spark.bigquery.DefaultSource$$anonfun$createRelation$2$$anonfun$apply$4.apply(DefaultSource.scala:63), com.miraisolutions.spark.bigquery.DefaultSource$TypeParameter.foldType(DefaultSource.scala:218), com.miraisolutions.spark.bigquery.DefaultSource$$anonfun$createRelation$2.apply(DefaultSource.scala:63), com.miraisolutions.spark.bigquery.DefaultSource$$anonfun$createRelation$2.apply(DefaultSource.scala:62), com.miraisolutions.spark.bigquery.DefaultSource$.com$miraisolutions$spark$bigquery$DefaultSource$$withBigQueryClient(DefaultSource.scala:202), com.miraisolutions.spark.bigquery.DefaultSource.createRelation(DefaultSource.scala:62), org.apache.spark.sql.execution.datasources.DataSource.write(DataSource.scala:469), org.apache.spark.sql.execution.datasources.SaveIntoDataSourceCommand.run(SaveIntoDataSourceCommand.scala:50), org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:58), org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:56), org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute(commands.scala:74), org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:117), org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:117), org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:138), org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151), org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:135), org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:116), org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:92), org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:92), org.apache.spark.sql.DataFrameWriter.runCommand(DataFrameWriter.scala:609), org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:233),