You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
cannot resolve 'CAST(avg(CAST(NULL AS DOUBLE)) AS NULL)' due to data type mismatch: cannot cast double to null;;
'Aggregate [cast(avg(cast(null as double)) as null) AS FRESHID27 __ NULL#48721]
+- Join Inner, (target(rel __ RELATIONSHIP @ session_tmp190) __ NODE#48685 = b __ NODE @ session_tmp190#48703)
:- Join Inner, (node __ NODE @ session_tmp190#48689 = source(rel __ RELATIONSHIP @ session_tmp190) __ NODE#48684)
: :- Project [___id#48681 AS node __ NODE @ session_tmp190#48689]
: : +- Project [cast(cast(___id#48678L as string) as binary) AS ___id#48681]
: : +- Project [___id#48676L AS ___id#48678L]
: : +- LocalRelation [___id#48676L]
: +- LocalRelation <empty>, [rel __ RELATIONSHIP @ session_tmp190#48683, source(rel __ RELATIONSHIP @ session_tmp190) __ NODE#48684, target(rel __ RELATIONSHIP @ session_tmp190) __ NODE#48685]
+- Project [node __ NODE @ session_tmp190#48689 AS b __ NODE @ session_tmp190#48703]
+- Project [___id#48681 AS node __ NODE @ session_tmp190#48689]
+- Project [cast(cast(___id#48678L as string) as binary) AS ___id#48681]
+- Project [___id#48676L AS ___id#48678L]
+- LocalRelation [___id#48676L]
org.apache.spark.sql.AnalysisException: cannot resolve 'CAST(avg(CAST(NULL AS DOUBLE)) AS NULL)' due to data type mismatch: cannot cast double to null;;
'Aggregate [cast(avg(cast(null as double)) as null) AS FRESHID27 __ NULL#48721]
+- Join Inner, (target(rel __ RELATIONSHIP @ session_tmp190) __ NODE#48685 = b __ NODE @ session_tmp190#48703)
:- Join Inner, (node __ NODE @ session_tmp190#48689 = source(rel __ RELATIONSHIP @ session_tmp190) __ NODE#48684)
: :- Project [___id#48681 AS node __ NODE @ session_tmp190#48689]
: : +- Project [cast(cast(___id#48678L as string) as binary) AS ___id#48681]
: : +- Project [___id#48676L AS ___id#48678L]
: : +- LocalRelation [___id#48676L]
: +- LocalRelation <empty>, [rel __ RELATIONSHIP @ session_tmp190#48683, source(rel __ RELATIONSHIP @ session_tmp190) __ NODE#48684, target(rel __ RELATIONSHIP @ session_tmp190) __ NODE#48685]
+- Project [node __ NODE @ session_tmp190#48689 AS b __ NODE @ session_tmp190#48703]
+- Project [___id#48681 AS node __ NODE @ session_tmp190#48689]
+- Project [cast(cast(___id#48678L as string) as binary) AS ___id#48681]
+- Project [___id#48676L AS ___id#48678L]
+- LocalRelation [___id#48676L]
at org.apache.spark.sql.catalyst.analysis.package$AnalysisErrorAt.failAnalysis(package.scala:42)
at org.apache.spark.sql.catalyst.analysis.CheckAnalysis$$anonfun$$nestedInanonfun$checkAnalysis$1$2.applyOrElse(CheckAnalysis.scala:117)
at org.apache.spark.sql.catalyst.analysis.CheckAnalysis$$anonfun$$nestedInanonfun$checkAnalysis$1$2.applyOrElse(CheckAnalysis.scala:108)
at org.apache.spark.sql.catalyst.trees.TreeNode.$anonfun$transformUp$2(TreeNode.scala:280)
at org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:72)
at org.apache.spark.sql.catalyst.trees.TreeNode.transformUp(TreeNode.scala:280)
at org.apache.spark.sql.catalyst.trees.TreeNode.$anonfun$transformUp$1(TreeNode.scala:277)
at org.apache.spark.sql.catalyst.trees.TreeNode.$anonfun$mapChildren$1(TreeNode.scala:328)
at org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:189)
at org.apache.spark.sql.catalyst.trees.TreeNode.mapChildren(TreeNode.scala:326)
at org.apache.spark.sql.catalyst.trees.TreeNode.transformUp(TreeNode.scala:277)
at org.apache.spark.sql.catalyst.plans.QueryPlan.$anonfun$transformExpressionsUp$1(QueryPlan.scala:94)
at org.apache.spark.sql.catalyst.plans.QueryPlan.$anonfun$mapExpressions$1(QueryPlan.scala:106)
at org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:72)
at org.apache.spark.sql.catalyst.plans.QueryPlan.transformExpression$1(QueryPlan.scala:106)
at org.apache.spark.sql.catalyst.plans.QueryPlan.recursiveTransform$1(QueryPlan.scala:117)
at org.apache.spark.sql.catalyst.plans.QueryPlan.$anonfun$mapExpressions$3(QueryPlan.scala:122)
at scala.collection.TraversableLike.$anonfun$map$1(TraversableLike.scala:237)
at scala.collection.immutable.List.foreach(List.scala:392)
at scala.collection.TraversableLike.map(TraversableLike.scala:237)
at scala.collection.TraversableLike.map$(TraversableLike.scala:230)
at scala.collection.immutable.List.map(List.scala:298)
at org.apache.spark.sql.catalyst.plans.QueryPlan.recursiveTransform$1(QueryPlan.scala:122)
at org.apache.spark.sql.catalyst.plans.QueryPlan.$anonfun$mapExpressions$4(QueryPlan.scala:127)
at org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:189)
at org.apache.spark.sql.catalyst.plans.QueryPlan.mapExpressions(QueryPlan.scala:127)
at org.apache.spark.sql.catalyst.plans.QueryPlan.transformExpressionsUp(QueryPlan.scala:94)
at org.apache.spark.sql.catalyst.analysis.CheckAnalysis.$anonfun$checkAnalysis$1(CheckAnalysis.scala:108)
at org.apache.spark.sql.catalyst.analysis.CheckAnalysis.$anonfun$checkAnalysis$1$adapted(CheckAnalysis.scala:86)
at org.apache.spark.sql.catalyst.trees.TreeNode.foreachUp(TreeNode.scala:129)
at org.apache.spark.sql.catalyst.analysis.CheckAnalysis.checkAnalysis(CheckAnalysis.scala:86)
at org.apache.spark.sql.catalyst.analysis.CheckAnalysis.checkAnalysis$(CheckAnalysis.scala:83)
at org.apache.spark.sql.catalyst.analysis.Analyzer.checkAnalysis(Analyzer.scala:95)
at org.apache.spark.sql.catalyst.analysis.Analyzer.$anonfun$executeAndCheck$1(Analyzer.scala:109)
at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.markInAnalyzer(AnalysisHelper.scala:201)
at org.apache.spark.sql.catalyst.analysis.Analyzer.executeAndCheck(Analyzer.scala:106)
at org.apache.spark.sql.execution.QueryExecution.$anonfun$analyzed$1(QueryExecution.scala:62)
at org.apache.spark.sql.catalyst.QueryPlanningTracker.measurePhase(QueryPlanningTracker.scala:111)
at org.apache.spark.sql.execution.QueryExecution.analyzed$lzycompute(QueryExecution.scala:60)
at org.apache.spark.sql.execution.QueryExecution.analyzed(QueryExecution.scala:60)
at org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:52)
at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:79)
at org.apache.spark.sql.RelationalGroupedDataset.toDF(RelationalGroupedDataset.scala:65)
at org.apache.spark.sql.RelationalGroupedDataset.agg(RelationalGroupedDataset.scala:224)
at org.apache.spark.sql.Dataset.agg(Dataset.scala:1810)
at org.apache.spark.graph.cypher.SparkTable$DataFrameTable.$anonfun$group$12(SparkTable.scala:167)
at scala.util.Either.fold(Either.scala:191)
at org.apache.spark.graph.cypher.SparkTable$DataFrameTable.group(SparkTable.scala:167)
at org.apache.spark.graph.cypher.SparkTable$DataFrameTable.group(SparkTable.scala:21)
at org.opencypher.okapi.relational.impl.operators.Aggregate._table$lzycompute(RelationalOperator.scala:346)
at org.opencypher.okapi.relational.impl.operators.Aggregate._table(RelationalOperator.scala:344)
at org.opencypher.okapi.relational.impl.operators.RelationalOperator.table(RelationalOperator.scala:77)
at org.opencypher.okapi.relational.impl.operators.RelationalOperator._table(RelationalOperator.scala:60)
at org.opencypher.okapi.relational.impl.operators.RelationalOperator.table(RelationalOperator.scala:77)
at org.opencypher.okapi.relational.impl.operators.Select._table$lzycompute(RelationalOperator.scala:320)
at org.opencypher.okapi.relational.impl.operators.Select._table(RelationalOperator.scala:317)
at org.opencypher.okapi.relational.impl.operators.RelationalOperator.table(RelationalOperator.scala:77)
at org.opencypher.okapi.relational.impl.operators.Select._table$lzycompute(RelationalOperator.scala:320)
at org.opencypher.okapi.relational.impl.operators.Select._table(RelationalOperator.scala:317)
at org.opencypher.okapi.relational.impl.operators.RelationalOperator.table(RelationalOperator.scala:77)
at org.opencypher.okapi.relational.api.planning.RelationalCypherResult.$anonfun$getRecords$1(RelationalCypherResult.scala:71)
at scala.Option.flatMap(Option.scala:188)
at org.opencypher.okapi.relational.api.planning.RelationalCypherResult.getRecords(RelationalCypherResult.scala:65)
at org.opencypher.okapi.api.graph.CypherResult.records(CypherResult.scala:71)
at org.opencypher.okapi.api.graph.CypherResult.records$(CypherResult.scala:71)
at org.opencypher.okapi.relational.api.planning.RelationalCypherResult.records(RelationalCypherResult.scala:39)
at org.opencypher.okapi.tck.test.TCKGraph.execute(TCKFixture.scala:96)
at org.opencypher.tools.tck.api.Scenario$ScenarioExecutionContext.execute(Scenario.scala:168)
at org.opencypher.tools.tck.api.Scenario.$anonfun$executeOnGraph$1(Scenario.scala:66)
at scala.collection.LinearSeqOptimized.foldLeft(LinearSeqOptimized.scala:126)
at scala.collection.LinearSeqOptimized.foldLeft$(LinearSeqOptimized.scala:122)
at scala.collection.immutable.List.foldLeft(List.scala:89)
at org.opencypher.tools.tck.api.Scenario.executeOnGraph(Scenario.scala:60)
at org.opencypher.tools.tck.api.Scenario$$anon$1.execute(Scenario.scala:54)
at org.apache.spark.graph.cypher.tck.SparkCypherTckSuite.$anonfun$new$2(SparkCypherTckSuite.scala:32)
at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
at org.scalatest.OutcomeOf.outcomeOf(OutcomeOf.scala:85)
at org.scalatest.OutcomeOf.outcomeOf$(OutcomeOf.scala:83)
at org.scalatest.OutcomeOf$.outcomeOf(OutcomeOf.scala:104)
at org.scalatest.Transformer.apply(Transformer.scala:22)
at org.scalatest.Transformer.apply(Transformer.scala:20)
at org.scalatest.FunSuiteLike$$anon$1.apply(FunSuiteLike.scala:186)
at org.apache.spark.SparkFunSuite.withFixture(SparkFunSuite.scala:104)
at org.scalatest.FunSuiteLike.invokeWithFixture$1(FunSuiteLike.scala:184)
at org.scalatest.FunSuiteLike.$anonfun$runTest$1(FunSuiteLike.scala:196)
at org.scalatest.SuperEngine.runTestImpl(Engine.scala:289)
at org.scalatest.FunSuiteLike.runTest(FunSuiteLike.scala:196)
at org.scalatest.FunSuiteLike.runTest$(FunSuiteLike.scala:178)
at org.apache.spark.graph.cypher.tck.SparkCypherTckSuite.org$scalatest$BeforeAndAfterEach$$super$runTest(SparkCypherTckSuite.scala:18)
at org.scalatest.BeforeAndAfterEach.runTest(BeforeAndAfterEach.scala:221)
at org.scalatest.BeforeAndAfterEach.runTest$(BeforeAndAfterEach.scala:214)
at org.apache.spark.graph.cypher.tck.SparkCypherTckSuite.runTest(SparkCypherTckSuite.scala:18)
at org.scalatest.FunSuiteLike.$anonfun$runTests$1(FunSuiteLike.scala:229)
at org.scalatest.SuperEngine.$anonfun$runTestsInBranch$1(Engine.scala:396)
at scala.collection.immutable.List.foreach(List.scala:392)
at org.scalatest.SuperEngine.traverseSubNodes$1(Engine.scala:384)
at org.scalatest.SuperEngine.runTestsInBranch(Engine.scala:379)
at org.scalatest.SuperEngine.runTestsImpl(Engine.scala:461)
at org.scalatest.FunSuiteLike.runTests(FunSuiteLike.scala:229)
at org.scalatest.FunSuiteLike.runTests$(FunSuiteLike.scala:228)
at org.scalatest.FunSuite.runTests(FunSuite.scala:1560)
at org.scalatest.Suite.run(Suite.scala:1147)
at org.scalatest.Suite.run$(Suite.scala:1129)
at org.scalatest.FunSuite.org$scalatest$FunSuiteLike$$super$run(FunSuite.scala:1560)
at org.scalatest.FunSuiteLike.$anonfun$run$1(FunSuiteLike.scala:233)
at org.scalatest.SuperEngine.runImpl(Engine.scala:521)
at org.scalatest.FunSuiteLike.run(FunSuiteLike.scala:233)
at org.scalatest.FunSuiteLike.run$(FunSuiteLike.scala:232)
at org.apache.spark.SparkFunSuite.org$scalatest$BeforeAndAfterAll$$super$run(SparkFunSuite.scala:53)
at org.scalatest.BeforeAndAfterAll.liftedTree1$1(BeforeAndAfterAll.scala:213)
at org.scalatest.BeforeAndAfterAll.run(BeforeAndAfterAll.scala:210)
at org.scalatest.BeforeAndAfterAll.run$(BeforeAndAfterAll.scala:208)
at org.apache.spark.SparkFunSuite.run(SparkFunSuite.scala:53)
at org.scalatest.tools.SuiteRunner.run(SuiteRunner.scala:45)
at org.scalatest.tools.Runner$.$anonfun$doRunRunRunDaDoRunRun$13(Runner.scala:1340)
at org.scalatest.tools.Runner$.$anonfun$doRunRunRunDaDoRunRun$13$adapted(Runner.scala:1334)
at scala.collection.immutable.List.foreach(List.scala:392)
at org.scalatest.tools.Runner$.doRunRunRunDaDoRunRun(Runner.scala:1334)
at org.scalatest.tools.Runner$.$anonfun$runOptionallyWithPassFailReporter$24(Runner.scala:1031)
at org.scalatest.tools.Runner$.$anonfun$runOptionallyWithPassFailReporter$24$adapted(Runner.scala:1010)
at org.scalatest.tools.Runner$.withClassLoaderAndDispatchReporter(Runner.scala:1500)
at org.scalatest.tools.Runner$.runOptionallyWithPassFailReporter(Runner.scala:1010)
at org.scalatest.tools.Runner$.run(Runner.scala:850)
at org.scalatest.tools.Runner.run(Runner.scala)
at org.jetbrains.plugins.scala.testingSupport.scalaTest.ScalaTestRunner.runScalaTest2(ScalaTestRunner.java:131)
at org.jetbrains.plugins.scala.testingSupport.scalaTest.ScalaTestRunner.main(ScalaTestRunner.java:28)
The text was updated successfully, but these errors were encountered:
The text was updated successfully, but these errors were encountered: