I am getting key error. I checked and there are no empty partitions.
File "zoo_train_model_ray_rec.py", line 64, in <module>
results = est.predict(data,feature_cols=['encoded_input'])
File "/dfs/5/yarn/nm/usercache//appcache/application_1621089916366_56414/container_e41_1621089916366_56414_01_000001/analytics-zoo-bigdl_0.12.1-spark_2.4.3-0.10.0-python-api.zip/zoo/orca/learn/tf2/estimator.py", line 384, in predict
File "/dfs/5/yarn/nm/usercache//appcache/application_1621089916366_56414/container_e41_1621089916366_56414_01_000001/analytics-zoo-bigdl_0.12.1-spark_2.4.3-0.10.0-python-api.zip/zoo/orca/learn/tf2/estimator.py", line 345, in _predict_spark_xshards
File "/dfs/5/yarn/nm/usercache//appcache/application_1621089916366_56414/container_e41_1621089916366_56414_01_000001/analytics-zoo-bigdl_0.12.1-spark_2.4.3-0.10.0-python-api.zip/zoo/orca/data/ray_xshards.py", line 148, in to_spark_xshards
File "/dfs/5/yarn/nm/usercache//appcache/application_1621089916366_56414/container_e41_1621089916366_56414_01_000001/analytics-zoo-bigdl_0.12.1-spark_2.4.3-0.10.0-python-api.zip/zoo/orca/data/shard.py", line 143, in __init__
File "/dfs/5/yarn/nm/usercache//appcache/application_1621089916366_56414/container_e41_1621089916366_56414_01_000001/analytics-zoo-bigdl_0.12.1-spark_2.4.3-0.10.0-python-api.zip/zoo/orca/data/shard.py", line 207, in compute
File "/opt/cloudera/parcels/CDH-6.3.3-1.cdh6.3.3.p4462.8166904/lib/spark/python/lib/pyspark.zip/pyspark/rdd.py", line 1055, in count
File "/opt/cloudera/parcels/CDH-6.3.3-1.cdh6.3.3.p4462.8166904/lib/spark/python/lib/pyspark.zip/pyspark/rdd.py", line 1046, in sum
File "/opt/cloudera/parcels/CDH-6.3.3-1.cdh6.3.3.p4462.8166904/lib/spark/python/lib/pyspark.zip/pyspark/rdd.py", line 917, in fold
File "/opt/cloudera/parcels/CDH-6.3.3-1.cdh6.3.3.p4462.8166904/lib/spark/python/lib/pyspark.zip/pyspark/rdd.py", line 816, in collect
File "/opt/cloudera/parcels/CDH-6.3.3-1.cdh6.3.3.p4462.8166904/lib/spark/python/lib/py4j-0.10.7-src.zip/py4j/java_gateway.py", line 1257, in __call__
File "/opt/cloudera/parcels/CDH-6.3.3-1.cdh6.3.3.p4462.8166904/lib/spark/python/lib/pyspark.zip/pyspark/sql/utils.py", line 63, in deco
File "/opt/cloudera/parcels/CDH-6.3.3-1.cdh6.3.3.p4462.8166904/lib/spark/python/lib/py4j-0.10.7-src.zip/py4j/protocol.py", line 328, in get_return_value
py4j.protocol.Py4JJavaError: An error occurred while calling z:org.apache.spark.api.python.PythonRDD.collectAndServe.
: org.apache.spark.SparkException: Job aborted due to stage failure: Task 167 in stage 6.0 failed 4 times, most recent failure: Lost task 167.3 in stage 6.0 (TID 617,
hdp2stl020061.mastercard.int, executor 2): org.apache.spark.api.python.PythonException: Traceback (most recent call last):
File "/opt/cloudera/parcels/CDH-6.3.3-1.cdh6.3.3.p4462.8166904/lib/spark/python/lib/pyspark.zip/pyspark/worker.py", line 372, in main
process()
File "/opt/cloudera/parcels/CDH-6.3.3-1.cdh6.3.3.p4462.8166904/lib/spark/python/lib/pyspark.zip/pyspark/worker.py", line 367, in process
serializer.dump_stream(func(split_index, iterator), outfile)
File "/dfs/5/yarn/nm/usercache/e112008/appcache/application_1621089916366_56414/container_e41_1621089916366_56414_01_000001/analytics-zoo-bigdl_0.12.1-spark_2.4.3-0.10.0-python-api.zip/zoo/orca/data/ray_xshards.py", line 141, in <lambda>
File "/dfs/12/yarn/nm/usercache/e112008/appcache/application_1621089916366_56414/container_e41_1621089916366_56414_01_000005/analytics-zoo-bigdl_0.12.1-spark_2.4.3-0.10.0-python-api.zip/zoo/orca/data/ray_xshards.py", line 97, in get_from_ray
local_store_handle = ray.get_actor(idx_to_store_name[idx])
KeyError: 167