i trying to write, time based buckets in hdfs. but It is failing job, after added TimeBasedWriterPartitioner properties. please check what is wrong in properties file. but it is working SimpleDataWriterBuilder.
2015-11-29 21:57:21 PST INFO [main] org.quartz.impl.StdSchedulerFactory 1184 - Using default implementation for ThreadExecutor
2015-11-29 21:57:21 PST INFO [main] org.quartz.core.SchedulerSignalerImpl 61 - Initialized Scheduler Signaller of type: class org.quartz.core.SchedulerSignalerImpl
2015-11-29 21:57:21 PST INFO [main] org.quartz.core.QuartzScheduler 240 - Quartz Scheduler v.2.2.1 created.
2015-11-29 21:57:21 PST INFO [main] org.quartz.simpl.RAMJobStore 155 - RAMJobStore initialized.
2015-11-29 21:57:21 PST INFO [main] org.quartz.core.QuartzScheduler 305 - Scheduler meta-data: Quartz Scheduler (v2.2.1) 'LocalJobScheduler' with instanceId 'NON_CLUSTERED'
Scheduler class: 'org.quartz.core.QuartzScheduler' - running locally.
NOT STARTED.
Currently in standby mode.
Number of jobs executed: 0
Using thread pool 'org.quartz.simpl.SimpleThreadPool' - with 3 threads.
Using job-store 'org.quartz.simpl.RAMJobStore' - which does not support persistence. and is not clustered.
2015-11-29 21:57:21 PST INFO [main] org.quartz.impl.StdSchedulerFactory 1339 - Quartz scheduler 'LocalJobScheduler' initialized from specified file: '/Users/Kalyan/gobblin-dist/conf/quartz.properties'
2015-11-29 21:57:21 PST INFO [main] org.quartz.impl.StdSchedulerFactory 1343 - Quartz scheduler version: 2.2.1
2015-11-29 21:57:21 PST INFO [main] gobblin.scheduler.SchedulerDaemon 107 - Starting the scheduler daemon
2015-11-29 21:57:21 PST INFO [JobScheduler STARTING] gobblin.scheduler.JobScheduler 137 - Starting the job scheduler
2015-11-29 21:57:21 PST INFO [JobScheduler STARTING] org.quartz.core.QuartzScheduler 575 - Scheduler LocalJobScheduler_$_NON_CLUSTERED started.
2015-11-29 21:57:21 PST INFO [JobScheduler STARTING] gobblin.scheduler.JobScheduler 356 - Scheduling locally configured jobs
2015-11-29 21:57:21 PST INFO [JobScheduler STARTING] gobblin.scheduler.JobScheduler 369 - Loaded 2 job configurations
2015-11-29 21:57:21 PST WARN [JobScheduler STARTING] gobblin.scheduler.JobScheduler 203 - Job SHP_KafkaToHdfs has already been scheduled
2015-11-29 21:58:00 PST WARN [LocalJobScheduler_Worker-1] org.apache.hadoop.util.NativeCodeLoader 62 - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
2015-11-29 21:58:00 PST WARN [LocalJobScheduler_Worker-1] gobblin.runtime.JobContext 197 - Property task.data.root.dir is missing.
2015-11-29 21:58:00 PST INFO [TaskExecutor STARTING] gobblin.runtime.TaskExecutor 119 - Starting the task executor
2015-11-29 21:58:00 PST INFO [LocalTaskStateTracker STARTING] gobblin.runtime.AbstractTaskStateTracker 64 - Starting the task state tracker
2015-11-29 21:58:00 PST INFO [LocalJobScheduler_Worker-1] gobblin.metrics.GobblinMetrics 429 - Not reporting metrics to JMX
2015-11-29 21:58:00 PST INFO [LocalJobScheduler_Worker-1] gobblin.metrics.GobblinMetrics 379 - Not reporting metrics to log files
2015-11-29 21:58:00 PST INFO [LocalJobScheduler_Worker-1] gobblin.metrics.GobblinMetrics 440 - Not reporting metrics to Kafka
2015-11-29 21:58:00 PST INFO [LocalJobScheduler_Worker-1] gobblin.source.extractor.extract.kafka.KafkaWrapper$KafkaOldAPI 233 - Fetching topic metadata from broker localhost:9091
2015-11-29 21:58:01 PST WARN [LocalJobScheduler_Worker-1] gobblin.source.extractor.extract.kafka.KafkaSource 288 - Previous offset for partition shp.shipment:0 does not exist. This partition will start from the earliest offset: 0
2015-11-29 21:58:01 PST INFO [LocalJobScheduler_Worker-1] gobblin.source.extractor.extract.kafka.KafkaSource 407 - Created workunit for partition shp.shipment:0: lowWatermark=0, highWatermark=16, range=16
2015-11-29 21:58:01 PST WARN [LocalJobScheduler_Worker-1] gobblin.source.extractor.extract.kafka.KafkaSource 288 - Previous offset for partition shp.shipment:1 does not exist. This partition will start from the earliest offset: 0
2015-11-29 21:58:01 PST INFO [LocalJobScheduler_Worker-1] gobblin.source.extractor.extract.kafka.KafkaSource 407 - Created workunit for partition shp.shipment:1: lowWatermark=0, highWatermark=58, range=58
2015-11-29 21:58:01 PST WARN [LocalJobScheduler_Worker-1] gobblin.source.extractor.extract.kafka.KafkaSource 288 - Previous offset for partition shp.cgomvmtspec:0 does not exist. This partition will start from the earliest offset: 0
2015-11-29 21:58:01 PST INFO [LocalJobScheduler_Worker-1] gobblin.source.extractor.extract.kafka.KafkaSource 407 - Created workunit for partition shp.cgomvmtspec:0: lowWatermark=0, highWatermark=29, range=29
2015-11-29 21:58:01 PST WARN [LocalJobScheduler_Worker-1] gobblin.source.extractor.extract.kafka.KafkaSource 288 - Previous offset for partition shp.cgomvmtspec:1 does not exist. This partition will start from the earliest offset: 0
2015-11-29 21:58:01 PST INFO [LocalJobScheduler_Worker-1] gobblin.source.extractor.extract.kafka.KafkaSource 407 - Created workunit for partition shp.cgomvmtspec:1: lowWatermark=0, highWatermark=0, range=0
2015-11-29 21:58:01 PST INFO [LocalJobScheduler_Worker-1] gobblin.source.extractor.extract.kafka.workunit.packer.KafkaAvgRecordTimeBasedWorkUnitSizeEstimator 145 - For all topics not pulled in the previous run, estimated avg time to pull a record is 1.0 milliseconds
2015-11-29 21:58:01 PST INFO [LocalJobScheduler_Worker-1] gobblin.source.extractor.extract.kafka.workunit.packer.KafkaWorkUnitPacker 219 - Created MultiWorkUnit for partitions [shp.cgomvmtspec:1]
2015-11-29 21:58:01 PST INFO [LocalJobScheduler_Worker-1] gobblin.source.extractor.extract.kafka.workunit.packer.KafkaWorkUnitPacker 293 - MultiWorkUnit 0: estimated load=0.003010, partitions=[]
2015-11-29 21:58:01 PST INFO [LocalJobScheduler_Worker-1] gobblin.source.extractor.extract.kafka.workunit.packer.KafkaWorkUnitPacker 293 - MultiWorkUnit 1: estimated load=17.459740, partitions=[[shp.shipment:1]]
2015-11-29 21:58:01 PST INFO [LocalJobScheduler_Worker-1] gobblin.source.extractor.extract.kafka.workunit.packer.KafkaWorkUnitPacker 293 - MultiWorkUnit 2: estimated load=0.003010, partitions=[]
2015-11-29 21:58:01 PST INFO [LocalJobScheduler_Worker-1] gobblin.source.extractor.extract.kafka.workunit.packer.KafkaWorkUnitPacker 293 - MultiWorkUnit 3: estimated load=4.816480, partitions=[[shp.shipment:0]]
2015-11-29 21:58:01 PST INFO [LocalJobScheduler_Worker-1] gobblin.source.extractor.extract.kafka.workunit.packer.KafkaWorkUnitPacker 293 - MultiWorkUnit 4: estimated load=0.003010, partitions=[[shp.cgomvmtspec:1]]
2015-11-29 21:58:01 PST INFO [LocalJobScheduler_Worker-1] gobblin.source.extractor.extract.kafka.workunit.packer.KafkaWorkUnitPacker 293 - MultiWorkUnit 5: estimated load=0.003010, partitions=[]
2015-11-29 21:58:01 PST INFO [LocalJobScheduler_Worker-1] gobblin.source.extractor.extract.kafka.workunit.packer.KafkaWorkUnitPacker 293 - MultiWorkUnit 6: estimated load=0.003010, partitions=[]
2015-11-29 21:58:01 PST INFO [LocalJobScheduler_Worker-1] gobblin.source.extractor.extract.kafka.workunit.packer.KafkaWorkUnitPacker 293 - MultiWorkUnit 7: estimated load=8.729870, partitions=[[shp.cgomvmtspec:0]]
2015-11-29 21:58:01 PST INFO [LocalJobScheduler_Worker-1] gobblin.source.extractor.extract.kafka.workunit.packer.KafkaWorkUnitPacker 279 - Min load of multiWorkUnit = 0.003010; Max load of multiWorkUnit = 17.459740; Diff = 99.982759%
2015-11-29 21:58:01 PST INFO [LocalJobScheduler_Worker-1] gobblin.util.ExecutorsUtils 108 - Attempting to shutdown ExecutorSerivce: java.util.concurrent.ThreadPoolExecutor@74b83dac[Terminated, pool size = 0, active threads = 0, queued tasks = 0, completed tasks = 0]
2015-11-29 21:58:01 PST INFO [LocalJobScheduler_Worker-1] gobblin.util.ExecutorsUtils 127 - Successfully shutdown ExecutorService: java.util.concurrent.ThreadPoolExecutor@74b83dac[Terminated, pool size = 0, active threads = 0, queued tasks = 0, completed tasks = 0]
2015-11-29 21:58:01 PST INFO [LocalJobScheduler_Worker-1] gobblin.runtime.AbstractJobLauncher 234 - Starting job job_SHP_KafkaToHdfs_1448863080018
2015-11-29 21:58:01 PST INFO [LocalJobScheduler_Worker-1] gobblin.runtime.TaskExecutor 152 - Executing task task_SHP_KafkaToHdfs_1448863080018_0
2015-11-29 21:58:01 PST INFO [LocalJobScheduler_Worker-1] gobblin.runtime.TaskExecutor 152 - Executing task task_SHP_KafkaToHdfs_1448863080018_1
2015-11-29 21:58:01 PST INFO [LocalJobScheduler_Worker-1] gobblin.runtime.TaskExecutor 152 - Executing task task_SHP_KafkaToHdfs_1448863080018_2
2015-11-29 21:58:01 PST INFO [LocalJobScheduler_Worker-1] gobblin.runtime.TaskExecutor 152 - Executing task task_SHP_KafkaToHdfs_1448863080018_3
2015-11-29 21:58:01 PST INFO [LocalJobScheduler_Worker-1] gobblin.runtime.local.LocalJobLauncher 119 - Waiting for submitted tasks of job job_SHP_KafkaToHdfs_1448863080018 to complete...
2015-11-29 21:58:01 PST INFO [LocalJobScheduler_Worker-1] gobblin.runtime.local.LocalJobLauncher 121 - 4 out of 4 tasks of job job_SHP_KafkaToHdfs_1448863080018 are running
2015-11-29 21:58:01 PST INFO [TaskExecutor-1] gobblin.source.extractor.extract.kafka.KafkaExtractor 190 - Pulling topic shp.shipment
2015-11-29 21:58:01 PST INFO [TaskExecutor-0] gobblin.source.extractor.extract.kafka.KafkaExtractor 190 - Pulling topic shp.shipment
2015-11-29 21:58:01 PST INFO [TaskExecutor-1] gobblin.source.extractor.extract.kafka.KafkaExtractor 210 - Pulling partition shp.shipment:0 from offset 0 to 16, range=16
2015-11-29 21:58:01 PST INFO [TaskExecutor-0] gobblin.source.extractor.extract.kafka.KafkaExtractor 210 - Pulling partition shp.shipment:1 from offset 0 to 58, range=58
2015-11-29 21:58:01 PST ERROR [TaskExecutor-0] gobblin.runtime.Task 254 - Task task_SHP_KafkaToHdfs_1448863080018_0 failed
java.lang.IllegalStateException: Fork 0 of task task_SHP_KafkaToHdfs_1448863080018_0 has failed and is no longer running
at gobblin.runtime.Fork.putRecord(Fork.java:218)
at gobblin.runtime.Task.processRecord(Task.java:464)
at gobblin.runtime.Task.run(Task.java:173)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
2015-11-29 21:58:01 PST ERROR [ForkExecutor-0] gobblin.runtime.Fork 180 - Fork 0 of task task_SHP_KafkaToHdfs_1448863080018_1 failed to process data records
java.lang.RuntimeException: java.lang.InstantiationException
at gobblin.runtime.TaskContext.getDataWriterBuilder(TaskContext.java:302)
at gobblin.runtime.Fork.buildWriter(Fork.java:357)
at gobblin.runtime.Fork.buildWriterIfNotPresent(Fork.java:371)
at gobblin.runtime.Fork.processRecords(Fork.java:391)
at gobblin.runtime.Fork.run(Fork.java:166)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.InstantiationException
at sun.reflect.InstantiationExceptionConstructorAccessorImpl.newInstance(InstantiationExceptionConstructorAccessorImpl.java:48)
at java.lang.reflect.Constructor.newInstance(Constructor.java:422)
at java.lang.Class.newInstance(Class.java:442)
at gobblin.runtime.TaskContext.getDataWriterBuilder(TaskContext.java:298)
... 11 more
2015-11-29 21:58:01 PST ERROR [TaskExecutor-1] gobblin.runtime.Task 254 - Task task_SHP_KafkaToHdfs_1448863080018_1 failed
java.lang.IllegalStateException: Fork 0 of task task_SHP_KafkaToHdfs_1448863080018_1 has failed and is no longer running
at gobblin.runtime.Fork.putRecord(Fork.java:218)
at gobblin.runtime.Task.processRecord(Task.java:464)
at gobblin.runtime.Task.run(Task.java:173)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
2015-11-29 21:58:01 PST ERROR [ForkExecutor-1] gobblin.runtime.Fork 180 - Fork 0 of task task_SHP_KafkaToHdfs_1448863080018_0 failed to process data records
java.lang.RuntimeException: java.lang.InstantiationException
at gobblin.runtime.TaskContext.getDataWriterBuilder(TaskContext.java:302)
at gobblin.runtime.Fork.buildWriter(Fork.java:357)
at gobblin.runtime.Fork.buildWriterIfNotPresent(Fork.java:371)
at gobblin.runtime.Fork.processRecords(Fork.java:391)
at gobblin.runtime.Fork.run(Fork.java:166)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.InstantiationException
at sun.reflect.InstantiationExceptionConstructorAccessorImpl.newInstance(InstantiationExceptionConstructorAccessorImpl.java:48)
at java.lang.reflect.Constructor.newInstance(Constructor.java:422)
at java.lang.Class.newInstance(Class.java:442)
at gobblin.runtime.TaskContext.getDataWriterBuilder(TaskContext.java:298)
... 11 more
2015-11-29 21:58:01 PST INFO [TaskExecutor-1] gobblin.source.extractor.extract.kafka.KafkaExtractor 266 - Actual high watermark for partition shp.shipment:0=7, expected=16
2015-11-29 21:58:01 PST INFO [TaskExecutor-0] gobblin.source.extractor.extract.kafka.KafkaExtractor 266 - Actual high watermark for partition shp.shipment:1=5, expected=58
2015-11-29 21:58:01 PST INFO [TaskExecutor-1] gobblin.source.extractor.extract.kafka.KafkaExtractor 278 - Avg time to pull a record for partition shp.shipment:0 not recorded
2015-11-29 21:58:01 PST INFO [TaskExecutor-0] gobblin.source.extractor.extract.kafka.KafkaExtractor 278 - Avg time to pull a record for partition shp.shipment:1 not recorded
2015-11-29 21:58:01 PST INFO [TaskExecutor-0] gobblin.runtime.Task 277 - publish.data.at.job.level is true. Will publish data at the job level.
2015-11-29 21:58:01 PST INFO [TaskExecutor-1] gobblin.runtime.Task 277 - publish.data.at.job.level is true. Will publish data at the job level.
2015-11-29 21:58:01 PST INFO [TaskExecutor-1] gobblin.runtime.TaskExecutor 205 - Scheduled retry of failed task task_SHP_KafkaToHdfs_1448863080018_1 to run in 0 seconds
2015-11-29 21:58:01 PST INFO [TaskExecutor-0] gobblin.runtime.TaskExecutor 205 - Scheduled retry of failed task task_SHP_KafkaToHdfs_1448863080018_0 to run in 0 seconds
2015-11-29 21:58:01 PST INFO [TaskRetryExecutor-0] gobblin.source.extractor.extract.kafka.KafkaExtractor 190 - Pulling topic shp.shipment
2015-11-29 21:58:01 PST INFO [TaskExecutor-1] gobblin.source.extractor.extract.kafka.KafkaExtractor 190 - Pulling topic shp.cgomvmtspec
2015-11-29 21:58:01 PST INFO [TaskRetryExecutor-0] gobblin.source.extractor.extract.kafka.KafkaExtractor 210 - Pulling partition shp.shipment:1 from offset 0 to 58, range=58
2015-11-29 21:58:01 PST INFO [TaskExecutor-1] gobblin.source.extractor.extract.kafka.KafkaExtractor 210 - Pulling partition shp.cgomvmtspec:1 from offset 0 to 0, range=0
2015-11-29 21:58:01 PST INFO [TaskExecutor-1] gobblin.source.extractor.extract.kafka.KafkaExtractor 177 - Finished pulling partition shp.cgomvmtspec:1
2015-11-29 21:58:01 PST INFO [TaskExecutor-1] gobblin.source.extractor.extract.kafka.KafkaExtractor 165 - Finished pulling topic shp.cgomvmtspec
2015-11-29 21:58:01 PST INFO [TaskExecutor-0] gobblin.source.extractor.extract.kafka.KafkaExtractor 190 - Pulling topic shp.cgomvmtspec
2015-11-29 21:58:01 PST INFO [TaskExecutor-1] gobblin.runtime.Task 177 - Extracted 0 data records
2015-11-29 21:58:01 PST INFO [TaskExecutor-0] gobblin.source.extractor.extract.kafka.KafkaExtractor 210 - Pulling partition shp.cgomvmtspec:0 from offset 0 to 29, range=29
2015-11-29 21:58:01 PST INFO [TaskExecutor-1] gobblin.runtime.Task 178 - Row quality checker finished with results:
2015-11-29 21:58:01 PST ERROR [ForkExecutor-1] gobblin.runtime.Fork 180 - Fork 0 of task task_SHP_KafkaToHdfs_1448863080018_0 failed to process data records
java.lang.RuntimeException: java.lang.InstantiationException
at gobblin.runtime.TaskContext.getDataWriterBuilder(TaskContext.java:302)