Hi
I am trying to ingest data both from a S3 bucket and a local filesystem (separately using different ingestion config
files and to different data sources). I am using druid-0.12.0
I have the following set up in common.runtime,properties in both conf/druid/_common and conf-quickstart/druid/_common
directories I have druid_s3_extensions in loadList.
# For S3:
druid.storage.type=s3
druid.storage.bucket=<S3 bucket>
druid.storage.baseKey=druid/segments
druid.s3.accessKey=<access key>
druid.s3.secretKey=<secret key>
But when I have the above the my S3 file data gets ingested properly but I get the below exception when I try reading data from
local file system.
java.lang.Exception: java.io.IOException: No FileSystem for scheme: s3n at org.apache.hadoop.mapred.LocalJobRunner$Job.runTasks(LocalJobRunner.java:462) ~[hadoop-mapreduce-client-common-2.7.3.jar:?] at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:529) [hadoop-mapreduce-client-common-2.7.3.jar:?] Caused by: java.io.IOException: No FileSystem for scheme: s3n at org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2660) ~[hadoop-common-2.7.3.jar:?]
at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2667) ~[hadoop-common-2.7.3.jar:?]
When I comment the above S3 section in both the common.runtime.properties file the ingestion for local file works as expected.
Is this how it is supposed to be? Is there a way to have both work without making any changes to the
common.runtime.properties file?
My S3 file ingestion config file looks like -
"ioConfig" : {
"type" : "index",
"firehose" : {
"type" : "static-s3",
"uris" : [
"s3://<bucket>/<file>",
"s3://<bucket>/<file>",
"s3://<bucket>/<file>"
],
"prefixes" : []
},
"appendToExisting" : true
}My local file ingestion config file looks like this -"ioConfig" : {
"type" : "hadoop",
"inputSpec" : {
"type" : "multi",
"children": [
{
"type" : "static",
"paths" : "<file>"
},
{
"type" : "static",
"paths" : "<file>"
},
{
"type" : "static",
"paths" : "<file>"
}
]
}
}Can someone please let me know what I am doing wrong?Thank you
--
You received this message because you are subscribed to the Google Groups "Druid User" group.
To unsubscribe from this group and stop receiving emails from it, send an email to druid-user+...@googlegroups.com.
To post to this group, send email to druid...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/druid-user/7ad6d345-4790-4af4-ba8e-4bd6697ae76c%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
2018-05-04T00:39:16,002 ERROR [task-runner-0-priority-0] io.druid.indexing.overlord.ThreadPoolTaskRunner - Exception while running task[IndexTask{id=index_04132018_S3_2018-05-04T00:39:10.878Z, type=index, dataSource=04132018_S3}] java.lang.NullPointerException at com.google.common.base.Preconditions.checkNotNull(Preconditions.java:213) ~[guava-16.0.1.jar:?] at io.druid.segment.realtime.firehose.LocalFirehoseFactory.initObjects(LocalFirehoseFactory.java:83) ~[druid-server-0.12.0.jar:0.12.0] at io.druid.data.input.impl.AbstractTextFilesFirehoseFactory.connect(AbstractTextFilesFirehoseFactory.java:57) ~[druid-api-0.12.0.jar:0.12.0] at io.druid.data.input.impl.AbstractTextFilesFirehoseFactory.connect(AbstractTextFilesFirehoseFactory.java:46) ~[druid-api-0.12.0.jar:0.12.0
To view this discussion on the web visit https://groups.google.com/d/msgid/druid-user/43ac1294-5b4c-465c-8e2d-73878096790b%40googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/druid-user/eb3368c8-cb03-409a-abb3-20f01e5263e8%40googlegroups.com.
"ioConfig" : {
"type" : "index",
"firehose" : {
"type": "local",
"inputFormat": "io.druid.data.input.avro.AvroValueInputFormat",
"baseDir" : "<directory>",
"filter" : "<someFile>.avro"
},
"appendToExisting" : true
}
......
......
"tuningConfig" : {
"type" : "index",
"partitionsSpec" : {
"type" : "hashed",
"targetPartitionSize" : 5000000
},
"jobProperties" : {
"avro.schema.input.value.path" : "<path>/<sameSchemaFile_Used_To_Convert_Parquet_To_Avro>.avsc"
}
}The exception I am getting is -java.lang.UnsupportedOperationException: makeParser not supported at io.druid.data.input.avro.AvroParseSpec.makeParser(AvroParseSpec.java:64) ~[?:?] at io.druid.data.input.impl.StringInputRowParser.initializeParser(StringInputRowParser.java:135) ~[druid-api-0.12.0.jar:0.12.0]at io.druid.data.input.impl.StringInputRowParser.startFileFromBeginning(StringInputRowParser.java:141) ~[druid-api-0.12.0.jar:0.12.0] at io.druid.data.input.impl.FileIteratingFirehose.getNextLineIterator(FileIteratingFirehose.java:91) ~[druid-api-0.12.0.jar:0.12.0] at io.druid.data.input.impl.FileIteratingFirehose.hasMore(FileIteratingFirehose.java:67) ~[druid-api-0.12.0.jar:0.12.0] at io.druid.indexing.common.task.IndexTask.generateAndPublishSegments(IndexTask.java:660) ~[druid-indexing-service-0.12.0.jar:0.12.0]