R62 enrichment fails at Enrich Raw Events step

101 views
Skip to first unread message

Robert Kingston

unread,
Apr 17, 2015, 12:05:33 AM4/17/15
to snowpl...@googlegroups.com
Hi guys,

We're running r62 with EMR ETL Enrich 0.13.0 and are seeing job failures since upgrading, at the following step:

Elasticity Scalding Step: Enrich Raw Events

stderr is reporting the following error:

Exception in thread "main" cascading.flow.FlowException: step failed: (1/2) .../snowplow/enriched-events, with job id: job_201504170232_0002, please see cluster logs for failure messages
	at cascading.flow.planner.FlowStepJob.blockOnJob(FlowStepJob.java:221)
	at cascading.flow.planner.FlowStepJob.start(FlowStepJob.java:149)
	at cascading.flow.planner.FlowStepJob.call(FlowStepJob.java:124)
	at cascading.flow.planner.FlowStepJob.call(FlowStepJob.java:43)
	at java.util.concurrent.FutureTask.run(FutureTask.java:262)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

Trawling through Hadoop logs, I find the following warnings but no errors or fatal exceptions:

2015-04-17 02:34:23,574 WARN org.apache.hadoop.metrics2.impl.MetricsSystemImpl (main): Source name ugi already exists!
2015-04-17 02:34:23,721 WARN org.apache.hadoop.io.compress.snappy.LoadSnappy (main): Snappy native library is available
2015-04-17 02:34:32,797 WARN org.apache.hadoop.metrics2.impl.MetricsSystemImpl (main): ReduceTask metrics system already initialized!
2015-04-17 02:34:32,797 WARN org.apache.hadoop.metrics2.impl.MetricsSystemImpl (main): Source name jvm already exists!
2015-04-17 02:34:32,815 WARN org.apache.hadoop.metrics2.impl.MetricsSystemImpl (main): Source name ShuffleClientMetrics already exists!
2015-04-17 02:34:38,903 WARN com.hadoop.compression.lzo.LzoCodec (s3distcp-simpler-executor-worker-4): Could not find build properties file with revision hash


Our normal setup allows us to run enrichment within 30 mins, but with 0.13.0, it spiked to 2hrs before failing:

Master: m1.small
Core: m1.small
Task: m3.xlarge x1 @ spot instance

We amped it up to m3.xlarge instances and it failed after just 8 minutes. Our config file is:

:logging:
  :level: DEBUG
:aws:
  :access_key_id: ################
  :secret_access_key: #######################
:s3:
  :region: ap-southeast-2
  :buckets:
    :assets: s3://snowplow-hosted-assets
    :log: s3://#####-logs
    :raw:
      :in: s3://#####-in
:processing: s3://#####-process
:archive: s3://#####-archive/cf
:enriched: :good: s3://#####-out/enriched/good
:bad: s3://#####-out/enriched/bad
:errors: :shredded: :good: s3://#####-out/shredded/good
:bad: s3://#####-out/shredded/bad
:errors: :emr: :ami_version: 2.4.2 :region: ap-southeast-2 :placement: ap-southeast-2b :ec2_subnet_id: :ec2_key_name: ##### :bootstrap: [] :software: :hbase: :lingual: :jobflow: :master_instance_type: m3.xlarge :core_instance_count: 1 :core_instance_type: m3.xlarge :task_instance_count: 1 :task_instance_type: m3.xlarge :task_instance_bid: 0.08 :etl: :job_name: Snowplow Cloudfront Enrichment :versions: :hadoop_enrich: 0.13.0 :hadoop_shred: 0.2.1 :collector_format: cloudfront # Or 'clj-tomcat' for the Clojure Collector :continue_on_unexpected_error: false ...

We have the following enrichments enabled - switching off campaign_attribution.json shaved a good 30% off the failing run time but didn't affect errors:

ip_lookups.json
referer_parser.json

We've also been using our own iglu repository and are sending a few unstructured events through.

Not sure what might be causing this. Have rechecked I've followed the upgrade steps as I've done in the past.

Can anyone see anything wrong with the above?

Cheers,
Rob

Robert Kingston

unread,
Apr 17, 2015, 12:37:17 AM4/17/15
to snowpl...@googlegroups.com
OK, I noticed I needed to update hadoop_shred to 0.3.0. But still no cigar.

Alex Dean

unread,
Apr 17, 2015, 2:59:02 AM4/17/15
to snowpl...@googlegroups.com
Hi Rob,

Which version did you upgrade from?

Thanks,

Alex

--
You received this message because you are subscribed to the Google Groups "Snowplow" group.
To unsubscribe from this group and stop receiving emails from it, send an email to snowplow-use...@googlegroups.com.
For more options, visit https://groups.google.com/d/optout.



--
Co-founder
Snowplow Analytics
The Roma Building, 32-38 Scrutton Street, London EC2A 4RQ, United Kingdom
+44 (0)203 589 6116
+44 7881 622 925
@alexcrdean

Rob Kingston

unread,
Apr 17, 2015, 3:12:37 AM4/17/15
to snowpl...@googlegroups.com

Hi Alex,

We came from 0.9.13.

Cheers,
Rob


You received this message because you are subscribed to a topic in the Google Groups "Snowplow" group.
To unsubscribe from this topic, visit https://groups.google.com/d/topic/snowplow-user/JuNn4BIz_EM/unsubscribe.
To unsubscribe from this group and all its topics, send an email to snowplow-use...@googlegroups.com.

Alex Dean

unread,
Apr 17, 2015, 3:36:01 AM4/17/15
to snowpl...@googlegroups.com
Right, so it sounds like you were upgrading from 0.10.1 to 0.13.0 for the Hadoop Enrich. A few things:
  1. It would be worth trying with each enrichment switched off to see if it's tied to a single enrichment
  2. You can use the --debug flag on EmrEtlRunner for more log output
  3. This page should help you find the Hadoop logs containing the errors: https://github.com/snowplow/snowplow/wiki/Troubleshooting-jobs-on-Elastic-MapReduce
  4. It would be worth trying 0.11.0 and 0.12.0 to see if you can isolate the problem to a version upgrade
  5. A missing schema in your Iglu can cause the job to go very slowly, but I haven't heard of it causing a crash

Hope this helps, let us know what you find out...

A

Robert Kingston

unread,
Apr 17, 2015, 4:22:59 AM4/17/15
to snowpl...@googlegroups.com
Awesome - thanks Alex!

Rolling back to 0.12.0 worked. I'll re-enable my enrichments and report back. 

Will also have a look through the more detailed --debug output for your use here.

Cheers,
Rob

Alex Dean

unread,
Apr 17, 2015, 4:26:44 AM4/17/15
to snowpl...@googlegroups.com
Keep us posted!

Robert Kingston

unread,
Apr 17, 2015, 5:07:53 AM4/17/15
to snowpl...@googlegroups.com
Right - all enrichments switched on (minus anon_ip) works swimmingly under 0.12.0. 

Failed Jobs seem to point to enrichment manager. Hadoop logs show errors like this:

2015-04-17 08:08:08,417 ERROR cascading.flow.stream.TrapHandler (main): caught Throwable, no trap available, rethrowing
cascading.pipe.OperatorException: [com.twitter.scalding.M...][com.twitter.scalding.RichPipe.each(RichPipe.scala:471)] operator Each failed executing operation
at cascading.flow.stream.FunctionEachStage.receive(FunctionEachStage.java:107)
at cascading.flow.stream.FunctionEachStage.receive(FunctionEachStage.java:39)
at cascading.flow.stream.SourceStage.map(SourceStage.java:102)
at cascading.flow.stream.SourceStage.run(SourceStage.java:58)
at cascading.flow.hadoop.FlowMapper.run(FlowMapper.java:130)
at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:441)
at org.apache.hadoop.mapred.MapTask.run(MapTask.java:377)
at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1132)
at org.apache.hadoop.mapred.Child.main(Child.java:249)
Caused by: java.lang.NullPointerException
at com.snowplowanalytics.snowplow.enrich.common.utils.JsonUtils$.stripInstanceEtc(JsonUtils.scala:240)
at com.snowplowanalytics.snowplow.enrich.common.utils.JsonUtils$.extractJson(JsonUtils.scala:204)
at com.snowplowanalytics.snowplow.enrich.common.utils.JsonUtils$.validateAndReformatJson(JsonUtils.scala:189)
at com.snowplowanalytics.snowplow.enrich.common.utils.JsonUtils$$anonfun$1.apply(JsonUtils.scala:59)
at com.snowplowanalytics.snowplow.enrich.common.utils.JsonUtils$$anonfun$1.apply(JsonUtils.scala:58)
at com.snowplowanalytics.snowplow.enrich.common.enrichments.EnrichmentManager$$anonfun$3.apply(EnrichmentManager.scala:106)
at com.snowplowanalytics.snowplow.enrich.common.enrichments.EnrichmentManager$$anonfun$3.apply(EnrichmentManager.scala:106)
at com.snowplowanalytics.snowplow.enrich.common.utils.MapTransformer$$anonfun$1.apply(MapTransformer.scala:158)
at com.snowplowanalytics.snowplow.enrich.common.utils.MapTransformer$$anonfun$1.apply(MapTransformer.scala:155)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244)
at scala.collection.immutable.HashMap$HashMap1.foreach(HashMap.scala:224)
at scala.collection.immutable.HashMap$HashTrieMap.foreach(HashMap.scala:403)
at scala.collection.TraversableLike$class.map(TraversableLike.scala:244)
at scala.collection.AbstractTraversable.map(Traversable.scala:105)
at com.snowplowanalytics.snowplow.enrich.common.utils.MapTransformer$.com$snowplowanalytics$snowplow$enrich$common$utils$MapTransformer$$_transform(MapTransformer.scala:155)
at com.snowplowanalytics.snowplow.enrich.common.utils.MapTransformer$TransformableClass.transform(MapTransformer.scala:132)
at com.snowplowanalytics.snowplow.enrich.common.enrichments.EnrichmentManager$.enrichEvent(EnrichmentManager.scala:191)
at com.snowplowanalytics.snowplow.enrich.common.EtlPipeline$$anonfun$1$$anonfun$apply$1$$anonfun$apply$2$$anonfun$apply$3.apply(EtlPipeline.scala:81)
at com.snowplowanalytics.snowplow.enrich.common.EtlPipeline$$anonfun$1$$anonfun$apply$1$$anonfun$apply$2$$anonfun$apply$3.apply(EtlPipeline.scala:80)
at scalaz.NonEmptyList$class.map(NonEmptyList.scala:29)
at scalaz.NonEmptyListFunctions$$anon$4.map(NonEmptyList.scala:164)
at com.snowplowanalytics.snowplow.enrich.common.EtlPipeline$$anonfun$1$$anonfun$apply$1$$anonfun$apply$2.apply(EtlPipeline.scala:80)
at com.snowplowanalytics.snowplow.enrich.common.EtlPipeline$$anonfun$1$$anonfun$apply$1$$anonfun$apply$2.apply(EtlPipeline.scala:78)
at scalaz.Validation$class.map(Validation.scala:114)
at scalaz.Success.map(Validation.scala:329)
at com.snowplowanalytics.snowplow.enrich.common.EtlPipeline$$anonfun$1$$anonfun$apply$1.apply(EtlPipeline.scala:78)
at com.snowplowanalytics.snowplow.enrich.common.EtlPipeline$$anonfun$1$$anonfun$apply$1.apply(EtlPipeline.scala:76)
at scala.Option.map(Option.scala:145)
at com.snowplowanalytics.snowplow.enrich.common.EtlPipeline$$anonfun$1.apply(EtlPipeline.scala:76)
at com.snowplowanalytics.snowplow.enrich.common.EtlPipeline$$anonfun$1.apply(EtlPipeline.scala:74)
at scalaz.Validation$class.map(Validation.scala:114)
at scalaz.Success.map(Validation.scala:329)
at com.snowplowanalytics.snowplow.enrich.common.EtlPipeline$.processEvents(EtlPipeline.scala:74)
at com.snowplowanalytics.snowplow.enrich.hadoop.EtlJob$$anonfun$7.apply(EtlJob.scala:172)
at com.snowplowanalytics.snowplow.enrich.hadoop.EtlJob$$anonfun$7.apply(EtlJob.scala:171)
at com.twitter.scalding.MapFunction.operate(Operations.scala:58)
at cascading.flow.stream.FunctionEachStage.receive(FunctionEachStage.java:99)
... 11 more





2015-04-17 08:08:08,650 WARN org.apache.hadoop.mapred.Child (main): Error running child
cascading.pipe.OperatorException: [com.twitter.scalding.M...][com.twitter.scalding.RichPipe.each(RichPipe.scala:471)] operator Each failed executing operation
at cascading.flow.stream.FunctionEachStage.receive(FunctionEachStage.java:107)
at cascading.flow.stream.FunctionEachStage.receive(FunctionEachStage.java:39)
at cascading.flow.stream.SourceStage.map(SourceStage.java:102)
at cascading.flow.stream.SourceStage.run(SourceStage.java:58)
at cascading.flow.hadoop.FlowMapper.run(FlowMapper.java:130)
at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:441)
at org.apache.hadoop.mapred.MapTask.run(MapTask.java:377)
at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1132)
at org.apache.hadoop.mapred.Child.main(Child.java:249)
Caused by: java.lang.NullPointerException
at com.snowplowanalytics.snowplow.enrich.common.utils.JsonUtils$.stripInstanceEtc(JsonUtils.scala:240)
at com.snowplowanalytics.snowplow.enrich.common.utils.JsonUtils$.extractJson(JsonUtils.scala:204)
at com.snowplowanalytics.snowplow.enrich.common.utils.JsonUtils$.validateAndReformatJson(JsonUtils.scala:189)
at com.snowplowanalytics.snowplow.enrich.common.utils.JsonUtils$$anonfun$1.apply(JsonUtils.scala:59)
at com.snowplowanalytics.snowplow.enrich.common.utils.JsonUtils$$anonfun$1.apply(JsonUtils.scala:58)
at com.snowplowanalytics.snowplow.enrich.common.enrichments.EnrichmentManager$$anonfun$3.apply(EnrichmentManager.scala:106)
at com.snowplowanalytics.snowplow.enrich.common.enrichments.EnrichmentManager$$anonfun$3.apply(EnrichmentManager.scala:106)
at com.snowplowanalytics.snowplow.enrich.common.utils.MapTransformer$$anonfun$1.apply(MapTransformer.scala:158)
at com.snowplowanalytics.snowplow.enrich.common.utils.MapTransformer$$anonfun$1.apply(MapTransformer.scala:155)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244)
at scala.collection.immutable.HashMap$HashMap1.foreach(HashMap.scala:224)
at scala.collection.immutable.HashMap$HashTrieMap.foreach(HashMap.scala:403)
at scala.collection.TraversableLike$class.map(TraversableLike.scala:244)
at scala.collection.AbstractTraversable.map(Traversable.scala:105)
at com.snowplowanalytics.snowplow.enrich.common.utils.MapTransformer$.com$snowplowanalytics$snowplow$enrich$common$utils$MapTransformer$$_transform(MapTransformer.scala:155)
at com.snowplowanalytics.snowplow.enrich.common.utils.MapTransformer$TransformableClass.transform(MapTransformer.scala:132)
at com.snowplowanalytics.snowplow.enrich.common.enrichments.EnrichmentManager$.enrichEvent(EnrichmentManager.scala:191)
at com.snowplowanalytics.snowplow.enrich.common.EtlPipeline$$anonfun$1$$anonfun$apply$1$$anonfun$apply$2$$anonfun$apply$3.apply(EtlPipeline.scala:81)
at com.snowplowanalytics.snowplow.enrich.common.EtlPipeline$$anonfun$1$$anonfun$apply$1$$anonfun$apply$2$$anonfun$apply$3.apply(EtlPipeline.scala:80)
at scalaz.NonEmptyList$class.map(NonEmptyList.scala:29)
at scalaz.NonEmptyListFunctions$$anon$4.map(NonEmptyList.scala:164)
at com.snowplowanalytics.snowplow.enrich.common.EtlPipeline$$anonfun$1$$anonfun$apply$1$$anonfun$apply$2.apply(EtlPipeline.scala:80)
at com.snowplowanalytics.snowplow.enrich.common.EtlPipeline$$anonfun$1$$anonfun$apply$1$$anonfun$apply$2.apply(EtlPipeline.scala:78)
at scalaz.Validation$class.map(Validation.scala:114)
at scalaz.Success.map(Validation.scala:329)
at com.snowplowanalytics.snowplow.enrich.common.EtlPipeline$$anonfun$1$$anonfun$apply$1.apply(EtlPipeline.scala:78)
at com.snowplowanalytics.snowplow.enrich.common.EtlPipeline$$anonfun$1$$anonfun$apply$1.apply(EtlPipeline.scala:76)
at scala.Option.map(Option.scala:145)
at com.snowplowanalytics.snowplow.enrich.common.EtlPipeline$$anonfun$1.apply(EtlPipeline.scala:76)
at com.snowplowanalytics.snowplow.enrich.common.EtlPipeline$$anonfun$1.apply(EtlPipeline.scala:74)
at scalaz.Validation$class.map(Validation.scala:114)
at scalaz.Success.map(Validation.scala:329)
at com.snowplowanalytics.snowplow.enrich.common.EtlPipeline$.processEvents(EtlPipeline.scala:74)
at com.snowplowanalytics.snowplow.enrich.hadoop.EtlJob$$anonfun$7.apply(EtlJob.scala:172)
at com.snowplowanalytics.snowplow.enrich.hadoop.EtlJob$$anonfun$7.apply(EtlJob.scala:171)
at com.twitter.scalding.MapFunction.operate(Operations.scala:58)
at cascading.flow.stream.FunctionEachStage.receive(FunctionEachStage.java:99)
... 11 more

Alex Dean

unread,
Apr 17, 2015, 5:13:16 AM4/17/15
to snowpl...@googlegroups.com
Hey Rob,

Aha! Could you do a run with the error bucket switched on, and then share the row(s) in the error bucket (that will be the one(s) causing the NPE.

Are you saying those NPEs are in 0.13.0 only or in 0.12.0 as well?

A

Robert Kingston

unread,
Apr 17, 2015, 7:07:10 AM4/17/15
to snowpl...@googlegroups.com
Here we go... Look at these ghastly referers. 

The NPEs were only in 0.13.0. They passed through 0.12.0 like water through a sponge. There is no record of the event below showing in enriched output.


2015-04-15 07:55:22 SFO5 474 150.70.173.58 GET d1warv1mtvh6n4.cloudfront.net /i 200 - Mozilla/4.0%2520(compatible;%2520MSIE%25208.0;%2520Windows%2520NT%25205.1;%2520Trident/4.0) e=pv&page=Southbank%2520Hotels%2520Melbourne%2520%257C%2520Travelodge%2520Southbank%2520-%2520Hotel%2520Overview%2520%257C%2520Hotels%2520in%2520Southbank&dtm=1429084433162&tid=591837&vp=1920x916&ds=1920x916&vid=1&duid=2fc6905c75e3628c&p=web&tv=js-2.0.2&fp=247302420&aid=tfe&lang=en-US&cs=UTF-8&tz=Australia%252FSydney&tna=cf&refr=http%253A%252F%252Fwww.google.com%252Fafs%252Fads%253Fq%253Dtravel%252520lodge%252520southbank%2526lo%253Dtrue%2526adpage%253D1%2526adrep%253D3%2526po%253Dtrue%2526r%253Dm%2526fexp%253D21404%25252C7000108%2526client%253Daj-fot-hp%2526channel%253DAPN11737%25252Cgeo-AU%25252Cbrow_FF%25252Csp_testWB%25252Clang_en%25252Ctest_15608%25252Ctpr_1%2526hl%253Den%2526adtest%253Doff%2526adsafe%253Dmedium%2526type%253D0%2526oe%253DUTF-8%2526ie%253DUTF-8%2526jsei%253D4%2526format%253Dp7%25257Cn3%2526ad%253Dn3p7%2526nocache%253D6441429084397534%2526num%253D0%2526output%253Duds_ads_only%2526v%253D3%2526adlh%253Don%2526adext%253Das1%25252Cl1%25252Csr1%2526bsl%253D8%2526u_his%253D2%2526u_tz%253D600%2526dt%253D1429084397535%2526u_w%253D1920%2526u_h%253D1080%2526biw%253D1903%2526bih%253D916%2526psw%253D1903%2526psh%253D916%2526frm%253D0%2526uio%253Duv3cs1va1sl1sr1lo1po1cc0st18sv13sd13lH-ff2lt22-ff2lt22%2526rurl%253Dhttp%25253A%25252F%25252Fwww.search.ask.com%25252Fweb%25253Fq%25253Dtravel%25252Blodge%25252Bsouthbank%252526apn_dtid%25253D%2525255Eaaa204%2525255EYY%2525255EAU%252526apn_dbr%25253Dff_36.0%252526psv%25253D%252526itbv%25253D12.27.0.937%252526p2%25253D%2525255EBPQ%2525255Eaaa204%2525255EYY%2525255EAU%252526apn_ptnrs%25253D%2525255EBPQ%252526o%25253DAPN11737%252526gct%25253Dhp%252526tpid%25253DAVGSP-LS%252526pf%25253DV7%252526trgb%25253DFF%252526pt%25253Dtb%252526apn_uid%25253D265A4847-ACC4-46EC-A11D-79CD594A90BC%252526tpr%25253D1%252526doi%25253D2015-04-04%252526ts%25253D1429084395801%2526referer%253Dhttp%25253A%25252F%25252Fwww.search.ask.com%25252F%25253Ftpid%25253DAVGSP-LS%252526o%25253DAPN11737%252526pf%25253DV7%252526trgb%25253DFF%252526p2%25253D%2525255EBPQ%2525255Eaaa204%2525255EYY%2525255EAU%252526gct%25253Dhp%252526apn_ptnrs%25253D%2525255EBPQ%252526apn_dtid%25253D%2525255Eaaa204%2525255EYY%2525255EAU%252526apn_dbr%25253Dff_36.0%252526apn_uid%25253D265A4847-ACC4-46EC-A11D-79CD594A90BC%252526itbv%25253D12.27.0.937%252526doi%25253D2015-04-04%252526psv%25253D%252526pt%25253Dtb&f_pdf=1&f_qt=0&f_realp=0&f_wma=0&f_dir=0&f_fla=1&f_java=0&f_gears=0&f_ag=1&res=1920x1080&cd=24&co - Hit SANbuapY8H_BvX9Uop-3l1fUdTcZT7dopy17k-8TfcLjEKwIP9TxBQ== d1warv1mtvh6n4.cloudfront.net http 2191 0.000

2015-04-15 07:55:22 SFO5 474 150.70.173.58 GET d1warv1mtvh6n4.cloudfront.net /i 200 - Mozilla/4.0%2520(compatible;%2520MSIE%25208.0;%2520Windows%2520NT%25205.1;%2520Trident/4.0) e=pv&page=Southbank%2520Hotels%2520Melbourne%2520%257C%2520Travelodge%2520Southbank%2520-%2520Hotel%2520Overview%2520%257C%2520Hotels%2520in%2520Southbank&dtm=1429084433162&tid=591837&vp=1920x916&ds=1920x916&vid=1&duid=2fc6905c75e3628c&p=web&tv=js-2.0.2&fp=247302420&aid=tfe&lang=en-US&cs=UTF-8&tz=Australia%252FSydney&tna=cf&refr=http%253A%252F%252Fwww.google.com%252Fafs%252Fads%253Fq%253Dtravel%252520lodge%252520southbank%2526lo%253Dtrue%2526adpage%253D1%2526adrep%253D3%2526po%253Dtrue%2526r%253Dm%2526fexp%253D21404%25252C7000108%2526client%253Daj-fot-hp%2526channel%253DAPN11737%25252Cgeo-AU%25252Cbrow_FF%25252Csp_testWB%25252Clang_en%25252Ctest_15608%25252Ctpr_1%2526hl%253Den%2526adtest%253Doff%2526adsafe%253Dmedium%2526type%253D0%2526oe%253DUTF-8%2526ie%253DUTF-8%2526jsei%253D4%2526format%253Dp7%25257Cn3%2526ad%253Dn3p7%2526nocache%253D6441429084397534%2526num%253D0%2526output%253Duds_ads_only%2526v%253D3%2526adlh%253Don%2526adext%253Das1%25252Cl1%25252Csr1%2526bsl%253D8%2526u_his%253D2%2526u_tz%253D600%2526dt%253D1429084397535%2526u_w%253D1920%2526u_h%253D1080%2526biw%253D1903%2526bih%253D916%2526psw%253D1903%2526psh%253D916%2526frm%253D0%2526uio%253Duv3cs1va1sl1sr1lo1po1cc0st18sv13sd13lH-ff2lt22-ff2lt22%2526rurl%253Dhttp%25253A%25252F%25252Fwww.search.ask.com%25252Fweb%25253Fq%25253Dtravel%25252Blodge%25252Bsouthbank%252526apn_dtid%25253D%2525255Eaaa204%2525255EYY%2525255EAU%252526apn_dbr%25253Dff_36.0%252526psv%25253D%252526itbv%25253D12.27.0.937%252526p2%25253D%2525255EBPQ%2525255Eaaa204%2525255EYY%2525255EAU%252526apn_ptnrs%25253D%2525255EBPQ%252526o%25253DAPN11737%252526gct%25253Dhp%252526tpid%25253DAVGSP-LS%252526pf%25253DV7%252526trgb%25253DFF%252526pt%25253Dtb%252526apn_uid%25253D265A4847-ACC4-46EC-A11D-79CD594A90BC%252526tpr%25253D1%252526doi%25253D2015-04-04%252526ts%25253D1429084395801%2526referer%253Dhttp%25253A%25252F%25252Fwww.search.ask.com%25252F%25253Ftpid%25253DAVGSP-LS%252526o%25253DAPN11737%252526pf%25253DV7%252526trgb%25253DFF%252526p2%25253D%2525255EBPQ%2525255Eaaa204%2525255EYY%2525255EAU%252526gct%25253Dhp%252526apn_ptnrs%25253D%2525255EBPQ%252526apn_dtid%25253D%2525255Eaaa204%2525255EYY%2525255EAU%252526apn_dbr%25253Dff_36.0%252526apn_uid%25253D265A4847-ACC4-46EC-A11D-79CD594A90BC%252526itbv%25253D12.27.0.937%252526doi%25253D2015-04-04%252526psv%25253D%252526pt%25253Dtb&f_pdf=1&f_qt=0&f_realp=0&f_wma=0&f_dir=0&f_fla=1&f_java=0&f_gears=0&f_ag=1&res=1920x1080&cd=24&co - Hit SANbuapY8H_BvX9Uop-3l1fUdTcZT7dopy17k-8TfcLjEKwIP9TxBQ== d1warv1mtvh6n4.cloudfront.net http 2191 0.000

Alex Dean

unread,
Apr 17, 2015, 7:29:14 AM4/17/15
to snowpl...@googlegroups.com
Ah! Those referers are causing the context (that &co= at the end) to truncate (IE GET limitations). That is then tripping this bug:

https://github.com/snowplow/snowplow/issues/1622

causing an NPE which causes the load to fail. Temporary workaround for you Rob: keep the error bucket enabled. We'll fix that NPE in a release soon.

Also, if you are sending rich contexts with events and they are getting truncated, consider switching to the Clojure Collector and POST...

Cheers,

Alex
Reply all
Reply to author
Forward
0 new messages