Hey Chris, this sounds pretty interesting - I've been searching the forum.
We have some data, e.g. from 2 sources with 2 different schemas, but with at least a bare minimum common set of columns that we want to GroupBy (A, B). All the other columns are different. I'm trying to imagine how I could use this in:
1. Convert data into a common/merged result set - e.g. maybe with a flexible schema like (type, A, B, originalRecord) -- where type might be a String "someType1" or "someOtherType2" and originalRecord is perhaps JSON representing the whole record of data with all columns in it. Essentially I'm imagining extracting the common columns A, B into a structured schema, and used type as a static field to describe how I can access the arbitrary data in originalRecord and deal with the different schemas there. e.g.
[ "mother", A1, B1, "{'A':'A1','B':'B1','fieldX':value, 'fieldY':value}" ]
[ "father", A1, B1, "{'A':A1', 'B':'B1', 'fieldZ':'somethingElse', 'fieldAAAAA':{object}}" ]
2. During my Every() (now walking in groups of A,B) -- based on the "type" of a record, use various type-specific accessors into the originalRecord data to accomplish what I need. E.g.
if type==mother then extract fieldX
if type==father then etract fieldAAAAA.subtype
#2 sounds like it's addressed in the document. Did it ever get implemented? Also do you have any tips on how to convert a whole record (of Avro) into a single column value ala originalRecord that would be accessible in an Every?
Thanks,
Dusty