Best way to validate custom context data?


#1

We rolled out some complex custom contexts today, and our testing process devolved into our engineers firing events, and then us running the EMR ETL and checking the enriched-bad folder on S3 for any problems.
I’m sure there’s a better way - how do you solve this problem?


#2

You might be able to use the Schema validation service for this. While I haven’t used it, it appears to validate not just schema format, but instances of a schema against a definition.


#3

Another option (in addition to what @njenkins has mentioned above) we use is to test and validate all schemas by first pushing them to Iglu on a Snowplow Mini instance (using igluctl) before pushing them into production. This allows us to validate that the data will be enriched/processed correctly and that the result that is populated in Elasticsearch is what we expect.


#4

Thanks for the answers so far. We’re going to at least start using http://www.jsonschemavalidator.net for manual testing.


#5

@dweitzenfeld, please bear in mind that jsonschemavalidator.net doesn’t strictly follows JSON Schema 4 specification. Particularly it will try to validate strings against unexisting date format (date-time is correct). You can find more information here (although another online validator is mentioned).


#6

Thanks for the heads up @anton