Empty s3 shredded logs after successful EmrEtlRunner job


#1

I have successfully run the EmrEtlRunner job but after the successful completion of the job when I look into the s3 buckets they are empty.

Can anyone please help me out that why I am getting empty shredded, enriched and raw files in the s3 bucket I created.

This is my config.yml file

aws:
  # Credentials can be hardcoded or set in environment variables
  access_key_id: ********
  secret_access_key: ******
  s3:
    region: us-east-2
    buckets:
      assets: s3://snowplow-hosted-assets # DO NOT CHANGE unless you are hosting the jarfiles etc yourself in your own bucket
      jsonpath_assets: # If you have defined your own JSON Schemas, add the s3:// path to your own JSON Path files in your own bucket here
      log: s3://collector-snowplow-log
      encrypted: false # Whether the buckets below are enrcrypted using server side encryption (SSE-S3)
      raw:
        in:                  # This is a YAML array of one or more in buckets - you MUST use hyphens before each entry in the array, as below
          - s3://elasticbeanstalk-us-east-2-969763981762/resources/environments/logs/publish/e-q3ewmvvjq2        # e.g. s3://my-old-collector-bucket
                # e.g. s3://my-new-collector-bucket
        processing: s3://collector-snowplow-data/processing
        archive: s3://collector-snowplow-data/archive/raw    # e.g. s3://my-archive-bucket/raw
      enriched:
        good: s3://collector-snowplow-data/enriched/good       # e.g. s3://my-out-bucket/enriched/good
        bad: s3://collector-snowplow-data/enriched/bad       # e.g. s3://my-out-bucket/enriched/bad
        errors:               # Leave blank unless :continue_on_unexpected_error: set to true below
        archive: s3://collector-snowplow-data/archive/enriched   # Where to archive enriched events to, e.g. s3://my-archive-bucket/enriched
      shredded:
        good: s3://collector-snowplow-data/shredded/good       # e.g. s3://my-out-bucket/shredded/good
        bad: s3://collector-snowplow-data/enriched/bad        # e.g. s3://my-out-bucket/shredded/bad
        errors:              # Leave blank unless :continue_on_unexpected_error: set to true below
        archive: s3://collector-snowplow-data/archive/shredded   # Where to archive shredded events to, e.g. s3://my-archive-bucket/shredded
  emr:
    ami_version: 5.9.0
    region: us-east-2       # Always set this
    jobflow_role: EMR_EC2_DefaultRole # Created using $ aws emr create-default-roles
    service_role: EMR_DefaultRole     # Created using $ aws emr create-default-roles
    placement:              # Set this if not running in VPC. Leave blank otherwise
    ec2_subnet_id: subnet-16112f6d         # Set this if running in VPC. Leave blank otherwise
    ec2_key_name: finalsnowplowkeypair
    bootstrap: []           # Set this to specify custom boostrap actions. Leave empty otherwise
    software:
      hbase:                # Optional. To launch on cluster, provide version, "0.92.0", keep quotes. Leave empty otherwise.
      lingual:              # Optional. To launch on cluster, provide version, "1.1", keep quotes. Leave empty otherwise.
    # Adjust your Hadoop cluster below
    jobflow:
      job_name: Snowplow ETL-2 # Give your job a name
      master_instance_type: m4.large
      core_instance_count: 2
      core_instance_type: m4.large
      core_instance_ebs:    # Optional. Attach an EBS volume to each core instance.
        volume_size: 100    # Gigabytes
        volume_type: "gp2"
        volume_iops: 400    # Optional. Will only be used if volume_type is "io1"
        ebs_optimized: false # Optional. Will default to true
      task_instance_count: 0 # Increase to use spot instances
      task_instance_type: m4.large
      task_instance_bid: 0.015 # In USD. Adjust bid, or leave blank for non-spot-priced (i.e. on-demand) task instances
    bootstrap_failure_tries: 3 # Number of times to attempt the job in the event of bootstrap failures
    configuration:
      yarn-site:
        yarn.resourcemanager.am.max-attempts: "1"
      spark:
        maximizeResourceAllocation: "true"
    additional_info:        # Optional JSON string for selecting additional features
collectors:
  format: clj-tomcat # For example: 'clj-tomcat' for the Clojure Collector, 'thrift' for Thrift records, 'tsv/com.amazon.aws.cloudfront/wd_access_log' for Cloudfront access logs or 'ndjson/urbanairship.connect/v1' for UrbanAirship Connect events
enrich:
  versions:
    spark_enrich: 1.14.0 # Version of the Spark Enrichment process
  continue_on_unexpected_error: false # Set to 'true' (and set :out_errors: above) if you don't want any exceptions thrown from ETL
  output_compression: NONE # Compression only supported with Redshift, set to NONE if you have Postgres targets. Allowed formats: NONE, GZIP
storage:
  versions:
    rdb_loader: 0.14.0
    rdb_shredder: 0.13.1        # Version of the Spark Shredding process
    hadoop_elasticsearch: 0.1.0 # Version of the Hadoop to Elasticsearch copying process
monitoring:
  tags: {} # Name-value pairs describing this job
  logging:
    level: DEBUG # You can optionally switch to INFO for production

#2

Do you see your imported data in s3://collector-snowplow-data/archive/raw,s3://collector-snowplow-data/archive/shredded or s3://collector-snowplow-data/archive/enriched? If you didn’t skip the database import your data gets imported into the specified database and will be moved to the archive if everything went smooth. The folders and $ files are just garbage leftovers from the run if they’re empty


#3

Heey tcclass, Thank you for your response.

I am getting some logs in s3://collector-snowplow-data/archive/raw
Enriched and shredded folders have files in them but they are empty.

One more thing that I found is that the POST Request to my Clojure collector is being aborted.
There can be the situation that data is not even being sent by the tracker to the collector.

Can you please guide me about it further.

Thankyou
Arsalan Mehmood


#4

I never worked with the clojure collector so I can just guess

  • you didn’t setup SSL for your site is that correct?
  • you have your adblocker deactivated?
  • the data that you POST is not too big?

#5

Hey tclass,

I checked my SSL certificate was not properly configured. Now Configured it properly but now getting this timeout issue.

Can you please guide me a bit about it?

Thank you
Arsalan Mehmood


#6

I can’t really help you from here with the SSL setup and you’re already discussing that in a different topic POST request to Clojure collector from JS tracker is timed out let’s close this discussion