Description

AWS Data Lake Technical Lead - Remote, but occasional travel to Waterbury, VT when required.

Remote Job |   2024-05-13 10:45:03

Apply Now

Share Job 

Job Code : 2024-MY3TECH0934

Role: AWS Data Lake Technical Lead

Location: Remote, but occasional travel to Waterbury, VT when required.

Duration: Long-term Contract

 

Background:

Client is seeking to improve statewide law enforcement data access. The purpose is to design and implement a state-controlled system to access de-identified, aggregated law enforcement and related data currently housed in a record management system.

Client in collaboration with the Agency of Digital (ADS) Services, is seeking to procure Amazon Web Service (AWS) professional services to work with the ADS Technical Lead to build out the Public Services Lakehouse environment.

 

Existing Technology Environment

Client currently uses a Computer Aided Dispatch Records Management System (CAD RMS) system that is currently running in a MySQL instance that will be used in the first phase of the Data Lake build. The supplemental data ranges from SQL server instances and flat file sources that are currently housed internally either in SharePoint or internal file stores.

 

Requirements:

Design and implement the Clients’ Data Lake in the AWS Environment

  • Store Data in AWS CJIS compliant environment.
  • Implement and design the lake house technologies with the IT Tech Lead
  • assigned to the project.
  • Design and implement using latest AWS Lake House standards.
  • Data Security Layer
  • Implement and design security Identity and Access Management roles and processes.
  • Design and Create security IAM templates.
  • Data Ingestion Layer
  • Ingestion Design for variety of sources
    • Operational Database Sources
      •  
        • MySQL, SQL Server
  • SaaS Applications
  • File Shares (SharePoint and one drive)
  • Stream Data Sources
  • System Templates of ingestion processes
  • Data Storage Layer
  • AWS S3 CJIS compliant S3 buckets
  • AWS Redshift Infrastructure
    • Power BI Connector process
  • Data Processing Layer
  • Data Extract, Load and Transform (ELT) process for loading from source to S3 and Transforming data from S3 to Redshift for reporting and analytics.
  • Templates for creating ELT processes for future processes.
  • Data Catalog Layer
  • Design and implement solution to solve data schema drift in AWS Glue for use with reporting and analytical needs.
  • Design and build crawlers for schema and build catalog that stores schema information.
  • Meta data store in catalog for consumption in data warehouse
  • Create Catalog crawler templates for data sources.
  • Create Templates for expansion of the data lake for future agencies.
  • Base data warehouse implemented environment for reporting/dashboard use.
  • Template designs for future data Lakehouse implementations.
  • Lake House design needs to be able to be used by Power BI
  • Design each layer for scaling based on usage.

 

Professional Service Requirements:

  • AWS Lakehouse Certification
  • AWS Design and Implementation Lake House Technologies
  • AWS IAM
  • AWS Lakehouse Technology
  • AWS Glue Knowledge
  • Data Catalog and Crawler
  • AWS Redshift
  • AWS Athena
  • CJIS certified data storage
  • AWS CJIS Data location
  • CJIS Security Background Check (See Additional Attachments / Assurances)

Key Skills
Education

Any Graduate