AWS Site Reliability Engineer - Remote

COLUMBUS, OH

100K - 115K

Job Description:

Summary:

The IS Technical Specialist provides technical and consultative support on the most complex technical matters.

Duties and Responsibilities:

  • Migrate data located in a multitude of data stores, into the Data Lake
  • Orchestrate processes to ETL that data, slice it into the various data marts
  • Manage access to the data through Lake Formation
  • Build a data delivery pipeline to ingest a high volume of the real-time streams, detect anomalies, slice into the window analytics, and put those results in the Elastic search system for the further dashboard consumption
  • Analyze, scope, and estimate tasks, identify technology stack and tools
  • Design and implement optimal architecture and migration plan
  • Develop new and re-architecture solution modules, re-design and re-factor program code
  • Specify the infrastructure and assist DevOps engineers with provisioning
  • Examine performance and advise necessary infrastructure changes
  • Communicate with the client on project-related issues
  • Collaborate with in-house and external development and analytical team
  • Analyzes, designs, and develops systems based on user specifications.
  • Provides technical assistance in solving hardware or software problems.
  • Possesses an in-depth knowledge of and works with the technical tools available for systems development and support.
  • Maintains and demonstrates knowledge of technology industry trends, particularly as they apply to Huntington.
  • May assist with identifying training needs or with the training of less experienced staff.
  • May serve as project leader for specific projects.
  • Performs other duties as assigned.

Please Note: This position is available for remote work, however, the expectation is to be available to work during Eastern or Central time zone hours.

Basic Qualifications:

  • Bachelor's Degree
  • 5 years of experience with AWS operations

Preferred Qualifications:

  • Hands-on experience designing efficient architectures for high-load enterprise-scale applications or ‘big data’ pipelines
  • Hands-on experience utilizing AWS data toolsets including but not limited to DMS, Glue, Data Brew, EMR, SCT
  • Practical experience in implementing big data architecture and pipelines
  • Hands-on experience with message queuing, stream processing, and highly scalable ‘big data’ stores
  • Advanced knowledge and experience working with SQL and NoSQL databases
  • Proven experience in re-design and re-architecting the large complex business applications
  • Strong self-management and self-organizational skills
  • Successful candidates should have experience with any of the following software/tools (not all required at the same time):
    • Python and PySpark - strong knowledge especially with developing Glue jobs
    • Big data tools: Kafka, Spark, Hadoop (HDFS3, YARN2, Tez, Hive, HBase)
    • Stream-processing systems: Kinesis Streaming, Spark-Streaming, Kafka Streams, Kinesis Analytics
    • AWS cloud services: EMR, RDS, MSK, Redshift, DocumentDB, Lambda
    • Message queue systems: ActiveMQ, RabbitMQ, AWS SQS
    • Federated identity services (SSO): Okta, AWS Cognito
  • Experience working in multi-platform environment
  • Ability to balance both development and support roles
  • Experience in working on projects that involve business segments
  • We are looking for a candidate with 3+ years of experience in Data, Cloud, or Software Engineer role, who has attained a degree in Computer Science, Statistics, Informatics, Information Systems, or another quantitative field
  • Usage of HUDI with AWS Data Lakes
  • Graph databases development and optimization 3+ years
  • Neo4j, GREMLIN, Amazon Neptune, Knowledge Graphs
  • Valid AWS certificates would be a great plus
  • Strong interpersonal skills, focus on customer service, and the ability to work well with other IT, vendor, and business groups


Share Profile