Hadoop / Data / Java Engineer Santa Clara, CA
Job Description:
Hadoop/Data/Java Engineer
Location: Santa Clara, CA
Full Time
Position description:
Client is looking for a highly experienced Hadoop/Data/Java Engineer with
experience in working with large-scale, distributed data pipelines to join their
office in Santa Clara, California. Responsibilities span the full data
engineering lifecycle from architecture and design, data analysis, software
development, quality assurance, capacity planning and managing the analytics
environment as a whole.
What Youâ??ll Do:
· Build distributed, scalable, and reliable data pipelines that ingest and process data at scale and in real-time.
· Collaborate with other teams to design and develop data tools that support both operations and product use cases.
· Perform offline analysis of large data sets using components from the Hadoop ecosystem.
· Evaluate and advise on technical aspects of open work requests in the product backlog with the project lead.
· Own product features from the development phase through to production deployment.
· Evaluate Big Data technologies and prototype solutions to improve our data processing architecture.
What Youâ??ll Need:
· BS in Computer Science or related area.
- 12 â?? 15 years of software development experience.
- Proficiency with Java, Python, Scala, HBase, Hive, MapReduce, ETL, Postgres, Visualization technologies, etcâ?¦
- Flair for data, schema, data model, how to bring efficiency in Big Data related life cycle.
- Understanding of automated QA needs related to Big Data.
- Understanding of various Visualization platforms (Tableau, D3JS, others).
- Experience in building data analytical applications in retail, hospitality, or restaurant vertical is a big plus.
- Proficiency with agile or lean development practices.
- Strong object-oriented design and analysis skills.
- Excellent technical and organizational skills.
- Excellent written and verbal communication skills.
Top 4 Skill Sets/Technologies Youâ??ll Have:
- Java
- Experience working in an Agile environment
- Data Integration / Data Management experience
- Hadoop / Hive / Spark / Tableau / ETL / Data Integration
Technologies that we use include:
Java, Hadoop/MapReduce, Flume, Storm, Kafka, MemSQL, Pig, Hive, Tableau
Integration, ETL