August 29
🏡 Remote – New York
Apache
AWS
Azure
Cloud
ETL
Google Cloud Platform
Hadoop
Informatica
Kafka
MySQL
NoSQL
Oracle
Python
RDBMS
Spark
SQL
• Responsible for design, development and maintenance of data pipelines to enable data analysis and reporting. • Builds, evolves and scales out infrastructure to ingest, process and extract meaning out data. • Write complex SQL queries or python code to support analytics needs. • Manage projects / processes, working independently with limited supervision. • Work with structured and unstructured data from a variety of data stores, such as data lakes, relational database management systems, and/or data warehouses. • Combines, optimizes, and manages multiple big data sources. • Builds data infrastructure and determines proper data formats to ensure data is ready for use.
• Strong background in cloud computing, software engineering and data processing. • Data management experience. • Experience in ETL Tools such as Pentaho, Talend, Informatica, Azure Data Factory, Apache Kafka and Apache Camel. • Experience designing and implementing analysis solutions on Hadoop-based platforms such as Cloudera Hadoop, or Hortonworks Data Platform or Spark based platforms such as Databricks. • Proficient in RDBMS such as Oracle, SQL Server, DB2, MySQL etc. • Strong analytical and problem-solving skills. • Strong verbal and written communication skills. • Proficient programming skills in Python, SQL NoSQL, and Spark. • Ability to manage multiple projects essential. • Ability to work independently or in groups. • Ability to prioritize time. • Ability to adapt to a rapidly changing environment.
Apply Now