Acerca de
Senior Data Engineer
Responsibilities
-
Lead data engineers and complex projects to completion.
-
Design, plan, prioritize and be responsible for the project.
-
Build scalable, maintainable data pipelines (batch/streaming-ingestion/ELT/ETL/data products) and ensure their quality/reliability/integrity.
-
Collaborate with analytics and business teams to improve data models that feed business intelligence tools, increase data accessibility and foster data-driven decision-making across the organization.
-
Coach and mentor junior data engineers to develop their skills.
Qualifications
-
Bachelor’s Degree in Computer Science, Software Engineering, Information Technology, or equivalent industry experience.
-
6 years experience in Big Data technologies and their ecosystem.
-
Proficient in SQL, Python or Linux & Unix.
-
Experience in the Hadoop ecosystem such as HDFS, Spark, Hive, Sqoop, Airflow, Oozie, Ranger, Ambari, Flink.
-
Experience in cloud computing technologies such as AWS, Azure, GCP.
-
Experience working with relational databases such as MySQL, PostgreSQL, SQL Server, Oracle.
-
Experience working with NoSQL databases such as MongoDB, HBase, Cassandra, Bigtable, DynamoDB, and Cosmos DB.
-
Experience working with search engines tools like ElasticSearch.
-
Experience in end to end data management solutions.
-
Experience in data migration tools such as Fivetran, Informatica, database migration tools.
-
Ability to design data lake, data warehouse, data mart based on AWS, Azure, GCP and on-premise.
-
Understanding of data lake management such as life cycle management, storage class design, and access control.
-
Ability to optimize data warehouse and data mart such as indexing, clustering, and partitioning.
-
Ability to design data modeling (schema design) such as star schema, snowflake schema, fact table, dimensional table.
-
Experience in ETL/ ELT solutions for both on cloud and on-premise.
-
Ability to develop ETL/ELT solutions in Python, Spark, SQL.
-
Experience in container management - Docker and Kubernetes.
-
Understanding of real-time and batch processing.
-
Experience in real-time processing (streaming) tools such as Apache Kafka, RabbitMQ, Cloud Pub/Sub, Azure Event Hubs, Amazon Kinesis.
-
Experience in workflow orchestration, monitoring or data pipeline tools such as Apache Airflow, Azure Data Factory, Luigi, NiFi, AWS Step Function.
-
Innovative problem-solving skills with the ability to identify and resolve complex architectural issues.
-
Ability to communicate clearly and work closely with cross-functional teams such as Data Analyst, Data Visualization, Software Engineering, and businesses functions.
-
Good command of English.
-
Excellent organizational and leadership skills.
-
A good team player.