Macrosoft Pakistan is looking for a Data Engineer
Job Title: Data Engineer
Location: Remote
Experience – 5+ years
Openings: 2
About the Role:
We are looking for highly skilled Data Engineers to join the our team. The ideal candidates will have strong experience in backend development, automation, big data processing, and cloud platforms, with the ability to design, build, and optimize large-scale data pipelines and workflows.
Key Responsibilities:
โข Develop and maintain ETL pipelines and distributed data processing systems.
โข Build workflows integrating Rails apps, Java code, Spark jobs, and Iceberg data.
โข Work on big data technologies like Apache Spark (batch & streaming jobs, performance tuning).
โข Manage and optimize data storage solutions, including Apache Iceberg and SQL (large datasets, query optimization).
โข Automate processes and maintain Linux-based systems.
โข Deploy and manage solutions on AWS (EC2, S3, EMR/Glue, IAM).
โข Collaborate with cross-functional teams to gather requirements, design solutions, and ensure smooth integration.
Required Skills & Experience:
โข Strong experience in scripting languages (Python, Ruby, or others).
โข Proficiency in Apache Spark and SQL.
โข Experience with Apache Iceberg (tables, schema evolution, partitioning).
โข Working knowledge of AWS (EC2, S3, EMR/Glue, IAM).
โข Strong Linux automation and system administration skills.
โข Good understanding of ETL pipelines and distributed data processing.
โข Excellent communication, requirement gathering, and cross-team collaboration skills.
Good to Have:
โข Familiarity with Git, CI/CD pipelines, Docker.
โข Exposure to Terraform/CloudFormation.
โข Understanding of unit and integration testing.
Priority Attributes:
โข Must-Have: Strong backend development, automation, Spark, Iceberg, SQL, AWS, workflow orchestration, and soft skills.
โข Good/Nice to Have: DevOps/CI-CD knowledge, Infrastructure as Code, testing exposure.