Data Engineer

Job Title: Data Engineer

Job Description:

Seeking a skilled Data Engineer with a robust background in PySpark and extensive experience with AWS services, including Athena and EMR. The ideal candidate will be responsible for designing, developing, and optimizing large-scale data processing systems, ensuring efficient and reliable data flow and transformation.

Key Responsibilities:

• Data Pipeline Development: Design, develop, and maintain scalable data pipelines using PySpark to process and transform large datasets.

• AWS Integration: Utilize AWS services, including Athena and EMR, to manage and optimize data workflows and storage solutions.

• Data Management: Implement data quality, data governance, and data security best practices to ensure the integrity and confidentiality of data.

• Performance Optimization: Optimize and troubleshoot data processing workflows for performance, reliability, and scalability.

• Collaboration: Work closely with data scientists, analysts, and other stakeholders to understand data requirements and deliver solutions that meet business needs.

• Documentation: Create and maintain comprehensive documentation of data pipelines, ETL processes, and data architecture.

Required Skills and Qualifications:

• Education: Bachelor's or Master’s degree in Computer Science, Engineering, or a related field.

• Experience: 5+ years of experience as a Data Engineer or in a similar role, with a strong emphasis on PySpark.

• Technical Expertise:

o Proficient in PySpark for data processing and transformation.

o Extensive experience with AWS services, specifically Athena and EMR.

o Strong knowledge of SQL and database technologies.

o Experience with Apache Airflow is a plus

o Familiarity with other AWS services such as S3, Lambda, and Redshift.

• Programming: Proficiency in Python; experience with other programming languages is a plus.

• Problem-Solving: Excellent analytical and problem-solving skills with attention to detail.

• Communication: Strong verbal and written communication skills to effectively collaborate with team members and stakeholders.

• Agility: Ability to work in a fast-paced, dynamic environment and adapt to changing priorities.

Preferred Qualifications:

• Experience with data warehousing solutions and BI tools.

• Knowledge of other big data technologies such as Hadoop, Hive, and Kafka.

• Understanding of data modeling, ETL processes, and data warehousing concepts.

• Experience with DevOps practices and tools for CI/CD.

Axiom is a global information technology, consulting and outsourcing company and services provider. Our IT solutions empower organizations and individuals throughout the world to maximize value and quality to succeed in today's challenging business environment. As a fast-growing new economy company, we focus our strengths to offer world-class solutions and services through the convergence of technology, innovation, expertise and experience. We provide software consulting, development and IT-enabled services to clients across the globe. We work towards delivering sustained value creation for customers, employees, industries and society at large. Core offerings include data warehousing, middleware development, product development and web-enablement of legacy applications in verticals like telecom, finance, healthcare, manufacturing, energy & utilities, retail & distribution, enablement of legacy Relentless exploration of technology horizons and a Global Delivery Model that is a judicious combination of onsite, offsite and offshore development, offer a complete range of high-ROI business solutions spanning the consulting, technology, operations and process outsourcing value chain.

View all jobs
Get hired quicker

Be the first to apply. Receive an email whenever similar jobs are posted.

Ace your job interview

Understand the required skills and qualifications, anticipate the questions you may be asked, and study well-prepared answers using our sample responses.

Data Engineer Q&A's
Report this job
Apply for this job