Data Engineer

Location: Remote.           Type: Full-time

MUST submit a Resume in English, Work hours are in the US PST time zone, California.

As a Senior Data Engineer, you will be responsible for designing, building, and maintaining scalable data pipelines on an AWS cloud platform to support our client's data-driven initiatives. Your expertise in ETL data ingestion frameworks/tools will play a critical role in ensuring efficient data processing and integration

 

Mandatory skills:

  • 5 to 8 years of experience.
  • Python, Pyspark, Airflow, AWS Glue.

 

Accountabilities :

  • Create and maintain data ingestion pipelines, models, and architectures required to support a growing Data Marketing business 
  • Work with Product Management, business partners, and the Data Science team members to understand and create solutions to meet their needs 
  • Work with the Quality Engineers to validate that solutions are meeting requirements. 
  • Implement automation processes as the opportunities present. 

Basic Qualifications:

  • Familiarity with Data Pipeline Management Frameworks on Cloud (AWS Preferred, Azure, Google): As a Senior Data Engineer, you should have a strong understanding of data pipeline management frameworks offered by major cloud providers like AWS, Azure, and Google. Your expertise in working with these platforms will enable you to design and implement robust data pipelines to extract, transform, and load data from various sources. 
  • Familiarity with ETL Data Ingestion Framework/Tools. You should be well-versed in ETL (Extract, Transform, Load) data ingestion frameworks/tools, such as Azure Data Factory, Google Data Fusion, and SSIS. Your knowledge of these tools will facilitate seamless data integration and ensure data quality throughout the pipeline. 
  • Hands-on Experience with Python: Proficiency in Python is essential for this role. You should have hands-on experience using Python to develop data processing scripts, data manipulation, and transformation tasks, as well as implementing data engineering solutions. 
  • Knowledge of Source Control and Scrum Agile Software Development Methodologies: A strong foundation in source control practices, such as Bit Bucket, is required. Moreover, you should be familiar with Scrum Agile software development methodologies to effectively collaborate with cross-functional teams and deliver high-quality data engineering solutions. 
  • Familiarity with AWS Ecosystem: Having a deep understanding of the AWS ecosystem, including training jobs, processing jobs, and Sagemaker, will be a significant advantage. This knowledge will allow you to leverage AWS services efficiently and optimize data workflows. 
  • Good Exposure and hands-on working on the following skills o Glue, Glue Catalog, Crawler, Lambda, Airflow, IAM, S3, Athena, RedShift, Python, PySpark, SQL knowledge, Dynamo DB, Extensive knowledge on applying data transformations, GIT/Bit Bucket 

 

Preferred Qualifications:  

  • Experience in large-data solutions is highly desirable. 
  • Excellent verbal, written, and interpersonal communication skills. 
  • Experience with Scikit-learn, PyTorch, and Huggingface, and Building Transformer and Sentence Transformer Models: Your expertise in working with popular machine learning libraries like Scikit-learn, PyTorch, and Huggingface will be critical for developing and deploying transformer and sentence transformer models. Experience in building and fine-tuning these models will further enhance your role as a Senior Data Engineer.

CKCODECONNECT is an Equal Opportunity Employer and does not discriminate based on race, age, color, religion, sex, sexual orientation, gender identity, national origin, veteran, disability status, or any other characteristic protected by applicable law.

We will guide you in your career journey!

  Apply Now

Apply for this position

*
*
* Attach your resume. Max size 2mb Allowed Type(s): pdf