Guardian Life Logo

Guardian Life

Data Engineer

Posted 6 Days Ago
Be an Early Applicant
In-Office
Chennai, Tamil Nadu
Mid level
In-Office
Chennai, Tamil Nadu
Mid level
As a Data Engineer, you will analyze raw data, create data pipelines, collaborate with teams, and develop solutions for machine learning and business intelligence.
The summary above was generated by AI
Job Description:

Job Description – Data Engineer

As a Data Engineer, you will play a key role in this exciting journey. Your contributions will go beyond coding, as you'll help bring life to ideas, transforming innovative ideas into tangible solutions that directly impact our business and customers.

You'll work in an innovative, fast-paced environment, collaborating with bright minds while enjoying a balance between strategic and hands-on work. We value continuous learning, and you will have the chance to expand your skillset, mastering new tools and technologies that advance our company's goals.

We look forward to welcoming a committed team player who thrives on creating value through innovative solutions and is eager to make a significant impact.

You will

  • Perform detailed analysis of raw data sources by applying business context and collaborate with cross-functional teams to transform raw data into curated & certified data assets to be used for ML and BI use cases. Create scalable and trusted data pipelines which generates curated data assets in centralized data lake / data warehouse ecosystem.
  • Monitor and troubleshoot data pipeline performance, identifying and resolving bottlenecks and issues.
  • Extract text data from variety of sources like documents (Word, PDFs, Text Files, JSON etc.), logs, text notes stored in databases, using Web scrapping method from web pages to support development of NLP / LLM solutions.
  • Collaborate with data science and data engineering team to build scalable and reproducible machine learning pipelines for inference.
  • Leverage different public / private APIs for the purpose of extracting data, invoking functionalities as required for the use cases.
  • Develop real time data solutions by developing new API endpoints or streaming frameworks.
  • Develop, test, and maintain robust tools, frameworks, and libraries that standardize and streamline the data & machine learning lifecycle.
  • Implement robust data drift and model monitoring frameworks to use them across pipelines.
  • Collaborate with cross-functional teams of Data Science, Data Engineering, business units and various IT teams.
  • Create and maintain effective documentation for project and practices ensuring transparency and effective team communication.
  • Stay up to date with the latest trends in modern data engineering, machine learning & AI.

You Have

  • Bachelor’s or master’s degree with 3+ years of experience in Computer Science, Data Science, Engineering, or a related field.
  • 3+ years of experience in working with Python, SQL, PySpark and bash scripts. Proficient in software development lifecycle and software engineering practices.
  • 3+ years of experience in developing and maintaining robust data pipelines for both structured and unstructured data to be used by Data Scientists to build ML Models.
  • 3+ years of experience working with Cloud Data Warehousing (Redshift, Snowflake, Databricks SQL or equivalent) platforms and experience in working with distributed frameworks like Spark.
  • 2+ years of hands-on experience in using Databricks platform for data engineering. Detailed knowledge of Delta Lake, Databricks Workflow, Job Clusters, Databricks CLI, Databricks Workspace etc.
  • Solid understanding of machine learning life cycle, data mining, and ETL techniques.
  • Familiarity with commonly used machine learning libraries (like scikit-learn, xgboost) in terms of exposure and handling of code base which makes use of these libraries for model training & scoring.
  • Proficiency in understanding of REST APIs, experience in using different types of APIs to either extract data or perform a functionality exposed by APIs.
  • Familiarity in Pythonic API development frameworks like Flask / FastAPI. Experience in using containerization frameworks like Docker / Kubernetes.
  • Hands-on experience in building and maintaining tools and libraries which have been used by multiple teams across organization. e.g. Creating Data Engineering common utility libraries, DQ Libraries etc.
  • Proficient in understanding and incorporating software engineering principles in design & development process.
  • Hands on experience with using CI/CD tools (e.g., Jenkins or equivalent), version control (Github, Bitbucket), Orchestration (Airflow, Prefect or equivalent)
  • Excellent communication skills and ability to work and collaborate with cross functional teams across technology and business.

Life at Guardian: https://youtu.be/QEtkY6EkEuQ

Location:

This position can be based in any of the following locations:

Chennai

Current Guardian Colleagues: Please apply through the internal Jobs Hub in Workday

Top Skills

Airflow
Bash Scripts
Ci/Cd Tools
Databricks Sql
Docker
Fastapi
Flask
Jenkins
Kubernetes
Prefect
Pyspark
Python
Redshift
Rest Apis
Snowflake
Spark
SQL

Guardian Life Chennai, Tamil Nadu, IND Office

D Block, 5th Floor, Tidel Park, #4 Rajiv Gandhi Salai Taramani, , Chennai, Tamil Nadu, India, 600 113

Similar Jobs

6 Hours Ago
In-Office
5 Locations
Senior level
Senior level
Software
The Data Engineer role involves designing and deploying ETL processes, managing data integrations from GA4 into BigQuery/Snowflake, optimizing data queries, and collaborating with stakeholders on data strategies.
Top Skills: AlteryxAws GlueBigQueryCloud FunctionsCloudrunDataflowDataprocDbtFivetranGCPGitGoogle Analytics 4LookerPower BIPub/SubPythonSnowflakeSQLTerraform
8 Hours Ago
In-Office
Chennai, Tamil Nadu, IND
Entry level
Entry level
Digital Media • Gaming • Internet of Things • News + Entertainment • Retail • Business Intelligence • Cybersecurity
The Data Engineer 1 is responsible for designing and building data pipelines, ensuring data quality, and developing information systems for data storage and access, collaborating with partners to optimize data operations and processes.
Top Skills: Aws S3DatabricksKubernetesRedshiftTeradata
8 Hours Ago
In-Office
Chennai, Tamil Nadu, IND
Internship
Internship
Digital Media • Gaming • Internet of Things • News + Entertainment • Retail • Business Intelligence • Cybersecurity
Design, build and operate technology architecture for data management. Develop data pipelines, ensure quality and collaborate on data solutions across platforms.
Top Skills: Application Programming Interfaces (Api)Aws S3DatabricksKubernetesRedshiftTeradata

What you need to know about the Chennai Tech Scene

To locals, it's no secret that South India is leading the charge in big data infrastructure. While the environmental impact of data centers has long been a concern, emerging hubs like Chennai are favored by companies seeking ready access to renewable energy resources, which provide more sustainable and cost-effective solutions. As a result, Chennai, along with neighboring Bengaluru and Hyderabad, is poised for significant growth, with a projected 65 percent increase in data center capacity over the next decade.

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account