Barclays Logo

Barclays

Data Engineer - PySpark Developer

Reposted 5 Days Ago
Be an Early Applicant
In-Office
Pune, Mahārāshtra
Mid level
In-Office
Pune, Mahārāshtra
Mid level
The Data Engineer - PySpark Developer will build and maintain data architectures and pipelines, ensuring data accuracy and security. Responsibilities include collaborating with data scientists, developing algorithms, and driving continuous improvement.
The summary above was generated by AI
Job Description

Purpose of the role

To build and maintain the systems that collect, store, process, and analyse data, such as data pipelines, data warehouses and data lakes to ensure that all data is accurate, accessible, and secure. 

Accountabilities

  • Build and maintenance of data architectures pipelines that enable the transfer and processing of durable, complete and consistent data.
  • Design and implementation of data warehoused and data lakes that manage the appropriate data volumes and velocity and adhere to the required security measures.
  • Development of processing and analysis algorithms fit for the intended data complexity and volumes.
  • Collaboration with data scientist to build and deploy machine learning models.

Analyst Expectations

  • To perform prescribed activities in a timely manner and to a high standard consistently driving continuous improvement.
  • Requires in-depth technical knowledge and experience in their assigned area of expertise
  • Thorough understanding of the underlying principles and concepts within the area of expertise
  • They lead and supervise a team, guiding and supporting professional development, allocating work requirements and coordinating team resources.
  • If the position has leadership responsibilities, People Leaders are expected to demonstrate a clear set of leadership behaviours to create an environment for colleagues to thrive and deliver to a consistently excellent standard. The four LEAD behaviours are: L – Listen and be authentic, E – Energise and inspire, A – Align across the enterprise, D – Develop others.
  • OR for an individual contributor, they develop technical expertise in work area, acting as an advisor where appropriate.
  • Will have an impact on the work of related teams within the area.
  • Partner with other functions and business areas.
  • Takes responsibility for end results of a team’s operational processing and activities.
  • Escalate breaches of policies / procedure appropriately.
  • Take responsibility for embedding new policies/ procedures adopted due to risk mitigation.
  • Advise and influence decision making within own area of expertise.
  • Take ownership for managing risk and strengthening controls in relation to the work you own or contribute to. Deliver your work and areas of responsibility in line with relevant rules, regulation and codes of conduct.
  • Maintain and continually build an understanding of how own sub-function integrates with function, alongside knowledge of the organisations products, services and processes within the function.
  • Demonstrate understanding of how areas coordinate and contribute to the achievement of the objectives of the organisation sub-function.
  • Make evaluative judgements based on the analysis of factual information, paying attention to detail.
  • Resolve problems by identifying and selecting solutions through the application of acquired technical experience and will be guided by precedents.
  • Guide and persuade team members and communicate complex / sensitive information.
  • Act as contact point for stakeholders outside of the immediate function, while building a network of contacts outside team and external to the organisation.

All colleagues will be expected to demonstrate the Barclays Values of Respect, Integrity, Service, Excellence and Stewardship – our moral compass, helping us do what we believe is right. They will also be expected to demonstrate the Barclays Mindset – to Empower, Challenge and Drive – the operating manual for how we behave.

Join us as a Data Engineer - PySpark Developer at Barclays, where you'll take part in the evolution of our digital landscape, driving innovation and excellence. You'll harness cutting-edge technology to revolutionize our digital offerings, ensuring unparalleled customer experiences. As a part of the team, you will deliver technology stack, using strong analytical and problem solving skills to understand the business requirements and deliver quality solutions. You'll be working on complex technical problems that will involve detailed analytical skills and analysis. This will be done in conjunction with fellow engineers, business analysts and business stakeholders. 

To be successful as Data Engineer - PySpark Developer you should have experience with:

  • Hands-on programming experience in a Big Data Hadoop ecosystem.
  • Proficiency in PySpark, Hive, and Impala.
  • Exposure to Mongo DB or any other NoSQL database.
  • Solid experience with Unix shell.
  • Experience with scheduling tools like AutoSys, airflow.
  • Strong understanding of Agile methodologies and tools (JIRA, Confluence).
  • Experience with CI/CD tools such as Jenkins, TeamCity, or GitLab.
  • Excellent communication and collaboration skills.
  • Ability to work independently and drive delivery with minimal supervision.

Some other highly valued skills include:

  • Bachelor’s degree in Computer Science, Engineering, or a related field.
  • Relevant certifications in Big Data or cloud technologies are a plus.

You may be assessed on key critical skills relevant for success in role, such as risk and controls, change and transformation, business acumen, strategic thinking and digital and technology, as well as job-specific technical skills.

This role is based in Pune.

Top Skills

Airflow
Autosys
Ci/Cd
Confluence
Gitlab
Hive
Impala
Jenkins
JIRA
MongoDB
NoSQL
Pyspark
Teamcity
Unix

Barclays Chennai, Tamil Nadu, IND Office

Chennai, India, 600004

Similar Jobs

3 Days Ago
Hybrid
Pune, Mahārāshtra, IND
Senior level
Senior level
Blockchain • Fintech • Payments • Consulting • Cryptocurrency • Cybersecurity • Quantum Computing
Lead the development of high-quality, scalable data pipelines and manage data warehousing projects in a Big Data environment. Optimize performance and troubleshoot issues in Spark applications, while collaborating in Agile teams to enhance data workflows and standards.
Top Skills: AirflowApache NifiAWSAzureClouderaDatabricksGitHadoopNetezzaOraclePysparkPythonSQLUnix
13 Days Ago
In-Office
Pune, Mahārāshtra, IND
Mid level
Mid level
Fintech • Financial Services
The Data Engineer will build and maintain data architectures, develop algorithms, and collaborate on machine learning models while ensuring data accuracy and security.
Top Skills: AWSHadoopLinux Shell ScriptingPysparkSQLTeradata
5 Days Ago
In-Office
Pune, Mahārāshtra, IND
Senior level
Senior level
Fintech • Financial Services
As a Data Engineer, you will develop high-quality data products for banking operations, focusing on data architecture, warehousing, and agile decision-making processes.
Top Skills: Big DataHadoopHivePysparkPythonSnowflake

What you need to know about the Chennai Tech Scene

To locals, it's no secret that South India is leading the charge in big data infrastructure. While the environmental impact of data centers has long been a concern, emerging hubs like Chennai are favored by companies seeking ready access to renewable energy resources, which provide more sustainable and cost-effective solutions. As a result, Chennai, along with neighboring Bengaluru and Hyderabad, is poised for significant growth, with a projected 65 percent increase in data center capacity over the next decade.

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account