Join phData, a dynamic and innovative leader in the modern data stack. We partner with major cloud data platforms like Snowflake, AWS, Azure, GCP, Fivetran, Pinecone, Glean and dbt to deliver cutting-edge services and solutions. We're committed to helping global enterprises overcome their toughest data challenges.
phData is a remote-first global company with employees based in the United States, Latin America and India. We celebrate the culture of each of our team members and foster a community of technological curiosity, ownership and trust. Even though we're growing extremely fast, we maintain a casual, exciting work environment. We hire top performers and allow you the autonomy to deliver results.
- 5x Snowflake Partner of the Year (2020, 2021, 2022, 2023, 2024)
- Fivetran, dbt, Atlation, Matillion Partner of the Year
- #1 Partner in Snowflake Advanced Certifications
- 600+ Expert Cloud Certifications (Sigma, AWS, Azure, Dataiku, etc)
- Recognized as an award-winning workplace in US, India and LATAM
In this hands-on role, you will provide Hadoop Administration and ensure performance, reliability, and optimization for big data clusters as well as recommend resources required to deploy and optimize big data technologies.
- Installation, administration, and configuration of big data clusters
- Hadoop Administration; managing and monitoring distributed systems and middleware application performance; recommend, configure and optimize Hadoop ecosystem
- Configuration, troubleshooting and performance tuning of Java applications
- Hadoop security; LDAP, Active Directory and Kerberos (KDC) administration
- Hadoop encryption; HDFS transparent data encryption, LUKS and PKI methodologies
Qualifications:
- 2-4 years of Linux OS installation, configuration, administration and performance optimization as a Linux System or Java Middleware Engineer with emphasis on distributed computing.
- Experience integrating Linux OS with user authentication backend (LDAP/Active Directory)
- Hadoop experience including:
- Hadoop distribution (Cloudera preferred) including cluster installation and configuration
- Core Hadoop (HDFS, Hive, YARN) and on one or more ecosystem products/languages such as HBase, Spark, Impala, Search, Kudu, etc.
- Java application configuration and performance tuning
- Support for cloud-based big data clusters; AWS preferred
- Infrastructure automation; experience with Ansible and Git is a plus
- Experience scoping activities on large scale, complex technology infrastructure projects
- Proven experience working with key stakeholders and customers with the ability to translate “big picture” business requirements and use cases into a Hadoop solution, including ingestion of many data sources, ETL processing, data access, and consumption, as well as custom analytics.
- Excellent communication skills and customer relationship management including project escalations, and participating in executive steering meetings
Perks and Benefits
- Medical Insurance for Self & Family
- Medical Insurance for Parents
- Term Life & Personal Accident
- Wellness Allowance
- Broadband Reimbursement
- Professional Development Allowance
- Reimbursement of Skill Upgrade Certifications
- Certification Reimbursement
#LI-DNI
phData celebrates diversity and is committed to creating an inclusive environment for all employees. Our approach helps us to build a winning team that represents a variety of backgrounds, perspectives, and abilities. So, regardless of how your diversity expresses itself, you can find a home here at phData. We are proud to be an equal opportunity employer. We prohibit discrimination and harassment of any kind based on race, color, religion, national origin, sex (including pregnancy), sexual orientation, gender identity, gender expression, age, veteran status, genetic information, disability, or other applicable legally protected characteristics. If you would like to request an accommodation due to a disability, please contact us at People Operations.