Enable Data Incorporated Logo

Enable Data Incorporated

Data Scientist (3 to 7 Years)

Posted 14 Days Ago
Remote
Hiring Remotely in India
Mid level
Remote
Hiring Remotely in India
Mid level
The Data Scientist role involves predictive analytics, NLP, and Deep Learning with a focus on feature engineering for structured and text data.
The summary above was generated by AI

Job Description – Data Scientist (3–7 years)

Location:
Role Type: Full‑time
Category: Data Science – Structured Data / Text Data (NLP & GenAI)

About the Role

We are seeking a highly skilled Data Scientist (3–7 years of experience) to join our team and work across two major data science domains:

  1. Structured Data (80–90%) – Predictive analytics, forecasting, cost estimation, likelihood modeling, and batch‑oriented machine learning pipelines.
  2. Text / Unstructured Data (NLP & GenAI) – Building low‑latency real‑time systems using deep learning, LLMs, prompt engineering, and agentic AI frameworks.

This role requires strong expertise in Big Data processing, modern ML tools, and the ability to build scalable, production-ready data science solutions.

Key Responsibilities

Structured Data – Machine Learning & Analytics

  • Build, deploy, and optimize ML models for predictive analytics, forecasting, classification, and regression.
  • Perform large-scale feature engineering using PySpark and Big Data tools.
  • Work on batch pipelines, model versioning, and experiment tracking.
  • Develop cost estimation and risk/likelihood models using statistical and ML techniques.

Text Data / NLP / GenAI

  • Build NLP pipelines using deep learning frameworks such as PyTorch, TensorFlow, or similar.
  • Develop real‑time, low‑latency inference systems for text classification, embeddings, semantic search, summarization, and retrieval.
  • Create prompts, context graphs, and agentic workflows for LLM-based systems.
  • Apply knowledge of prompt engineering, context engineering, and autonomous agent frameworks to production systems.

Core Data Science Engineering & MLOps

  • Work in Databricks for ETL, feature engineering, ML training, and orchestration.
  • Use Azure services for model deployment, data pipelines, and infrastructure.
  • Collaborate using Git-based workflows; leverage tools like GitHub Copilot, Claude Code, etc.
  • Implement model monitoring, observability, drift detection, and performance tracking.

Required Skills & Experience

✅ Core Skills

  • Strong hands-on experience with Databricks (Delta Lake, MLflow, Job Orchestration).
  • Excellent PySpark skills for large-scale distributed data processing.
  • Proficiency in Azure cloud services (ADF, Azure ML, AKS, Databricks on Azure).
  • Strong understanding of ML algorithms, statistical methods, and data analysis.
  • Experience with deep learning frameworks:
    • PyTorch
    • TensorFlow
    • Transformers (HuggingFace)
  • Experience with model monitoring and ML observability.
  • Ability to write clean, optimized code and leverage AI code assistants.

✅ NLP / GenAI Specific Skills

  • Prompt engineering (task prompts, chain of thought, tool calling, retrieval prompts).
  • Context engineering (retrieval pipelines, RAG, memory management, context structuring).
  • Knowledge of LLM-based agentic frameworks (LangChain, Semantic Kernel, CrewAI, AutoGen, etc.).
  • Experience with vector databases and embedding models is a plus.

Good to Have Skills

  • Experience with containerization (Docker, Kubernetes, AKS).
  • Experience deploying models to production (REST APIs, real-time endpoints).
  • Knowledge of streaming technologies (Kafka, EventHub, Spark Streaming).
  • Understanding of CI/CD for ML (Azure DevOps / GitHub Actions).

Who You Are

  • A problem solver who is comfortable working with both structured and unstructured data.
  • Someone who enjoys using modern AI tools to accelerate development.
  • A data scientist who writes clean, production-grade code.
  • A collaborator who thrives in cross-functional teams and fast-paced environments.

Top Skills

Azure
Databricks
Pyspark
PyTorch
TensorFlow

Similar Jobs

6 Days Ago
Remote or Hybrid
Senior level
Senior level
Big Data • Information Technology
The Senior Data Scientist at NielsenIQ drives data-driven solutions, manages multiple projects, and ensures data quality and effective communication across teams.
Top Skills: PythonSQL
32 Minutes Ago
Remote
Gujarat, IND
Mid level
Mid level
Artificial Intelligence • Hardware • Information Technology • Machine Learning
Responsible for lab equipment maintenance, calibration, troubleshooting, and team leadership while ensuring compliance with regulations and conducting failure analysis.
Top Skills: 3D X-RayBend TesterCsamFibFtirHast ChamberIso/Iec 17025LinuxReflow OvenSemShock TesterSoak ChamberTemp CycleTemperature ChamberTesterThbWindowsX-Section
32 Minutes Ago
Remote
Gujarat, IND
Mid level
Mid level
Artificial Intelligence • Hardware • Information Technology • Machine Learning
The Technical Program Manager will identify and resolve assembly process issues, lead continuous yield improvement, and manage material suppliers while supporting audits and process validation initiatives.
Top Skills: ApcData ManagementFdcProcess OptimizationRmsSpc

What you need to know about the Chennai Tech Scene

To locals, it's no secret that South India is leading the charge in big data infrastructure. While the environmental impact of data centers has long been a concern, emerging hubs like Chennai are favored by companies seeking ready access to renewable energy resources, which provide more sustainable and cost-effective solutions. As a result, Chennai, along with neighboring Bengaluru and Hyderabad, is poised for significant growth, with a projected 65 percent increase in data center capacity over the next decade.

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account