Job Description
Position: Data Engineer
Location: No Location Constraint
Contract
Role Summary
We are seeking an experienced Data Engineer to design, build, and optimize scalable, high‑performance data pipelines using Databricks, Apache Airflow, Snowflake, Python, and SQL.
The role involves end‑to‑end ownership of data ingestion, transformation, orchestration, and optimization across cloud‑based data platforms, enabling analytics, reporting, and downstream data science use cases.
Key Responsibilities
Data Engineering & Pipeline Development
- Design, develop, and maintain batch and streaming data pipelines using Databricks (PySpark) and Snowflake.
- Build ETL / ELT frameworks to ingest data from multiple sources (RDBMS, APIs, flat files, cloud storage).
- Implement data transformation logic using Python and SQL for scalable and high‑volume datasets.
- Develop metadata‑driven and reusable pipelines following enterprise data engineering best practices.
Workflow Orchestration
- Create and manage complex workflows using Apache Airflow.
- Implement scheduling, dependency management, retries, alerts, and failure handling.
- Integrate Airflow with Databricks jobs, Snowflake tasks, and cloud services.
Databricks & Lakehouse Architecture
- Work on Databricks Lakehouse architecture including Bronze / Silver / Gold (Medallion) layers.
- Optimize Spark jobs using partitioning, caching, broadcast joins, and performance tuning.
- Manage Databricks jobs, clusters, notebooks, and workspace configurations.
Snowflake Development
- Design and optimize Snowflake schemas, tables, views, and warehouses.
- Implement Snowflake SQL transformations, performance tuning, and cost optimization.
- Work with Snowflake features such as Time Travel, Cloning, Tasks, Streams (where applicable).
Data Quality, Governance & Security
- Implement data quality checks, validation frameworks, and reconciliation logic.
- Ensure adherence to data governance, security, and compliance requirements.
- Collaborate with governance teams on metadata, lineage, and access controls.
CI/CD & Operations
- Implement CI/CD pipelines for data code using Git‑based version control systems.
- Support production deployments, monitoring, and incident resolution.
- Work closely with DevOps, Architecture, and Analytics teams.
Experience: 3-5 Years .
The expected compensation for this role ranges from $45,000 to $121,000 .
Final compensation will depend on various factors, including your geographical location, minimum wage obligations, skills, and relevant experience. Based on the position, the role is also eligible for Wipro's standard benefits including a full range of medical and dental benefits options, disability insurance, paid time off (inclusive of sick leave), other paid and unpaid leave options.
Applicants are advised that employment in some roles may be conditioned on successful completion of a post-offer drug screening, subject to applicable state law.
Wipro provides equal employment opportunities to all employees and applicants for employment and prohibits discrimination and harassment of any type without regard to race, color, religion, age, sex, national origin, disability status, genetics, protected veteran status, sexual orientation, gender identity or expression, or any other characteristic protected by federal, state, or local laws. Applications from veterans and people with disabilities are explicitly welcome.
Reinvent your world. We are building a modern Wipro. We are an end-to-end digital transformation partner with the boldest ambitions. To realize them, we need people inspired by reinvention. Of yourself, your career, and your skills. We want to see the constant evolution of our business and our industry. It has always been in our DNA - as the world around us changes, so do we. Join a business powered by purpose and a place that empowers you to design your own reinvention.