top of page

Data Engineer

Job Description :


Jash Data Sciences: Letting Data Speak!

Harnessing data for innovation, we specialize in data science, engineering, and Industry 4.0.

From predictive analytics to image/text intelligence, we empower growth with informed decisions. Partner with us for data-driven transformation.


Do you love solving real-world data problems with the latest and best techniques? And having fun while solving them in a team! Then come and join our high-energy team of passionate data people. Jash Data Sciences is the right place for you.


We are a cutting-edge Data Sciences and Data Engineering startup based in Pune, India. We believe in continuous learning and evolving together. And we let the data speak!


What will you be doing?


● You will be discovering trends in the data sets and developing algorithms to transform

raw data for further analytics

● Create Data Pipelines to bring in data from various sources, with different formats,

transform it, and finally load it to the target database.

● Implement ETL/ ELT processes in the cloud using tools like AirFlow, Glue, Stitch, Cloud

Data Fusion, and DataFlow.

● Design and implement Data Lake, Data Warehouse, and Data Marts in AWS, GCP, or

Azure using Redshift, BigQuery, PostgreSQL, etc.

● Creating efficient SQL queries and understanding query execution plans for tuning

queries on engines like PostgreSQL.

● Performance tuning of OLAP/ OLTP databases by creating indices, tables, and views.

● Write Python scripts for the orchestration of data pipelines

● Have thoughtful discussions with customers to understand their data engineering

requirements. Break complex requirements into smaller tasks for execution.


What do we need from you?


● Strong Python coding skills with basic knowledge of algorithms/data structures and

their application.

● Strong understanding of Data Engineering concepts including ETL, ELT, Data Lake, Data

Warehousing, and Data Pipelines.

● Experience designing and implementing Data Lakes, Data Warehouses, and Data Marts

that support terabytes of scale data.

● A track record of implementing Data Pipelines on public cloud environments

(AWS/GCP/Azure) is highly desirable

● A clear understanding of Database concepts like indexing, query performance

optimization, views, and various types of schemas.

● Hands-on SQL programming experience with knowledge of windowing functions,

subqueries, and various types of joins.

● Experience working with Big Data technologies like PySpark/ Hadoop

● A good team player with the ability to communicate with clarity

● Show us your git repo/ blog!


Qualification


● 1-2 years of experience working on Data Engineering projects for Data Engineer I

● 2-5 years of experience working on Data Engineering projects for Data Engineer II

● 1-5 years of Hands-on Python programming experience

● Bachelors/Masters' degree in Computer Science is good to have

● Courses or Certifications in the area of Data Engineering will be given a higher preference.

● Candidates who have demonstrated a drive for learning and keeping up to date with technology by continuing to do various courses/self-learning will be given high preference.

Apply Now To Get Your Job

bottom of page