About the Company:
Data Prism is on a mission to democratize data, helping startups and enterprises alike in an effort to make data-driven decisions. We provide data acquisition (and analysis) and data engineering services, adding huge value for our clients around the globe. Our robust and reliable solutions have helped clients improve both their top line as well as bottom line.
Job Designation: Data Engineer
About the Role:
You will be responsible for designing, developing, and maintaining scalable data pipelines to support data-driven initiatives. The role requires a strong understanding of data architecture, database management, and programming skills to handle large volumes of data.
Job Responsibilities:
● Writing, maintaining, and debugging web crawlers and scrapers to extract data on large scale
● Using APIs to fetch data and store the data in databases (SQL or NoSQL)
● Parsing data extracted from various sources and performing data cleaning and transformation
● Implement data quality and data validation processes to ensure accuracy, consistency, and reliability of data
● Processing extracted data to extract business insights and taking appropriate actions
● Develop and maintain robust, scalable, and high-performance data pipelines and ETL processes
● Design, build, and optimize data models, databases, and data warehouses for storage and retrieval of structured and unstructured data
● Maintaining repositories using version control tools, like Git, and deploying programs on servers
Job Qualifications:
● No degree requirement - we will test skills through our recruitment process
● 2-3 years of professional experience in Python and Data Engineering (unless you are exceptional).
● Experience with modern data warehousing tools, like Snowflake and BigQuery
● Advanced knowledge of different databases (such as PostgreSQL and MongoDB) and experience of writing complex SQL queries
● Experience with data orchestration and data transformation tools, like DBT, Airflow, and Fivetran, to design, implement, and maintain complex data models
● Knowledge of AWS (EC2, RDS, Glue, EMR, Lambda, S3, etc.), Azure (ADF, Databricks, etc.), and GCP (Compute Engine, Functions, etc.)
● Proficient in scraping libraries and frameworks of Python (like Selenium, BS4, and Scrapy) and Requests module
● Knowledge of API integration to implement complex workflow automations
● Experience with CI/CD and version control tools, like Git and Github
● Familiar with UNIX & Shell Scripting
Desired Soft Skills:
● Great communication skills
● Ability to work independently
● Fast learner
● Self-driven
Benefits We Offer:
● Market Competitive salary
● Bi-Annual Performance based increments
● Highly progressive career path
● Work on challenging projects
● Friendly work environment with minimal hierarchy
● Dedicated time for learning
● Paid Leaves (Casual, Medical, Maternity & Paternity)
● Activities for the refreshment/team building of employees
Where to Apply:
Send your resume at