Data Engineer I & II

  • Bengaluru
  • Tsworks

Who We Are

tsworks Technologies India Private Limited (subsidiary of The Software Works, Inc, USA) is a technology product and services company. Our mission is to provide domain expertise, innovative solutions, and thought leadership to empower businesses to thrive in a digital world. We value our employees and take pride in providing best value in customer engagements defined by your skills, ideas, and attitude.


About This Role

tsworks Technologies India Private Limited is seeking driven and motivated Data Engineers to join its Digital Services Team. You will get hands-on experience with projects employing industry-leading technologies. This would initially be focused on the operational readiness and maintenance of existing applications and would transition into a build and maintenance role in the long run.

 

  •  Position: Data Engineer
  • Experience: 3+ Years
  • Location: Bangalore, India


Mandatory Required Qualification

  • Strong proficiency in cloud-based data platforms such as AWS, Azure, or Google Cloud Platform
  • Strong programming skills like Python, Java, Scala etc.
  • Experience in DevOps and CI/CD implementation
  • Knowledge of data warehousing and ETL processes
  • Excellent Communication Skills


In This Role, You Will

  • Develops and maintains scalable data pipelines and builds out new API integrations to support continuing increases in data volume and complexity.
  • Collaborates with analytics and business teams to improve data models that feed business intelligence tools, increasing data accessibility and fostering data-driven decision making across the organization.
  • Implements processes and systems to monitor data quality, ensuring production data is always accurate and available for key stakeholders and business processes that depend on it.
  • Writes unit/integration tests, contributes to engineering wiki, and documents work.
  • Performs data analysis required to troubleshoot data related issues and assist in the resolution of data issues.
  • Works closely with a team of frontend and backend engineers, product managers, and analysts.
  • Defines company data assets (data models), spark, sparkSQL, and hiveSQL jobs to populate data models.
  • Designs data integrations and data quality framework.
  • Works closely with all business units and engineering teams to develop strategy for long term data platform architecture.
  • Develop and maintain data pipelines for efficient data extraction, transformation, and loading (ETL) processes.
  • Work on data transformations and processing using cloud data integration tools
  • Develop and maintain data models within cloud databases like snowflake and related tools to support reporting, analytics, and business intelligence needs.
  • Collaborate with cross-functional teams to understand data requirements and design appropriate data integration solutions.
  • Integrate data from various sources, both internal and external, ensuring data quality and consistency.
  • Ensure data models are designed for scalability, reusability, and flexibility.
  • Implement data quality checks, validations, and monitoring processes to ensure data accuracy and integrity across cloud environments.
  • Adhere to data governance standards and best practices to maintain data security and compliance.
  • Handling performance optimizations in the cloud environments
  • Maintain comprehensive documentation for data pipelines, processes, and architecture within both Azure and Snowflake environments including best practices, standards, and procedures.

 

Skills & Knowledge

  • Bachelor's or Master's degree in Computer Science, Engineering, or a related field.
  • 4 + Years of experience in Information Technology, designing, developing and executing solutions.
  • 3+ Years of hands-on experience in designing and executing data solutions on Azure cloud platforms as a Data Engineer.
  • Strong proficiency in cloud services
  • Familiarity with cloud databases like snowflake, Synapse etc.
  • Hands-on experience in data modelling, batch and real-time pipelines, using Python, Java or JavaScript and experience working with Restful APIs are required.
  • Expertise in DevOps and CI/CD implementation.
  • Hands-on experience with SQL and NoSQL databases.
  • Knowledge of big data technologies such as Hadoop, Spark, or Kafka.
  • Familiarity with machine learning concepts and integration of ML pipelines into data workflows
  • Hands-on experience working in an Agile setting.
  • Is self-driven, naturally curious, and able to adapt to a fast-paced work environment.
  • Can articulate, create, and maintain technical and non-technical documentation.
  • Public cloud certifications are desired.