ER - Data Engineer

Date: 5 Sep 2024

Location: Chennai, TN, IN, 600113

Company: Altimetrik

Primary Skill :Data Engineer

Secondary Skill : GCP, Big Query

Exp : 5 to 10 yrs

Job Description:

We are seeking a skilled Data Engineer with expertise in Google Cloud Platform (GCP), particularly in BigQuery and Pub/Sub, coupled with advanced proficiency in Python. As a Data Engineer, you will play a pivotal role in designing, implementing, and optimizing data pipelines, ensuring the smooth flow of data within our organization.

Responsibilities:

1. Architect and Develop Data Pipelines: Design, develop, and maintain scalable and efficient data pipelines using GCP services such as Pub/Sub for real-time data ingestion and BigQuery for storage and analysis.

2. Data Transformation and Processing: Implement data transformation processes to cleanse, enrich, and aggregate raw data from various sources, ensuring data quality and consistency.

3. Optimize Performance: Fine-tune data pipelines and queries to optimize performance and reduce latency, ensuring timely access to data for stakeholders.

4. Monitoring and Maintenance: Implement monitoring solutions to track pipeline performance and proactively address issues. Perform regular maintenance tasks to ensure the reliability and availability of data infrastructure.

5. Collaboration: Collaborate with cross-functional teams including Data Scientists, Software Engineers, and Business Analysts to understand data requirements and provide technical solutions to address business needs.

6. Documentation: Document data pipelines, processes, and best practices to ensure knowledge sharing and maintain a comprehensive understanding of data architecture.

Requirements:

1. Proficiency in GCP Services: Extensive hands-on experience with Google Cloud Platform services, particularly BigQuery for data storage and analysis, and Pub/Sub for real-time data streaming.

2. Python Programming: Strong programming skills in Python for data manipulation, scripting, and automation tasks. Experience with libraries such as Pandas, NumPy, and TensorFlow is highly desirable.

3. Data Modeling: Solid understanding of data modeling concepts and experience in designing efficient data models for analytics and reporting purposes.

4. SQL Skills: Proficiency in writing complex SQL queries for data extraction, transformation, and analysis within BigQuery.

5. Experience with Data Warehousing: Familiarity with data warehousing concepts and experience in implementing data warehouse solutions using GCP BigQuery or similar technologies.

6. Problem-solving Skills: Strong analytical and problem-solving skills with the ability to troubleshoot and resolve complex data-related issues.

7. Communication Skills: Excellent communication skills with the ability to effectively collaborate with cross-functional teams and articulate technical concepts to non-technical stakeholders.

8. Bachelor’s Degree: Bachelor’s degree in Computer Science, Engineering, Mathematics, or a related field. Advanced degree or relevant certifications in Data Engineering or GCP are a plus.

Company Overview</Company Overview>

Altimetrik delivers outcomes for our clients by rapidly enabling digital business & culture and infuse speed and agility into enterprise technology and connected solutions. We are practitioners of end-to-end business and technology transformation. We tap into an organization’s technology, people, and assets to fuel fast, meaningful results for global enterprise customers across financial services, payments, retail, automotive, healthcare, manufacturing, and other industries. Founded in 2012 and with offices across the globe, Altimetrik makes industries, leaders and Fortune 500 companies more agile, empowered and successful. 

Altimetrik helps get companies get “unstuck”.  We’re a technology company that lives organizations a process and context to solve problems in unconventional ways.  We’re a catalyst for organization’s talent and technology, helping teams push boundaries and challenge traditional approaches. We make delivery more bold, efficient, collaborative and even more enjoyable.

Role definition</Job Overview>
  • Should be an individual contributor as well as handling a team. Expertise in below mentioned domain.

  •  Strong fundamentals in Computer Science - programming, algorithms, data structures, sql querying

  •  Basic understanding of Data landscape (ETL, Data processing, Data Storage, Reporting)

  •  Working experience in Cloud would be a plus

  • Working experience in Spark /DataBricks /Scala/Java would be an added advantage

  •  Working experience in tools like NiFi, Kafka etc .

  •  Data Streaming and handling real time data load would be a big plus

  •  Excellent SQL/Data Analysis Skills

  •  Creating Data Mapping documents between Source to target and also documenting business rules on transformations is mandatroy

  •  Should have worked in traditional ETL tool and processes

  •  Should have experience in RDBMS / NoSQL /Big Data

  •  Should have basic experience in programming/scripting language (Python/JAVA/C++)

  •  Basics of Data Modelling and Data Architecture

  •  Triage Data Issues and do RCA

  •  Data Testing - Unit Testing, Functional Testing and end to end Data Validation

  •  Expertise in Performance Optimization

<Roles & Responsibilities style='font-size: 1em; margin: 0px'>Value you will deliver</Roles & Responsibilities>
  • Achieve top line and bottom line targets for the account/portfolio
  • Detailed account plan and achieving on the same on a quarter on quarter basis
  • Formal engagement plan for liaising with key senior customer stakeholders, and facilitating Altimetrik leadership connects with client executives via a multi-tier approach
  • Lead and influence an engaged and effective workforce and that is fully integrated with Altimetrik vision, values and purpose
Technical/Functional competency</Education and Experience Required>
  • IT software development, digital solutions, digital transformation background is mandatory.
  • Good understanding of technology, software development methodologies especially as relating to developing custom software for customers using agile principles

 

Behavioral competency

  • Strong problem solving skills
  • Ability to lead initiatives and people toward common goals
  • Working knowledge of systems infrastructure
  • Excellent oral and written communication, presentation, and analytical skills

Perks

  • Top 100 GPTW certified company with award winning employee practices
  • Industry-best compensation & benefits
  • Fast growing company with opportunity to work with Fortune league global brands and latest technologies
  • Extensive scope to learn and upskill
  • Heavy focus on work-life balance and employee well-being
  • Collaboration, team bonding and fun with work is given very high priority

Primary Skill :Data Engineer

Secondary Skill : GCP, Big Query

Exp : 5 to 10 yrs

Job Description:

We are seeking a skilled Data Engineer with expertise in Google Cloud Platform (GCP), particularly in BigQuery and Pub/Sub, coupled with advanced proficiency in Python. As a Data Engineer, you will play a pivotal role in designing, implementing, and optimizing data pipelines, ensuring the smooth flow of data within our organization.

Responsibilities:

1. Architect and Develop Data Pipelines: Design, develop, and maintain scalable and efficient data pipelines using GCP services such as Pub/Sub for real-time data ingestion and BigQuery for storage and analysis.

2. Data Transformation and Processing: Implement data transformation processes to cleanse, enrich, and aggregate raw data from various sources, ensuring data quality and consistency.

3. Optimize Performance: Fine-tune data pipelines and queries to optimize performance and reduce latency, ensuring timely access to data for stakeholders.

4. Monitoring and Maintenance: Implement monitoring solutions to track pipeline performance and proactively address issues. Perform regular maintenance tasks to ensure the reliability and availability of data infrastructure.

5. Collaboration: Collaborate with cross-functional teams including Data Scientists, Software Engineers, and Business Analysts to understand data requirements and provide technical solutions to address business needs.

6. Documentation: Document data pipelines, processes, and best practices to ensure knowledge sharing and maintain a comprehensive understanding of data architecture.

Requirements:

1. Proficiency in GCP Services: Extensive hands-on experience with Google Cloud Platform services, particularly BigQuery for data storage and analysis, and Pub/Sub for real-time data streaming.

2. Python Programming: Strong programming skills in Python for data manipulation, scripting, and automation tasks. Experience with libraries such as Pandas, NumPy, and TensorFlow is highly desirable.

3. Data Modeling: Solid understanding of data modeling concepts and experience in designing efficient data models for analytics and reporting purposes.

4. SQL Skills: Proficiency in writing complex SQL queries for data extraction, transformation, and analysis within BigQuery.

5. Experience with Data Warehousing: Familiarity with data warehousing concepts and experience in implementing data warehouse solutions using GCP BigQuery or similar technologies.

6. Problem-solving Skills: Strong analytical and problem-solving skills with the ability to troubleshoot and resolve complex data-related issues.

7. Communication Skills: Excellent communication skills with the ability to effectively collaborate with cross-functional teams and articulate technical concepts to non-technical stakeholders.

8. Bachelor’s Degree: Bachelor’s degree in Computer Science, Engineering, Mathematics, or a related field. Advanced degree or relevant certifications in Data Engineering or GCP are a plus.