AI & Data
In this age of disruption, organizations need to navigate the future with confidence, embracing decision making with clear, data-driven choices that deliver enterprise value in a dynamic business environment.
The AI & Data team leverages the power of data, analytics, robotics, science and cognitive technologies to uncover hidden relationships from vast troves of data, generate insights, and inform decision-making. Together with the offering portfolio helps clients transform their business by architecting organizational intelligence programs and differentiated strategies to win in their chosen markets.
AI & Data will work with our clients to:
- Implement large-scale data ecosystems including data management, governance and the integration of structured and unstructured data to generate insights leveraging cloud-based platforms
- Leverage automation, cognitive and science-based techniques to manage data, predict scenarios and prescribe actions
- Drive operational efficiency by maintaining their data ecosystems, sourcing analytics expertise and providing As-a-Service offerings for continuous insights and improvements
Education and Experience
- Bachelor’s or Master’s degree in Computer Science, Engineering, Information Technology, or related field.
- 3–6 years of hands-on experience in Scala development, preferably in a data engineering or data pipeline context.
Key Responsibilities
- Collaborate with business analysts and stakeholders to gather and analyze requirements for data pipeline solutions.
- Design, develop, and maintain scalable data pipelines using Scala and related technologies.
- Write clean, efficient, and well-documented Scala code for data ingestion, transformation, and processing.
- Develop and execute unit, integration, and end-to-end tests to ensure data quality and pipeline reliability.
- Orchestrate and schedule data pipelines using tools such as Apache Airflow, Oozie, or similar workflow schedulers.
- Monitor, troubleshoot, and optimize data pipelines for performance and reliability.
- Participate in code reviews, provide constructive feedback, and adhere to best practices in software development.
- Document technical solutions, data flows, and pipeline architectures.
- Work closely with DevOps and Data Engineering teams to deploy and maintain solutions in production environments.
- Stay current with emerging technologies and industry trends in big data and Scala development.
Required Skills & Qualifications
- Strong proficiency in Scala, including functional programming concepts.
- Experience building and maintaining ETL/data pipelines.
- Solid understanding of data structures, algorithms, and software engineering principles.
- Experience with workflow orchestration/scheduling tools (e.g., Apache Airflow, Oozie, Luigi, or similar).
- Familiarity with distributed data processing frameworks (e.g., Apache Spark, Kafka, Flink).
- Proficiency in writing unit and integration tests for data pipelines.
- Experience with version control systems (e.g., Git).
- Strong problem-solving skills and attention to detail.
- Excellent communication and collaboration skills.
Preferred Skills & Qualifications
- Experience with cloud platforms (AWS, Azure, or GCP) and related data services.
- Knowledge of SQL and NoSQL databases (e.g., PostgreSQL, Cassandra, MongoDB).
- Familiarity with containerization and orchestration tools (Docker, Kubernetes).
- Exposure to CI/CD pipelines and DevOps practices.
- Experience with data modeling and data warehousing concepts.
- Knowledge of other programming languages (e.g., Python, Java) is a plus.
- Experience working in Agile/Scrum environments.