World leader in gases, technologies and services for Industry and Health, Air Liquide is present in 80 countries with approximately 66,000 employees and serves more than 3,6 million customers and patients. Oxygen, nitrogen and hydrogen have been at the core of the company’s activities since its creation in 1902. Air Liquide’s ambition is to be the leader in its industry, delivering long-term performance and acting responsibly.
Candidate must have valid employment authorization in the U.S. and must not require visa sponsorship now or in the future. This position is not open for non-immigrant visa sponsorship.
Within Americas Hub Data & BI Team, and reporting to the Data & BI Product Manager, the AWS Big Data Technical Leader designs, develops and industrializes “Big Data” Cloud-based solutions, highly distributed which fuels strategic business decisions across the organisation. They act as design authority for Big Data developments on the data lake, and play an active role in engineering data pipelines, integrations, transformation workflows and managing data lake infrastructure and related toolings. They define technological roadmap, and fosters Agile and DevOps methodologies. They contribute in successful delivery of ongoing and future projects, contribute in scale, run and maintenance of existing and future products, continuous improvement and automation of the Americas Data Platform.
This position can be located in Houston Texas or Montreal Canada. No relocation benefits
Work with product manager in defining the strategy and implementation of the data lake, ETL, ingestion tools and big data solutions that fuels strategic business decisions across the organization
Assists in the execution of the Digital and IT roadmap
Play an active role in in defining technologies to be used for new Data & BI applications and improvements to existing applications
Creates and implements a technical strategy to enable the infrastructure to scale to larger projects and teams
Contributes to architecting complex big data solutions. Own the execution of the data lake engineering team, oversee process and delivery timelines to keep projects on schedule. Provide support for the data processing.
Establish best practices for naming, governing, managing, modeling, cleansing, transforming,
moving, storing, and searching all data. Define local norms for local purposes according to group recommendations
Participates in selection of Big Data Engineers, trains and lead their developments on project, run and maintenance of products
Leads a team of Big Data engineers who build data pipelines, integrations, transformation workflows and manage data lake/warehouse infrastructure and related tooling
Ensure high levels of data quality across the related business areas. Establish SLAs for all data and processes. Develop data quality checks for source and target data sets.
Establish SLAs for all data and processes.
Anticipate challenges by maintaining existing codebases and system infrastructures, as well as enhancing the development, staging, and production environments of our applications. Scale, maintain and improve the existing products.
Hand in code reviews before product roll-out aligned with data & BI standards, as well as in project planning and management, participate, contribute actively and provide input for the ongoing improvement of engineering practices and procedures. Design, implement, optimize , with data architects and devops, the data processing pipelines allowing data applications to be fed from various data sources
Foster an agile / devops methodology
Work Closely with Air Liquide Data & BI teams around the globe: for centralized support ( security, etc...), and alignment in terms of architecture and product orientation, to propose new features for the core product
Design and implement visualization and analysis tools, and automations, to facilitate the work of data developers, data scientists and data analysts
Works with his/her team as well as with cross-functional teams to continuously improve work processes. Continuously analyses the functioning of the activity, identifies causes of dysfunctions and proposes corrective measures accordingly.
BSC - Bachelor's of Science in Computer Science
4+ years of experience in Hadoop, Spark (PySpark, Scala), Python for data engineering and SQL
4 years of experience in Git development workflow, DevOps/CI-CD tools: JIRA, GitLab, GitLab CI, NEXUS, SonarQube
4+ years of experience in CloudFormation, IAM, Lambda, SQS, Athena, Airflow, EMR,
Glue, S3, Unix
4+ years of experience in AWS DevOps
3+ years of experience with Docker and REST APIs
AWS Certified Big Data - Specialty
AWS Certified Solutions Architect - Associate
AWS Certified Developer - Associate
AWS Certified SysOps Administrator - Associate
Experience in AI
Experience with Azure (Azure Data Warehouse, SQL database,..) or time series databases
Experience in selecting and implementing Machine Learning libraries and standard algorithms and metrics