Data Engineer - Big Data Developer
Apply now »Date: 10 May 2023
Location: Madrid, ES
Company: Alstom
Req ID:408877
Leading societies to a low carbon future, Alstom develops and markets mobility solutions that provide the sustainable foundations for the future of transportation. Our product portfolio ranges from high-speed trains, metros, monorail, and trams to integrated systems, customised services, infrastructure, signalling and digital mobility solutions. Joining us means joining a caring, responsible, and innovative company where more than 70,000 people lead the way to greener and smarter mobility, worldwide
JOB TITLE.
Data Engineer - Big Data developer
DURATION: Permanent
LOCATION: Madrid. Travels at national and international levels.
PURPOSE OF THE JOB.
The purpose of the job is to develop through iterative process data Integration and analytics solutions. Designing solutions answering to Projects/Customer needs, demonstrate the value of the solutions through "proof of concept" and "proof of value" phases in cooperation with Alstom customers and partners, and developing and deploying final solutions.
You will Build data analytics solutions, developing Data applications (Data ingestion, Cleaning, data processing, etc.…) to provide state-of-the-art advanced analytic and quantitative tools and modelling techniques to derive business insights, solve complex business problems and improve decisions.
Reporting to Data Engineering group leader, the role will develop, support, and advice on the day-to-day Data platform, Data processing and Data analytics requirements of Business areas and Deployment Projects key operational processes while continually improving the impact of these processes
Key Accountabilities:
• Support the design of solutions that would provide added value to our customers through the use data processing, data analytics and visualization techniques.
• Define to complete data storage, analytic, and visualization processes anticipating the constraints (regulations, IT policy, confidentiality, etc.) that will apply for the development and deployment of data applications for our customers.
• Identify what data is available and relevant, including internal and external data sources, leveraging innovative data collection processes (IoT, geo-location, social media, weather forecast).
• Participate in all steps of the Data integration process (Ingestion & Data Transformation E2E refining data towards a Master Model).
• Leverage scheduler components like Apache AirFlow.
• Work with the latest version of Apache Spark using strongly typed languages like Scala and the SparkSQL APIs for transformations.
• Understand and use the latest data open storage formats such as Delta Lake or Iceberg
• Be part of the definition of the workflow pipeline and tackle new challenges when dealing and transforming from source, raw, harmonized, or master data layers.
• Learn new skills and tackle new technologies, been an engineer in a cool Smart Cities project where you will work both On-Premises and Cloud (currently Azure).
• Be comfortable working with microservices and distributed architectures in environments such as Kubernetes.
• Help along with the software industrialization and automation processes, being active towards promoting the use of VCS, SW tools for project building and unit testing. Further help when we design which tools to integrate for a full DevOps experience.
• “Code, code, code”, but do it using the latest and best IDEs with a strong focus for good code and quality.
• Leverage best practices in continuous integration and delivery.
• Help drive optimization, testing and tooling to improve data quality.
• Get to know your data and extend the Data Model as we encounter new challenges within our Smart City.
• Be a part of Data Governance team, ensuring data competition and quality.
• You’ve had exposure to data modeling, data access, and data storage techniques.
• You understand the value of collaboration within teams.
• Work alongside Data Scientist and the business in suggesting the best decisions.
• Participate in a fast-growing project under Innovation where technological experimentation through PoCs and PoVs is encouraged.
• Development of Data Quality work plans and frameworks to guarantee data quality throughout its life cycle.
• Our data comes from several origins, some real-time, so be ready to use Kafka and its APIs.
• Along with the use of distributed storage like S3, we also integrate Cassandra, so it is useful you also know your NoSQL like Cassandra.
The job will be located in the Smart Mobility innovation center in Madrid, Spain.
Educational Requirements
Mandatory:
- A degree in Computer Science or Engineering is required for this role.
- You should have minimum 2-3 years of working experience in developing Data applications, developing software for different sub system. Theoretical knowledge would not be considered.
Skills and Experience
Mandatory:
Experience in Data processing applications based on Spark and Hadoop architectures.
Strong experience in programming is required for this role (Agile methodologies). Experience in big data, Data Integration or Data analytics application development. Implementing and deploying Data processing and advanced analytics (Machine Learning/statistics) solutions for large enterprise or commercial customers in complex heterogeneous environments. Hands-on experience in the Big Data stack and related technologies:
• Scala. Akka, akka streams, akka http, typesafe, slick.
• Apache Spark, Spark SQL, Hive.
• Kafka
• Delta Lake/ Iceberg.
• PostgreSQL.
• Kubernetes and Docker.
• Airflow
• Python
• Linux
Fluency in English is essential while knowledge of Spanish or French will be a plus.
Desirable:
Knowledge on Prometheus, Grafana, Kafka, NoSQL (Cassandra, Redis), and Cloud environment management (Azure). Master Data Model knowledge and experience will be a plus.
Experience with Azure data technologies such as Azure SQL Database, Azure Data Factory, Purview, Databricks, Azure Data Lake Storage Gen 2 will be highly valued.
You should be flexible and passionate about data analytics. You take initiative and are proactive with strong organizational and planning skills along with excellent communication and presentation skills. You are goal oriented and possess customer approach by understanding the customer's decision-making process, goals, objectives, and strategies.
Competencies & Skills
-
You should be flexible and passionate about data analytics.
-
You take initiative and are proactive with strong organizational and planning skills along with excellent communication and presentation skills.
-
You are goal-oriented and possess a customer approach by understanding the customer's decision-making process, goals, objectives and strategies.
- Flexibility to work successfully with both local and remote/international teams.
- Ability to pay close attention to detail, meet deadlines, balance multiple tasks, work well under pressure, and work with frequent interruptions.
- Desire to be a team player and contributor in a highly collaborative environment across diverse and distributed teams.
Alstom is the leading company in the mobility sector, solving the most interesting challenges for tomorrow’s mobility. That’s why we value inquisitive and innovative people who are passionate about working together to reinvent mobility, making it smarter and more sustainable. Day after day, we are building an agile, inclusive and responsible culture, where a diverse group of people are offered opportunities to learn, grow and advance in their careers, with options across functions and geographic locations. Are you ready to join a truly international community of great people on a challenging journey with a tangible impact and purpose?
Equal opportunity statement:
Alstom is an equal opportunity employer committed to creating an inclusive working environment where all our employees are encouraged to reach their full potential, and individual differences are valued and respected. All qualified applicants are considered for employment without regard to race, colour, religion, gender, sexual orientation, gender identity, age, national origin, disability status, or any other characteristic protected by local law.
Job Segment:
Cloud, Testing, Developer, Computer Science, Linux, Technology