Operations Data Engineer
TITAN Group established the Digital Center of Competence in 2020 to implement major digital transformation projects, through a core team of specialists, including Data Scientists and advanced Analytics Experts, who coordinate digital activities across the Group to enhance the new Digital operating model.
As an Operations Data Engineer, you will enable data-driven decision-making within the organization's digital infrastructure and system operations.
You will be responsible for the continuous improvement of the TITAN Cloud Integrated platform, focusing on architecting, designing, and developing Cloud solutions.
The position requires expert proficiency in technology and architecture, hands-on coding skills, and the ability to translate high-level business requirements into technology execution.
Are you passionate about working with big data sets advanced cloud analytic tools, bringing data sets together to support digital business requirements, initiatives and being part of TITAN’s vision for Digital Transformation?
Apply now to join our dynamic team at a very exciting time and have the opportunity to work with big data and advanced high-edge digital technologies in an international environment.
What you will be doing :
Architect, build, test, and manage our common integrated data platform infrastructure from data collection and data processing all the way to data storage and data manipulation.
Establish an ongoing improvement plan for Cloud extensibility support operations, including technology reviews, root cause analysis, action reports analysis, and Cloud portfolio review to ensure adequate infrastructure to provide a suitable operating environment.
Manage processes related to modernization, architecture, and transformation through Cloud technologies adoption.
Maintain, develop, and optimize our batch processing pipelines and ingestion large data volumes; create streaming processes for real-time applications.
Evaluate project-level architecture reviews and vendor selection recommendations.
Improve scalability, stability, automation, accuracy, speed, cost, and efficiency of our existing Data Systems by performing query optimization, database tuning, indexing, and partitioning.
Support existing and design / develop new ETL / ELT processes that move, integrate, ingest, transform and manage data on Cloud.
Work with the OT & IT Security Team, Solution Architects, and Infrastructure Team over any security-related issues that arise and assist in log reviews & maintenance of operational activities.
What you’ll have :
University degree in Computer Science, Mathematics, Engineering, Information Technology, Management Science and Technology or other relevant fields.
Two (2)- four (4) years’ experience in a relevant position.
Experience of at least two (2) years in Azure Data Bricks / Data Warehouse / Azure Data Factory / Azure Data Delta Lake is mandatory.
In addition, knowledge of Apache Spark will be considered an asset.
Critical Functional Skills :
Ability to navigate the following technology stack : Python, SQL, Spark, Scala, PySpark, and Docker / Kubernetes.
Strong experience in Relational Databases (MySQL, SQL Server, or PostgreSQL) non-relational databases (Mongo DB or Cassandra. etc.).
DevOps and Azure experience with CI / CD and Infrastructure as Code tools is a plus.
Experience in open-source distributed streaming platforms (i.e., Kafka) that can be used to build real-time streaming data pipelines and applications.
Experience with integrating data from multiple heterogeneous sources and various formats (CSV, XML, JSON, Avro, Parquet)
Ability to manipulate, process, and extract value from large and disconnected datasets.
Experience in data visualization tools (Tableau) and reports to explore data and find insights.
Experience in building and optimizing big data’ data pipelines, architectures, and data sets and in the development of Data Ingestion Pipelines and Data Analysis Applications.
Experience performing root cause analysis on internal and external data and processes to answer specific business questions and identify opportunities for improvement.
Proficient understanding of microservices architecture and distributed systems.
Understanding and implementation experience with both relational and data warehousing models.
What you need to succeed in this role :
Superior analytical skills and a strong sense of urgency and ownership in your work.
Excellent verbal and written communication as well as presentation skills in both Greek and English.
A team player attitude with solid problem-solving skills.
Ability to work with cross-functional teams in a dynamic environment.