What you will be doing
Architect, build, test, and manage our common integrated data platform infrastructure from data collection and data processing all the way to data storage and data manipulation.
Establish an ongoing improvement plan for Cloud extensibility support operations, including technology reviews, root cause analysis, action reports analysis, and Cloud portfolio review to ensure adequate infrastructure to provide a suitable operating environment.
Manage processes related to modernization, architecture, and transformation through Cloud technologies adoption.
Maintain, develop, and optimize our batch processing pipelines and ingestion large data volumes; create streaming processes for real-time applications.
Evaluate project-level architecture reviews and vendor selection recommendations.
Improve scalability, stability, automation, accuracy, speed, cost, and efficiency of our existing Data Systems by performing query optimization, database tuning, indexing, and partitioning.
Support existing and design / develop new ETL / ELT processes that move, integrate, ingest, transform and manage data on Cloud.
Work with the OT & IT Security Team, Solution Architects, and Infrastructure Team over any security-related issues that arise and assist in log reviews & maintenance of operational activities.
What you’ll have :
University degree in Computer Science, Mathematics, Engineering, Information Technology, Management Science and Technology or other relevant fields.
Two (2)- four (4) years’ experience in a relevant position.
Experience of at least two (2) years in Azure Data Bricks / Data Warehouse / Azure Data Factory / Azure Data Delta Lake is mandatory.
In addition, knowledge of Apache Spark will be considered an asset.
Critical Functional Skills :
Ability to navigate the following technology stack : Python, SQL, Spark, Scala, PySpark, and Docker / Kubernetes.
Strong experience in Relational Databases (MySQL, SQL Server, or PostgreSQL) non-relational databases (Mongo DB or Cassandra. etc.).
DevOps and Azure experience with CI / CD and Infrastructure as Code tools is a plus.
Experience in open-source distributed streaming platforms (i.e., Kafka) that can be used to build real-time streaming data pipelines and applications.
Experience with integrating data from multiple heterogeneous sources and various formats (CSV, XML, JSON, Avro, Parquet)
Ability to manipulate, process, and extract value from large and disconnected datasets.
Experience in data visualization tools (Tableau) and reports to explore data and find insights.
Experience in building and optimizing big data’ data pipelines, architectures, and data sets and in the development of Data Ingestion Pipelines and Data Analysis Applications.
Experience performing root cause analysis on internal and external data and processes to answer specific business questions and identify opportunities for improvement.
Proficient understanding of microservices architecture and distributed systems.
Understanding and implementation experience with both relational and data warehousing models.
What you need to succeed in this role :
Superior analytical skills and a strong sense of urgency and ownership in your work.
Excellent verbal and written communication as well as presentation skills in both Greek and English.
A team player attitude with solid problem-solving skills.
Ability to work with cross-functional teams in a dynamic environment.