Το LinkedIn και τρίτα μέρη χρησιμοποιούν ουσιώδη και μη ουσιώδη cookie για την παροχή, τη διασφάλιση, την ανάλυση και τη βελτίωση των Υπηρεσιών τους, καθώς και για την προβολή σχετικών διαφημίσεων (όπως επαγγελματικές διαφημίσεις και αγγελίες θέσεων εργασίας) εντός και εκτός LinkedIn. Μάθετε περισσότερα στην Πολιτική για τα Cookie μας.
Επιλέξτε Αποδοχή για να παράσχετε τη συναίνεσή σας ή Απόρριψη για να απορρίψετε τα μη ουσιώδη cookie για αυτήν τη χρήση. Μπορείτε να ενημερώσετε τις επιλογές σας οποιαδήποτε στιγμή από τις ρυθμίσεις σας.
Are you passionate about solving complex big data analytics problems using cutting-edge technologies?
EPAM is looking for a skilled Senior Palantir Data Engineer to join our growing, globally distributed team.
In this role, you’ll work on a high-impact Data Transformation project with our client from the insurance sector. This initiative leverages Big Data and Machine Learning technologies to shape data-driven decisions in the Property & Casualty business domain.
If you’re eager to apply your expertise in Python/PySpark, SQL and Palantir to design and implement complex data pipelines, while collaborating with a multicultural and dynamic team, we’d love to hear from you! We offer a hybrid work model with a mix of remote and on-site work at EPAM’s Nicosia office.
Responsibilities
Lead the design and implement robust, large-scale data pipelines and analytics solutions
Oversee the monitoring and optimization of data pipelines for performance and scalability using advanced tools and techniques, including Python/PySpark and structured query languages
Optimize data workflows to support critical decision-making processes
Harness state-of-the-art tools and technologies (including Palantir Foundry) to address new and emerging business challenges
Partner with cross-functional and globally distributed teams (e.g., data scientists, analysts, business stakeholders) to align project goals and execution strategies
Contribute to a global strategic initiative focused on enhancing the ability to make data-driven decisions across the Property & Casualty value chain
Stay ahead of emerging technologies and trends (e.g., Generative AI, Machine Learning) and recommend potential applications in the data ecosystem
Requirements
A Bachelor’s degree (or equivalent) in Computer Science, Data Science or a related discipline
5+ years of experience working with large-scale distributed computing systems
Proficiency in Python/PySpark to build and optimize complex data pipelines
Hands-on experience working with Databricks for large-scale data processing and analytics
Strong SQL skills (preferably Spark SQL) for data querying and manipulation
Deep understanding of data warehousing concepts and ELT techniques
Experience with Palantir Foundry is a must
Familiarity with Agile and Scrum development methodologies
Nice to have
Knowledge of HTML, CSS, JavaScript and Gradle
Experience in the Insurance domain or the financial industry
Familiarity with Microsoft Power BI
Exposure to Machine Learning or Generative AI technologies
We offer
Private healthcare insurance
Global travel medical and accident insurance
Regular performance assessments
Referral bonuses
Family friendly initiatives
Learning and development opportunities including in-house training and coaching, professional certifications, over 22,000 courses on LinkedIn Learning Solutions and much more
*All benefits and perks are subject to certain eligibility requirements
Επίπεδο αρχαιότητας
Μεσαίο-ανώτερο επίπεδο
Τύπος εργασίας
Πλήρους απασχόλησης
Αρμοδιότητες θέσης εργασίας
Βιομηχανίες
Οι παραπομπές αυξάνουν τις πιθανότητές σας για συνέντευξη κατά EPAM Systems 2x
Ξεκλειδώνουμε τη γνώση στην κοινότητα με έναν νέο τρόπο. Οι ειδικοί προσθέτουν χρήσιμες πληροφορίες απευθείας σε κάθε άρθρο, με τη βοήθεια της τεχνητής νοημοσύνης.