Post a job

Data Lead

Codility logo

Location
DE, GB + 1 more
Codility

Job Description

Codility is a leading platform for assessing and advancing technical skills in an AI-driven world.

Over the last 10+ years we pioneered technical assessments for engineers as part of their recruitment process and helped 5000+ customers including world’s biggest names like Microsoft, Google, and Amazon build their engineering teams with our technical screening and interviewing solutions.

The role

You will lead the Data Engineering team, which tackles complex, cross-regional, multi-domain challenges impacting Codility's business and its customers. The team is involved in everything related to the Data platform and analytics. It is spearheading Codility’s machine learning initiatives (ML) translated into value-bringing products and features for our customers.

As a team, we're never satisfied with maintaining the status quo; we're always seeking to innovate and improve. We believe in innovation through iteration, following a cycle of building, measuring, learning, and refining. Collaboration is central to our approach, and we strive to create a supportive, welcoming, and innovative environment where teamwork thrives. We work closely with various functions across the company to achieve our goals.

What will you do?

  • Lead a cross-functional team to develop new data products for the company and the company’s customers.
  • Enable the Codility departments to make data-driven decisions via easy access, adding transforming, and retrieving business-relevant data
  • Promote and Implement observability and monitoring within the data platform.
  • Contribute to the design and implementation of Codility’s data architecture, with a focus on real-time processing, reliability, scalability, and cost-efficiency.
  • Collaborate on the delivery, storage, and transformation of data, contributing to Codility’s data models while ensuring compliance with Data laws and regulations
  • Take full technical ownership of data products, from understanding business requirements to deployment and beyond.
  • Build and cultivate a strong, cohesive data engineering team where individuals continually develop their skills.
  • Establish strong foundations for data quality and promote them throughout the organization.

You have:

  • Proven experience in leading cross-functional Data teams (data platform and data engineering/science) with customer-centric, product mindset
  • Hands-on experience in deploying, maintaining, and updating data models in production
  • Deep understanding of SQL and NoSQL Databases
  • Experience in developing, evaluating, deploying, running, and maintaining ML models (AWS SageMaker, MLFlow).
  • Experience with Spark, and orchestration tools (Airflow, dbt, or similar)
  • Hands-on experience building ETL and ELT pipelines
  • Experience with Data Warehouse or Lakehouse solutions (eg., for example, Google BigQuery, Amazon Redshift, Snowflake, Databricks, AWS Data Lake)
  • Experience deploying and managing data technologies on AWS or GCP (preferably in an IaaC manner)
  • Solid programming skills in Python, Scala or R

And:

  • You are not afraid to ask questions, give or take feedback
  • You are conscious of security, governance, metrics, performance and deliverables
  • You possess a data-driven mindset
  • You have good communication skills
  • You take the driver’s seat, never shy away from challenges, and get things done!

Your ideal data tech stackexpertise:

  • Cloud: AWS services (S3, Glue, EMR, ECS, Lambdas, Redshift.. )
  • ML: Pandas, NumPy, Scikit-learn, Jupyter, Python, SQL, Tensorflo
  • Big Data Technologies: Presto, Spark
  • Orchestration: Airflow
  • BI tool: Tableau
  • Database: PostgreSQL
  • IaaC: Terraform, Atlantis
  • Monitoring: Prometheus Stack, Grafana

Strong Advantage:

  • Experience with LangChain or other Al agent frameworks.
  • Experience with LLM observability and the optimal use of LLMs, the tradeoff between fine-tuning and RAGs, etc.
  • Familiarity with data structures for ML and Al programs, including RAG and vector databases.

How we behave.

We're human. Our diversity and unique experience make us strong. We allow ourselves to be vulnerable and treat one another with grace. We take ownership. We expect one another to take initiative and trust each person to make decisions based on the best available data. We have passion, perseverance and urgency to hit our goals and work together to create the best outcomes. We think big. We innovate and challenge the status quo to maximize the value we deliver to our users. We constantly experiment with new ways to drive excellence. We're real. We are honest with ourselves and one another. We listen, speak up, and are ready to change our minds.

DISCLAIMER.

At Codility, we know that great work isn’t done without a phenomenal team. We are always looking to hire the absolute best talent and recognize that diversity in our experiences and backgrounds is what makes us stronger. We insist on an inclusive culture where everyone feels safe to contribute and help us innovate. We hire candidates of any race, color, ancestry, religion, national origin, sexual orientation, gender identity, age, marital or family status, disability, or veteran status. These differences are what enable us to work towards the future we envision for ourselves, our product, our customers, and our world.

Advice from our career coach

As someone who has led cross-functional Data teams, a successful applicant for the Data Engineering team at Codility should showcase their experience and skills in developing new data products and enabling data-driven decisions. Here are some specific tips to help you stand out as an applicant:

  • Highlight your experience in leading cross-functional Data teams with a customer-centric, product mindset.
  • Showcase your hands-on experience in deploying, maintaining, and updating data models in production.
  • Demonstrate your deep understanding of SQL and NoSQL Databases.
  • Emphasize your experience in developing, evaluating, deploying, running, and maintaining ML models using tools like AWS SageMaker and MLFlow.
  • Showcase your hands-on experience in building ETL and ELT pipelines and working with Data Warehouse or Lakehouse solutions like Google BigQuery, Amazon Redshift, or Snowflake.
  • Highlight your programming skills in Python, Scala, or R.
  • Emphasize your ability to take ownership of data products and drive them from understanding business requirements to deployment.
  • Showcase your experience with cloud services like AWS (S3, Glue, Redshift) and Big Data technologies like Presto and Spark.
  • Demonstrate your expertise in monitoring with tools like Prometheus Stack and Grafana.
  • Highlight your experience with BI tools like Tableau and orchestration tools like Airflow.

Apply for this job

Expired?

Please let Codility know you found this job with RemoteJobs.org. This helps us grow!

About the job

Aug 31, 2024

Full-time

  1. DE Germany
  2. GB United Kingdom
  3. PL Poland
RemoteJobs.org mascot