Job title: Data Science Engineer - EH
Job type: Permanent
Emp type: Full-time
Industry: Information Technology / IT
Salary: Negotiable
Location: Tokyo, Japan
Job published: 2025-01-21
Job ID: 46658

Job Description

Data Science Engineer 

 

■ Your Role and Responsibilities 

  •  Identify the data needs of data scientists at the company, document their requirements, and develop robust, secure, and scalable data pipelines to enable and accelerate their analyses.
  • Own critical data pipelines, help ensure their continued operations, and extend them to meet the needs of the business.
  • Be on the lookout for potential improvements to our data marts and aggregations that improve their scalability, make them easier for data scientists to use, and reduce potentially dangerous duplication.
  • Help with maintenance and extension of our internal ETL frameworks built on Python and Scala.
  • Contribute to the development of documentation and educational materials on tools and data pipelines owned by Data Science Foundations. Provide 1-on-1 project support to data scientists and help them get the most of our tools and data.

■ Work Location

・Tokyo, Japan 

 

■ Experience and Qualifications 

  • You enjoy problem-solving, learning new technologies, and helping others get their work done.
  • You are excited to work with data-scientists and business stakeholders to deliver real and visible business value.
  • You like to take ownership of your projects and independently build something new and immediately usable.
  • You have worked as a data engineer handling millions of records per day.
  • You have experience in SQL and Spark that goes beyond the basics and have worked with Python and/or Scala
  • You have experience with using a batch job tool such as Prefect or Airflow (we use Prefect)
  • You are comfortable working with existing code using git or another VCS in a team-setting.
  • You must be eligible to work in Japan and be able to conduct business in English to communicate with

 

■ Additional Preferred Qualifications

  • Experience with AWS or cloud computing and cloud infrastructure in general. Experience with SageMaker, Glue, EMR, S3, RDS, Redshift are a big plus.
  • You know how to maintain and optimize a PostgreSQL database.
  • Experience as a data scientist, scientific researcher, or in a data analytics role.
  • Worked with Terraform or similar Infrastructure as Code technologies.
  • You are familiar with the concept of CI/CD (CircleCI, Jenkins, ...).
  • You have worked on a payment platform or other financial technology field.
  • You have worked with and understand the concept of NoSQL databases and message brokers. Experience with ElasticSearch, Kafka, Cassandra would be useful.