Job title: Data Science Engineer - EH
Job type: Permanent
Emp type: Full-time
Industry: Information Technology / IT
Salary: Negotiable
Location: Tokyo, Japan
Job published: 2025-01-21
Job ID: 46658

Job Description

Data Science Engineer 

 

■ Your Role and Responsibilities 

  •  Identify the data needs of data scientists at the company, document their requirements, and develop robust, secure, and scalable data pipelines to enable and accelerate their analyses.
  • Own critical data pipelines, help ensure their continued operations, and extend them to meet the needs of the business.
  • Be on the lookout for potential improvements to our data marts and aggregations that improve their scalability, make them easier for data scientists to use, and reduce potentially dangerous duplication.
  • Help with maintenance and extension of our internal ETL frameworks built on Python and Scala.
  • Contribute to the development of documentation and educational materials on tools and data pipelines owned by Data Science Foundations. Provide 1-on-1 project support to data scientists and help them get the most of our tools and data.

■ Work Location

・Tokyo, Japan 

 

■ Experience and Qualifications 

  • You enjoy problem-solving, learning new technologies, and helping others get their work done.
  • You are excited to work with data-scientists and business stakeholders to deliver real and visible business value.
  • You like to take ownership of your projects and independently build something new and immediately usable.
  • You have worked as a data engineer handling millions of records per day.
  • You have experience in SQL and Spark that goes beyond the basics and have worked with Python and/or Scala
  • You have experience with using a batch job tool such as Prefect or Airflow (we use Prefect)
  • You are comfortable working with existing code using git or another VCS in a team-setting.
  • You must be eligible to work in Japan and be able to conduct business in English to communicate with

 

■ Additional Preferred Qualifications

  • Experience with AWS or cloud computing and cloud infrastructure in general. Experience with SageMaker, Glue, EMR, S3, RDS, Redshift are a big plus.
  • You know how to maintain and optimize a PostgreSQL database.
  • Experience as a data scientist, scientific researcher, or in a data analytics role.
  • Worked with Terraform or similar Infrastructure as Code technologies.
  • You are familiar with the concept of CI/CD (CircleCI, Jenkins, ...).
  • You have worked on a payment platform or other financial technology field.
  • You have worked with and understand the concept of NoSQL databases and message brokers. Experience with ElasticSearch, Kafka, Cassandra would be useful.

 

File types (doc, docx, pdf, rtf, png, jpeg, jpg, bmp, jng, ppt, pptx, csv, gif) size up to 5MB
File types (doc, docx, pdf, rtf, png, jpeg, jpg, bmp, jng, ppt, pptx, csv, gif) size up to 5MB