職務内容
Data Science Engineer
■ Your Role and Responsibilities
- Identify the data needs of data scientists at the company, document their requirements, and develop robust, secure, and scalable data pipelines to enable and accelerate their analyses.
- Own critical data pipelines, help ensure their continued operations, and extend them to meet the needs of the business.
- Be on the lookout for potential improvements to our data marts and aggregations that improve their scalability, make them easier for data scientists to use, and reduce potentially dangerous duplication.
- Help with maintenance and extension of our internal ETL frameworks built on Python and Scala.
- Contribute to the development of documentation and educational materials on tools and data pipelines owned by Data Science Foundations. Provide 1-on-1 project support to data scientists and help them get the most of our tools and data.
■ Work Location
・Tokyo, Japan
■ Experience and Qualifications
- You enjoy problem-solving, learning new technologies, and helping others get their work done.
- You are excited to work with data-scientists and business stakeholders to deliver real and visible business value.
- You like to take ownership of your projects and independently build something new and immediately usable.
- You have worked as a data engineer handling millions of records per day.
- You have experience in SQL and Spark that goes beyond the basics and have worked with Python and/or Scala
- You have experience with using a batch job tool such as Prefect or Airflow (we use Prefect)
- You are comfortable working with existing code using git or another VCS in a team-setting.
- You must be eligible to work in Japan and be able to conduct business in English to communicate with
■ Additional Preferred Qualifications
- Experience with AWS or cloud computing and cloud infrastructure in general. Experience with SageMaker, Glue, EMR, S3, RDS, Redshift are a big plus.
- You know how to maintain and optimize a PostgreSQL database.
- Experience as a data scientist, scientific researcher, or in a data analytics role.
- Worked with Terraform or similar Infrastructure as Code technologies.
- You are familiar with the concept of CI/CD (CircleCI, Jenkins, ...).
- You have worked on a payment platform or other financial technology field.
- You have worked with and understand the concept of NoSQL databases and message brokers. Experience with ElasticSearch, Kafka, Cassandra would be useful.