Skip to content
View TeoZosa's full-sized avatar

Highlights

  • Pro

Block or report TeoZosa

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
TeoZosa/README.md

Howdy! I'm Teo, great to meet you 👋

I build production ML systems that deliver customer impact at scale by any means necessary.

  • Sometimes this means getting in the weeds wrangling raw data; training and evaluating models; or setting up super reliable high-throughput, low-latency, highly-observable online/offline ML feature and model serving systems1.
  • Other times this means sitting in meetings with business, product, and technical stakeholders from all corners of the company, as well as partners and colleagues across the industry, to define the right problems to tackle and the right things to build.
  • No matter what, it always means listening to the people we're serving, making sure we're solving the most important problems, and thinking beyond what's possible to what is actually needed to add the most value to our customers2.

I'm currently focus-firing all of the above as a Staff Machine Learning Engineer at Mercari, Japan’s largest C2C e-commerce marketplace, to support the world-class AI teams and applications across the company.


Talks

2023

2024

2025

  • [Coming Soon] FOSSASIA Summit 2025: LLMOps for Eval-Driven Development at Scale5

Footnotes

  1. My current personal record was for an AI search ranking system at Mercari:

    1. Offline data & model training pipelines that processed petabytes of data.
    2. Online serving system that handled nearly 6K RPS of search traffic with average e2e latency <35ms (p95 <100ms) and five 9's uptime.
      1. Streaming real-time feature ingestion pipeline that processed over 100K client events per second and over 2.3 terabytes of data per day; significant optimization for a monthly operational cost of $567/mo.
      2. Feature store that served over 3.5M read/write RPS with average read latency <3ms (p95 <10ms).
      3. Model server that handled nearly 6K RPS with average read latency <13ms (p95 <38ms).
  2. If A/B test results are at least a decent proxy for customer value, the work I spearheaded at Mercari in AI search ranking led to a projected $50M/year revenue increase.

  3. w/ @ginstrom 2

  4. w/ @chingisooinar 2 3

  5. w/ @jehandadk

Pinned Loading

  1. structlog-sentry-logger structlog-sentry-logger Public

    A multi-purpose, pre-configured, performance-optimized structlog logger with (optional) Sentry integration via structlog-sentry

    Python 25 2

  2. cookiecutter-cruft-poetry-tox-pre-commit-ci-cd cookiecutter-cruft-poetry-tox-pre-commit-ci-cd Public

    A Modern DevSecOps-centric Cookiecutter template for Python packages and/or projects

    Makefile 18 4

  3. jupyterhub-fastbook jupyterhub-fastbook Public

    Deploy JupyterHub to your Kubernetes cluster pre-loaded with fast.ai's Practical Deep Learning for Coders course notebooks and all required dependencies for an all-in-one "It Just Works™" deployment.

    Makefile

  4. ci-docker-images ci-docker-images Public

    Pre-built containerized base environments for Python project CI pipelines. Emulates a local development environment, containing (pun intended) pyenv-managed Python versions (`3.7`, `3.8`, and `3.9`…

    Dockerfile

  5. pytudes pytudes Public

    Miscellaneous programming challenges for the purpose of personal/professional edification, inspired by Peter Norvig's pytudes repository

    Python 2