machine learning

I Quit My $130,000 ML Engineer Job After Learning 4 Lessons

working as a machine learning engineer at a Big Tech company. On paper, I had a dream job: Flexible working Smart and friendly colleagues Great perks and advantages Good work-life balance Barely any meetings And my compensation was well over...

Scaling ML Inference on Databricks: Liquid or Partitioned? Salted or Not?

Introduction a continuous variable for 4 different products. The machine learning pipeline was in-built Databricks and there are two major components.  Feature preparation in SQL with serverless compute. Inference on an ensemble of several hundred models using...

The Gap Between Junior and Senior Data Scientists Isn’t Code

five minutes on LinkedIn or X, you’ll notice a loud debate in the info science industry. It’s been out for some time now, but this week, it finally caught my attention. As much as...

A Generalizable MARL-LP Approach for Scheduling in Logistics

Introduction that always operates with surprising inefficiency: manual processes, piles of paperwork, legal complexities. Many corporations still run on paper or Excel and don’t even collect data on their shipments. But what if an organization...

Designing Data and AI Systems That Hold Up in Production

Do you see yourself as a full-stack developer? How does your experience across the entire stack (from frontend to database) change the way you view the information scientist role? I do, but not within the...

Scaling Feature Engineering Pipelines with Feast and Ray

project involving the construct of propensity models to predict customers’ prospective purchases, I encountered feature engineering issues that I had seen quite a few times before. These challenges might be broadly classified into two categories: 1)...

Construct Effective Internal Tooling with Claude Code

is incredibly effective at quickly build up recent applications. That is, in fact, super useful for any programming task, whether it's working on an existing legacy application or a brand new codebase. Nevertheless, from...

AI in Multiple GPUs: Gradient Accumulation & Data Parallelism

is an element of a series about distributed AI across multiple GPUs: Introduction Distributed Data Parallelism (DDP) is the primary parallelization method we’ll have a look at. It’s the baseline approach that’s all the time utilized in...

Recent posts

Popular categories

ASK ANA