Heute Mathe, morgen Lokad

Am 12. Juni 2018 bin ich auf Kurzbesuch in meiner Uni, an der TU Darmstadt und halte einen Vortrag im Seminar ‘Heute Mathe Morgen ?’ und spreche über meine in Paris ansässige Firma Lokad:

12.06.18 — 13:30 Uhr in S2|15 51

Wer Interesse hat mehr zu erfahren über:

  • Deep Learning / Machine Learning in der Industrie,
  • Quantitative Logistik Optimierung,
  • was endliche l_1-Folgen mit Logistik zu tun haben,
  • Arbeiten in Frankreich,

oder wer tendenziell Interesse hat

ist herzlich willkommen!

Bitte teilen!


On June 12th  I have the great honor to be back in Darmstadt! I’ll be speaking in the alumni seminar ‘Heute Mathe Morgen ?’

Deep Learning: What it is and how it relates to supply chains

Disclaimer: In this post I’m going to write about how we use Deep Learning in my company, Lokad.


When you follow the news about deep learning, you might have come across exciting breakthroughs such as algorithms which are able to colorize black and white photographs or automatic real-life translations of texts on pictures taken by a phone app .  While these are all pretty cool applications, they do not immediately give any direct use cases for most traditional businesses. At Lokad, our goal is to translate the stunning reach of deep learning capabilities into the real world, to optimize supply chains everywhere.

So, before going into detail how we do that, let me quickly and very roughly summarize what Deep Learning actually entails without going too much into technical details.

First of all, deep learning is a flavor of machine learning. Regular non-machine learning algorithms require full prior knowledge on the task (and no training data whatsoever). An expert-knowledge approach to demand forecasting would require you to specify in advance all specific rules and patterns such as

“All articles that have category=Spring will peak in May and slowly die down until October.”

This, however may only be true for some products of this category. It is also possible that  there might be subcategories that behave a bit differently and so on. Combining these with a moving average forecast already yields an overall understanding of future demand which is not so far from reality.

However it does have the following downsides:

  1. It does not embrace uncertainty — In our experience, risk and uncertainty are crucial for supply chains, since it’s mostly the boundary cases that can be either very profitable or very costly if ignored,
  2. You have to maintain and manage the complexity of your rule set – An approach is only as powerful as the set of rules that are applied to it. Maintaining rules is very costly i.e. for each rule in the algorithm, we calculate there is an initial cost of about 1 man-day of implementation, testing and proper documentation initially and about half a day of maintenance. Assuming you keep on refining your rules and therefore have to readjust the old ones this yields a cost of 8k € per rule for a five year period. It is worth noting that this only applies  for one rule and does not take into account the exponential increase in complexity that arises when dealing with more complex product portfolios. Even demand patterns for small businesses usually exhibit dozens of influences making their maintenance incredibly costly.

Now imagine that there is a technology that could, like a human child, learn on its own to deduce patterns from data and could thus independently predict how your portfolio of products develop throughout a year.

Just like a child in development, a deep learning algorithm will try to make sense of the world by trying to deduce correlations from observations. It will test them and discard those that do not make any sense for the remaining data.

Again following our analogy, like a child learning to makes sense of the world, a deep learning algorithm is consuming lots and lots of data and the key lies in grasping the information that is actually relevant. While a child in a big city might be completely overwhelmed with all the different colors, noises and smells, it will learn later that the traffic lights are the ones to watch out for in combination with noises coming from approaching cars that are most critical when crossing a street. The same mechanism is in place for deep learning. The algorithm may process a vast amount of data and needs to find out the essence of what drives demand.

The way to figure out what is important and what is not is carried out via repeating similar situations several times, like you would repeat correct traffic behavior with a child. A human brain is highly parallelizing its sensory input processing and reaction, so that it is able to react quickly to urgent new data such as a car that is approaching while crossing the street.

With the rise of big data, parallelization became also a key topic driving efficiency and, in fact, feasibility of a “human-like” autonomous learning process.

At Lokad, we actually use the parallelized computing power high end gaming graphic cards in our cloud servers, to efficiently run our optimization for our clients, processing for a portfolio of 10.000 products with five years of sales data in less half an hour while largely outperforming any conventional moving average based algorithms (or even Lokad’s own earlier generation machine learning forecasts) in accuracy.  

Lokad then uses the demand forecasting results which come in a probabilistic format to optimize the supply chain decisions taking into account economic drivers such as one’s stance on growth vs. profitability. With these analyses, Lokad directly delivers the best supply chain  decisions such as purchase orders or dispatching decisions. “Best” here refers to the economic driver set up (i.e. growth vs. profitability, opportunity costs etc.) that has been put in place supply chain decisions. It it will scale with the business as one’s portfolio and demand patterns become more complex making any hard coded demand forecasting rules which need to be maintained by a human completely obsolete.  



Average Developer salary in Germany 58k €, 261 working days – 30 days of vacation in 2018 yields a 250€ manday rate)

Repost: Can an algorithm beat VOGUE in predicting fashion trends?

What is Fashion’s next big trend is a three digit billion dollar question. It is however not a question, that, personally, I thought I wouldn’t have much to contribute to, given that I work in a technology company, where the biggest trend seems to be goofy meme t-shirts. So, naturally, our approach to fashion trends is completely non-traditional and does not rely on our own fashion expertise (which is probably a good thing).

Traditionally, trend scouting starts by closely following high fashion designers and fashion shows, which may be picked up later by the mainstream fashion industry.

Other trend sources can come from films and TV series, influences from YouTube and the blogosphere. To determine exactly what will sell in the next seasons used to be a question only industry experts, with decades of experience, could determine. Then, it was more of a guessing game, when a trend wave would pick up and how long it would last before slowly ebbing out or being completely consumed by another trend.

At Lokad, our data scientists have developed algorithms to revolutionize this trend scouting process, allowing for large scale reliable and accurate sales predictions based on prior seasons’ sales history. Furthermore, while trend-scouting may be the holy grail of purchasing decisions for fashion organisation, identifying statistical noise is equally important and usually completely overlooked. Excluding noise in demand, that is observed “links” between products that are happening at random without any particular trend connection, can boost the effectiveness of any buying decisions.

To find out what actually drives sales, that is what attributes of a product make customers buy a product is somewhat elusive. Defining a product for now is more of an art than a science. At Lokad, our quantitative approach allows us to study precisely what has triggered sales, may it be a fancy product name, a combination of price, value of material or may be a certain visible accessory, which made a customer choose one product over other similar products.

In particular, Lokad can leverage similarities between products to predict sales quantities of products that have never been sold anywhere by comparing it to sales of other products in the past. To look for similarities between products, Lokad analyses not only product attributes such as materials or colours, but also envisions leveraging the selling history of customers.

Identifying early trend adopters within a customer base allows Lokad to obtain a next season “preview”: for example, imagine you have that one very fashionable colleague who always wears a certain colour before everyone else does. With big data crunching tools, Lokad can identify this colleague and through their current shopping habits can pick up what many others will want to buy next season.

However, in the end, there is no way around a team of fashion experts who can distinguish the mega trends from the “ordinary” ones. However, with Lokad, this expert group can leverage their insights on a much higher scale with high statistically sound computational power – minimizing costs for any brand or market place. This is what we refer to as augmented human intelligence, putting the enablement of the expert group as the ultimate goal of our work.


This article was first published on LinkedIn on October 14, 2017. 

Disclaimer: Katharina currently works as a supply chain scientist at Lokad.