Depends on the definition

it's about machine learning, data science and more

How the LIME algorithm fails

You maybe know the LIME algorithm from some of my earlier blog posts. It can be quite useful to “debug” data sets and understand machine learning models better. But LIME is fooled very easily. We use the eli5 TextExplainer which is based on LIME and the 20newsgroup data set to show how LIME can fail.

Cluster discovery in german recipes

If you are dealing with a large collections of documents, you will often find yourself in the situation where you are looking for some structure and understanding what is contained in the documents. Here I’ll show you a convenient method… Continue Reading →

That was PyCon DE & PyData Berlin 2019

Last weekend PyCon DE and PyData Berlin joined in Berlin for a great conference event that I was lucky to attend. The speaker line-up was great and often it was hard to choose which talk or tutorial to attend. I… Continue Reading →

Model uncertainty in deep learning with Monte Carlo dropout in keras

Deep learning models have shown amazing performance in a lot of fields such as autonomous driving, manufacturing, and medicine, to name a few. However, these are fields in which representing model uncertainty is of crucial importance. The standard deep learning… Continue Reading →

Interpretable Named entity recognition with keras and LIME

In the previous posts, we saw how to build strong and versatile named entity recognition systems and how to properly evaluate them. But often you want to understand your model beyond the metrics. So in this tutorial I will show you how you can build an explainable and interpretable NER system with keras and the LIME algorithm.

Introduction to entity embeddings with neural networks

Since a lot of people recently asked me how neural networks learn the embeddings for categorical variables, for example words, I’m going to write about it today. In this article you will learn what an embedding layer really is and how neural nets can learn representations for categorical variables with it.

Introduction to n-gram language models

You might have heard, that neural language models power a lot of the recent advances in natural language processing. Namely large models like Bert and GPT-2. But there is a fairly old approach to language modelling that is quite successful… Continue Reading →

Learn to identify ingredients with neural networks

Today, I want to show you how you can build an NLP application without explicitly labeled data. I use the “German Recipes Dataset” I recently published on kaggle, to build a neural network model, that can identify ingredients in cooking… Continue Reading →

How to use magnitude with keras

This time we have a look into the magnitude library, a feature-packed Python package for utilizing vector embeddings in machine learning models in a fast, efficient, and simple manner. We want to utilize the embeddings magnitude provides and use them in keras.

Text analysis with named entities

This is the second post of my series about understanding text datasets. Here we use the named entities to get some information about our data set.

« Older posts

© 2019 Depends on the definition

Up ↑

STAY UP-TO-DATE

Sign up for my NEWSLETTER and always get more hands-on python machine learning news and my latest posts.