

Data Science at Home
Francesco Gadaleta
Cutting through AI bullsh*t.Come join the discussion on Discord! https://discord.gg/4UNKGf3
Episodes
Mentioned books

Apr 9, 2019 • 17min
Episode 55: Beyond deep learning
The successes that deep learning systems have achieved in the last decade in all kinds of domains are unquestionable. Self-driving cars, skin cancer diagnostics, movie and song recommendations, language translation, automatic video surveillance, digital assistants represent just a few examples of the ongoing revolution that affects or is going to disrupt soon our everyday life.
But all that glitters is not gold…
Read the full post on the Amethix Technologies blog

Mar 9, 2019 • 12min
Episode 54: Reproducible machine learning
In this episode I speak about how important reproducible machine learning pipelines are.
When you are collaborating with diverse teams, several tasks will be distributed among different individuals. Everyone will have good reasons to change parts of your pipeline, leading to confusion and definitely a number of options that soon explode.
In all those cases, tracking data and code is extremely helpful to build models that are reproducible anytime, anywhere.
Listen to the podcast and learn how.

Jan 23, 2019 • 15min
Episode 53: Estimating uncertainty with neural networks
Have you ever wanted to get an estimate of the uncertainty of your neural network? Clearly Bayesian modelling provides a solid framework to estimate uncertainty by design. However, there are many realistic cases in which Bayesian sampling is not really an option and ensemble models can play a role.
In this episode I describe a simple yet effective way to estimate uncertainty, without changing your neural network’s architecture nor your machine learning pipeline at all.
The post with mathematical background and sample source code is published here.

Jan 17, 2019 • 16min
Episode 52: why do machine learning models fail? [RB]
The success of a machine learning model depends on several factors and events. True generalization to data that the model has never seen before is more a chimera than a reality. But under specific conditions a well trained machine learning model can generalize well and perform with testing accuracy that is similar to the one performed during training.
In this episode I explain when and why machine learning models fail from training to testing datasets.

Jan 8, 2019 • 23min
Episode 51: Decentralized machine learning in the data marketplace (part 2)
In this episode I am completing the explanation about the integration fitchain-oceanprotocol that allows secure on-premise compute to operate in the decentralized data marketplace designed by Ocean Protocol.
As mentioned in the show, this is a picture that provides a 10000-feet view of the integration.
I hope you enjoy the show!

Dec 26, 2018 • 24min
Episode 50: Decentralized machine learning in the data marketplace
In this episode I briefly explain how two massive technologies have been merged in 2018 (work in progress :) - one providing secure machine learning on isolated data, the other implementing a decentralized data marketplace.
In this episode I explain:
How do we make machine learning decentralized and secure?
How can data owners keep their data private?
How can we benefit from blockchain technology for AI and machine learning?
I hope you enjoy the show!
References
fitchain.io decentralized machine learnin
Ocean protocol decentralized data marketplace

Dec 19, 2018 • 21min
Episode 49: The promises of Artificial Intelligence
It's always good to put in perspective all the findings in AI, in order to clear some of the most common misunderstandings and promises.
In this episode I make a list of some of the most misleading statements about what artificial intelligence can achieve in the near future.

Oct 21, 2018 • 29min
Episode 48: Coffee, Machine Learning and Blockchain
In this episode - which I advise to consume at night, in a quite place - I speak about private machine learning and blockchain, while I sip a cup of coffee in my home office.
There are several reasons why I believe we should start thinking about private machine learning...
It doesn't really matter what approach becomes successful and gets adopted, as long as it makes private machine learning possible. If people own their data, they should also own the by-product of such data.
Decentralized machine learning makes this scenario possible.

Sep 11, 2018 • 57min
Episode 47: Are you ready for AI winter? [Rebroadcast]
Today I am having a conversation with Filip Piękniewski, researcher working on computer vision and AI at Koh Young Research America.
His adventure with AI started in the 90s and since then a long list of experiences at the intersection of computer science and physics, led him to the conclusion that deep learning might not be sufficient nor appropriate to solve the problem of intelligence, specifically artificial intelligence.
I read some of his publications and got familiar with some of his ideas. Honestly, I have been attracted by the fact that Filip does not buy the hype around AI and deep learning in particular.
He doesn’t seem to share the vision of folks like Elon Musk who claimed that we are going to see an exponential improvement in self driving cars among other things (he actually said that before a Tesla drove over a pedestrian).

Sep 4, 2018 • 17min
Episode 46: why do machine learning models fail? (Part 2)
In this episode I continue the conversation from the previous one, about failing machine learning models.
When data scientists have access to the distributions of training and testing datasets it becomes relatively easy to assess if a model will perform equally on both datasets. What happens with private datasets, where no access to the data can be granted?
At fitchain we might have an answer to this fundamental problem.


