Podcasts by Category
- 448 - Rust in the Cosmos Part 2: testing software in space (Ep. 255)
In this episode of "Rust in the Cosmos" we delve into the challenge of testing software for... ehm ... space How can Rust help? Let's find out ;) Sponsors Arctic Wolf Learn what the new year holds for ransomware as a service, Active Directory, artificial intelligence and more when you download the 2024 Arctic Wolf Labs Predictions Report today at arcticwolf.com/datascience Intrepid AI (https://intrepid.ai) is an AI assisted all-in-one platform for robotics teams. Build robotics applications in minutes, not months. Amethix works to create and maximize the impact of the world’s leading corporations and startups, so they can create a better future for everyone they serve. We provide solutions in AI/ML, Fintech, Defense, Robotics and Predictive maintenance. Communities AeroRust, Intrepid, Bytenook AeroRust Discord invite: https://discord.com/invite/6jJyx5nEUq AeroRust website: AeroRust.org Intrepid AI Discord https://discord.gg/cSSzche6Ct Intrepid AI website: https://intrepid.ai References Open-source GitHub workflows/actions - https://docs.github.com/en/actions Gitlab CI - https://docs.gitlab.com/ee/ci/ Alternatives Circle ci - https://circleci.com/ Travis http://travis-ci.com Dagger.io - https://dagger.io Actionshttps://github.com/dtolnay/rust-toolchain https://github.com/taiki-e/install-action https://github.com/marketplace/actions/install-cargo-binstall Additional tools:https://github.com/nextest-rs/nextest https://github.com/cargo-bins/cargo-binstall Jon Gjengset - Towards Impeccable Rust Lechev.space: lechev.space
Fri, 19 Apr 2024 - 32min - 447 - Rust in the Cosmos: Decoding Communication Part I (Ep. 254)
In this inaugural episode of "Rust in the Cosmos," we delve into the intricacies of communication in space and some of the challenges in space application development. Sponsors Intrepid AI (https://intrepid.ai) is an AI assisted all-in-one platform for robotics teams. Build robotics applications in minutes, not months. Amethix works to create and maximize the impact of the world’s leading corporations and startups, so they can create a better future for everyone they serve. We provide solutions in AI/ML, Fintech, Defense, Robotics and Predictive maintenance.
Thu, 11 Apr 2024 - 26min - 446 - AI and Video Game Development: Navigating the Future Frontier (Ep. 253)
In this episode we delve into the dynamic realm of game development and the transformative role of artificial intelligence (AI). Join Frag, Jim and Mike as they explore the current landscape of game development processes, from initial creative ideation to the integration of AI-driven solutions. With Mike's expertise as a software executive and avid game developer, we uncover the potential of AI to revolutionize game design, streamline development cycles, and enhance player experiences. Discover insights into AI's applications in asset creation, code assistance, and even gameplay itself, as we discuss real-world implementations and cutting-edge research. From the innovative GameGPT framework to the challenges of balancing automation with human creativity, this episode offers valuable perspectives and practical advice for developers looking to harness the power of AI in their game projects. Don't miss out on this insightful exploration at the intersection of technology and entertainment! Sponsors Intrepid AI (https://intrepid.ai) is an AI assisted all-in-one platform for robotics teams. Build robotics applications in minutes, not months. References https://mikeyoung44.itch.io/spacefreighter - Itch.io link to play the game https://discord.gg/yaXgYZ5Ymn - Discord server for the game https://aimodels.substack.com/ - Mike’s newsletter (links to website too)
Sun, 31 Mar 2024 - 47min - 445 - Kaggle Kommando's Data Disco: Laughing our Way Through AI Trends (Ep. 252)
In this episode, join me and the Kaggle Grand Master, Konrad Banachewicz, for a hilarious journey into the zany world of data science trends. From algorithm acrobatics to AI, creativity, Hollywood movies, and music, we just can't get enough. It's the typical episode with a dose of nerdy comedy you didn't know you needed. Buckle up, it's a data disco, and we're breaking down the binary! Sponsors Intrepid AI is an AI assisted all-in-one platform for robotics teams. Build robotics applications in minutes, not months. Learn what the new year holds for ransomware as a service, Active Directory, artificial intelligence and more when you download the 2024 Arctic Wolf Labs Predictions Report today at arcticwolf.com/datascience 🔗 Links Mentioned in the Episode: Generative AI for time series: TimeGPT Documentation Lag-llama: GitHub (Note: The benchmark results on this one are pretty horrible) Open source LLM: Olmo Blog Post Quantization for LLM: Hugging Face Guide And finally, don't miss Konrad's Substack for more nerdy goodness! (If you're there already, be there again! 😄)
Thu, 07 Mar 2024 - 42min - 444 - Revolutionizing Robotics: Embracing Low-Code Solutions (Ep. 251)
In this episode of Data Science at Home, we explore the game-changing impact of low-code solutions in robotics development. Discover how these tools bridge the coding gap, simplify integration, and enable trial-and-error development. We'll also uncover challenges with traditional coding methods using ROS. Join us for a concise yet insightful discussion on the future of robotics! Sponsors Intrepid AI is an AI assisted all-in-one platform for robotics teams. Build robotics applications in minutes, not months. Learn what the new year holds for ransomware as a service, Active Directory, artificial intelligence and more when you download the 2024 Arctic Wolf Labs Predictions Report today at arcticwolf.com/datascience
Fri, 16 Feb 2024 - 19min - 443 - Is SQream the fastest big data platform? (Ep. 250)
Join us in a dynamic conversation with Yori Lavi, Field CTO at SQream, as we unravel the data analytics landscape. From debunking the data lakehouse hype to SQream's GPU-based magic, discover how extreme data challenges are met with agility. Yori shares success stories, insights into SQream's petabyte-scale capabilities, and a roadmap to breaking down organizational bottlenecks in data science. Dive into the future of data analytics with SQream's commitment to innovation, leaving legacy formats behind and leading the charge in large-scale, cost-effective data projects. Tune in for a dose of GPU-powered revolution! References SQream - GPU-based Big Data Platform Patents Assigned to SQREAM TECHNOLOGIES LTD
Tue, 30 Jan 2024 - 57min - 442 - OpenAI CEO Shake-up: Decoding December 2023 (Ep. 249)
In this episode from a month ago, join me as we unravel the controversial CEO firing at OpenAI in December 2023. I share my insights on the events, decode the intricacies, and explore what lies ahead for this influential organization. Don't miss this concise yet insightful take on the intersection of leadership and artificial intelligence innovation. Sponsor Learn what the new year holds for ransomware as a service, Active Directory, artificial intelligence and more when you download the 2024 Arctic Wolf Labs Predictions Report today at arcticwolf.com/datascience
Sun, 21 Jan 2024 - 27min - 441 - Careers, Skills, and the Evolution of AI (Ep. 248)
!!WARNING!! Due to some technical issues the volume is not always constant during the show. I sincerely apologise for any inconvenienceFrancesco In this episode, I speak with Richie Cotton, Data Evangelist at DataCamp, as he delves into the dynamic intersection of AI and education. Richie, a seasoned expert in data science and the host of the podcast, brings together a wealth of knowledge and experience to explore the evolving landscape of AI careers, the skills essential for generative AI technologies, and the symbiosis of domain expertise and technical skills in the industry. References Become a generative AI developer in this FREE code-along series. Learn to build a chatbot using the OpenAI API, the Pinecone API, and LangChain, and learn to build NLP and image applications with Hugging Face. https://www.datacamp.com/ai-code-alongs Learn to use ChatGPT and the OpenAI API in the OpenAI Fundamentals skill track. https://www.datacamp.com/tracks/openai-fundamentals Get started with deep learning using PyTorch in the Introduction to Deep Learning with PyTorch course. https://www.datacamp.com/courses/introduction-to-deep-learning-with-pytorch
Mon, 08 Jan 2024 - 32min - 440 - Open Source Revolution: AI’s Redemption in Data Science (Ep. 247)
Dive into the world of Data Science at Home with our latest episode, where we explore the dynamic relationship between Artificial Intelligence and the redemption of open source software. In this thought-provoking discussion, I share my insights on why now, more than ever, is the opportune moment for open source to leave an indelible mark on the field of AI. Join me as I unpack my opinions and set expectations for the near future, discussing the pivotal role open source is set to play in shaping the landscape of data science and artificial intelligence. Don't miss out—tune in to gain a deeper understanding of this revolutionary intersection! This episode is available as YouTube stream at https://www.youtube.com/live/0Enenz1HqIs?si=woyYdjJVz656BneH&t=915
Tue, 19 Dec 2023 - 36min - 439 - Destroy your toaster before it kills you. Drama at OpenAI and other stories (Ep. 244)
Brace yourselves, dear friends! In this episode, we delve into the earth-shattering revelation that OpenAI might have stumbled upon AGI (lol) and we're all just seconds away from being replaced by highly sophisticated toasters (lol lol). Spoiler alert: OpenAI's CEO is just playing 7D chess with the entire human race. So, sit back, relax, and enjoy this totally not ominous exploration into the 'totally not happening' future of AI!
Fri, 01 Dec 2023 - 26min - 438 - Concurrent is not parallel - Part 1 (Ep. 142)
In plain English, concurrent and parallel are synonyms. Not for a CPU. And definitely not for programmers. In this episode I summarize the ways to parallelize on different architectures and operating systems. Rock-star data scientists must know how concurrency works and when to use it IMHO. Our Sponsors This episode is supported by Chapman’s Schmid College of Science and Technology, where master’s and PhD students join in cutting-edge research as they prepare to take the next big leap in their professional journey.To learn more about the innovative tools and collaborative approach that distinguish the Chapman program in Computational and Data Sciences, visit chapman.edu/datascience Amethix use advanced Artificial Intelligence and Machine Learning to build data platforms and predictive engines in domain like finance, healthcare, pharmaceuticals, logistics, energy. Amethix provide solutions to collect and secure data with higher transparency and disintermediation, and build the statistical models that will support your business.
Wed, 10 Mar 2021 - 32min - 437 - Backend technologies for machine learning in production (Ep. 141)
This is one of the most dynamic and fascinating topics: API technologies for machine learning. It's always fun to build ML models. But how about serving them in the real world? In this episode I speak about three must-know technologies to place your model behind an API. Our Sponsors This episode is supported by Chapman’s Schmid College of Science and Technology, where master’s and PhD students join in cutting-edge research as they prepare to take the next big leap in their professional journey.To learn more about the innovative tools and collaborative approach that distinguish the Chapman program in Computational and Data Sciences, visit chapman.edu/datascience If building software is your passion, you’ll love ThoughtWorks Technology Podcast. It’s a podcast for techies by techies. Their team of experienced technologists take a deep dive into a tech topic that’s piqued their interest — it could be how machine learning is being used in astrophysics or maybe how to succeed at continuous delivery.
Tue, 02 Mar 2021 - 25min - 436 - How to reinvent banking and finance with data and technology (Ep. 139)
The financial system is changing. It is becoming more efficient and integrated with many more services making our life more... digital. Is the old banking system doomed to fail? Or will it just be disrupted by the smaller players of the fintech industry?In this episode we answer some of these fundamental questions with Alessandro E. Hatami from Pacemakers Subscribe to the Newsletter and come chat with us on the official Discord channel Our Sponsors This episode is supported by Chapman’s Schmid College of Science and Technology, where master’s and PhD students join in cutting-edge research as they prepare to take the next big leap in their professional journey.To learn more about the innovative tools and collaborative approach that distinguish the Chapman program in Computational and Data Sciences, visit chapman.edu/datascience Amethix use advanced Artificial Intelligence and Machine Learning to build data platforms and predictive engines in domain like finance, healthcare, pharmaceuticals, logistics, energy. Amethix provide solutions to collect and secure data with higher transparency and disintermediation, and build the statistical models that will support your business.
Mon, 15 Feb 2021 - 36min - 435 - What's up with WhatsApp? (Ep. 138)Sun, 07 Feb 2021 - 30min
- 434 - Is Rust flexible enough for a flexible data model? (Ep. 137)
In this podcast I get inspired by Paul Done's presentation about The Six Principles for Building Robust Yet Flexible Shared Data Applications, and show how powerful of a language Rust is while still maintaining the flexibility of less strict languages. Our Sponsor This episode is supported by Chapman’s Schmid College of Science and Technology, where master's and PhD students join in cutting-edge research as they prepare to take the next big leap in their professional journey. To learn more about the innovative tools and collaborative approach that distinguish the Chapman program in Computational and Data Sciences, visit chapman.edu/datascience
Mon, 01 Feb 2021 - 28min - 433 - Is Apple M1 good for machine learning? (Ep.136)
In this episode I explain the basics of computer architecture and introduce some features of the Apple M1 Is it good for Machine Learning tasks? References Computer architectures book https://www.amazon.com/Computer-Architecture-Quantitative-John-Hennessy/dp/012383872X Performance https://nod.ai/comparing-apple-m1-with-amx2-m1-with-neon/
Mon, 25 Jan 2021 - 28min - 432 - Rust and deep learning with Daniel McKenna (Ep. 135)
In this episode I speak with Daniel McKenna about Rust, machine learning and artificial intelligence. You can find Daniel from http://github.com/xd009642 https://twitter.com/xd009642 Don't forget to come join me in our Discord channel speaking about all things data science. Subscribe to the official Newsletter and never miss an episode
Mon, 18 Jan 2021 - 22min - 431 - Scaling machine learning with clusters and GPUs (Ep. 134)
Let's finish this year with an amazing episode about scaling ML with clusters and GPUs. Kind of as a continuation of Episode 112 I have a terrific conversation with Aaron Richter from Saturn Cloud about, well, making ML faster and scaling it to massive infrastructure. Aaron can be reached on his website https://rikturr.com and Twitter @rikturr Our Sponsor Saturn Cloud is a data science and machine learning platform for scalable Python analytics. Users can jump into cloud-based Jupyter and Dask to scale Python for big data using the libraries they know and love, while leveraging Docker and Kubernetes so that work is reproducible, shareable, and ready for production. Try Saturn Cloud for free at https://saturncloud.io Twitter: @saturn_cloud
Thu, 31 Dec 2020 - 30min - 430 - What is data ethics? (Ep. 133)
What is data ethics? In this episode I have an interesting chat with Denny Wong from FaqBot and Muna. Our Sponsor Amethix use advanced Artificial Intelligence and Machine Learning to build data platforms and predictive engines in domain like finance, healthcare, pharmaceuticals, logistics, energy. Amethix provide solutions to collect and secure data with higher transparency and disintermediation, and build the statistical models that will support your business. References Denny's Twitter profile The data ethics awareness workshop for AI practitioners
Sat, 19 Dec 2020 - 25min - 429 - A Standard for the Python Array API (Ep. 132)
Our Links Come join me in our Discord channel speaking about all things data science. Subscribe to the official Newsletter and never miss an episode Follow me on Twitch during my live coding sessions usually in Rust and Python Our Sponsors ProtonMail offers a simple and trusted solution to protect your internet connection and access blocked or restricted websites. All of ProtonMail and ProtonVPN’s apps are open source and have been inspected by cybersecurity experts, and Proton is based in Switzerland, home to some of the world’s strongest privacy laws Amethix use advanced Artificial Intelligence and Machine Learning to build data platforms and predictive engines in domain like finance, healthcare, pharmaceuticals, logistics, energy. Amethix provide solutions to collect and secure data with higher transparency and disintermediation, and build the statistical models that will support your business. References https://data-apis.org/blog/announcing_the_consortium https://data-apis.github.io/array-api/latest/ https://github.com/data-apis/python-record-api
Tue, 08 Dec 2020 - 33min - 428 - What happens to data transfer after Schrems II? (Ep. 131)
In this episode Adam Leon Smith, CTO of DragonFly and expert in data regulations explains some of the consequences of Schrems II and data transfers from EU to US. For very interesting references and a practical example, subscribe to our Newsletter
Fri, 04 Dec 2020 - 31min - 427 - Test-First Machine Learning [RB] (Ep. 130)
Our Links Come join me in our Discord channel speaking about all things data science. Subscribe to the official Newsletter and never miss an episode Follow me on Twitch during my live coding sessions usually in Rust and Python Our Sponsors ProtonMail offers a simple and trusted solution to protect your internet connection and access blocked or restricted websites. All of ProtonMail and ProtonVPN’s apps are open source and have been inspected by cybersecurity experts, and Proton is based in Switzerland, home to some of the world’s strongest privacy laws Amethix use advanced Artificial Intelligence and Machine Learning to build data platforms and predictive engines in domain like finance, healthcare, pharmaceuticals, logistics, energy. Amethix provide solutions to collect and secure data with higher transparency and disintermediation, and build the statistical models that will support your business.
Tue, 01 Dec 2020 - 20min - 426 - Similarity in Machine Learning (Ep. 129)
Come join me in our Discord channel speaking about all things data science. Follow me on Twitch during my live coding sessions usually in Rust and Python Subscribe to the official Newsletter and never miss an episode Our Sponsors ProtonMail offers a simple and trusted solution to protect your internet connection and access blocked or restricted websites. All of ProtonMail and ProtonVPN's apps are open source and have been inspected by cybersecurity experts, and Proton is based in Switzerland, home to some of the world’s strongest privacy laws Amethix use advanced Artificial Intelligence and Machine Learning to build data platforms and predictive engines in domain like finance, healthcare, pharmaceuticals, logistics, energy. Amethix provide solutions to collect and secure data with higher transparency and disintermediation, and build the statistical models that will support your business.
Tue, 24 Nov 2020 - 30min - 425 - Distill data and train faster, better, cheaper (Ep. 128)
Come join me in our Discord channel speaking about all things data science. Follow me on Twitch during my live coding sessions usually in Rust and Python Our Sponsors Amethix use advanced Artificial Intelligence and Machine Learning to build data platforms and predictive engines in domain like finance, healthcare, pharmaceuticals, logistics, energy. Amethix provide solutions to collect and secure data with higher transparency and disintermediation, and build the statistical models that will support your business. References Dataset distillation (official paper) GitHub repo
Tue, 17 Nov 2020 - 23min - 424 - Machine Learning in Rust: Amadeus with Alec Mocatta [RB] (ep. 127)
Come join me in our Discord channel speaking about all things data science. Follow me on Twitch during my live coding sessions usually in Rust and Python Our Sponsors ProtonVPN offers a simple and trusted solution to protect your internet connection and access blocked or restricted websites. All of ProtonVPN’s apps are open source and have been inspected by cybersecurity experts, and Proton is based in Switzerland, home to some of the world's strongest privacy laws Amethix use advanced Artificial Intelligence and Machine Learning to build data platforms and predictive engines in domain like finance, healthcare, pharmaceuticals, logistics, energy. Amethix provide solutions to collect and secure data with higher transparency and disintermediation, and build the statistical models that will support your business.
Wed, 11 Nov 2020 - 24min - 423 - Top-3 ways to put machine learning models into production (Ep. 126)
Come join me in our Discord channel speaking about all things data science. Follow me on Twitch during my live coding sessions usually in Rust and Python Our Sponsors physicspodcast.com is not just a physics podcast. But also interviews with scientists, scholars, authors and reflections on the history and future of science and technology are all in the wheelhouse. Amethix use advanced Artificial Intelligence and Machine Learning to build data platforms and predictive engines in domain like finance, healthcare, pharmaceuticals, logistics, energy. Amethix provide solutions to collect and secure data with higher transparency and disintermediation, and build the statistical models that will support your business.
Sat, 07 Nov 2020 - 20min - 422 - Remove noise from data with deep learning (Ep.125)
Come join me in our Discord channel speaking about all things data science. Follow me on Twitch during my live coding sessions usually in Rust and Python Our Sponsors ProtonMail is a secure and private email provider that protects yourmessages with end-to-end encryption and zero-access encryption so that besides you, noone can access them. Amethix use advanced Artificial Intelligence and Machine Learning to build data platforms and predictive engines in domain like finance, healthcare, pharmaceuticals, logistics, energy. Amethix provide solutions to collect and secure data with higher transparency and disintermediation, and build the statistical models that will support your business. References DeepInterpolation
Tue, 03 Nov 2020 - 23min - 421 - What is contrastive learning and why it is so powerful? (Ep. 124)
Come join me in our Discord channel speaking about all things data science. Follow me on Twitch during my live coding sessions usually in Rust and Python Our Sponsors The Monday Apps Challenge is bringing developers around the world together to compete in order to build apps that can improve the way teams work together on monday.com Amethix use advanced Artificial Intelligence and Machine Learning to build data platforms and predictive engines in domain like finance, healthcare, pharmaceuticals, logistics, energy. Amethix provide solutions to collect and secure data with higher transparency and disintermediation, and build the statistical models that will support your business. References A Simple Framework for Contrastive Learning of Visual Representations
Fri, 30 Oct 2020 - 26min - 420 - Neural search (Ep. 123)
Come join me in our Discord channel speaking about all things data science. Follow me on Twitch during my live coding sessions usually in Rust and Python This episode is supported by Monday.com The Monday Apps Challenge is bringing developers around the world together to compete in order to build apps that can improve the way teams work together on monday.com.
Fri, 23 Oct 2020 - 19min - 419 - Let's talk about federated learning (Ep. 122)
Let's talk about federated learning. Why is it important? Why large organizations are not ready yet? Come join me in our Discord channel speaking about all things data science. Follow me on Twitch during my live coding sessions usually in Rust and Python This episode is supported by Monday.com The Monday Apps Challenge is bringing developers around the world together to compete in order to build apps that can improve the way teams work together on monday.com.
Sun, 18 Oct 2020 - 30min - 418 - How to test machine learning in production (Ep. 121)
Come join me in our Discord channel speaking about all things data science. Follow me on Twitch during my live coding sessions usually in Rust and Python This episode is supported by Monday.com Monday.com bring teams together so you can plan, manage and track everything your team is working on in one centralized place The monday Apps Challenge is bringing developers around the world together to compete in order to build apps that can improve the way teams work together on monday.com.
Sun, 11 Oct 2020 - 28min - 417 - Why synthetic data cannot boost machine learning (Ep. 120)
Come join me in our Discord channel speaking about all things data science. Follow me on Twitch during my live coding sessions usually in Rust and Python This episode is supported by Women in Tech by Manning Conferences
Sat, 26 Sep 2020 - 23min - 416 - Machine learning in production: best practices [LIVE from twitch.tv] (Ep. 119)
Hey there! Having the best time of my life ;) This is the first episode I record while I am live on my new Twitch channel :) So much fun! Feel free to follow me for the next live streaming. You can also see me coding machine learning stuff in Rust :)) Don't forget to jump on the usual Discord and have a chat I'll see you there!
Wed, 16 Sep 2020 - 37min - 415 - Testing in machine learning: checking deeplearning models (Ep. 118)
In this episode I speak with Adam Leon Smith, CTO at DragonFly and expert in testing strategies for software and machine learning.We cover testing with deep learning (neuron coverage, threshold coverage, sign change coverage, layer coverage, etc.), combinatorial testing and their practical aspects. On September 15th there will be a live@Manning Rust conference. In one Rust-full day you will attend many talks about what's special about rust, building high performance web services or video game, about web assembly and much more.If you want to meet the tribe, tune in september 15th to the live@manning rust conference.
Fri, 04 Sep 2020 - 18min - 414 - Testing in machine learning: generating tests and data (Ep. 117)
In this episode I speak with Adam Leon Smith, CTO at DragonFly and expert in testing strategies for software and machine learning. On September 15th there will be a live@Manning Rust conference. In one Rust-full day you will attend many talks about what's special about rust, building high performance web services or video game, about web assembly and much more.If you want to meet the tribe, tune in september 15th to the live@manning rust conference.
Sat, 29 Aug 2020 - 20min - 413 - Why you care about homomorphic encryption (Ep. 116)
After deep learning, a new entry is about ready to go on stage. The usual journalists are warming up their keyboards for blogs, news feeds, tweets, in one word, hype.This time it's all about privacy and data confidentiality. The new words, homomorphic encryption. Join and chat with us on the official Discord channel. Sponsors This episode is supported by Amethix Technologies. Amethix works to create and maximize the impact of the world’s leading corporations, startups, and nonprofits, so they can create a better future for everyone they serve. They are a consulting firm focused on data science, machine learning, and artificial intelligence. References Towards a Homomorphic Machine Learning Big Data Pipeline for the Financial Services Sector IBM Fully Homomorphic Encryption Toolkit for Linux
Wed, 12 Aug 2020 - 18min - 412 - Test-First machine learning (Ep. 115)
In this episode I speak about a testing methodology for machine learning models that are supposed to be integrated in production environments. Don't forget to come chat with us in our Discord channel Enjoy the show! -- This episode is supported by Amethix Technologies. Amethix works to create and maximize the impact of the world’s leading corporations, startups, and nonprofits, so they can create a better future for everyone they serve. They are a consulting firm focused on data science, machine learning, and artificial intelligence.
Mon, 03 Aug 2020 - 19min - 411 - GPT-3 cannot code (and never will) (Ep. 114)
The hype around GPT-3 is alarming and gives and provides us with the awful picture of people misunderstanding artificial intelligence. In response to some comments that claim GPT-3 will take developers' jobs, in this episode I express some personal opinions about the state of AI in generating source code (and in particular GPT-3). If you have comments about this episode or just want to chat, come join us on the official Discord channel. This episode is supported by Amethix Technologies. Amethix works to create and maximize the impact of the world’s leading corporations, startups, and nonprofits, so they can create a better future for everyone they serve. They are a consulting firm focused on data science, machine learning, and artificial intelligence.
Sun, 26 Jul 2020 - 19min - 410 - Make Stochastic Gradient Descent Fast Again (Ep. 113)
There is definitely room for improvement in the family of algorithms of stochastic gradient descent. In this episode I explain a relatively simple method that has shown to improve on the Adam optimizer. But, watch out! This approach does not generalize well. Join our Discord channel and chat with us. References More descent, less gradient Taylor Series
Wed, 22 Jul 2020 - 20min - 409 - What data transformation library should I use? Pandas vs Dask vs Ray vs Modin vs Rapids (Ep. 112)
In this episode I speak about data transformation frameworks available for the data scientist who writes Python code. The usual suspect is clearly Pandas, as the most widely used library and de-facto standard. However when data volumes increase and distributed algorithms are in place (according to a map-reduce paradigm of computation), Pandas no longer performs as expected. Other frameworks play a role in such context. In this episode I explain the frameworks that are the best equivalent to Pandas in bigdata contexts. Don't forget to join our Discord channel and comment previous episodes or propose new ones. This episode is supported by Amethix Technologies Amethix works to create and maximize the impact of the world’s leading corporations, startups, and nonprofits, so they can create a better future for everyone they serve. Amethix is a consulting firm focused on data science, machine learning, and artificial intelligence. References Pandas a fast, powerful, flexible and easy to use open source data analysis and manipulation tool - https://pandas.pydata.org/ Modin - Scale your pandas workflows by changing one line of code - https://github.com/modin-project/modin Dask advanced parallelism for analytics https://dask.org/ Ray is a fast and simple framework for building and running distributed applications https://github.com/ray-project/ray RAPIDS - GPU data science https://rapids.ai/
Sun, 19 Jul 2020 - 21min - 408 - [RB] It’s cold outside. Let’s speak about AI winter (Ep. 111)
In this episode I speak with Filip Piekniewski about some of the most worth noting findings in AI and machine learning in 2019. As a matter of fact, the entire field of AI has been inflated by hype and claims that are hard to believe. A lot of the promises made a few years ago have revealed quite hard to achieve, if not impossible. Let's stay grounded and realistic on the potential of this amazing field of research, not to bring disillusion in the near future. Join us to our Discord channel to discuss your favorite episode and propose new ones. This episode is brought to you by Protonmail Click on the link in the description or go to protonmail.com/datascience and get 20% off their annual subscription.
Fri, 03 Jul 2020 - 36min - 407 - Rust and machine learning #4: practical tools (Ep. 110)
In this episode I make a non exhaustive list of machine learning tools and frameworks, written in Rust. Not all of them are mature enough for production environments. I believe that community effort can change this very quickly. To make a comparison with the Python ecosystem I will cover frameworks for linear algebra (numpy), dataframes (pandas), off-the-shelf machine learning (scikit-learn), deep learning (tensorflow) and reinforcement learning (openAI). Rust is the language of the future.Happy coding! Reference BLAS linear algebra https://en.wikipedia.org/wiki/Basic_Linear_Algebra_Subprograms Rust dataframe https://github.com/nevi-me/rust-dataframe Rustlearn https://github.com/maciejkula/rustlearn Rusty machine https://github.com/AtheMathmo/rusty-machine Tensorflow bindings https://lib.rs/crates/tensorflow Juice (machine learning for hackers) https://lib.rs/crates/juice Rust reinforcement learning https://lib.rs/crates/rsrl
Mon, 29 Jun 2020 - 24min - 406 - Rust and machine learning #3 with Alec Mocatta (Ep. 109)
In the 3rd episode of Rust and machine learning I speak with Alec Mocatta. Alec is a +20 year experience professional programmer who has been spending time at the interception of distributed systems and data analytics. He's the founder of two startups in the distributed system space and author of Amadeus, an open-source framework that encourages you to write clean and reusable code that works, regardless of data scale, locally or distributed across a cluster. Only for June 24th, LDN *Virtual* Talks June 2020 with Bippit (Alec speaking about Amadeus)
Mon, 22 Jun 2020 - 23min - 405 - Rust and machine learning #2 with Luca Palmieri (Ep. 108)
In the second episode of Rust and Machine learning I am speaking with Luca Palmieri, who has been spending a large part of his career at the interception of machine learning and data engineering. In addition, Luca contributed to several projects closer to the machine learning community using the Rust programming language. Linfa is an ambitious project that definitely deserves the attention of the data science community (and it's written in Rust, with Python bindings! How cool??!). References Series Announcement - Zero to Production in Rust https://www.lpalmieri.com/posts/2020-05-10-announcement-zero-to-production-in-rust/ Zero To Production #0: Foreword https://www.lpalmieri.com/posts/2020-05-24-zero-to-production-0-foreword/ Taking ML to production with Rust: a 25x speedup https://www.lpalmieri.com/posts/2019-12-01-taking-ml-to-production-with-rust-a-25x-speedup/
Fri, 19 Jun 2020 - 27min - 404 - Rust and machine learning #1 (Ep. 107)
This is the first episode of a series about the Rust programming language and the role it can play in the machine learning field. Rust is one of the most beautiful languages I have ever studied so far. I personally come from the C programming language, though for professional activities in machine learning I had to switch to the loved and hated Python language. This episode is clearly not providing you with an exhaustive list of the benefits of Rust, nor its capabilities. For this you can check the references and start getting familiar with what I think it's going to be the language of the next 20 years. Sponsored This episode is supported by Pryml Technologies. Pryml offers secure and cost effective data privacy solutions for your organisation. It generates a synthetic alternative without disclosing you confidential data. References The Rust Programming Language Cookin' with Rust
Wed, 17 Jun 2020 - 22min - 403 - Protecting workers with artificial intelligence (with Sandeep Pandya CEO Everguard.ai)(Ep. 106)
In this episode I have a chat with Sandeep Pandya, CEO at Everguard.ai a company that uses sensor fusion, computer vision and more to provide safer working environments to workers in heavy industry.Sandeep is a senior executive who can hide the complexity of the topic with great talent. This episode is supported by Pryml.io Pryml is an enterprise-scale platform to synthesise data and deploy applications built on that data back to a production environment.Test ideas. Launch new products. Fast. Secure.
Mon, 15 Jun 2020 - 16min - 402 - Compressing deep learning models: rewinding (Ep.105)
As a continuation of the previous episode in this one I cover the topic about compressing deep learning models and explain another simple yet fantastic approach that can lead to much smaller models that still perform as good as the original one. Don't forget to join our Slack channel and discuss previous episodes or propose new ones. This episode is supported by Pryml.io Pryml is an enterprise-scale platform to synthesise data and deploy applications built on that data back to a production environment. References Comparing Rewinding and Fine-tuning in Neural Network Pruninghttps://arxiv.org/abs/2003.02389
Mon, 01 Jun 2020 - 15min - 401 - Compressing deep learning models: distillation (Ep.104)
Using large deep learning models on limited hardware or edge devices is definitely prohibitive. There are methods to compress large models by orders of magnitude and maintain similar accuracy during inference. In this episode I explain one of the first methods: knowledge distillation Come join us on Slack Reference Distilling the Knowledge in a Neural Network https://arxiv.org/abs/1503.02531 Knowledge Distillation and Student-Teacher Learning for Visual Intelligence: A Review and New Outlooks https://arxiv.org/abs/2004.05937
Wed, 20 May 2020 - 22min - 400 - Pandemics and the risks of collecting data (Ep. 103)
Codiv-19 is an emergency. True. Let's just not prepare for another emergency about privacy violation when this one is over. Join our new Slack channel This episode is supported by Proton. You can check them out at protonmail.com or protonvpn.com
Fri, 08 May 2020 - 20min - 399 - Why average can get your predictions very wrong (ep. 102)
Whenever people reason about probability of events, they have the tendency to consider average values between two extremes. In this episode I explain why such a way of approximating is wrong and dangerous, with a numerical example. We are moving our community to Slack. See you there!
Sun, 19 Apr 2020 - 14min - 398 - Activate deep learning neurons faster with Dynamic RELU (ep. 101)
In this episode I briefly explain the concept behind activation functions in deep learning. One of the most widely used activation function is the rectified linear unit (ReLU). While there are several flavors of ReLU in the literature, in this episode I speak about a very interesting approach that keeps computational complexity low while improving performance quite consistently. This episode is supported by pryml.io. At pryml we let companies share confidential data. Visit our website. Don't forget to join us on discord channel to propose new episode or discuss the previous ones. References Dynamic ReLU https://arxiv.org/abs/2003.10027
Wed, 01 Apr 2020 - 22min - 397 - WARNING!! Neural networks can memorize secrets (ep. 100)
One of the best features of neural networks and machine learning models is to memorize patterns from training data and apply those to unseen observations. That's where the magic is. However, there are scenarios in which the same machine learning models learn patterns so well such that they can disclose some of the data they have been trained on. This phenomenon goes under the name of unintended memorization and it is extremely dangerous. Think about a language generator that discloses the passwords, the credit card numbers and the social security numbers of the records it has been trained on. Or more generally, think about a synthetic data generator that can disclose the training data it is trying to protect. In this episode I explain why unintended memorization is a real problem in machine learning. Except for differentially private training there is no other way to mitigate such a problem in realistic conditions.At Pryml we are very aware of this. Which is why we have been developing a synthetic data generation technology that is not affected by such an issue. This episode is supported by Harmonizely. Harmonizely lets you build your own unique scheduling page based on your availability so you can start scheduling meetings in just a couple minutes.Get started by connecting your online calendar and configuring your meeting preferences.Then, start sharing your scheduling page with your invitees! References The Secret Sharer: Evaluating and Testing Unintended Memorization in Neural Networkshttps://www.usenix.org/conference/usenixsecurity19/presentation/carlini
Mon, 23 Mar 2020 - 24min - 396 - Attacks to machine learning model: inferring ownership of training data (Ep. 99)
In this episode I explain a very effective technique that allows one to infer the membership of any record at hand to the (private) training dataset used to train the target model. The effectiveness of such technique is due to the fact that it works on black-box models of which there is no access to the data used for training, nor model parameters and hyperparameters. Such a scenario is very realistic and typical of machine learning as a service APIs. This episode is supported by pryml.io, a platform I am personally working on that enables data sharing without giving up confidentiality. As promised below is the schema of the attack explained in the episode. References Membership Inference Attacks Against Machine Learning Models
Sat, 14 Mar 2020 - 19min - 395 - Don't be naive with data anonymization (Ep. 98)
Masking, obfuscating, stripping, shuffling. All the above techniques try to do one simple thing: keeping the data private while sharing it with third parties. Unfortunately, they are not the silver bullet to confidentiality. All the players in the synthetic data space rely on simplistic techniques that are not secure, might not be compliant and risky for production. At pryml we do things differently.
Sun, 08 Mar 2020 - 13min - 394 - Why sharing real data is dangerous (Ep. 97)
There are very good reasons why a financial institution should never share their data. Actually, they should never even move their data. Ever.In this episode I explain you why.
Sun, 01 Mar 2020 - 10min - 393 - Building reproducible machine learning in production (Ep. 96)
Building reproducible models is essential for all those scenarios in which the lead developer is collaborating with other team members. Reproducibility in machine learning shall not be an art, rather it should be achieved via a methodical approach. In this episode I give a few suggestions about how to make your ML models reproducible and keep your workflow as smooth. Enjoy the show!Come visit us on our discord channel and have a chat
Sat, 22 Feb 2020 - 14min - 392 - Bridging the gap between data science and data engineering: metrics (Ep. 95)
Data science and data engineering are usually two different departments in organisations. Bridging the gap between the two is essential to success. Many times the brilliant applications created by data scientists don't find a match in production, just because they are not production-ready. In this episode I have a talk with Daan Gerits, co-founder and CTO at Pryml.io
Fri, 14 Feb 2020 - 13min - 391 - A big welcome to Pryml: faster machine learning applications to production (Ep. 94)
Why so much silence? Building a company! That's why :) I am building pryml, a platform that allows data scientists build their applications on data they cannot get access to. This is the first of a series of episodes in which I will speak about the technology and the challenges we are facing while we build it. Happy listening and stay tuned!
Fri, 07 Feb 2020 - 09min - 390 - It's cold outside. Let's speak about AI winter (Ep. 93)
In the last episode of 2019 I speak with Filip Piekniewski about some of the most worth noting findings in AI and machine learning in 2019. As a matter of fact, the entire field of AI has been inflated by hype and claims that are hard to believe. A lot of the promises made a few years ago have revealed quite hard to achieve, if not impossible. Let's stay grounded and realistic on the potential of this amazing field of research, not to bring disillusion in the near future. Join us to our Discord channel to discuss your favorite episode and propose new ones. I would like to thank all of you for supporting and inspiring us. I wish you a wonderful 2020!Francesco and the team of Data Science at Home
Tue, 31 Dec 2019 - 36min - 389 - The dark side of AI: bias in the machine (Ep. 92)
This is the fourth and last episode of mini series "The dark side of AI". I am your host Francesco and I’m with Chiara Tonini from London. The title of today’s episode is Bias in the machine C: Francesco, today we are starting with an infuriating discussion. Are you ready to be angry? F: yeah sure is this about brexit? No, I don’t talk about that. In 1986 the New York City’s Rockefeller University conducted a study on breast and uterine cancers and their link to obesity. Like in all clinical trials up to that point, the subjects of the study were all men. So Francesco, do you see a problem with this approach? F: No problem at all, as long as those men had a perfectly healthy uterus.In medicine, up to the end of the 20th century, medical studies and clinical trials were conducted on men, medicine dosage and therapy calculated on men (white men). The female body has historically been considered an exception, or variation, from a male body. F: Like Eve coming from Adam’s rib. I thought we were past that...When the female body has been under analysis, the focus was on the difference between it and the male body, the so-called “bikini approach”: the reproductive organs are different, therefore we study those, and those only. For a long time medicine assumed this was the only difference. Oh good ...This has led to a hugely harmful fallout across society. Because women had reproductive organs, they should reproduce, and all else about them was deemed uninteresting. Still today, they consider a woman without children somehow to have betrayed her biological destiny. This somehow does not apply to a man without children, who also has reproductive organs. F: so this is an example of a very specific type of bias in medicine, regarding clinical trials and medical studies, that is not only harmful for the purposes of these studies, but has ripple effects in all of societyOnly in the 2010 a serious conversation has started about the damage caused by not including women in clinical trials. There are many many examples (which we list in the references for this episode). Give me oneResearchers consider cardiovascular disease a male disease - they even call it “the widower”. They conduct studies on male samples. But it turns out, the symptoms of a heart attack, especially the ones leading up to one, are different in women. This led to doctors not recognising or dismissing the early symptoms in women. F: I was reading that women are also subject to chronic pain much more than men: for example migraines, and pain related to endometriosis. But there is extensive evidence now of doctors dismissing women’s pain, as either imaginary, or “inevitable”, like it is a normal state of being and does not need a cure at all. The failure of the medical community as a whole to recognise this obvious bias up to the 21st century is an example of how insidious the problem of bias is. There are 3 fundamental types of bias: One: Stochastic drift: you train your model on a dataset, and you validate the model on a split of the training set. When you apply your model out in the world, you systematically add bias in the predictions due to the training data being too specific Two: The bias in the model, introduced by your choice of the parameters of your model. Three: The bias in your training sample: people put training samples together, and people have culture, experience, and prejudice. As we will see today, this is the most dangerous and subtle bias. Today we’ll talk about this bias. Bias is a warping of our understanding of reality. We see reality through the lens of our experience and our culture. The origin of bias can date back to traditions going back centuries, and is so ingrained in our way of thinking, that we don’t even see it anymore. F: And let me add, when it comes to machine learning, we see reality through the lens of data. Bias is everywhere, and we could spend hours and hours talking about it. It’
Sat, 28 Dec 2019 - 20min - 388 - The dark side of AI: metadata and the death of privacy (Ep. 91)
Get in touch with us Join the discussion about data science, machine learning and artificial intelligence on our Discord server Episode transcript We always hear the word “metadata”, usually in a sentence that goes like this Your Honor, I swear, we were not collecting users data, just metadata. Usually the guy saying this sentence is Zuckerberg, but could be anybody from Amazon or Google. “Just” metadata, so no problem. This is one of the biggest lies about the reality of data collection. F: Ok the first question is, what the hell is metadata? Metadata is data about data. F: Ok… still not clear.Imagine you make a phone call to your mum. How often do you call your mum, Francesco?F: Every day of course! (coughing) Good boy! Ok, so let’s talk about today’s phone call. Let’s call “data” the stuff that you and your mum actually said. What did you talk about? F: She was giving me the recipe for her famous lasagna. So your mum’s lasagna is the DATA. What is the metadata of this phone call? The lasagna has data of its own attached to it: the date and time when the conversation happened, the duration of the call, the unique hardware identifiers of your phone and your mum’s phone, the identifiers of the two sim cards, the location of the cell towers that pinged the call, the GPS coordinates of the phones themselves. F: yeah well, this lasagna comes with a lot of data :) And this is assuming that this data is not linked to any other data like your Facebook account or your web browsing history. More of that later. F: Whoa Whoa Whoa, ok. Let’s put a pin in that. Going back to the “basic” metadata that you describe. I think we understand the concept of data about data. I am sure you did your research and you would love to paint me a dystopian nightmare, as always. Tell us why is this a big deal? Metadata is a very big deal. In fact, metadata is far more “useful” than the actual data, where by “useful” I mean that it allows a third party to learn about you and your whole life. What I am saying is, the fact that you talk with your mum every day for 15 minutes is telling me more about you than the content of the actual conversations. In a way, the content does not matter. Only the metadata matters. F: Ok, can you explain this point a bit more? Imagine this scenario: you work in an office in Brussels, and you go by car. Every day, you use your time in the car while you go home to call your mum. So every day around 6pm, a cell tower along the path from your office to your home pings a call from your phone to your mum’s phone. Someone who is looking at your metadata, knows exactly where you are while you call your mum. Every day you will talk about something different, and it doesn't really matter. Your location will come through loud and clear. A lot of additional information can be deduced from this too: for example, you are moving along a motorway, therefore you have a car. The metadata of a call to mum now becomes information on where you are at 6pm, and the way you travel. F: I see. So metadata about the phone call is, in fact, real data about me. Exactly. YOU are what is interesting, not your mum’s lasagna. F: you say so because you haven’t tried my mum’s lasagna. But I totally get your point. Now, imagine that one day, instead of going straight home, you decide to go somewhere else. Maybe you are secretly looking for another job. Your metadata is recording the fact that after work you visit the offices of a rival company. Maybe you are a journalist and you visit your anonymous source. Your metadata records wherever you go, and one of these places is your secret meeting with your source. Anyone’s metadata can be combined with yours. There will be someone who was with you at the time and place of your secret meeting. Anyone who comes in contact with you can be tagged and monitored. Now their anonymity has been reduced. F: I get it. So, compared to the content of my c
Mon, 23 Dec 2019 - 23min - 387 - The dark side of AI: recommend and manipulate (Ep. 90)
In 2017 a research group at the University of Washington did a study on the Black Lives Matter movement on Twitter. They constructed what they call a “shared audience graph” to analyse the different groups of audiences participating in the debate, and found an alignment of the groups with the political left and political right, as well as clear alignments with groups participating in other debates, like environmental issues, abortion issues and so on. In simple terms, someone who is pro-environment, pro-abortion, left-leaning, is also supportive of the Black Lives Matter movement, and viceversa. F: Ok, this seems to make sense, right? But… I suspect there is more to this story? So far, yes…. What they did not expect to find, though, was a pervasive network of Russian accounts participating in the debate, which turned out to be orchestrated by the Internet Research Agency, the not-so-secret Russian secret service agency of internet black ops. The same connected with the US election and Brexit referendum, allegedly. F: Are we talking about actual spies? Where are you going with this? Basically, the Russian accounts (part of them human and part of them bots) were infiltrating all aspects of the debate, both on the left and on the right side, and always taking the most extreme stances on any particular aspect of the debate. The aim was to radicalise the conversation, to make it more and more extreme, in a tactic of divide-and-conquer: turn the population against itself in an online civil war, push for policies that normally would be considered too extreme (for instance, give tanks to the police to control riots, force a curfew, try to ban Muslims from your country). Chaos and unrest have repercussions on international trade and relations, and can align to foreign interests. F: It seems like a pretty indirect and convoluted way of influencing a foreign power… You might think so, but you are forgetting social media. This sort of operation is directly exploiting a core feature of internet social media platforms. And that feature, I am afraid, is recommender systems. F: Whoa. Let’s take a step back. Let’s recap the general features of recommender systems, so we are on the same page. The main purpose of recommender systems is to recommend people the same items similar people show an interest in.Let’s think about books and readers. The general idea is to find a way to predict the best book to the best reader. Amazon is doing it, Netflix is doing it, probably the bookstore down the road does that too, just on a smaller scale.Some of the most common methods to implement recommender systems, use concepts such as cosine/correlation similarity, matrix factorization, neural autoencoders and sequence predictors. The major issue of recommender systems is in their validation. Even though validation occurs in a way that is similar to many machine learning methods, one should recommend a set of items first (in production) and measure the efficacy of such a recommendation. But, recommending is already altering the entire scenario, a bit in the flavour of the Heisenberg principle of uncertainty. F: In the attention economy, the business model is to monetise the time the user spends on a platform, by showing them ads. Recommender systems are crucial for this purpose.Chiara, you are saying that these algorithms have effects that are problematic? As you say, recommender systems exist because the business model of social media platforms is to monetise attention. The most effective way to keep users’ attention is to show them stuff they could show an interest in.In order to do that, one must segment the audience to find the best content for each user. But then, for each user, how do you keep them engaged, and make them consume more content? F: You’re going to say the word “filter bubble” very soon. Spot on. To keep the user on the platform, you start by showing them content that they are interested in, and that agrees with their opinion.
Wed, 11 Dec 2019 - 20min - 386 - The dark side of AI: social media and the optimization of addiction (Ep. 89)
Chamath Palihapitiya, former Vice President of User Growth at Facebook, was giving a talk at Stanford University, when he said this: “I feel tremendous guilt. The short-term, dopamine-driven feedback loops that we have created are destroying how society works ”. He was referring to how social media platforms leverage our neurological build-up in the same way slot machines and cocaine do, to keep us using their products as much as possible. They turn us into addicts. F: how many times do you check your Facebook in a day? I am not a fan of Facebook. I do not have it on my phone. Still, I check it in the morning on my laptop, and maybe twice more per day. I have a trick though: I do not scroll down. I only check the top bar to see if someone has invited me to an event, or contacted me directly. But from time to time, this resolution of mine slips, and I catch myself scrolling down, without even realising it! F: is it the first thing you check when you wake up? No because usually I have a message from you!! :) But yes, while I have my coffee I do a sweep on Facebook and twitter and maybe Instagram, plus the news. F: Check how much time you spend on Facebook And then sum it up to your email, twitter, reddit, youtube, instagram, etc. (all viable channels for ads to reach you) We have an answer. More on that later. Clearly in this episode there is some form of addiction we would like to talk about. So let’s start from the beginning: how does addiction work? Dopamine is a hormone produced by our body, and in the brain it works as a neurotransmitter, a chemical that neurons use to transmit signals to each other. One of the main functions of dopamine is to shape the “reward-motivated behaviour”: this is the way our brain learns through association, positive reinforcement, incentives, and positively-valenced emotions, in particular, pleasure. In other words, it makes our brain desire more of the things that make us feel good. These things can be for example good food, sex, and crucially, good social interactions, like hugging your friends or your baby, or having a laugh together. Because we are evolved to be social animals with complex social structures, successful social interactions are an evolutionary advantage, and therefore they trigger dopamine release in our brain, which makes us feel good, and reinforces the association between the action and the reward. This feeling motivates us to repeat the behaviour. F: now that you mention reinforcement, I recall that this mechanism is so powerful and effective that in fact we have been inspired by nature and replicated it in-silico with reinforcement learning. The idea is to motivate (and eventually create an addictive pattern) an agent to follow what is called the optimal policy by giving it positive rewards or punishing it when things don’t go the way we planned. In our brain, every time an action produces a reward, the connection between action and reward becomes stronger. Through reinforcement, a baby learns to distinguish a cat from a dog, or that fire hurts (that was me). F: and so this means that all the social interactions people get from social media platforms are in fact doing the same, right? Yes, but with a difference: smartphones in our pockets keep us connected to an unlimited reserve of constant social interactions. This constant flux of notifications - the rewards - flood our brain with dopamine. The mechanism of reinforcement can spin out of control. The reward pathways in our brain can malfunction, and this leads to addiction. F: you are saying that social media has LITERALLY the effect of a drug? Yes. In fact, social media platforms are DESIGNED to exploit the rewards systems in our brain. They are designed to work like a drug. Have you been to a casino and played roulette or the slot machines? F: ...maybe? Why is it fun to play roulette? The fun comes from the WAIT before the reward. You put a chip on a number, you don’t know how it’s goin
Tue, 03 Dec 2019 - 22min - 385 - How to improve the stability of training a GAN (Ep. 88)
Generative Adversarial Networks or GANs are very powerful tools to generate data. However, training a GAN is not easy. More specifically, GANs suffer of three major issues such as instability of the training procedure, mode collapse and vanishing gradients. In this episode I not only explain the most challenging issues one would encounter while designing and training Generative Adversarial Networks. But also some methods and architectures to mitigate them. In addition I elucidate the three specific strategies that researchers are considering to improve the accuracy and the reliability of GANs. The most tedious issues of GANs Convergence to equilibrium A typical GAN is formed by at least two networks: a generator G and a discriminator D. The generator's task is to generate samples from random noise. In turn, the discriminator has to learn to distinguish fake samples from real ones. While it is theoretically possible that generators and discriminators converge to a Nash Equilibrium (at which both networks are in their optimal state), reaching such equilibrium is not easy. Vanishing gradients Moreover, a very accurate discriminator would push the loss function towards lower and lower values. This in turn, might cause the gradient to vanish and the entire network to stop learning completely. Mode collapse Another phenomenon that is easy to observe when dealing with GANs is mode collapse. That is the incapability of the model to generate diverse samples. This in turn, leads to generated data that are more and more similar to the previous ones. Hence, the entire generated dataset would be just concentrated around a particular statistical value. The solution Researchers have taken into consideration several approaches to overcome such issues. They have been playing with architectural changes, different loss functions and game theory. Listen to the full episode to know more about the most effective strategies to build GANs that are reliable and robust. Don't forget to join the conversation on our new Discord channel. See you there!
Mon, 18 Nov 2019 - 28min - 384 - What if I train a neural network with random data? (with Stanisław Jastrzębski) (Ep. 87)
What happens to a neural network trained with random data? Are massive neural networks just lookup tables or do they truly learn something? Today’s episode will be about memorisation and generalisation in deep learning, with Stanislaw Jastrzębski from New York University. Stan spent two summers as a visiting student with Prof. Yoshua Bengio and has been working on Understanding and improving how deep network generalise Representation Learning Natural Language Processing Computer Aided Drug Design What makes deep learning unique? I have asked him a few questions for which I was looking for an answer for a long time. For instance, what is deep learning bringing to the table that other methods don’t or are not capable of? Stan believe that the one thing that makes deep learning special is representation learning. All the other competing methods, be it kernel machines, or random forests, do not have this capability. Moreover, optimisation (SGD) lies at the heart of representation learning in the sense that it allows finding good representations. What really improves the training quality of a neural network? We discussed about the accuracy of neural networks depending pretty much on how good the Stochastic Gradient Descent method is at finding minima of the loss function. What would influence such minima?Stan's answer has revealed that training set accuracy or loss value is not that interesting actually. It is relatively easy to overfit data (i.e. achieve the lowest loss possible), provided a large enough network, and a large enough computational budget. However, shape of the minima, or performance on validation sets are in a quite fascinating way influenced by optimisation. Optimisation in the beginning of the trajectory, steers such trajectory towards minima of certain properties that go much further than just training accuracy. As always we spoke about the future of AI and the role deep learning will play. I hope you enjoy the show! Don't forget to join the conversation on our new Discord channel. See you there! References Homepage of Stanisław Jastrzębski https://kudkudak.github.io/ A Closer Look at Memorization in Deep Networks https://arxiv.org/abs/1706.05394 Three Factors Influencing Minima in SGD https://arxiv.org/abs/1711.04623 Don't Decay the Learning Rate, Increase the Batch Size https://arxiv.org/abs/1711.00489 Stiffness: A New Perspective on Generalization in Neural Networks https://arxiv.org/abs/1901.09491
Tue, 12 Nov 2019 - 19min - 383 - Deeplearning is easier when it is illustrated (with Jon Krohn) (Ep. 86)
In this episode I speak with Jon Krohn, author of Deeplearning Illustrated a book that makes deep learning easier to grasp. We also talk about some important guidelines to take into account whenever you implement a deep learning model, how to deal with bias in machine learning used to match jobs to candidates and the future of AI. You can purchase the book from informit.com/dsathome with code DSATHOME and get 40% off books/eBooks and 60% off video training
Tue, 05 Nov 2019 - 44min - 382 - More powerful deep learning with transformers (Ep. 84) (Rebroadcast)
Some of the most powerful NLP models like BERT and GPT-2 have one thing in common: they all use the transformer architecture. Such architecture is built on top of another important concept already known to the community: self-attention.In this episode I explain what these mechanisms are, how they work and why they are so powerful. Don't forget to subscribe to our Newsletter or join the discussion on our Discord server References Attention is all you need https://arxiv.org/abs/1706.03762 The illustrated transformer https://jalammar.github.io/illustrated-transformer Self-attention for generative models http://web.stanford.edu/class/cs224n/slides/cs224n-2019-lecture14-transformers.pdf
Wed, 27 Nov 2019 - 37min - 381 - What is wrong with reinforcement learning? (Ep. 82)
Join the discussion on our Discord server After reinforcement learning agents doing great at playing Atari video games, Alpha Go, doing financial trading, dealing with language modeling, let me tell you the real story here.In this episode I want to shine some light on reinforcement learning (RL) and the limitations that every practitioner should consider before taking certain directions. RL seems to work so well! What is wrong with it? Are you a listener of Data Science at Home podcast? A reader of the Amethix Blog? Or did you subscribe to the Artificial Intelligence at your fingertips newsletter? In any case let’s stay in touch! https://amethix.com/survey/ References Emergence of Locomotion Behaviours in Rich Environments https://arxiv.org/abs/1707.02286 Rainbow: Combining Improvements in Deep Reinforcement Learning https://arxiv.org/abs/1710.02298 AlphaGo Zero: Starting from scratch https://deepmind.com/blog/article/alphago-zero-starting-scratch
Tue, 15 Oct 2019 - 21min - 380 - Have you met Shannon? Conversation with Jimmy Soni and Rob Goodman about one of the greatest minds in history (Ep. 81)
Join the discussion on our Discord server In this episode I have an amazing conversation with Jimmy Soni and Rob Goodman, authors of “A mind at play”, a book entirely dedicated to the life and achievements of Claude Shannon. Claude Shannon does not need any introduction. But for those who need a refresh, Shannon is the inventor of the information age. Have you heard of binary code, entropy in information theory, data compression theory (the stuff behind mp3, mpg, zip, etc.), error correcting codes (the stuff that makes your RAM work well), n-grams, block ciphers, the beta distribution, the uncertainty coefficient? All that stuff has been invented by Claude Shannon :) Articles: https://medium.com/the-mission/10-000-hours-with-claude-shannon-12-lessons-on-life-and-learning-from-a-genius-e8b9297bee8f https://medium.com/the-mission/on-claude-shannons-103rd-birthday-here-are-103-memorable-claude-shannon-quotes-maxims-and-843de4c716cf?source=your_stories_page--------------------------- http://nautil.us/issue/51/limits/how-information-got-re_invented http://nautil.us/issue/50/emergence/claude-shannon-the-las-vegas-cheat Claude's papers: https://medium.com/the-mission/a-genius-explains-how-to-be-creative-claude-shannons-long-lost-1952-speech-fbbcb2ebe07f http://www.math.harvard.edu/~ctm/home/text/others/shannon/entropy/entropy.pdf A mind at play (book links): http://amzn.to/2pasLMz -- Hardcover https://amzn.to/2oCfVL0 -- Audio
Thu, 10 Oct 2019 - 32min - 379 - Attacking machine learning for fun and profit (with the authors of SecML Ep. 80)
Join the discussion on our Discord server As ML plays a more and more relevant role in many domains of everyday life, it’s quite obvious to see more and more attacks to ML systems. In this episode we talk about the most popular attacks against machine learning systems and some mitigations designed by researchers Ambra Demontis and Marco Melis, from the University of Cagliari (Italy). The guests are also the authors of SecML, an open-source Python library for the security evaluation of Machine Learning (ML) algorithms. Both Ambra and Marco are members of research group PRAlab, under the supervision of Prof. Fabio Roli. SecML Contributors Marco Melis (Ph.D Student, Project Maintainer, https://www.linkedin.com/in/melismarco/)Ambra Demontis (Postdoc, https://pralab.diee.unica.it/it/AmbraDemontis) Maura Pintor (Ph.D Student, https://it.linkedin.com/in/maura-pintor)Battista Biggio (Assistant Professor, https://pralab.diee.unica.it/it/BattistaBiggio) References SecML: an open-source Python library for the security evaluation of Machine Learning (ML) algorithms https://secml.gitlab.io/. Demontis et al., “Why Do Adversarial Attacks Transfer? Explaining Transferability of Evasion and Poisoning Attacks,” presented at the 28th USENIX Security Symposium (USENIX Security 19), 2019, pp. 321–338. https://www.usenix.org/conference/usenixsecurity19/presentation/demontis W. Koh and P. Liang, “Understanding Black-box Predictions via Influence Functions,” in International Conference on Machine Learning (ICML), 2017. https://arxiv.org/abs/1703.04730 Melis, A. Demontis, B. Biggio, G. Brown, G. Fumera, and F. Roli, “Is Deep Learning Safe for Robot Vision? Adversarial Examples Against the iCub Humanoid,” in 2017 IEEE International Conference on Computer Vision Workshops (ICCVW), 2017, pp. 751–759. https://arxiv.org/abs/1708.06939 Biggio and F. Roli, “Wild Patterns: Ten Years After the Rise of Adversarial Machine Learning,” Pattern Recognition, vol. 84, pp. 317–331, 2018. https://arxiv.org/abs/1712.03141 Biggio et al., “Evasion attacks against machine learning at test time,” in Machine Learning and Knowledge Discovery in Databases (ECML PKDD), Part III, 2013, vol. 8190, pp. 387–402. https://arxiv.org/abs/1708.06131 Biggio, B. Nelson, and P. Laskov, “Poisoning attacks against support vector machines,” in 29th Int’l Conf. on Machine Learning, 2012, pp. 1807–1814. https://arxiv.org/abs/1206.6389 Dalvi, P. Domingos, Mausam, S. Sanghai, and D. Verma, “Adversarial classification,” in Tenth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD), Seattle, 2004, pp. 99–108. https://dl.acm.org/citation.cfm?id=1014066 Sundararajan, Mukund, Ankur Taly, and Qiqi Yan. "Axiomatic attribution for deep networks." Proceedings of the 34th International Conference on Machine Learning-Volume 70. JMLR. org, 2017. https://arxiv.org/abs/1703.01365 Ribeiro, Marco Tulio, Sameer Singh, and Carlos Guestrin. "Model-agnostic interpretability of machine learning." arXiv preprint arXiv:1606.05386 (2016). https://arxiv.org/abs/1606.05386 Guo, Wenbo, et al. "Lemna: Explaining deep learning based security applications." Proceedings of the 2018 ACM SIGSAC Conference on Computer and Communications Security. ACM, 2018. https://dl.acm.org/citation.cfm?id=3243792 Bach, Sebastian, et al. "On pixel-wise explanations for non-linear classifier decisions by layer-wise relevance propagation." PloS one 10.7 (2015): E0130140. https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0130140
Tue, 01 Oct 2019 - 34min - 378 - [RB] How to scale AI in your organisation (Ep. 79)
Join the discussion on our Discord server Scaling technology and business processes are not equal. Since the beginning of the enterprise technology, scaling software has been a difficult task to get right inside large organisations. When it comes to Artificial Intelligence and Machine Learning, it becomes vastly more complicated. In this episode I propose a framework - in five pillars - for the business side of artificial intelligence.
Thu, 26 Sep 2019 - 13min - 377 - [RB] Replicating GPT-2, the most dangerous NLP model (with Aaron Gokaslan) (Ep. 83)
Join the discussion on our Discord server In this episode, I am with Aaron Gokaslan, computer vision researcher, AI Resident at Facebook AI Research. Aaron is the author of OpenGPT-2, a parallel NLP model to the most discussed version that OpenAI decided not to release because too accurate to be published. We discuss about image-to-image translation, the dangers of the GPT-2 model and the future of AI. Moreover, Aaron provides some very interesting links and demos that will blow your mind! Enjoy the show! References Multimodal image to image translation (not all mentioned in the podcast but recommended by Aaron) Pix2Pix: https://phillipi.github.io/pix2pix/ CycleGAN: https://junyanz.github.io/CycleGAN/ GANimorph Paper: https://arxiv.org/abs/1808.04325 Code: https://github.com/brownvc/ganimorph UNIT:https://arxiv.org/abs/1703.00848 MUNIT:https://github.com/NVlabs/MUNIT DRIT: https://github.com/HsinYingLee/DRIT GPT-2 and related Try OpenAI's GPT-2: https://talktotransformer.com/ Blogpost: https://blog.usejournal.com/opengpt-2-we-replicated-gpt-2-because-you-can-too-45e34e6d36dc The Original Transformer Paper: https://arxiv.org/abs/1706.03762 Grover: The FakeNews generator and detector: https://rowanzellers.com/grover/
Fri, 18 Oct 2019 - 37min - 376 - [RB] How to generate very large images with GANs (Ep. 85)
Join the discussion on our Discord server In this episode I explain how a research group from the University of Lubeck dominated the curse of dimensionality for the generation of large medical images with GANs. The problem is not as trivial as it seems. Many researchers have failed in generating large images with GANs before. One interesting application of such approach is in medicine for the generation of CT and X-ray images.Enjoy the show! References Multi-scale GANs for Memory-efficient Generation of High Resolution Medical Images https://arxiv.org/abs/1907.01376
Mon, 04 Nov 2019 - 14min - 375 - How to cluster tabular data with Markov Clustering (Ep. 73)
In this episode I explain how a community detection algorithm known as Markov clustering can be constructed by combining simple concepts like random walks, graphs, similarity matrix. Moreover, I highlight how one can build a similarity graph and then run a community detection algorithm on such graph to find clusters in tabular data. You can find a simple hands-on code snippet to play with on the Amethix Blog Enjoy the show! References [1] S. Fortunato, “Community detection in graphs”, Physics Reports, volume 486, issues 3-5, pages 75-174, February 2010. [2] Z. Yang, et al., “A Comparative Analysis of Community Detection Algorithms on Artificial Networks”, Scientific Reports volume 6, Article number: 30750 (2016) [3] S. Dongen, “A cluster algorithm for graphs”, Technical Report, CWI (Centre for Mathematics and Computer Science) Amsterdam, The Netherlands, 2000. [4] A. J. Enright, et al., “An efficient algorithm for large-scale detection of protein families”, Nucleic Acids Research, volume 30, issue 7, pages 1575-1584, 2002.
Wed, 21 Aug 2019 - 20min - 374 - Waterfall or Agile? The best methodology for AI and machine learning (Ep. 72)
The two most widely considered software development models in modern project management are, without any doubt, the Waterfall Methodology and the Agile Methodology. In this episode I make a comparison between the two and explain what I believe is the best choice for your machine learning project. An interesting post to read (mentioned in the episode) is How businesses can scale Artificial Intelligence & Machine Learning https://amethix.com/how-businesses-can-scale-artificial-intelligence-machine-learning/
Wed, 14 Aug 2019 - 14min - 373 - Training neural networks faster without GPU [RB] (Ep. 77)
Join the discussion on our Discord server Training neural networks faster usually involves the usage of powerful GPUs. In this episode I explain an interesting method from a group of researchers from Google Brain, who can train neural networks faster by squeezing the hardware to their needs and making the training pipeline more dense. Enjoy the show! References Faster Neural Network Training with Data Echoinghttps://arxiv.org/abs/1907.05550
Tue, 17 Sep 2019 - 22min - 372 - [RB] Validate neural networks without data with Dr. Charles Martin (Ep. 74)
In this episode, I am with Dr. Charles Martin from Calculation Consulting a machine learning and data science consulting company based in San Francisco. We speak about the nuts and bolts of deep neural networks and some impressive findings about the way they work. The questions that Charles answers in the show are essentially two: Why is regularisation in deep learning seemingly quite different than regularisation in other areas on ML? How can we dominate DNN in a theoretically principled way? References The WeightWatcher tool for predicting the accuracy of Deep Neural Networks https://github.com/CalculatedContent/WeightWatcher Slack channel https://weightwatcherai.slack.com/ Dr. Charles Martin Blog http://calculatedcontent.com and channel https://www.youtube.com/c/calculationconsulting Implicit Self-Regularization in Deep Neural Networks: Evidence from Random Matrix Theory and Implications for Learning - Charles H. Martin, Michael W. Mahoney
Tue, 27 Aug 2019 - 44min - 371 - [RB] Complex video analysis made easy with Videoflow (Ep. 75)
In this episode I am with Jadiel de Armas, senior software engineer at Disney and author of Videflow, a Python framework that facilitates the quick development of complex video analysis applications and other series-processing based applications in a multiprocessing environment. I have inspected the videoflow repo on Github and some of the capabilities of this framework and I must say that it’s really interesting. Jadiel is going to tell us a lot more than what you can read from Github References Videflow Github official repository https://github.com/videoflow/videoflow
Thu, 29 Aug 2019 - 30min - 370 - Episode 68: AI and the future of banking with Chris Skinner [RB]
In this episode I have a wonderful conversation with Chris Skinner. Chris and I recently got in touch at The banking scene 2019, fintech conference recently held in Brussels. During that conference he talked as a real trouble maker - that’s how he defines himself - saying that “People are not educated with loans, credit, money” and that “Banks are failing at digital”. After I got my hands on his last book Digital Human, I invited him to the show to ask him a few questions about innovation, regulation and technology in finance.
Tue, 09 Jul 2019 - 41min - 369 - Episode 67: Classic Computer Science Problems in Python
Today I am with David Kopec, author of Classic Computer Science Problems in Python, published by Manning Publications. His book deepens your knowledge of problem solving techniques from the realm of computer science by challenging you with interesting and realistic scenarios, exercises, and of course algorithms. There are examples in the major topics any data scientist should be familiar with, for example search, clustering, graphs, and much more. Get the book from https://www.manning.com/books/classic-computer-science-problems-in-python and use coupon code poddatascienceathome19 to get 40% discount. References Twitter https://twitter.com/davekopec GitHub https://github.com/davecom classicproblems.com
Tue, 02 Jul 2019 - 28min - 368 - Episode 66: More intelligent machines with self-supervised learning
In this episode I talk about a new paradigm of learning, which can be found a bit blurry and not really different from the other methods we know of, such as supervised and unsupervised learning. The method I introduce here is called self-supervised learning. Enjoy the show! Don't forget to subscribe to our Newsletter at amethix.com and get the latest updates in AI and machine learning. We do not spam. Promise! References Deep Clustering for Unsupervised Learning of Visual Features Self-supervised Visual Feature Learning with Deep Neural Networks: A Survey
Tue, 25 Jun 2019 - 18min - 367 - Episode 65: AI knows biology. Or does it?
The successes of deep learning for text analytics, also introduced in a recent post about sentiment analysis and published here are undeniable. Many other tasks in NLP have also benefitted from the superiority of deep learning methods over more traditional approaches. Such extraordinary results have also been possible due to the neural network approach to learn meaningful character and word embeddings, that is the representation space in which semantically similar objects are mapped to nearby vectors. All this is strictly related to a field one might initially find disconnected or off-topic: biology. Don't forget to subscribe to our Newsletter at amethix.com and get the latest updates in AI and machine learning. We do not spam. Promise! References [1] Rives A., et al., “Biological structure and function emerge from scaling unsupervised learning to 250 million protein sequences”, biorxiv, doi: https://doi.org/10.1101/622803 [2] Vaswani A., et al., “Attention is all you need”, Advances in neural information processing systems, pp. 5998–6008, 2017. [3] Bahdanau D., et al., “Neural machine translation by jointly learning to align and translate”, arXiv, http://arxiv.org/abs/1409.0473.
Sun, 23 Jun 2019 - 12min - 366 - Episode 64: Get the best shot at NLP sentiment analysis
The rapid diffusion of social media like Facebook and Twitter, and the massive use of different types of forums like Reddit, Quora, etc., is producing an impressive amount of text data every day. There is one specific activity that many business owners have been contemplating over the last five years, that is identifying the social sentiment of their brand, by analysing the conversations of their users. In this episode I explain how one can get the best shot at classifying sentences with deep learning and word embedding. Additional material Schematic representation of how to learn a word embedding matrix E by training a neural network that, given the previous M words, predicts the next word in a sentence. Word2Vec example source code https://gist.github.com/rlangone/ded90673f65e932fd14ae53a26e89eee#file-word2vec_example-py References [1] Mikolov, T. et al., "Distributed Representations of Words and Phrases and their Compositionality", Advances in Neural Information Processing Systems 26, pages 3111-3119, 2013. [2] The Best Embedding Method for Sentiment Classification, https://medium.com/@bramblexu/blog-md-34c5d082a8c5 [3] The state of sentiment analysis: word, sub-word and character embedding https://amethix.com/state-of-sentiment-analysis-embedding/
Fri, 14 Jun 2019 - 12min - 365 - Episode 63: Financial time series and machine learning
In this episode I speak to Alexandr Honchar, data scientist and owner of blog https://medium.com/@alexrachnogAlexandr has written very interesting posts about time series analysis for financial data. His blog is in my personal list of best tutorial blogs. We discuss about financial time series and machine learning, what makes predicting the price of stocks a very challenging task and why machine learning might not be enough.As usual, I ask Alexandr how he sees machine learning in the next 10 years. His answer - in my opinion quite futuristic - makes perfect sense. You can contact Alexandr on Twitter https://twitter.com/AlexRachnog Facebook https://www.facebook.com/rachnog Medium https://medium.com/@alexrachnog Enjoy the show!
Tue, 04 Jun 2019 - 21min - 364 - Episode 62: AI and the future of banking with Chris Skinner
In this episode I have a wonderful conversation with Chris Skinner. Chris and I recently got in touch at The banking scene 2019, fintech conference recently held in Brussels. During that conference he talked as a real trouble maker - that’s how he defines himself - saying that “People are not educated with loans, credit, money” and that “Banks are failing at digital”. After I got my hands on his last book Digital Human, I invited him to the show to ask him a few questions about innovation, regulation and technology in finance.
Tue, 28 May 2019 - 42min - 363 - Episode 61: The 4 best use cases of entropy in machine learning
It all starts from physics. The entropy of an isolated system never decreases… Everyone at school, at some point of his life, learned this in his physics class. What does this have to do with machine learning? To find out, listen to the show. References Entropy in machine learning https://amethix.com/entropy-in-machine-learning/
Tue, 21 May 2019 - 21min - 362 - Episode 60: Predicting your mouse click (and a crash course in deeplearning)
Deep learning is the future. Get a crash course on deep learning. Now! In this episode I speak to Oliver Zeigermann, author of Deep Learning Crash Course published by Manning Publications at https://www.manning.com/livevideo/deep-learning-crash-course Oliver (Twitter: @DJCordhose) is a veteran of neural networks and machine learning. In addition to the course - that teaches you concepts from prototype to production - he's working on a really cool project that predicts something people do every day... clicking their mouse. If you use promo code poddatascienceathome19 you get a 40% discount for all products on the Manning platform Enjoy the show! References: Deep Learning Crash Course (Manning Publications) https://www.manning.com/livevideo/deep-learning-crash-course?a_aid=djcordhose&a_bid=e8e77cbf Companion notebooks for the code samples of the video course "Deep Learning Crash Course" https://github.com/DJCordhose/deep-learning-crash-course-notebooks/blob/master/README.md Next-button-to-click predictor source code https://github.com/DJCordhose/ux-by-tfjs
Thu, 16 May 2019 - 39min - 361 - Episode 59: How to fool a smart camera with deep learning
In this episode I met three crazy researchers from KULeuven (Belgium) who found a method to fool surveillance cameras and stay hidden just by holding a special t-shirt. We discussed about the technique they used and some consequences of their findings. They published their paper on Arxiv and made their source code available at https://gitlab.com/EAVISE/adversarial-yolo Enjoy the show! References Fooling automated surveillance cameras: adversarial patches to attack person detection Simen Thys, Wiebe Van Ranst, Toon Goedemé Eavise Research Group KULeuven (Belgium)https://iiw.kuleuven.be/onderzoek/eavise
Tue, 07 May 2019 - 24min - 360 - Episode 58: There is physics in deep learning!
There is a connection between gradient descent based optimizers and the dynamics of damped harmonic oscillators. What does that mean? We now have a better theory for optimization algorithms.In this episode I explain how all this works. All the formulas I mention in the episode can be found in the post The physics of optimization algorithms Enjoy the show.
Tue, 30 Apr 2019 - 19min - 359 - Episode 57: Neural networks with infinite layers
How are differential equations related to neural networks? What are the benefits of re-thinking neural network as a differential equation engine? In this episode we explain all this and we provide some material that is worth learning. Enjoy the show! Residual Block References [1] K. He, et al., “Deep Residual Learning for Image Recognition”, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 770-778, 2016 [2] S. Hochreiter, et al., “Long short-term memory”, Neural Computation 9(8), pages 1735-1780, 1997. [3] Q. Liao, et al.,”Bridging the gaps between residual learning, recurrent neural networks and visual cortex”, arXiv preprint, arXiv:1604.03640, 2016. [4] Y. Lu, et al., “Beyond Finite Layer Neural Networks: Bridging Deep Architectures and Numerical Differential Equation”, Proceedings of the 35th International Conference on Machine Learning (ICML), Stockholm, Sweden, 2018. [5] T. Q. Chen, et al., ” Neural Ordinary Differential Equations”, Advances in Neural Information Processing Systems 31, pages 6571-6583}, 2018
Tue, 23 Apr 2019 - 16min - 358 - Episode 56: The graph network
Since the beginning of AI in the 1950s and until the 1980s, symbolic AI approaches have dominated the field. These approaches, also known as expert systems, used mathematical symbols to represent objects and the relationship between them, in order to depict the extensive knowledge bases built by humans. The opposite of the symbolic AI paradigm is named connectionism, which is behind the machine learning approaches of today
Tue, 16 Apr 2019 - 16min - 357 - Episode 55: Beyond deep learning
The successes that deep learning systems have achieved in the last decade in all kinds of domains are unquestionable. Self-driving cars, skin cancer diagnostics, movie and song recommendations, language translation, automatic video surveillance, digital assistants represent just a few examples of the ongoing revolution that affects or is going to disrupt soon our everyday life.But all that glitters is not gold…Read the full post on the Amethix Technologies blog
Tue, 09 Apr 2019 - 17min - 356 - Episode 54: Reproducible machine learning
In this episode I speak about how important reproducible machine learning pipelines are. When you are collaborating with diverse teams, several tasks will be distributed among different individuals. Everyone will have good reasons to change parts of your pipeline, leading to confusion and definitely a number of options that soon explode. In all those cases, tracking data and code is extremely helpful to build models that are reproducible anytime, anywhere. Listen to the podcast and learn how.
Sat, 09 Mar 2019 - 11min - 355 - Episode 53: Estimating uncertainty with neural networks
Have you ever wanted to get an estimate of the uncertainty of your neural network? Clearly Bayesian modelling provides a solid framework to estimate uncertainty by design. However, there are many realistic cases in which Bayesian sampling is not really an option and ensemble models can play a role. In this episode I describe a simple yet effective way to estimate uncertainty, without changing your neural network’s architecture nor your machine learning pipeline at all. The post with mathematical background and sample source code is published here.
Wed, 23 Jan 2019 - 15min - 354 - Episode 52: why do machine learning models fail? [RB]
The success of a machine learning model depends on several factors and events. True generalization to data that the model has never seen before is more a chimera than a reality. But under specific conditions a well trained machine learning model can generalize well and perform with testing accuracy that is similar to the one performed during training. In this episode I explain when and why machine learning models fail from training to testing datasets.
Thu, 17 Jan 2019 - 15min - 353 - Episode 51: Decentralized machine learning in the data marketplace (part 2)
In this episode I am completing the explanation about the integration fitchain-oceanprotocol that allows secure on-premise compute to operate in the decentralized data marketplace designed by Ocean Protocol. As mentioned in the show, this is a picture that provides a 10000-feet view of the integration. I hope you enjoy the show!
Tue, 08 Jan 2019 - 23min - 352 - Episode 50: Decentralized machine learning in the data marketplace
In this episode I briefly explain how two massive technologies have been merged in 2018 (work in progress :) - one providing secure machine learning on isolated data, the other implementing a decentralized data marketplace. In this episode I explain: How do we make machine learning decentralized and secure? How can data owners keep their data private? How can we benefit from blockchain technology for AI and machine learning? I hope you enjoy the show! References fitchain.io decentralized machine learnin Ocean protocol decentralized data marketplace
Wed, 26 Dec 2018 - 24min - 351 - Episode 49: The promises of Artificial Intelligence
It's always good to put in perspective all the findings in AI, in order to clear some of the most common misunderstandings and promises. In this episode I make a list of some of the most misleading statements about what artificial intelligence can achieve in the near future.
Wed, 19 Dec 2018 - 21min - 350 - Episode 48: Coffee, Machine Learning and Blockchain
In this episode - which I advise to consume at night, in a quite place - I speak about private machine learning and blockchain, while I sip a cup of coffee in my home office.There are several reasons why I believe we should start thinking about private machine learning...It doesn't really matter what approach becomes successful and gets adopted, as long as it makes private machine learning possible. If people own their data, they should also own the by-product of such data. Decentralized machine learning makes this scenario possible.
Sun, 21 Oct 2018 - 28min - 349 - Episode 47: Are you ready for AI winter? [Rebroadcast]
Today I am having a conversation with Filip Piękniewski, researcher working on computer vision and AI at Koh Young Research America. His adventure with AI started in the 90s and since then a long list of experiences at the intersection of computer science and physics, led him to the conclusion that deep learning might not be sufficient nor appropriate to solve the problem of intelligence, specifically artificial intelligence. I read some of his publications and got familiar with some of his ideas. Honestly, I have been attracted by the fact that Filip does not buy the hype around AI and deep learning in particular. He doesn’t seem to share the vision of folks like Elon Musk who claimed that we are going to see an exponential improvement in self driving cars among other things (he actually said that before a Tesla drove over a pedestrian).
Tue, 11 Sep 2018 - 56min
Podcasts similar to Data Science at Home
- Global News Podcast BBC World Service
- El Partidazo de COPE COPE
- Herrera en COPE COPE
- The Dan Bongino Show Cumulus Podcast Network | Dan Bongino
- Es la Mañana de Federico esRadio
- La Noche de Dieter esRadio
- Hondelatte Raconte - Christophe Hondelatte Europe 1
- Curiosidades de la Historia National Geographic National Geographic España
- Dateline NBC NBC News
- 財經一路發 News98
- La rosa de los vientos OndaCero
- Más de uno OndaCero
- La Zanzara Radio 24
- L'Heure Du Crime RTL
- El Larguero SER Podcast
- Nadie Sabe Nada SER Podcast
- SER Historia SER Podcast
- Todo Concostrina SER Podcast
- 安住紳一郎の日曜天国 TBS RADIO
- アンガールズのジャンピン[オールナイトニッポンPODCAST] ニッポン放送
- 辛坊治郎 ズーム そこまで言うか! ニッポン放送
- 飯田浩司のOK! Cozy up! Podcast ニッポン放送
- 吳淡如人生實用商學院 吳淡如
- 武田鉄矢・今朝の三枚おろし 文化放送PodcastQR