• Home
  • Fresh Content
  • Courses
  • Resources
  • Podcast
  • Talks
  • Publications
  • Sponsorship
  • Testimonials
  • Contact
  • Menu

Jon Krohn

  • Home
  • Fresh Content
  • Courses
  • Resources
  • Podcast
  • Talks
  • Publications
  • Sponsorship
  • Testimonials
  • Contact
Jon Krohn

Open-Source LLM Libraries and Techniques, with Dr. Sebastian Raschka

Added on March 22, 2024 by Jon Krohn.

Today's superhuman guest is Dr. Sebastian Raschka,, author of the bestselling "ML with PyTorch and sklearn" book, iconic technical blogger (>350k followers) and Staff Research Engineer at Lightning AI. Hear him detail open-source libraries for LLMs.

More on Sebastian:

• Is Staff Research Engineer at Lightning AI, the company behind the popular PyTorch Lightning open-source library for training and deploying PyTorch models, including Large Language Models (LLMs), with ease.
• Iconic technical blogger (50k subscribers) and social-media contributor (>350k combined followers across LinkedIn and Twitter)
• Was previously Assistant Professor of Statistics at University of Wisconsin-Madison.
• Holds a PhD in statistical data mining from Michigan State University.

Today’s episode is technical and will primarily be of interest to hands-on practitioners like data scientists, software developers and machine learning engineers.

In it, Sebastian details:

• The many super-helpful open-source libraries that PyTorch Lightning leads development of.
• Dora parameter-efficient fine-tuning.
• Google’s “open-source” Gemma models.
• Multi-query attention.
• The leading alternatives to RLHF.
• Where he sees the next big opportunities in LLM development.

The SuperDataScience podcast is available on all major podcasting platforms, YouTube, and at SuperDataScience.com.

Vonnegut's Player Piano (1952): An Eerie Novel on the Current AI Revolution

Added on March 17, 2024 by Jon Krohn.

Player Piano, despite being written seven decades ago, could not be more relevant to the AI revolution that’s accelerated dramatically in the past year.

Read More

NumPy, SciPy and the Economics of Open-Source, with Dr. Travis Oliphant

Added on March 12, 2024 by Jon Krohn.

Huge episode today with iconic Dr. Travis Oliphant, creator of NumPy and SciPy, the standard libraries for numeric operations (downloaded 8 million and 3 million times PER DAY, respectively). Hear about the future of open-source, including the impact of GenAI.

More on Travis:

• Founded Anaconda, Inc., the company behind the also-ubiquitous Python package manager.

• Founded the massive PyData conferences and communities as well as its associated non-profit foundation, NumFOCUS.

• Currently serves as the CEO of two firms: OpenTeams and Quansight.

• Holds a PhD in biomedical engineering from the Mayo Clinic in Minnesota.

Today’s episode will primarily be of interest to hands-on practitioners like data scientists, software developers and machine learning engineers.

In it, Travis details:

• How his journey creating open-source software began and how NumPy and SciPy grew to become the most popular foundational Python libraries for working with data.

• How he identifies commercial opportunities to support his vast open-source efforts and communities.

• How AI, particularly generative AI, is transforming open-source development.

• Where open-source innovation is headed in the years to come.

The SuperDataScience podcast is available on all major podcasting platforms, YouTube, and at SuperDataScience.com.

In Data Science, Podcast, SuperDataScience, YouTube Tags superdatascience, datascience, numpy, scipy, python, opensource

The Top 10 Episodes of 2023

Added on March 11, 2024 by Jon Krohn.

In 2023, we had a new record of 4 million combined podcast downloads and YouTube views. That’s up from 3.3 million a year earlier; thank you for your support listening, rating, sharing, liking, commenting on episodes and so on!

Read More

The Best A.I. Startup Opportunities, with venture capitalist Rudina Seseri

Added on March 5, 2024 by Jon Krohn.

How should an A.I. startup find product-market fit? How do some A.I. startups become spectacularly successful? The renowned (and highly technical!) A.I. venture-capital investor Rudina Seseri answers these questions and more in today's episode.

Rudina:

• Founder and Managing Partner of Glasswing Ventures in Boston.

• Led investments and/or served on the Board of Directors of more than a dozen SaaS startups, many of which were acquired.

• Was named Startup Boston's 2022 "Investor of the Year" amongst many other formal recognitions.

• Is a sought-after keynote speaker on investing in A.I. startups.

• Executive Fellow at Harvard Business School.

• Holds an MBA from Harvard University.

Today’s episode will be interesting to anyone who’s keen on scaling their impact with A.I., particularly through A.I. startups or investment.

In this episode, Rudina details:

• How data are used to assess venture capital investments.

• What makes particular AI startups so spectacularly successful.

• Her "A.I. Palette" for examining categories of machine learning models and mapping them to categories of training data.

• How Generative AI isn’t a fad, but it is still only a component of the impact that AI more broadly can make.

• The automated systems she has built for staying up to date on all of the most impactful AI developments.

The SuperDataScience podcast is available on all major podcasting platforms, YouTube, and at SuperDataScience.com.

In Podcast, Professional Development, SuperDataScience, YouTube Tags superdatascience, machinelearning, ai, startups, venturecapital, aistartup

Gemini 1.5 Pro, the Million-Token-Context LLM

Added on March 1, 2024 by Jon Krohn.

In episode, #761, we detailed the public release of Google’s Gemini Ultra, the only LLM that is in the same class as OpenAI’s GPT-4 in terms of capabilities. Well, hot on the heels of that announcement, is the release of Gemini Pro 1.5.

Read More

Gemini Ultra: How to Release an A.I. Product for Billions of Users, with Google’s Lisa Cohen

Added on February 27, 2024 by Jon Krohn.

Google recently released Gemini Ultra, their largest language model. I love Ultra and now use it instead of GPT-4 on many tasks. Today's guest, Lisa Cohen, leads Gemini's rollout; hear from her how a company with billions of users rolls out new A.I. products.

More on Gemini Ultra:

• The only LLM with comparable capabilities to GPT-4 (in my experience as well as on benchmark evaluations, although I know benchmarking has plenty of issues!)

• Ultra maintains attention across large context windows (Gemini 1.5 Pro has a million-token context, btw!), competently generating natural language and code.

• Like GPT-4V, Ultra is multi-modal and so accepts both an image and text as input at the same time.

• Piggybacking on Google's excellence at search, I’ve found Gemini Ultra to be particularly effective at tasks that involve real-time search (the Google "Bard" project that focused on real-time information retrieval was renamed "Gemini" when Gemini Ultra was released).

Lisa Cohen is perhaps the best person on the planet to be speaking to about the momentous Gemini releases because Lisa is Director of Data Science & Engineering for Google's Gemini, Assistant and Search Platforms. In addition, she:

• Was previously Senior Director of Data Science at Twitter and Principal Director of Data Science at Microsoft.

• Holds a Master's in Applied Math from Harvard University.

In this episode, Lisa details:

• The three LLMs in Google’s Gemini family and how the largest one, Gemini Ultra, fits in.

• The many ways you can access Gemini models today.

• How absolutely enormous LLM projects are carried out and how they’re rolled out safely and confidently to literally billions of users.

• How LLMs like Gemini Ultra are transforming life and work for everyone from data scientists to educators to children, and how this transformation will continue in the coming years.

The SuperDataScience podcast is available on all major podcasting platforms, YouTube, and at SuperDataScience.com.

In Data Science, Interview, Podcast, SuperDataScience, YouTube Tags superdatascience, machinelearning, geminiai, ai, geminiultra, llms

Humans Love A.I.-Crafted Beer

Added on February 23, 2024 by Jon Krohn.

I recently recorded tipplers' reactions as they had their first taste of the A.I.-crafted "Krohn&Borg" lager I co-developed. Today's episode illustrates the result: Humans love A.I. beer! There's also cool content on using CRISPR-Cas9 to modify yeast genes.

Thanks again to Beau Warren, Head Brewer at Species X Beer Project, for the opportunity to collaborate on this delicious project. You can check out Episode #755 for tons of detail on the ML packages used and the models developed to craft beer with A.I.

And thanks to all of the guests/judges in today's episode:

• Rehgan Avon of AlignAI

• Alexandra Hagmeyer (Dauterman) of Path Robotics

• Kelsey Dingelstedt of Women in Analytics (WIA)

• William McFarland of Omega Yeast

• Jim Lachey of the Super Bowl XXVI-winning Washington Commanders

The SuperDataScience podcast is available on all major podcasting platforms, YouTube, and at SuperDataScience.com.

Tags superdatascience, machinelearning, ai, beer, yeast, crispr, crisprcas9

Full Encoder-Decoder Transformers Fully Explained, with Kirill Eremenko

Added on February 20, 2024 by Jon Krohn.

Last month, Kirill Eremenko was on the show to detail Decoder-Only Transformers (like the GPT series). It was our most popular episode ever, so he's come right back today to detail an even more sophisticated architecture: Encoder-Decoder Transformers.

If you don’t already know him, Kirill:

• Is Founder and CEO of SuperDataScience, an e-learning platform that is the namesake of this podcast.

• Founded the Super Data Science Podcast in 2016 and hosted the show until he passed me the reins a little over three years ago.

• Has reached more than 2.7 million students through the courses he’s published on Udemy, making him Udemy’s most popular data science instructor.

Kirill was most recently on the show for Episode #747 to provide a technical introduction to the Transformer module that underpins all the major modern Large Language Models (LLMs) like the GPT, Gemini, Llama and BERT architectures. We received an unprecedented amount of positive feedback from that episode, demanding more! So here we are.

That episode, #747, as well as today’s, are perhaps the two most technical episodes of this podcast ever so they probably appeal mostly to hands-on practitioners like data scientists and ML engineers, particularly those who already have some understanding of deep neural networks.

In this episode, Kirill:

• Reviews the key Transformer theory that we covered in Episode #747, namely the individual neural-network components of the Decoder-Only architecture that prevails in generative LLMs like the GPT series models.

• Builds on that to detail the full, Encoder-Decoder Transformer architecture that is used in the original Transformer by Google, in their “Attention is All You Need” paper, as well as in other models that excel at both natural-language understanding and generation such as T5 and BART.

• Discusses the performance and capability pros and cons of full Encoder-Decoder architectures relative to Decoder-Only architectures like GPT and Encoder-Only architectures like BERT.

The SuperDataScience podcast is available on all major podcasting platforms, YouTube, and at SuperDataScience.com.

In Data Science, Podcast, SuperDataScience, YouTube Tags superdatascience, machinelearning, deeplearning, ai, llms, transformermodels

The Mamba Architecture: Superior to Transformers in LLMs

Added on February 16, 2024 by Jon Krohn.

Modern, cutting-edge A.I. basically depends entirely on the Transformer. But now, the first serious contender to the Transformer has emerged and it’s called Mamba; we’ve got the full paper—called "Mamba: Linear-TimeSequence Modeling with Selective State Spaces" and written by researchers at Carnegie Mellon and Princeton.

Read More
In Data Science, Podcast, Five-Minute Friday, YouTube, SuperDataScience Tags superdatascience, machinelearning, deeplearning, ai, llms, transformermodels

How to Speak so You Blow Listeners’ Minds, with Cole Nussbaumer Knaflic

Added on February 13, 2024 by Jon Krohn.

Cole Nussbaumer Knaflic's book, "storytelling with data", has sold over 500k copies... wild! In today's episode, Cole details the best tricks from her latest book, "storytelling with you" — a goldmine on how to inform and profoundly engage people.

Cole:

• Is the author of “storytelling with data”, which has sold half a million copies, been translated into over 20 languages and is used by more than 100 universities. Nearly a decade old, the book is the #1 bestseller still today in several Amazon categories.

• Also wrote the follow-on, hands-on “storytelling with data: let’s practice!” a bestseller in its own right.

• Serves as the Founder and CEO of the storytelling with data company, which provides data-storytelling workshops and other resources.

• Previously she was a People Analytics Manager at Google.

• Holds a degree in math as well as an MBA from the University of Washington.

Today’s episode will be of interest to anyone who’d like to communicate so effectively and compellingly that people are blown away.

In this episode, Cole details:

• Her top tips for planning, creating and delivering an incredible presentation.

• A few special tips for communicating data effectively for all of you data nerds like me.

The SuperDataScience podcast is available on all major podcasting platforms, YouTube, and at SuperDataScience.com.

In Data Science, Podcast, SuperDataScience, YouTube Tags superdatascience, data, datacommunication, presentationskills

AlphaGeometry: AI is Suddenly as Capable as the Brightest Math Minds

Added on February 9, 2024 by Jon Krohn.

Google DeepMind's open-sourced AlphaGeometry blends "fast thinking" (like intuition) with "slow thinking" (like careful, conscious reasoning) to enable a big leap forward in A.I. capability and match human Math Olympiad gold medalists on geometry problems.

KEY CONTEXT
• A couple weeks ago, DeepMind published on AlphaGeometry in the prestigious journal peer-reviewed Nature.
• DeepMind focused on geometry due to its demand for high-level reasoning and logical deduction, posing a unique challenge that traditional ML models struggle with.

MASSIVE RESULTS
• AlphaGeometry tackled 30 International Mathematical Olympiad problems, solving 25. This outperforms human Olympiad bronze and silver medalists' averages (who solved 19.3 and 22.9, respectively) and closely rivals gold medalists (who solved 25.9).
• This new system crushes the previous state-of-the-art A.I., which solved only 10 out of 30 problems.
• Beyond solving problems, AlphaGeometry also generates understandable proofs, making A.I.-generated solutions more accessible to humans.

HOW?
• AlphaGeometry uses a new method of generating synthetic theorems and proofs, simulating 100 million unique examples to overcome the limitations of (expensive, laborious) human-generated proofs.
• It combines a neural (deep learning) language model for intuitive guesswork with a symbolic deduction engine for logical problem-solving, mirroring "fast" and "slow thinking" processes akin to human cognition (per Daniel Kahneman's "Thinking, Fast and Slow" book).

IMPACT
• A.I. that can "think fast and slow" like AlphaGeometry could generalize across mathematical fields and potentially other scientific disciplines, pushing the boundaries of human knowledge and problem-solving capabilities.

The SuperDataScience podcast is available on all major podcasting platforms, YouTube, and at SuperDataScience.com.

In Five-Minute Friday, Data Science, Podcast, SuperDataScience, YouTube Tags superdatascience, DeepMind, machine learning, deeplearning, ai, math, podcast

Brewing Beer with A.I., with Beau Warren

Added on February 6, 2024 by Jon Krohn.

In today's episode, Beau Warren of the innovative "Species X" brewery, details how we collaborated together on an A.I. model to craft the perfect beer. Dubbed "Krohn&Borg" lager, you can join us in Columbus, Ohio on Thursday night to try it yourself! 🍻

Read More
In Data Science, Podcast, Professional Development, SuperDataScience, YouTube Tags superdatascience, python, automl, machinelearning, ai, beer, brewing

A Code-Specialized LLM Will Realize AGI, with Jason Warner

Added on February 2, 2024 by Jon Krohn.

Don't miss this mind-blowing episode with Jason Warner, who compellingly argues that code-specialized LLMs will bring about AGI. His firm, poolside, was launched to achieve this and facilitate an "AI-led, developer-assisted" coding paradigm en route.

Jason:
• Is Co-Founder and CEO of poolside, a hot venture capital-backed startup that will shortly be launching its code-specialized Large Language Model and accompanying interface that is designed specifically for people who code like software developers and data scientists.
• Previously was Managing Director at the renowned Bay-Area VC Redpoint Ventures.
• Before that, held a series of senior software-leadership roles at major tech companies including being CTO of GitHub and overseeing the Product Engineering of Ubuntu.
• Holds a degree in computer science from Penn State University and a Master's in CS from Rensselaer Polytechnic Institute.

Today’s episode should be fascinating to anyone keen to stay abreast of the state of the art in A.I. today and what could happen in the coming years.

In today’s episode, Jason details:
• Why a code-generation-specialized LLM like poolside’s will be far more valuable to humans who code than generalized LLMs like GPT-4 or Gemini.
• Why he thinks AGI itself will be brought about by a code-specialized ML model like poolside’s.


The SuperDataScience podcast is available on all major podcasting platforms, YouTube, and at SuperDataScience.com.

In Data Science, Computer Science, Interview, Podcast, Professional Development, SuperDataScience, YouTube Tags superdatascience, machine learning, ai, llms, software development, software

Blend Any Programming Languages in Your ML Workflows, with Dr. Greg Michaelson

Added on January 30, 2024 by Jon Krohn.

The revolutionary Zerve IDE for data science launches today! Zerve gives ML teams a unified space to collaborate, build and deploy projects... and is free for most use-cases! In today's episode, Zerve co-founder Dr Greg Michaelson explains all the details.

Greg is a super-insightful, crisp and clear communicator; I think you'll really enjoy our conversation. More on him:
• He's Co-Founder and Chief Product Officer of Zerve, which just raised $3.8m in pre-seed funding.
• Previously spent 7 years as DataRobot’s Chief Customer Officer and 4 years as Senior Director of Analytics & Research for Travelers.
• Was a baptist pastor while he obtained his PhD in Applied Statistics from The University of Alabama — that perhaps explains some of the variance in how he’s such a silver-tongued communicator!

Today’s episode will appeal most to hands-on practitioners like data scientists, machine learning engineers and software developers, but may also be of interest to anyone who wants to stay on top of the latest approaches to developing and deploying ML models.

In this episode, Greg details: 
• Why his swish new Zerve IDE is so sorely needed.
• How their open-source Pypelines project uniquely generates Python code for Automated Machine Learning.
• Why AutoML is not suitable to most commercial use cases.
• Why most commercial A.I. projects fail and how to ensure they succeed.
• The straightforward way you can develop speaking skills as slick as his.

The SuperDataScience podcast is available on all major podcasting platforms, YouTube, and at SuperDataScience.com.

Tags superdatascience, datascience, machinelearning, software development

AI is Disadvantaging Job Applicants, But You Can Fight Back

Added on January 26, 2024 by Jon Krohn.

In today's important episode, the author, professor and journalist Hilke Schellmann details how specific HR-tech firms misuse A.I. to facilitate biased hiring, promotion, and firing decisions. She also covers how you can fight back and how A.I. can be done right!

Hilke’s book, "The Algorithm: How A.I. Decides Who Gets Hired, Monitored, Promoted, and Fired and Why We Need to Fight Back Now", was published earlier this month. In the exceptionally clear and well-written book, Hilke draws on exclusive information from whistleblowers, internal documents and real‑world tests to detail how many of the algorithms making high‑stakes decisions are biased, racist, and do more harm than good.

In addition to her book, Hilke:

• Is Assistant Professor of Journalism and A.I. at New York University.

• Previously worked in journalism roles at The Wall Street Journal, The New York Times and VICE Media.

• Holds a Master’s in investigative reporting from Columbia University.

Today’s episode will be accessible and interesting to anyone. In it, Hilke details:

• Examples of specific HR-technology firms that employ misleading Theranos-like tactics.

• How A.I. *can* be used ethically for hiring and throughout the employment lifecycle.

• What you can do to fight back if you suspect you’ve been disadvantaged by an automated process.

The SuperDataScience podcast is available on all major podcasting platforms, YouTube, and at SuperDataScience.com.

In Data Science, Interview, Podcast, Professional Development, SuperDataScience, YouTube Tags superdatascience, artificial intelligence, machine learning, hrtech, tatech

How to Found and Fund Your Own A.I. Startup, with Dr. Rasmus Rothe

Added on January 23, 2024 by Jon Krohn.

Fresh off the heels of his triumphant AI House Davos debut at the World Economic Forum last week, Dr. Rasmus Rothe is my guest today, discussing how to successfully found your own A.I. startup, attract venture capital and scale up!

Rasmus is a co-Founder of Merantix, the comprehensive ecosystem that finances, incubates and scales A.I. companies, transforming existing industries and spawning new ones. Merantix includes:

• A "venture studio" that builds transformative A.I. startups from the founding team up.

• A venture-capital fund that invests in A.I. startups.

• Merantix Momentum, a consulting partner for A.I. development and operation.

• The Merantix AI Campus, a slick physical location in Berlin that is Europe's largest A.I. co-working hub. It houses over 1000 entrepreneurs, researchers, investors, and policymakers.

In addition to Merantix, Rasmus:

• Co-founded and co-leads the German A.I. Association, a role that has him regularly providing policy guidance to Europe’s top politicians.

• Scaled, to 150 million users, a deep-learning powered service that analyzes faces.

• Studied computer science at Oxford, Princeton and ETH Zürich, culminating in a PhD in machine vision.

Today’s episode will be of great interest to anyone interested in commercializing and scaling A.I.

In this episode, Rasmus details:

• What makes a great A.I. entrepreneur.

• How to best raise capital for your own A.I. company.

• How to ensure your A.I. company is well-defended from competitors.

• What the future of work could look like in the coming decades as A.I. and robotics overhaul industry after industry.

The SuperDataScience podcast is available on all major podcasting platforms, YouTube, and at SuperDataScience.com.

Tags superdatascience, machinelearning, ai, startup, venturecapital

How A.I. is Transforming Science

Added on January 19, 2024 by Jon Krohn.

A.I. is not just a tool, but a driving force in reshaping the landscape of science. In today's episode, I dive into the profound implications A.I. holds for scientific discovery, citing applications across nuclear fusion, medicine, self-driving labs and more.

Here are some of the ways A.I. is transforming science that are covered in today's episode:

• Antibiotics: MIT researchers uncovered two new antibiotics in a single year (antibiotic discovery is very rare so this is crazy!) by using an ML model trained on the efficacy of known antibiotics to sift through millions of potential antibiotic compounds.

• Batteries: Similar sifting was carried out by A.I. at the University of Liverpool to narrow down the search for battery materials from 200,000 candidates to just five highly promising ones.

• Weather: Huawei's Pangu-Weather and NVIDIA's FourCastNet use ML to offer faster and more accurate forecasts than traditional super-compute-intensive weather simulations — crucial for predicting and managing natural disasters.

• Nuclear Fusion: AI is simplifying the once-daunting task of controlling plasma in tokamak reactors, thereby contributing to advancements in clean energy production.

• Self-Driving Labs: Automate research by planning, executing, and analyzing experiments autonomously, thereby speeding up scientific experimentation and unveiling new possibilities for discovery.

• Generative A.I.: Large Language Models (LLMs) tools are pioneering new frontiers in scientific research. From improving image resolution to designing novel molecules, these tools are yielding tangible results, with several A.I.-designed drugs currently in clinical trials. Tools like Elicit are streamlining the process of scientific literature review over vast corpora, allowing connections within or between fields to be uncovered automatically and suggesting new research directions.

The SuperDataScience podcast is available on all major podcasting platforms, YouTube, and at SuperDataScience.com.

Tags superdatascience, artificialintelligence, science, innovation, machinelearning

Data Science for Clean Energy, with Emily Pastewka

Added on January 16, 2024 by Jon Krohn.

How can data science and machine learning power the transition toward a sustainable global economy? The ML leader (and exceptional communicator of technical concepts!) Emily Pastewka is my guest today to fill us in on Green Data Science.

Emily:

• Leads the data function at Palmetto, a cleantech startup focused on home electrification.

• Prior to Palmetto, spent more than 10 years building consumer data products and solving marketplace problems as a data science and ML leader at huge fast-growing tech companies like Uber and Rent The Runway.

• Holds a Masters degree in Computer Science from Columbia University and undergraduate degrees in Economics and Environmental Policy from Duke.

Today’s episode should be accessible to technical and non-technical folks alike because whenever Emily got technical, she did an exquisite job of explaining the concepts.

In this episode, Emily details:

• How data science and A.I. can make the world greener by shifting us to clean energy.

• The team of people needed to bring cleantech data solutions to life.

• How econometrics plays a key role in nudging consumers toward greener decisions.

• Her top tips for excelling as a data leader.

• What she looks for in the scientists and engineers she hires.

The SuperDataScience podcast is available on all major podcasting platforms, YouTube, and at SuperDataScience.com.

Tags superdatascience, datascience, machinelearning, cleanenergy, greenenergy

The Five Levels of AGI

Added on January 12, 2024 by Jon Krohn.

Artificial General Intelligence (AGI) is a term thrown around a lot, but it's been poorly defined. Until now!

Read More
In Five-Minute Friday, Podcast, SuperDataScience, YouTube, Data Science Tags superdatascience, artificialintelligence, AGI, machinelearning, DeepMind
← Newer Posts Older Posts →
Back to Top