• Home
  • Fresh Content
  • Courses
  • Resources
  • Podcast
  • Talks
  • Publications
  • Sponsorship
  • Testimonials
  • Contact
  • Menu

Jon Krohn

  • Home
  • Fresh Content
  • Courses
  • Resources
  • Podcast
  • Talks
  • Publications
  • Sponsorship
  • Testimonials
  • Contact
Jon Krohn

AI Emits Far Less Carbon Than Humans (Doing the Same Task)

Added on October 13, 2023 by Jon Krohn.

There's been a lot of press about Large Language Models (LLMs), such as those behind ChatGPT, using vast amounts of energy per query. In fact, however, a person doing the same work emits 12x to 45x more carbon from their laptop alone.

Today’s "Five-Minute Friday" episode is a quick one on how “The Carbon Emissions of Writing and Illustrating Are Lower for AI than for Humans”. Everything in today’s episode is based on an ArXiV preprint paper with that title by researchers from UC Irvine, the Massachusetts Institute of Technology and other universities.

For writing a page of text, for example, the authors estimate:
• BLOOM open-source LLM (including training) produces ~1.6g CO2/query.
• OpenAI's GPT-3 (including training) produces ~2.2g CO2/query.
• Laptop usage for 0.8 hours (average time to write page) emits ~27g CO2 (that's 12x GPT-3).
• Desktop for same amount of writing time emits ~72g CO2 (32 x GPT-3).

For creating a digital illustration:
• Midjourney (including training) produces ~1.9g CO2/query.
• DALL-E 2 produces ~2.2g CO2/query.
• Human takes ~3.2 hours for the same work, emitting ~100g CO2 (45 x DALL-E 2) on a laptop or ~280g CO2 (127 x DALL-E 2) on a desktop.

There are complexities here, such as what humans do with their time instead of writing or illustrating; if it’s spent driving, for example, then the net impact would be worse. As someone who’d love to see the world at net negative carbon emissions ASAP through innovations like nuclear fusion and carbon capture, however, I have been getting antsy about how much energy state-of-the-art LLMs use, but this simple article turned that perspective upside down. I’ll continue to use A.I. to augment my work wherever I can... and hopefully get my day done earlier so I can get away from my machine and enjoy some time outdoors.

Hear more detail in today's episode or check out the video version to see figures as well.


The SuperDataScience podcast is available on all major podcasting platforms, YouTube, and at SuperDataScience.com.

In Five-Minute Friday, Data Science, SuperDataScience, YouTube, Podcast Tags dalle2, DALLE3, AI, LLM, data science, SuperDataScience

Quantum Machine Learning, with Dr. Amira Abbas

Added on October 10, 2023 by Jon Krohn.

Brilliant, eloquent Dr. Amira Abbas introduces us to Quantum Machine Learning in today's episode. She details the key concepts (like qubits), what's possible today (Quantum SVMs) and what the future holds (e.g., Quantum Neural Networks).

Amira: 
• Is a postdoctoral researcher at the University of Amsterdam as well as QuSoft, a world-leading quantum-computing research institution also in the Netherlands.
• Was previously on the Google Quantum A.I. team and did Quantum ML research at IBM.
• Holds a PhD in Quantum ML from the University of KwaZulu-Natal, during which she was a recipient of Google's PhD fellowship.

Much of today’s episode will be fascinating to anyone interested in how quantum computing is being applied to machine learning; there are, however, some relatively technical parts of the conversation that might be best-suited to folks who already have some familiarity with ML.

In this episode, Amira details: 
• What Quantum Computing is, how it’s different from the classical computing that dominates the world today, and where quantum computing excels relative to its classical cousin.
• Key terms such as qubits, quantum entanglement, quantum data and quantum memory.
• Where Quantum ML shows promise today and where it might in the coming years.
• How to get started in Quantum ML research yourself.
• Today’s leading software libraries for Quantum ML.

The SuperDataScience podcast is available on all major podcasting platforms, YouTube, and at SuperDataScience.com.

In Data Science, Interview, Podcast, YouTube, SuperDataScience Tags data science, SuperDataScience, ML, quantummachinelearning, AI

OpenAI’s DALL-E 3, Image Chat and Web Search

Added on October 6, 2023 by Jon Krohn.

Today's episode details three big releases from OpenAI: (1) DALL-E 3 text-to-image model, which "exactly" adheres to your prompt. (2) Image-to-text chat. (3) Real-time web search integrated into ChatGPT (which seems to lag behind Google's Bard).

So, first, DALL-E 3 text-to-image generation:
• Appears to generate images that are on par with Midjourney V5, the current state-of-the-art.
• The big difference is that apparently DALL-E 3 will actually generate images that adhere “exactly” to the text you provide.
• In contrast, the incumbent models in the state of the art typically ignore words or key parts of the description even though the quality is typically stunning.
• This adherence to prompts extends even to language that you’d like to include in the image, which is mega.
• Watch today's YouTube version for examples of all the above.

In addition, using Midjourney is a really bizarre user experience because it's done through Discord where you provide prompts and get results alongside dozens of other people at the same time. DALL-E 3, in contrast, will be within the slick ChatGPT Plus environment, which could completely get rid of the need to develop text-to-image prompt-engineering expertise in order to get great results. Instead, you can simply have an iterative back-and-forth conversation with ChatGPT to produce the image of your dreams.

Next up is image-to-text chat in ChatGPT Plus:
• We've known this was coming for a while.
• Works stunningly well in the test I've done so far.
• Today's YouTube version also shows an example of this.

Finally, real-time web search with Bing is now integrated into ChatGPT Plus:
• In my personal (anecdotal tests), this lagged behind Google's Bard.
• Bard is also free, so if real-time web search is what you're after, there doesn't seem to be a reason to pay for ChatGPT Plus. That said, for state-of-the-art general chat plus now image generation and text-to-image chat (per the above), ChatGPT Plus is well worth the price tag.


The SuperDataScience podcast is available on all major podcasting platforms, YouTube, and at SuperDataScience.com.

In Data Science, Five-Minute Friday, Podcast, SuperDataScience, YouTube Tags OpenAI, DALLE3, text to image, LLM, data science

Computational Mathematics and Fluid Dynamics, with Prof. Margot Gerritsen

Added on October 3, 2023 by Jon Krohn.

Today, the extremely intelligent and super delightful Prof. Margot Gerritsen returns to the show to introduce what Computational Mathematics is, detail countless real-world applications of it, and relate it to the field of data science.

Margot:
• Has been faculty at Stanford University for more than 20 years, including eight years as Director of the Institute for Computational and Mathematical Engineering.
• In 2015, co-founded Women in Data Science (WiDS) Worldwide, an organization that supports, inspires and lowers barriers to entry for women across over 200 chapters in over 160 countries.
• Hosts the corresponding Women in Data Science podcast.
• Holds a PhD from Stanford in which she focused on Computational Fluid Dynamics — a passion she has retained throughout her academic career.

Today’s episode should appeal to anyone.

In it this episode, Margot details:
• What computational mathematics is.
• How computational math is used to study fluid dynamics, with fascinating in-depth examples across traffic, water, oil, sailing, F1 racing, the flight of pterodactyls and more.
• Synaesthesia, a rare perceptual phenomenon, which in her case means she sees numbers in specific colors and how this relates to her lifelong interest in math.
• The genesis of her Women in Data Science organization and the impressive breadth of its global impact today.


The SuperDataScience podcast is available on all major podcasting platforms, YouTube, and at SuperDataScience.com.

In Data Science, Interview, Podcast, SuperDataScience, YouTube Tags Computational mathematics, data science, fluid dynamics

ChatGPT Custom Instructions: A Major, Easy Hack for Data Scientists

Added on September 29, 2023 by Jon Krohn.

Thanks to Shaan Khosla for tipping me off to a crazy easy hack to get markedly better results from GPT-4: providing Custom Instructions that prompt the algorithm to iterate upon its own output while critically evaluating and improving it.

Here's Shaan's full Custom Instructions text, which he himself has been iterating on in recent months:

"I need you to help me with a task. To help me with the task, first come up with a detailed outline of how you think you should respond, then critique the ideas in this outline (mention the advantages, disadvantages, and ways it could be improved), then use the original outline and the critiques you made to come up with your best possible solution.

"Overall, your tone should not be overly dramatic. It should be clear, professional, and direct. Don't sound robotic or like you're trying to sell something. You don't need to remind me you're a large language model, get straight to what you need to say to be as helpful as possible. Again, make sure your tone is clear, professional, and direct - not overly like you're trying to sell something."

Try it out! If you haven't used Custom Instructions before, in today's episode I talk you through how to set it up and explain why this approach is so effective. In the video version, I provide a screenshare that makes getting started foolproof.


The SuperDataScience podcast is available on all major podcasting platforms, YouTube, and at SuperDataScience.com.

In Data Science, Five-Minute Friday, Podcast, SuperDataScience, YouTube, Professional Development Tags ai, data science, CHATGPT, hack

Overcoming Adversaries with A.I. for Cybersecurity, with Dr. Dan Shiebler

Added on September 26, 2023 by Jon Krohn.

Recently in Detroit, my hotel randomly had a podcast studio complete with "ON AIR" sign haha. From there, I interviewed the wildly intelligent Dr. Dan Shiebler on how machine learning is used to tackle cybercrime.

Dan:
• As Head of Machine Learning at Abnormal Security, a cybercrime detection firm that has grown to over $100m in annual recurring revenue in just four years, manages a team of over 50 engineers.
• Previously worked at Twitter, first as a Staff ML Engineer and then as an ML Engineering Manager.
• Holds a PhD in A.I. Theory from the University of Oxford and obtained a perfect 4.0 GPA in his Computer Science and Neuroscience joint Bachelor’s from Brown University.

Today’s episode is on the technical side so might appeal most to hands-on practitioners like data scientists and ML engineers, but anyone who’d like to understand the state-of-the-art in cybersecurity should give it a listen.

In this episode, Dan details:
• The machine learning approaches needed to tackle the uniquely adversarial application of cybercrime detection.
• How to carry out real-time ML modeling.
• What his PhD research on Category Theory entailed and how it applies to the real world.
• The major problems facing humanity in the coming decades that he thinks A.I. will be able to help with… and those that he thinks A.I. won’t.


The SuperDataScience podcast is available on all major podcasting platforms, YouTube, and at SuperDataScience.com.

In Data Science, Interview, Podcast, SuperDataScience, YouTube Tags AI, cybersecurity, LLM, data science

Happiness and Life-Fulfillment Hacks

Added on September 22, 2023 by Jon Krohn.

Today, my 94-year-old grandmother shares the secrets behind her radiant happiness. Annie talks about the importance of community, relationships and setting daily intentions, blending time-tested wisdom with forward-thinking optimism.

In today’s episode, Annie discusses: 
• Her secret to happiness.
• How she maintains flourishing long-term relationships.
• The routines and mindset she has to still be living independently, including driving herself everywhere, at 94 years old.
• The pace of technological progress in her lifetime and how A.I. could enrich her life in the years to come.

This episode is something different from the usual pure tech focus so I encourage you to provide feedback if you had strong feelings on this episode one way or another. As always, your feedback is invaluable for shaping the direction of the show.


The SuperDataScience podcast is available on all major podcasting platforms, YouTube, and at SuperDataScience.com.

In YouTube, SuperDataScience, Personal Improvement Tags Happiness, Future, old age

Make Better Decisions with Data, with Dr. Allen Downey

Added on September 19, 2023 by Jon Krohn.

Today's episode with many-time bestselling author Allen Downey is incredible. Learn a ton from him about making better decisions with data, including how to prepare for Black Swan events and how your core beliefs will shift over your life.

Allen:
• Is a Professor Emeritus at Olin College and Curriculum Designer at the learning platform Brilliant.org.
• He was previously a Visiting Professor of Computer Science at Harvard University and a Visiting Scientist at Google.
• He has written 18 books (which he has made available for free online but which are also published in hard copy by major publishers. For example, his books "Think Python" and "Think Bayes" were bestsellers published by O'Reilly).
• His next book, "Probably Overthinking It", is available for pre-order now.
• Holds a PhD in Computer Science from University of California, Berkeley and Bachelor's and Masters degrees from the Massachusetts Institute of Technology.

Today’s episode focuses largely on content from Allen’s upcoming book — his first book intended for a lay audience — and so should appeal to anyone who’s keen to learn from an absolutely brilliant writer and speaker on “How to Use Data to Answer Questions, Avoid Statistical Traps, and Make Better Decisions.”

In this episode, Allen details: 
• Underused techniques like Survival Analysis that can be uniquely powerful in lots of ordinary circumstances.
• How to better prepare for rare “Black Swan” events.
• How to wrap your head around common data paradoxes such as Preston’s Paradox, Berkson’s Paradox and Simpson’s Paradox.
• What the Overton Window is and how our core beliefs shift relative to it over the course of our lifetime (this is extra trippy).

The SuperDataScience podcast is available on all major podcasting platforms, YouTube, and at SuperDataScience.com.

In Data Science, Podcast, SuperDataScience, YouTube Tags black swan events, MIT, Data Science, AI

Using A.I. to Overcome Blindness and Thrive as a Data Scientist

Added on September 15, 2023 by Jon Krohn.

Today's guest is the remarkable Tim Albiges, who lost the ability to see as an adult. Thanks to A.I. tools, as well as learning how to learn by sound and touch, he is now thriving as a data scientist and pursuing a fascinating PhD!

Tim was working as a restaurant manager eight years ago when he tragically lost his sight.

In the face of countless alarming and discriminatory acts against him on account of his blindness, he taught himself Braille and auditory learning techniques (and to raise math equations and diagrams using a special thermoform machine so that he can feel them) in order to be able to return to college and study computing and data science.

Not only did he succeed in obtaining a Bachelor’s degree in computing (with First-Class Honours), he is now pursuing a PhD at Bournemouth University full-time, in which he’s applying machine learning to solve medical problems. His first paper was published in the peer-reviewed journal Sensors earlier this year.

Today’s inspiring episode is accessible to technical and non-technical listeners alike.

In it, Tim details:
• Why a career in data science can be ideal for a blind person.
• How he’s using ML to automate the diagnosis of chronic respiratory diseases.
• The techniques he employs to live a full and independent life, with a particular focus on the A.I. tools that assist him both at work and at leisure.
• A keen athlete, how he’s adapted his approach to fitness in order to run the London marathon and enjoy a gripping team sport called goalball.


The SuperDataScience podcast is available on all major podcasting platforms, YouTube, and at SuperDataScience.com.

In Five-Minute Friday, Data Science, Interview, Podcast, SuperDataScience, YouTube Tags AI, data careers, ML, tech, blindness

Llama 2, Toolformer and BLOOM: Open-Source LLMs with Meta’s Dr. Thomas Scialom

Added on September 12, 2023 by Jon Krohn.

Thomas Scialom, PhD is behind many of the most popular Generative A.I. projects including Llama 2, the world's top open-source LLM. Today, the Meta A.I. researcher reveals the stories behind Llama 2 and what's in the works for Llama 3.

Thomas:
• Is an A.I. Research Scientist at Meta.
• Is behind some of the world’s best-known Generative A.I. projects including Llama 2, BLOOM, Toolformer and Galactica.
• Is contributing to the development of Artificial General Intelligence (AGI).
• Has lectured at many of the top A.I. labs (e.g., Google, Stanford, MILA).
• Holds a PhD from Sorbonne University, where he specialized in Natural-Language Generation with Reinforcement Learning.

Today’s episode should be equally appealing to hands-on machine learning practitioners as well as folks who may not be hands on but are nevertheless keen to understand the state-of-the-art in A.I. from someone who’s right on the cutting edge of it all.

In this episode, Thomas details: 
• Llama 2, today’s top open-source LLM, including what is what like behind the scenes developing it and what we can expect from the eventual Llama 3 and related open-source projects.
• The Toolformer LLM that learns how to use external tools.
• The Galactica science-specific LLM, why it was brought down after a few days, and how it might eventually re-emerge in a new form.
• How RLHF — reinforcement learning from human feedback — shifts the distribution of generative A.I. outputs from approximating the average of human responses to excellent, often superhuman quality.
• How soon he thinks AGI — artificial general intelligence — will be realized and how.
• How to make the most of the Generative A.I. boom as an entrepreneur.

The SuperDataScience podcast is available on all major podcasting platforms, YouTube, and at SuperDataScience.com.

In Data Science, Interview, Professional Development, SuperDataScience, YouTube Tags LLaMA2, GPT4, AGI, opensource, LLM, Meta, Toolformer

Code Llama

Added on September 8, 2023 by Jon Krohn.

Meta's Llama 2 offered state-of-the-art performance for an "open-source"* LLM... except on tasks involving code. Now Code Llama is here and it magnificently fills that gap by outperforming all other open-source LLMs on coding benchmarks.

Read More
In Data Science, SuperDataScience, YouTube, Podcast, Five-Minute Friday Tags code, Llama, Meta, AI, LLM, GPT4

Image, Video and 3D-Model Generation from Natural Language, with Dr. Ajay Jain

Added on September 5, 2023 by Jon Krohn.

Today, brilliant ML researcher Ajay Jain, Ph.D explains how a full-length feature film could be created using Stable-Diffusion-style generative A.I. — these models can now output flawless 3D models and compelling video clips.

Ajay:
• Is a Co-Founder of Genmo AI, a platform for using natural language to generate stunning state-of-the-art images, videos and 3D models.
• Prior to Genmo, he worked as a researcher on the Google Brain team in California, in the Uber Advanced Technologies Group in Toronto and on the Applied Machine Learning team at Facebook.
• Holds a degree in Computer Science and Engineering from MIT and did his PhD within the world-class Berkeley A.I. Research (BAIR) lab, where he specialized in deep generative models.
• Has published highly influential papers at all of the most prestigious ML conferences, including NeurIPS, ICML and CVPR.

Today’s episode is on the technical side so will likely appeal primarily to hands-on practitioners, but we did our best to explain concepts so that anyone who’d like to understand the state of the art in image, video and 3D-model generation can get up to speed.

In this episode, Ajay details:
• How the Creative General Intelligence he’s developing will allow humans to express anything in natural language and get it.
• How feature-length films could be created today using generative A.I. alone.
• How the Stable Diffusion approach to text-to-image generation differs from the Generative Adversarial Network approach.
• How neural nets can represent all the aspects of a visual scene so that the scene can be rendered as desired from any perspective.
• Why a self-driving vehicle forecasting pedestrian behavior requires similar modeling capabilities to text-to-video generation.
• What he looks for in the engineers and researchers he hires.

The SuperDataScience podcast is available on all major podcasting platforms, YouTube, and at SuperDataScience.com.

In YouTube, SuperDataScience Tags 3D, AI, data science, AGI

LangChain: Create LLM Applications Easily in Python

Added on September 2, 2023 by Jon Krohn.

Today's episode is a fun intro to the powerful, versatile LLM-development framework LangChain. In it, Kris Ograbek talks us through how to use LangChain to chat with previous episodes of SuperDataScience! 😎

Kris:
• Is a content creator who specializes in creating LLM-based projects — with Python libraries like LangChain and the Hugging Face Transformers library — and then using the projects to teach these LLM techniques.
• Previously, he worked as a software engineer in Germany.
• He holds a Master’s in Electrical and Electronics Engineering from the Wroclaw University of Science and Technology.

In this episode, Kris details:
• The exceptionally popular LangChain framework for developing LLM applications.
• Specifically, he introduces how LangChain is so powerful by walking us step-by-step through a chatbot he built that interactively answers questions about episodes of the SuperDataScience podcast.

Having been listening to the podcast for years, at the end of the episode Kris flips the script on me and asks me some of the burning questions he has for me — questions that perhaps many other listeners also have wondered the answers to.

The SuperDataScience podcast is available on all major podcasting platforms, YouTube, and at SuperDataScience.com.

In YouTube, SuperDataScience, Data Science Tags SuperDataScience, data science, LangChain, LLM, Python

Big A.I. R&D Risks Reap Big Societal Rewards, with Meta’s Dr. Laurens van der Maaten

Added on August 29, 2023 by Jon Krohn.

By making big research bets, the prolific Meta Senior Research Director Dr. Laurens van der Maaten has devised or supported countless world-changing machine-learning innovations across healthcare, climate change, privacy and more.

Laurens:
• Is a Senior Research Director at Meta, overseeing swathes of their high-risk, high-reward A.I. projects with application areas as diverse as augmented reality, biological protein synthesis and tackling climate change.
• Developed the "CrypTen" privacy-preserving ML framework.
• Pioneered web-scale weakly supervised training of image-recognition models.
• Along with the iconic Geoff Hinton, created the t-SNE dimensionality reduction technique (this paper alone has been cited over 36,000 times).
• In aggregate, his works have been cited nearly 100,000 times!
• Holds a PhD in machine learning from Tilburg University in the Netherlands.

Today’s episode will probably appeal primarily to hands-on data science practitioners, but there is tons of content in this episode for anyone who’d like to appreciate the state of the art in A.I. across a broad range of socially impactful, super-cool applications.

In this episode, Laurens details: 
• How he pioneered learning across billions of weakly labeled images to create a state-of-the-art machine-vision model.
• How A.I. can be applied to the synthesis of new biological proteins with implications for both medicine and agriculture.
• Specific ways A.I. is being used to tackle climate change as well as to simulate wearable materials for enhancing augmented-reality interactivity.
• A library just like PyTorch but where all the computations are encrypted.
• The wide range of applications of his ubiquitous dimensionality-reduction approach.
• His vision for the impact of A.I. on society in the coming decades.

The SuperDataScience podcast is available on all major podcasting platforms, YouTube, and at SuperDataScience.com.

In Data Science, Podcast, SuperDataScience, YouTube Tags Ai, Meta, PyTorch, R&D, ML

ChatGPT Code Interpreter: 5 Hacks for Data Scientists

Added on August 25, 2023 by Jon Krohn.

The ChatGPT Code Interpreter is surreal: It creates and executes Python code for whatever task you describe, debugs its own runtime errors, displays charts, does file uploads/downloads, and suggests sensible next steps all along the way.

Whether you write code yourself today or not, you can take advantage of GPT-4's stellar natural-language input/output capabilities to interact with the Code Interpreter. The mind-blowing experience is equivalent to having an expert data analyst, data scientist or software developer with you to instantaneously respond to your questions or requests.

As an example of these jaw-dropping capabilities (and given the data science-focused theme of my show), I use today's episode demonstrate the ChatGPT Code Interpreter's full automation of data analysis and machine learning. If you watch the episode on YouTube, you can even see the Code Interpreter hands-on in action while I interact with it solely with natural language.

Over the course of today's episode/video, the Code Interpreter:
1. Receives a sample data file that I provide it.
2. Uses natural language to describe all of the variables that are in the file.
3. Performs a four-step Exploratory Data Analysis (EDA), including histograms, scatterplots that compare key variables and key summary statistics (all explained in natural language).
4. Preprocesses all of my variables for machine learning.
5. Selects an appropriate baseline ML model, trains it and quantitatively evaluates its performance.
6. Suggests alternative models and approaches (e.g., grid search) to get even better performance and then automatically carries these out.
7. Optionally provides Python code every step of the way and is delighted to answer any questions I have about the code.

The whole process is a ton of fun and, again, requires no coding abilities to use (the "Code Interpreter" moniker could be misleadingly intimidating to non-coding folks). Even as an experienced data scientist, however, I would estimate that in many everyday situations use of the Code Interpreter could decrease my development time by a crazy 90% or more.

The big caveat with all of this is whether you're comfortable sharing your code with OpenAI. I wouldn't provide proprietary company code to it without clearing it with your firm first and — if you do use proprietary code with it — turn "Chat history & training" off in your ChatGPT Plus settings. To circumnavigate the data-privacy issue entirely, you could alternatively try Meta's newly-released "Code Llama — Instruct 34B" Large Language Model on your own infrastructure. Code Llama won't, however, be as good as the Code Interpreter in many circumstances and will require some technical savvy to get it up and running.

The SuperDataScience podcast is available on all major podcasting platforms, YouTube, and at SuperDataScience.com.

In Five-Minute Friday, Data Science, Podcast, SuperDataScience, YouTube Tags CHATGPT, code, ML, AI

Vicuña, Gorilla, Chatbot Arena and Socially Beneficial LLMs, with Prof. Joey Gonzalez

Added on August 22, 2023 by Jon Krohn.

Vicuna, Gorilla and the Chatbot Arena are all critical elements of the new open-source LLM ecosystem — the extremely knowledgeable and innovative Prof. Joseph Gonzalez is behind all of them. Get the details in today's episode

Joey:
• Is an Associate Professor of Electrical Engineering and Computer Science at the University of California, Berkeley.
• Co-directs the Berkeley RISE Lab, which studies Real-time, Intelligent, Secure and Explainable systems.
• Co-founded Turi (acquired by Apple for $200m) and more recently Aqueduct.
• His research is integral to major software systems including Apache Spark, Ray (for scaling Python ML), GraphLab (a high-level interface for distributed ML) and Clipper (low-latency ML serving).
• His papers—published in top ML journals—have been cited over 24,000 times.
• Developed Berkeley's upper-division data science class, which he now teaches to over 1000 students per semester.

Today’s episode will probably appeal primarily to hands-on data science practitioners but we made an effort to break down technical terms so that anyone who’s interested in staying on top of the latest in open-source Generative A.I. can enjoy the episode.

In it, Prof. Gonzalez details:
• How his headline-grabbing LLM, Vicuña, came to be and how it arose as one of the leading open-source alternatives to ChatGPT.
• How his Chatbot Arena became the leading proving ground for commercial and open-source LLMs alike.
• How his Gorilla project enables open-source LLMs to call APIs making it an open-source alternative to ChatGPT’s powerful plugin functionality.
• The race for longer LLM context windows.
• How both proprietary and open-source LLMs will thrive alongside each other in the coming years.
• His vision for how A.I. will have a massive, positive societal impact over the coming decades.

The SuperDataScience podcast is available on all major podcasting platforms, YouTube, and at SuperDataScience.com.

In Data Science, Podcast, SuperDataScience, YouTube Tags Vicuna, open source, LLMs, data science, AI, ML

Large Language Model Leaderboards and Benchmarks

Added on August 19, 2023 by Jon Krohn.

Llamas, Alpacas, Koalas, Falcons... there is a veritable zoo of LLMs out there! In today's episode, Caterina Constantinescu breaks down the LLM Leaderboards and evaluation benchmarks to help you pick the right LLM for your use case.

Caterina:
• Is a Principal Data Consultant at GlobalLogic, a full-lifecycle software development services provider with over 25,000 employees worldwide.
• Previously, she worked as a data scientist for financial services and marketing firms.
• Is a key player in data science conferences and Meetups in Scotland.
• Holds a PhD from The University of Edinburgh.

In this episode, Caterina details:
• The best leaderboards (e.g., HELM, Chatbot Arena and the Hugging Face Open LLM Leaderboard) for comparing the quality of both open-source and proprietary Large Language Models (LLMs).
• The advantages and issues associated with LLM evaluation benchmarks (e.g., evaluation dataset contamination is an big issue because the top-performing LLMs are often trained on all the publicly available data they can find... including benchmark-evaluation datasets).

The SuperDataScience podcast is available on all major podcasting platforms, YouTube, and at SuperDataScience.com.

In Data Science, Interview, SuperDataScience, YouTube Tags LLMs, hugging face, Data Science, ai

Feeding the World with ML-Powered Precision Agriculture

Added on August 15, 2023 by Jon Krohn.

Earth's population may peak around 10 billion later this century. To feed everyone while also avoiding climate disaster, A.I. is essential. Today, three leaders from Syngenta detail how ML is transforming agriculture and assuring our future.

Read More
In Podcast, SuperDataScience, YouTube Tags Syngenta, Agriculture, ML, data science, sustatinability, ai

Jon’s “Generative A.I. with LLMs” Hands-on Training

Added on August 11, 2023 by Jon Krohn.

Today's episode introduces my two-hour "Generative A.I with LLMs" training, which is packed with hands-on Python demos in Colab notebooks. It details open-source LLM (Hugging Face; PyTorch Lightning) and commercial (OpenAI API) options.

Read More
In Data Science, Five-Minute Friday, SuperDataScience, YouTube Tags hands-on learning, AI, LLMs, hugging face, open ai

How Data Happened: A History, with Columbia Prof. Chris Wiggins

Added on August 8, 2023 by Jon Krohn.

Today, Chris Wiggins — Chief Data Scientist at The New York Times and faculty at Columbia University — provides an enthralling, witty and rich History of Data Science. Chris is an extraordinarily gifted orator; don't miss this episode!

Chris:
• Is an Associate Professor of Applied Math at Columbia University.
• Has been CDS at The NY Times for nearly a decade.
• Co-authored two fascinating recently-published books: "How Data Happened: A History from the Age of Reason to the Age of Algorithms" and "Data Science in Context: Foundations, Challenges, Opportunities"

The vast majority of this episode will be accessible to anyone. There are just a couple of questions near the end that cover content on tools and programming languages that are primarily intended for hands-on practitioners.

In the episode, Chris magnificently details: 
• The history of data and statistics from its infancy centuries ago to the present.
• Why it’s a problem that most data scientists have limited exposure to the humanities.
• How and when Bayesian statistics became controversial.
• What we can do to address the key issues facing data science and ML today.
• His computational biology research at Columbia.
• The tech stack used for data science at the globally revered New York Times.

The SuperDataScience podcast is available on all major podcasting platforms, YouTube, and at SuperDataScience.com.

In Interview, Data Science, Podcast, SuperDataScience, YouTube Tags data science, Columbia, NY Times, bayesianstatistics
← Newer Posts Older Posts →
Back to Top