2024 Data Science Trend Predictions

Data Science

What are the big A.I. trends going to be in 2024? In this episode of SuperDataScience, hosted by our Chief Data Scientist, Jon Krohn, the magnificent data-science leader and futurist Sadie St. Lawrence fill us in by methodically making her way from the hardware layer (e.g., GPUs) up to the application layer (e.g., GenAI apps).

More on Sadie:

• Machine Learning instructor who’s had over 500,000 students.

• Founder/CEO of Women In Data™️, a community of over 60,000 across 55 countries.

• Head of A.I. at SSL Innovations.

• Hosts the Data Bytes podcast.

• Serves on multiple start-up boards.

Sadie was previously the guide to both the 2022 and the 2023 predictions episodes. Those are two of the all-time most popular episodes of this podcast and, if you listened to them, you already know that you’re in for a treat again today.

This episode will appeal to technical and non-technical folks alike — anyone who’d like to understand the trends that will shape the field of data science and the broader world not only in 2024 but also in the years beyond.

They start the episode off by looking back at how our predictions panned out from a year ago and then we’ll dive into our predictions for the year ahead. Starting from the hardware layer then moving gradually toward the application layer, specific 2024 trends Sadie leads discussion of include:

• A.I. hardware accelerators

• Large language models (LLMs) as “operating systems”

• Models like Q* that solve problems along non-linear paths

• Consolidation of enterprise systems thanks to function-calling LLMs

• Whether generative A.I. will replace data analysts

The SuperDataScience podcast is available on all major podcasting platforms, YouTube, and at SuperDataScience.com.


Getting Value From A.I.

In February 2023, our Chief Data Scientist, Jon Krohn, delivered this keynote on “Getting Value from A.I.” to open the second day of Hg Capital’s “Digital Forum” in London.

read full post

The Chinchilla Scaling Laws

The Chinchilla Scaling Laws dictate the amount of training data needed to optimally train a Large Language Model (LLM) of a given size. For Five-Minute Friday, our Chief Data Scientist, Jon Krohn, covers this ratio and the LLMs that have arisen from it.

read full post

StableLM: Open-Source “ChatGPT”-Like LLMs You Can Fit on One GPU

The folks who open-sourced Stable Diffusion have now released “StableLM”, their first Language Models. Pre-trained on an unprecedented amount of data for single-GPU LLMs (1.5 trillion tokens!), these are small but mighty.

read full post