35th issue! If you missed them, you can read the previous issues of the Machine Learning Monthly newsletter here.
Hey everyone!
Daniel here, I’m a machine learning engineer who teaches the following beginner-friendly machine learning courses:
I also write regularly about machine learning on my own blog as well as make videos on the topic on YouTube.
Since there's a lot going on, the utmost care has been taken to keep things to the point.
Enough about me!
You're here for this month's Machine Learning Monthly Newsletter. Typically a 500ish (+/-1,000ish, usually +) word post detailing some of the most interesting things on machine learning I've found in the last month.
OpenAI just released a new version of their GPT-3 model called ChatGPT with a focus on dialogue.
There’s no point trying to describe what’s going on with it, better to try it out for yourself.
All I can say is that it’s by far the best chatbot I’ve ever used.
Goes to show that amazing AI applications are on the horizon.
It’s a design and product problem now.
The models are there and they’re good enough, it’s up to what you build with them.
Check out this example response when I asked ChatGPT to create an upper body workout for me in the form of a folklore story:
text-davinci-003
Two OpenAI releases within a couple of days of each other!
As if ChatGPT wasn’t enough, there’s now a new model powering the original GPT-3 API, text-davinci-003
.
Trained using reinforcement learning to improve results, OpenAI claims several improvements over the previous text-davinci-002
model.
See the comparison Scale did on their blog comparing the two models across several different problems for more.
If you’ve ever tried to learn how to use cloud resources, you know there’s a pretty steep learning curve.
Modal changes this.
Built by Erik Bernhardsson (whose work has been in plenty of previous ML monthly’s + designed the recommendation engine at Spotify) and team, they’re setting the standard for future cloud development.
Check out their docs for a series of fantastic examples of how to train/deploy/use machine learning models in the Model cloud in minutes:
Google AI introduces two new papers on how to train language models with significantly less compute.
Example of the instruction fine-tuning task. Taking a series of existing NLP tasks and using them as instructions to prepare a model for future use on similar (but unseen) tasks. Source: Google AI Blog.
It looks like some of the weights of Google’s (smaller) versions of the models are available to use on GitHub as well as the Hugging Face models repository under the name “FLAN” for “fine-tuning language models”.
However, it seems Google may have missed the boat a little with making their models available to people to use.
Depending on what they’re doing internally with them (Google’s language models are no doubt powering several of their search services), it seems OpenAI has the leg up in providing usable language models (even if they perform worse than Google’s on paper).
It goes to show having a good product that’s available to use is better than having a great product that isn’t available.
Airbnb is one of my favourite apps.
I used it all throughout a recent Europe trip.
So I love seeing how they use one of my other loves, machine learning, to power their services.
Two new articles this month from their tech blog:
Ever since the release of the Zero to Mastery PyTorch course, I’ve been learning more and more about different libraries in the PyTorch ecosystem.
The PyTorch Image Models library (timm
for short) is one of the most popular and useful and often appears in citations for several research papers.
I’ve been using timm
to build new models for my AI project called Nutrify (take a photo of food and learn about it).
And the Practitioner’s Guide by Chris Hughes offers a fantastic walkthrough of many of the features.
I’ve even printed out the timm
train.py
script to study it and improve my own scripts.
Text-to-image models have exploded over the last several months.
On the surface, we put in a line of text, such as “a flamingo performing a bench press in a weightlifting” and we get an image back.
Using OpenAI’s DALL•E 2 with the prompt “a flamingo performing a bench press in a weightlifting”.
But what happens behind the scenes?
Eugene writes that the modern image generation models use a combination of four techniques:
The incredible Goku Mohandas has just updated the amazing Made with ML website with the heart of machine learning: the data stack.
Alongside several other new lessons such as machine learning orchestration (putting together several pieces of a machine learning pipeline) and machine learning testing the data stack lesson explains what kinds of data storage types go into a machine learning problem:
Knowing the different data stacks and how they interact with each other is a fantastic way to level up your knowledge as a data engineer and in turn, a machine learning engineer.
John McDonnell writes about where the future of AI is headed with all of the latest releases of language models.
In 2022 language models got good, really good.
Now in 2023 and onwards the trick will be combining them in some way to take actions in the real world and then using the results from those actions to update their steps.
Using large language models (LLMs) to start providing a service and then updating (fine-tuning) the model to provide an even better service. Source: John McDonnell Substack.
TorchMultimodal Beta is out!
With all the talk about vision and language models, you can now use them baked directly into the PyTorch library as a domain library.
There are already a bunch of pretrained models built-in such as ALBEF for visual question answering (asking a question of an image and getting a text result) and MDETR (detecting an object class on an image defined with natural language, e.g. a class of “pelican” even though you don’t have any labelled examples of pelican).
What a massive month for the ML world in November!
As always, let me know if there's anything you think should be included in a future post.
In the meantime, keep learning, keep creating, keep dancing.
See you next month,
Daniel
By the way, I'm a full-time instructor with Zero To Mastery Academy teaching people Machine Learning in the most efficient way possible. You can see a few of our courses below or check out all Zero To Mastery courses.