šŸ—žļøAI highlights from this week (1/27/23)

Generative music, language models as backend servers, AI Family Guy and moreā€¦

šŸ—žļøAI highlights from this week (1/27/23)

[Update: the previous version of this post had an incorrect sub header]

Hi readers,

Here are my highlights from the last week in AI!

P.S. Donā€™t forget to hit subscribe if youā€™re new to AI and want to learn more about the space.


Highlights

1/ Google makes leap in Generative Music

One of the areas that has most excited me about AI is its ability to democratize the creative process. As a musician myself, when I first started playing with generative AI products like Dall-E, my immediate thought was ā€œThis would be amazing for musicā€.

There have been a few different projects attempting to achieve generative music, including HarmonyAI which is able to generate new music that sounds like the input music and Riffusion which does short text-to-audio using Stable Diffusion by turning audio into images. OpenAI also published a paper on a model they call JukeBox that generates music in particular genres and styles.

In my opinion though, the holy grail is for a user to describe any kind of music or sound and for a model to generate, and it looks like Google just achieved this with MusicML!

Check out their research website where they shared lots of examples of MusicML in action, including longer songs, audio journeys with multiple parts, turning paintings into music and even generating specific instrument sounds! As of yet, thereā€™s no tools for you to try out MusicML with your own prompts but hereā€™s hoping that this research will be available in a product by Google later this year.

2/ Using a Large-scale Language Model as a backend

Last weekend ScaleAI1 hosted an AI hackathon in San Francisco. The winning teamā€™s project ā€œGPT is all your need for backendā€2, might pique the curiosity of any engineers reading this post, as they were able to show how a large-scale language model, in this case GPT, could be used instead of a traditional database and server-based backend3:

Hereā€™s how one of the team members described what they were aiming for:

What was so impressive about what the team achieved is that they were able to completely remove the need for a server or database to store data for their example application, a To Do app. Instead they just taught GPT4, what app they were building and how it should respond to requests, as well as providing examples of the type of data the frontend part of the To Do app might request e.g. a list of to do items. Once, this is done, the frontend can just describe the functions it wants to call, without them ever being defined!

Hereā€™s a more detailed description of how ā€œbackend-GPTā€ works from their Github Repository:

We basically used GPT to handle all the backend logic for a todo-list app. We represented the state of the app as a json with some prepopulated entries which helped define the schema. Then we pass the prompt, the current state, and some user-inputted instruction/API call in and extract a response to the client + the new state. So the idea is that instead of writing backend routes, the LLM can handle all the basic CRUD logic for a simple app so instead of writing specific routes, you can input commands like add_five_housework_todos() or delete_last_two_todos() or sort_todos_alphabetically() . It tends to work better when the commands are expressed as functions/pseudo function calls but natural language instructions like delete last todos also work.

Iā€™ve discussed in previous posts about the concept of emergent behavior, whereby a language model which is trained on a large enough dataset is able to carry out tasks and perform logic that is unexpected. This idea of a large-scale language model acting as as general purpose backend is a great example of emergent behavior!

3/ Atomic AI raises $35M to use AI for RNA-based drug discovery

With all the hype around chatbots and generative art, itā€™s great to also hear that AI companies are being created to save lives too. One such company is Atomic AI, a biotech startup that raised $35M in series A funding to do generative AI-based drug discovery focused on RNA molecules. Hereā€™s how Raphael Townshend, CEO of Atomic AI describes the opportunity his startup is going after in an interview with TechCrunch:

ā€œThereā€™s this central dogma that DNA goes to RNA, which goes to proteins. But itā€™s emerged in recent years that it does much more than just encode information,ā€¦ If you look at the human genome, about 2% becomes protein at some point. But 80 percent becomes RNA. And itā€™s doingā€¦ who knows what? Itā€™s vastly underexplored.ā€

Check out Michael Spencerā€™s post for more on Atomic AI and the intersection of AI and biotech:

What is Atomic AI?
Hey Everyone, I really like covering A.I. startups at the intersection of computational biology, biotech, genomics, and drug development. A.I. is increasingly becoming implicated in mRNA and RNA medicines. This week, Atomic AI, a biotechnology company fusing cutting-edge machine learning with state-of-the-art structural biology to unlock RNA drug discovā€¦

4/ Yann LeCun throws shade on ChatGPT!

The legendary AI researcher Yann LeCun, who was one of a few researchers pushing forward advancements in deep learning during the 70s-90s5 tweeted that he thought ChatGPT was overhyped:

I think Yann might be overestimating the general publicā€™s understanding of deep learning, AI and the progress weā€™ve made in the last few decades. Until ChatGPT, most people simple had not experience AI in a tangible and impressive product, as I shared in AI: Donā€™t believe the hype?:

Unlike itā€™s predecessors (e.g. Google Assistant, Echo, Siri), ChatGPT is really the first time an AI assistant truly seems like it could pass the Turing Test. There have been many impressive examples of ChatGPT in action and if you havenā€™t tried it yourself you should. ChatGPT successfully wrote a blog post for me and turned it into a twitter thread, gave me a recipe for pancakes that tasted delicious and helped me pick a Christmas present for my wife!

OpenAI are capturing attention not because of the sophistication of their models but because they are shipping great products, as pointed out by Dr. Jim Fan, an AI scientist who previously worked at OpenAI and Google:

Itā€™s also hard not to take Yannā€™s sentiment with a grain of salt given that he leads AI research at Meta. Maybe Yann should spend less time throwing shade and more time persuading Zuck to burn the virtual boats and join the AI race?

Or, maybe we should all just be friends and work on this togetherā€¦

5/ Family guy and generative AI

Wrapping up with this fun take on what Family Guy might have looked like as an 80s live action sitcom using images created with Midjourney!

Everything elseā€¦

Finally, in case you missed it, I also shared Part 3 of my series on the origins of Deep Learning:

šŸ¤“A Deep Dive into Deep Learning: Part 3
Hi Readers! Thank you for subscribing to my newsletter. Hereā€™s the final part of my deep dive into the origins of deep learning. In case you missed it, here are Part 1 and Part 2. The field of deep learning is filled with lots of jargon. When you see the šŸ¤“ emoji, thatā€™s where I go a

Thatā€™s all for this week!


Thanks for reading The Hitchhikers Guide to AI! Subscribe for free to receive new posts and support my work.


  1. Scale AI provides infrastructure and resources to label large datasets for machine learning for many different use cases including robotics, AR/VR, AI and autonomous vehicles. ā†©

  2. The projectā€™s title ā€œGPT is all you need for backendā€, is a play on words on ā€œAttention is all you need,ā€ the famous Google research paper that introduced the Transformer architecture used by large-scale language models. If you want to learn more about what Transformers are, read my latest post on the origins of Deep Learning. ā†©

  3. A ā€œbackendā€ is the part of a web application that stores and serves data to the ā€œfrontendā€ that you interact with as a user. For example this web page is the frontend of substack and the backend is what stores and serves all the text in this post. ā†©

  4. GPT or General Pre-trained Transformer is OpenAIā€™s large-scale language model that powers ChatGPT. ā†©

  5. If you want to learn more about Yann LeCunn and his work curing the ā€œAI Winterā€ read part 2 in my series on the origins of Deep Learning. ā†©