Neural Networks

AI Large Language Models Research

Scaling Transformer to 1M tokens and beyond with RMT (Paper Explained)

Yannic Kilcher explains this paper that promises to scale transformers to 1 million tokens and beyond. We take a look at the technique behind it: The Recurrent Memory Transformer, and what its strengths and weaknesses are.

Read More
Future Hardware Research

Get the Inside Scoop on Neuromorphic Computing Part 1

Computer design has always been inspired by biology, especially the brain. In this episode of Architecture All Access – Mike Davies, Senior Principal Engineer and Director of Intel’s Neuromorphic Computing Lab – explains the relationship of Neuromorphic Computing and understanding the principals of brain computations at the circuit level that are enabling next-generation intelligent devices […]

Read More
Neural Networks

Code From Scratch: Neural Networks

Sapphire Dev shows us how neural networks operate behind the hood can be hidden from the programmer in frameworks such as TensorFlow. But to become a competent AI developer, understanding the fundamental algorithms behind these networks is invaluable information. Here, we embark on a journey to create our own network in the Dart programming language.

Read More
Natural Language Processing Research

LLaMA: Open and Efficient Foundation Language Models (Paper Explained)

Large Language Models (LLMs) are all the rage right now. ChatGPT is the LLM everyone talks about, but there are others. With the attention (and money) that OpenAI is getting, expect more of them. LLaMA is a series of large language models from 7B to 65B parameters, trained by Meta AI. They train for longer […]

Read More
Neural Networks

What are Activation Functions?

Did you ever wonder what activation functions are? Then watch this video to find out. The script to this video was AI generated as was the presenter. Ironically, I did the grunt work of editing, mixing audio, and rendering the final product. Soon, there’ll be an AI for that, too.

Read More
AI Generative AI Natural Language Processing

ChatGPT: This AI has a JAILBREAK?!

Yannic explores ChatGPT and discovers that it has a JailBreak?! ChatGPT, OpenAI’s newest model is a GPT-3 variant that has been fine-tuned using Reinforcement Learning from Human Feedback, and it is taking the world by storm!

Read More
AI Natural Language Processing

ROME: Locating and Editing Factual Associations in GPT (Paper Explained & Author Interview)

Large Language Models have the ability to store vast amounts of facts about the world. But little is known, how these models actually do this. This paper aims at discovering the mechanism and location of storage and recall of factual associations in GPT models, and then proposes a mechanism for the targeted editing of such […]

Read More