-
AI and the New Faith: How the Singularity Became a Modern Religion
Artificial intelligence is advancing at an ever-growing pace, reshaping industries and capturing the world’s attention. While often framed as an economic and technological revolution, there’s a deeper, more unexpected transformation underway. In this context, people often refer to the technological singularity, a concept that has existed for over 60 years, dating back to John von… — read more
-
Looking Back On 2023 And Predictions for 2024
As we close the chapter on 2023, it’s time to revisit the predictions I laid out at the beginning of the year. It was a year marked by technological strides and societal challenges. Let’s evaluate how my forecasts stood against the unfolding of 2023. Let’s start with my predictions about AI: “AI will continue to… — read more
-
Episode 11: GPT-4 Leak, Wolfram Alpha und AGI.
In dieser Episode reden Florian und Ich über den GPT-4 Leak, den Nvidia deal mit China, alternative KI Ideen, kleine Roboter und vieles mehr. Mehr informationen auf dem Discord serverhttps://discord.gg/3YzyeGJHthoder auf https://mkannen.tech/ — read more
-
AI helps with AI Understanding
One of the main problems of LLMs is that they are black boxes and how they produce an output is not understandable for humans. Understanding what different neurons are representing and how they influence the model is important to make sure they are reliable and do not contain dangerous trends. OpenAI applied GPT-4 to find… — read more
-
Study Extends BERT’s Context Length to 2 Million Tokens
Researchers have made a breakthrough in the field of artificial intelligence, successfully extending the context length of BERT, a Transformer-based natural language processing model, to two million tokens. The team achieved this feat by incorporating a recurrent memory into BERT using the Recurrent Memory Transformer (RMT) architecture. The researchers’ method increases the model’s effective context… — read more
-
Google and DeepMind Team Up
Google and DeepMind just announced that they will unite Google Brain and Deepmind into Google DeepMind. This is a good step for both sites since Deepmind really needs the computing power of Google to make further progress on AGI and Google needs the Manpower and knowledge of the Deepmind team to quickly catch up to… — read more
-
MiniGPT-4 is an Open-Source Multimodal Model
MiniGPT-4, is an open-source multimodal model similar to the version of GPT-4 that was shown during OpenAI’s presentation. It combines a Visual encoder with an LLM. They used Vicuna which is a fine-tuned version of LLaMA. In the future, I hope more teams try to add new ideas to their models instead of creating more… — read more
-
Stanford and Google let AI roleplay
In a new research paper, Google and Stanford University created a sandbox world where they let 25 AI agents role-play. The agents are based on chatGPT-3.5 and behave more believably than real humans. Future agents based on GPT-4 will be able to act even more realistically and intelligently. This could not only mean that we… — read more
-
The New Wave of GPT Agents
Since GPT-3.5 and GPT-4 APIs are available many companies and start-ups have implemented them into their products. Now developers have started to do it the other way around. They build systems around GPT-4 to enable it to search, use APIs, execute code, and interact with itself. Examples are HuggingGPT or AutoGPT. They are based on… — read more
-
Open Letter to pause bigger AI models
A group of researchers and notable people released an open letter in which they call for a 6 month stop from developing models that are more advanced than GPT-4. Some of the notable names are researchers from competing companies like Deepmind, Google, and Stability AI like Victoria Krakovna, Noam Shazeer, and Emad Mostaque. But also… — read more
-
Listen to OpenAI
Many people saw the new episode of the Lex Friedman Podcast with Sam Altman, where he talks about some social and political implications of GPT-4. But fewer people saw the podcast with Ilya Sutskever, the Chief Scientist at OpenAI, which is way more technical and in my opinion even more exciting and enjoyable. I really… — read more
-
Sparks of Artificial General Intelligence: Early experiments with GPT-4
Microsoft researchers have conducted an investigation on an early version of OpenAI’s GPT-4, and they have found that it exhibits more general intelligence than previous AI models. The model can solve novel and difficult tasks spanning mathematics, coding, vision, medicine, law, psychology, and more, without needing any special prompting. Furthermore, in all of these tasks,… — read more
-
Learning to Grow Pretrained Models for Efficient Transformer Training
A new research paper proposes a method to accelerate the training of large-scale transformers, called the Linear Growth Operator (LiGO). By utilizing the parameters of smaller, pre-trained models to initialize larger models, LiGO can save up to 50% of the computational cost of training from scratch while achieving better performance. This approach could have important… — read more
-
ChatGPT’s biggest update jet
OpenAI announced that they will introduce plugins to ChatGPT. Two of them developed by OpenAi themself allow the model to search the web for information and run generated python code. Other third-party plugins like Wolfram allow the model to use other APIs to perform certain tasks. the future capabilities of a model enhanced this way… — read more
-
From GPT-4 to Proto-AGI
Deutsche Version Artificial General Intelligence (AGI) is the ultimate goal of many AI researchers and enthusiasts. It refers to the ability of a machine to perform any intellectual task that a human can do, such as reasoning, learning, creativity, and generalization. However, we are still far from achieving AGI with our current AI systems. One… — read more
-
FlexGen Enables High-Throughput Inference of Large Language Models on Single GPUs
FlexGen is a new generation engine that enables high-throughput inference of large language models on a single commodity GPU. It uses a linear programming optimizer to efficiently store and access tensors and compresses weights and attention cache to 4 bits. FlexGen achieves significantly higher throughput than state-of-the-art offloading systems, reaching a generation throughput of 1… — read more
-
MathPrompter: Mathematical Reasoning using Large Language Models
Microsoft published a new paper in which they present the language model MathPrompter which uses the Zero-shot chain-of-thought prompting technique to generate multiple Algebraic expressions or Python functions to solve the same math problem in different ways and thereby raise the confidence level in the output results. This led to a score of 92.5 on… — read more
-
Jan Leike Shares His Ideas on Alignment
Jan Leike, the alignment team lead at OpenAI, has published a blog post where he presents possible solutions for AI aligning to group preferences. He proposes a simulated deliberative democracy, where groups of randomly selected people agree on a stand on a specific topic. The decision is then used to fine-tune the AI. — read more
-
GPT-4 Next Week
In a small german information event today, four Microsoft employees talked about the potential of LLMs and mentioned that they are going to release GPT-4 next week. They implied that GPT-4 will be able to work with video data, which implies a multimodal model comparable to PaLM-E. Read more here. — read more