Apart Research
Subscribe
Sign in
Home
Apart Newsletter
Hackathons
Product development
Organization news
🌐 Versión en español
Latest
Top
Large language models might always be slightly misaligned
Apart Newsletter #28
Apr 25, 2023
•
Apart Research
3
Share this post
Large language models might always be slightly misaligned
news.apartresearch.com
Copy link
Facebook
Email
Note
Other
Apart Newsletter #27
ML safety research in interpretability and model sharing, game-playing language models, and critiques of AGI risk research
Apr 18, 2023
•
Apart Research
3
Share this post
Apart Newsletter #27
news.apartresearch.com
Copy link
Facebook
Email
Note
Other
Identifying semantic neurons, mechanistic circuits & interpretability web apps
Results from the 4th Alignment Jam
Apr 13, 2023
•
CC
2
Share this post
Identifying semantic neurons, mechanistic circuits & interpretability web apps
news.apartresearch.com
Copy link
Facebook
Email
Note
Other
Ethics or Reward?
This week we take a look at LLMs that need therapists, governance of machine learning hardware, and benchmarks for dangerous behaviour. Read to the end…
Apr 9, 2023
•
CC
5
Share this post
Ethics or Reward?
news.apartresearch.com
Copy link
Facebook
Email
Note
Other
Governing AI & Evaluating Danger
We might need to shut it all down, AI governance seems more important than ever and technical research is challenged. Welcome to this week's update…
Apr 3, 2023
Share this post
Governing AI & Evaluating Danger
news.apartresearch.com
Copy link
Facebook
Email
Note
Other
March 2023
What a Week! GPT-4 & Japanese Alignment
What a week. There was already a lot to cover Monday when I came in for work and I was going to do a special feature on the Japan Alignment Conference…
Mar 15, 2023
Share this post
What a Week! GPT-4 & Japanese Alignment
news.apartresearch.com
Copy link
Facebook
Email
Note
Other
Perspectives on AI Safety
This week, we take a look at interpretability used on a Go-playing neural network, glitchy tokens and the opinions and actions of top AI labs and…
Mar 6, 2023
Share this post
Perspectives on AI Safety
news.apartresearch.com
Copy link
Facebook
Email
Note
Other
February 2023
Bing Wants to Kill Humanity W07
Welcome to this week’s ML & AI safety update where we look at Bing going bananas, see that certification mechanisms can be exploited and that scaling…
Feb 21, 2023
Share this post
Bing Wants to Kill Humanity W07
news.apartresearch.com
Copy link
Facebook
Email
Note
Other
Will Microsoft and Google start an AI arms race? W06
We would not be an AI newsletter without covering the past week’s releases from Google and Microsoft but we will use this chance to introduce the…
Feb 10, 2023
Share this post
Will Microsoft and Google start an AI arms race? W06
news.apartresearch.com
Copy link
Facebook
Email
Note
Other
Extreme AI Risk W05
In this week's newsletter, we explore the topic of modern large models’ alignment and examine criticisms of extreme AI risk arguments. Of course, don't…
Feb 6, 2023
Share this post
Extreme AI Risk W05
news.apartresearch.com
Copy link
Facebook
Email
Note
Other
January 2023
Was ChatGPT a good idea? W04
In this week’s ML & AI Safety Update, we hear Paul Christiano’s take on one of OpenAI’s main alignment strategies, dive into the second round winners of…
Jan 28, 2023
2
Share this post
Was ChatGPT a good idea? W04
news.apartresearch.com
Copy link
Facebook
Email
Note
Other
Compiling code to neural networks? W03
Welcome to this week’s ML & AI Safety Report where we dive into overfitting and look at a compiler for Transformer architectures! This week is a bit…
Jan 20, 2023
2
Share this post
Compiling code to neural networks? W03
news.apartresearch.com
Copy link
Facebook
Email
Note
Other
Share
Copy link
Facebook
Email
Note
Other
This site requires JavaScript to run correctly. Please
turn on JavaScript
or unblock scripts