2025-03-17 08:00:00
Future LLMs have the potential to cause significant harm due to their ruthless effiency. I'm worried this will happen, and discuss the ways in which it might.
2025-03-13 08:00:00
I have very wide error bars on the potential future of large language models, and I think you should too. It's possible LLMs basically lead to AGI, and it's also possible they platteau.
2025-03-11 08:00:00
My work on privacy-preserving machine learning is often cited by lawyers arguing for or against how generative AI models violate copyright. This maybe isn't the right work to be citing.
2025-03-05 08:00:00
I have decided to leave Google, and will be joining Anthropic to continue my work on machine learning security
2025-02-09 08:00:00
A one-year review of people's predictions on an AI-forecasting survey I made last year. Most people were over-confident in their predictions.
2025-01-05 08:00:00
I wrote a (list of) regular expressions that will play a (not very good) chess game by running a 2-ply minimax search.