NASA Let AI Drive on Mars for the First Time
This one is genuinely wild. NASA's Perseverance rover just completed the first Mars drives ever planned entirely by artificial intelligence. Over two drives totaling 456 meters, Claude's vision-language models handled the planning that human operators had been doing by hand for 28 years.
Think about that for a second. For nearly three decades, a team of people sat down every day and figured out where a rover on another planet should drive. Now an AI does it.
This is one of those moments that sounds incremental but isn't. Autonomous planning on Mars isn't just a cool demo. It means faster science, less downtime waiting for Earth-side approvals, and a template for every deep-space mission that comes after this. Space exploration just got a lot more interesting.
Snap Fired 1,000 People and Said AI Did It
Snap CEO Evan Spiegel announced layoffs of around 1,000 employees and closed over 300 open roles. The reason he gave was pretty direct: AI now writes more than 65% of Snap's new code, so they just don't need as many people to do the same work.
This is the moment a lot of people have been nervous about for years. Not robots taking blue-collar jobs in some hypothetical future, but AI quietly making whole categories of white-collar tech work cheaper and faster right now. Snap is not a small company running some experiment. They shipped this at scale.
The honest take: this is going to keep happening. Companies that figure out how to run lean with AI-assisted teams are going to have a real cost advantage over the ones that don't. If you write code for a living, the question isn't whether AI is coming for some of your work. It already is. The question is what you do about it.
An Attorney Got Suspended for 57 Fake AI Citations
Omaha attorney Greg Lake just got suspended from practicing law after his appellate brief contained 57 defective citations out of 63 total. Twenty of those were full-on AI hallucinations. Cases that don't exist. Courts that never ruled. Made up out of thin air.
This is not the first time this has happened. U.S. courts handed out at least $145,000 in sanctions against attorneys for AI citation errors in just the first quarter of 2026. That's a lot of lawyers learning the hard way that ChatGPT does not double-check its sources.
Here's the thing. AI tools are genuinely useful for legal research. But they are not reliable for citation. They will invent plausible-sounding cases with confidence, and if you don't verify every single one, you are gambling your license. The tool is powerful. It is not trustworthy without a human in the loop checking the work.
Use the tools. Just don't trust them blindly. That lesson is coming for a lot of professions, and law is just the one where the consequences are most visible right now.