Ben and Ryan are joined by Robin Gupta for a conversation about benchmarking and testing AI systems. They talk through the lack of trust and confidence in AI, the inherent challenges of nondeterministic systems, the role of human verification, and whether we can (or should) expect an AI to be reliable. https://stackoverflow.blog/2024/05/24/would-you-board-a-plane-safety-tested-by-genai/
Войдите, чтобы добавить комментарий
Другие сообщения в этой группе

Today’s episode is a roundup of spontaneous, on-the-ground conversations from HumanX 2025, featuring guests from CodeConductor, DDN, Cloudflare, and Galileo. https://stackoverflow.blog/2025/04/25/grab

In this episode of Leaders of Code, we chat with guests from Lloyds Banking Group about their focus on engineering excellence and the need for organizations to adapt to new technologies while ensuring

An update on recent launches and the upcoming roadmap https://stackoverflow.blog/2025/04/23/community-products-roadmap-update-april-2025/

Ryan chats with Dataiku CEO and cofounder Florian Douetteau about the complexities of the genAI data stack and how his company is orchestrating it. https://stackoverflow.blog/2025/04/22/visually-orch

On today’s episode, Ben and Ryan chat with Laly Bar-Ilan, Chief Scientist at Bit. https://stackoverflow.blog/2025/04/18/generating-components-not-tokens/

Is “agentic AI” just a buzzword, or is it the sea change it seems? https://stackoverflow.blog/2025/04/17/wait-what-is-agentic-ai/

Kyle chats with Jesse Tomchak a software engineer at ClickUp about all the spicy backend takes they could find. https://stackoverflow.blog/2025/04/09/wbit-6-be-curious-ask-questions-and-don-t-argue-w