InfoQ Homepage Artificial Intelligence Content on InfoQ
-
GenAI Security: Defending Against Deepfakes and Automated Social Engineering
In this episode, QCon AI New York 2025 Chair Wes Reisz speaks with Reken CEO and Google Trust & Safety founder Shuman Ghosemajumder about the erosion of digital trust. They explore how deepfakes and automated social engineering are scaling cybercrime and argues defenders must move beyond default trust, utilizing behavioral telemetry and game theory to counter attacks that simulate human behavior.
-
Cloud Security Challenges in the AI Era - How Running Containers and Inference Weaken Your System
Marina Moore, a security researcher and the co-chair of the security and compliance TAG of CNCF, shares her concerns about the security vulnerabilities of containers. She explains where the issues originate, providing solutions and discussing alternative routes to using micro-VMs rather than containers. Additionally, she highlights the risks associated with AI inference.
-
If You Can’t Test It, Don’t Deploy It: The New Rule of AI Development?
Magdalena Picariello reframes how we think about AI, moving the conversation from algorithms and metrics to business impact and outcomes. She champions evaluation systems that don't just measure accuracy but also demonstrate real-world business value, and advocates for iterative development with continuous feedback to build optimal applications.
-
Mental Models in Architecture and Societal Views of Technology: a Conversation with Nimisha Asthagiri
In this podcast, Michael Stiefel spoke with Nimisha Asthagiri about the importance of system thinking, multi-agent systems, the consequences of society applying a technology into an area for which it was not designed, and whether we can ever have a healthy relationship with artificial intelligence.
-
Elena Samuylova on Large Language Model (LLM)-Based Application Evaluation and LLM as a Judge
In this podcast, InfoQ spoke with Elena Samuylova from Evidently AI, on best practices in evaluating Large Language Model (LLM)-based applications. She also discussed the tools for evaluating, testing and monitoring applications powered by AI technologies.
-
AI Amplifies Team Strengths and Weaknesses in Software Development
In this podcast, Shane Hastie, Lead Editor for Culture & Methods, spoke to Jon Kern and Anita Zbieg about how AI amplifies both delivery efficiency and weaknesses in development teams, the importance of fundamental collaboration practices, and maintaining holistic system thinking.
-
Building a Product-First Engineering Culture in the Age of AI
In this podcast, Shane Hastie, Lead Editor for Culture & Methods, spoke to Zach Lloyd about building a product-first engineering culture, and the critical importance of developers learning to effectively use AI tools while maintaining responsibility for code quality and understanding fundamental programming principles.
-
GitHub Next: how their research and prototyping team operates
In this podcast, Shane Hastie, Lead Editor for Culture & Methods spoke to Idan Gazit and Eddie Aftandilian from GitHub Next how their research and prototyping team operates as a "department of fool around and find out", exploring AI-powered developer tools through rapid experimentation and user feedback.
-
Technology Radar and the Reality of AI in Software Development
Shane Hastie, Lead Editor for Culture & Methods spoke to Rachel Laycock, Global CTO of Thoughtworks, about how the company's Technology Radar process captures technology trends around the globe. She is sceptical of the current AI efficiency hype, emphasizing that real value of generative AI tools lies in solving complex problems like legacy code comprehension rather than just writing code faster.
-
Using AI Code Generation to Migrate 20000 Tests
In this podcast, Shane Hastie, Lead Editor for Culture & Methods spoke to Sergii Gorbachov, a staff engineer at Slack, about how they successfully used AI combined with traditional coding approaches to migrate 20,000 tests in 10 months, discovering that AI alone was insufficient and required human oversight and conventional tools to work effectively.