InfoQ Homepage AI, ML & Data Engineering Content on InfoQ
-
Google Report Reveals How Threat Actors Are Currently Using Generative AI
Google's Threat Intelligence Group (GTIG) recently released a report on the adversarial misuse of generative AI. The team investigated prompts used by advanced persistent threat (APT) and coordinated information operations (IO) actors, finding that they have so far achieved productivity gains but have not yet developed novel capabilities.
-
Microsoft Launches New AI Chat Web App Template for .NET Development
Last week, Microsoft announced a new AI Chat Web App template, available in preview, designed to simplify AI development with .NET. This template is part of Microsoft's ongoing efforts to make AI more accessible, offering scaffolding and guidance in Visual Studio, Visual Studio Code, and the .NET CLI.
-
Google Introduces AI Co-Scientist System to Aid Scientific Research
Google has announced the development of an AI co-scientist system designed to assist scientists in generating hypotheses and research proposals. Built using Gemini 2.0, the system aims to accelerate scientific and biomedical discoveries by emulating the scientific method and fostering collaboration between humans and AI.
-
Google Enhances Data Privacy with Confidential Federated Analytics
Google has announced Confidential Federated Analytics (CFA), a technique designed to increase transparency in data processing while maintaining privacy. Building on federated analytics, CFA leverages confidential computing to ensure that only predefined and inspectable computations are performed on user data without exposing raw data to servers or engineers.
-
OpenAI Introduces Software Engineering Benchmark
OpenAI has introduced the SWE-Lancer benchmark, to evaluate the capabilities of advanced AI language models in real-world freelance software engineering tasks.
-
Google DeepMind Enhances AMIE for Long-Term Disease Management
Google DeepMind has extended the capabilities of its Articulate Medical Intelligence Explorer (AMIE) beyond diagnosis to support longitudinal disease management. The system is now designed to assist clinicians in monitoring disease progression, adjusting treatments, and adhering to clinical guidelines across multiple patient visits.
-
How Engineering Teams Are Tackling AI, Platform Engineering & DevEx: InfoQ Dev Summit Boston 2025
The InfoQ Dev Summit Boston 2025 conference (June 9-10) will bring together senior software practitioners to share proven strategies for integrating AI, scaling resilient architectures, and optimizing developer experience - three key areas that will define engineering success in the next 18 months.
-
instructlab.ai Uses Synthetic Data to Reduce Complexity of Fine-Tuning LLMs
InstructLab.ai implements the large-scale alignment for the chatbots concept(LAB), which intends to overcome the scalability challenges in the instruction-tuning phase of a large language model (LLM). Its approach leverages a synthetic data-based alignment tuning method for LLMs. Crafted taxonomies deliver the synthesization seeds for training data, reducing the need for human-annotated data.
-
Mistral AI Introduces Saba: Regional Language Model for Arabic and South Indian Language
Mistral AI has introduced Mistral Saba, a 24-billion-parameter language model designed to improve AI performance in Arabic and several Indian-origin languages, particularly South Indian languages like Tamil.
-
Google Enhances AlloyDB Vector Search with Inline Filtering and Enterprise Observability
Google enhanced its AlloyDB service with inline filtering and enterprise observability for vector search. This fully-managed PostgreSQL-compatible database now allows direct filtering during queries, offering improved speed and efficiency. Enhanced monitoring features provide deep insights, addressing scaling vector search operations challenges.
-
Hugging Face Publishes Guide on Efficient LLM Training across GPUs
Hugging Face has published the Ultra-Scale Playbook: Training LLMs on GPU Clusters, an open-source guide that provides a detailed exploration of the methodologies and technologies involved in training LLMs across GPU clusters.
-
Google Cloud Launches Gen AI Toolbox for Databases
Google Cloud has announced the public beta launch of Gen AI Toolbox for Databases, an open-source server developed in collaboration with LangChain. This new tool is designed to help developers seamlessly integrate production-grade, agent-based generative AI applications with databases while ensuring secure access, scalability, and observability.
-
IBM Granite 3.2 Brings New Vision Language Model, Chain of Thought Reasoning, Improved TimeSeries
IBM has introduced its new Granite 3.2 multi-modal and reasoning model. Granite 3.2 features experimental chain-of-thought reasoning capabilities that significantly improve its predecessor's performance, a new vision language model (VLM) outperforming larger models on several benchmarks, and smaller models for more efficient deployments.
-
Microsoft Releases BioEmu-1: a Deep Learning Model for Protein Structure Prediction
Microsoft Research has introduced BioEmu-1, a deep-learning model designed to predict the range of structural conformations that proteins can adopt. Unlike traditional methods that provide a single static structure, BioEmu-1 generates structural ensembles, offering a broader view of protein dynamics.
-
GitHub Copilot Extensions Integrate IDEs with External Services
Now generally available, GitHub Copilot Extensions allow developers to use natural language to query documentation, generate code, retrieve data, and execute actions on external services without leaving their IDEs. Besides using public extensions from companies like Docker, MongoDB, Sentry, and many more, developers can create their own extensions to work with internal libraries or APIs.