InfoQ Homepage Machine Learning Content on InfoQ
-
AI for Food Image Generation in Production: How & Why
Iaroslav Amerkhanov discusses how his team at Delivery Hero leveraged GenAI to generate food images, detailing the architecture, optimization, and business impact.
-
10 Reasons Your Multi-Agent Workflows Fail and What You Can Do about It
Victor Dibia discusses multi-agent systems, detailing how to build them with AutoGen, common failure points, and strategic approaches for senior software developers and engineering leaders.
-
Maximizing Deep Learning Performance on CPUs using Modern Architectures
Bibek Bhattarai demystifies Intel AMX, explaining how this CPU architecture accelerates deep learning workloads via low-precision matrix multiplication and efficient data handling.
-
A Framework for Building Micro Metrics for LLM System Evaluation
Denys Linkov discusses critical lessons for senior developers and leaders on building robust LLM systems and actionable metrics that prevent production issues and drive business value.
-
Supporting Diverse ML Systems at Netflix
David Berg and Romain Cledat discuss Metaflow, Netflix's ML infrastructure for diverse use cases from computer vision to recommendations.
-
From "Simple" Fine-Tuning to Your Own Mixture of Expert Models Using Open-Source Models
Sebastiano Galazzo shares practical tips and mistakes in creating custom LLMs for cost-effective AI. Learn LoRA, merging, MoE & optimization.
-
How Green is Green: LLMs to Understand Climate Disclosure at Scale
Leo Browning explains the journey of developing a Retrieval Augmented Generation (RAG) system at a climate-focused startup.
-
LLM and Generative AI for Sensitive Data - Navigating Security, Responsibility, and Pitfalls in Highly Regulated Industries
Stefania Chaplin and Azhir Mahmood discuss responsible, secure, and explainable AI in regulated industries. Learn MLOps, legislation, and future trends.
-
Unleashing Llama's Potential: CPU-Based Fine-Tuning
Anil Rajput and Rema Hariharan detail CPU-based LLM (Llama) optimization strategies for performance and TCO reduction.
-
Navigating LLM Deployment: Tips, Tricks, and Techniques
Meryem Arik shares best practices for self-hosting LLMs in corporate environments, highlighting the importance of cost efficiency and performance optimization.
-
The Harsh Reality of Building a Real-Time ML Feature Platform
Ivan Burmistrov shares how ShareChat built their own Real-Time Feature Platform serving more than 1 billion features per second, and how they managed to make it cost efficient.
-
Recommender and Search Ranking Systems in Large Scale Real World Applications
Moumita Bhattacharya overviews the industry search and recommendations systems, goes into modeling choices, data requirements and infrastructural requirements, while highlighting challenges.