Unlock efficient model deployment: Simplified Inference Operator setup on Amazon SageMaker HyperPod
Simplifying the deployment and scheduling of machine learning inference workloads across multiple instances and instance types on Amazon SageMaker HyperPod.
How Meta Used AI to Map Tribal Knowledge in Large-Scale Data Pipelines
AI coding assistants were ineffective at making useful edits in large-scale data pipelines because they lacked sufficient understanding of complex, multi-repository codebases spanning multiple languages and thousands of files.
Automate safety monitoring with computer vision and generative AI
Detecting safety hazards in real-time across hundreds of distributed operational sites using video feeds while maintaining low latency and managing the computational complexity of processing multiple camera streams.
How Aigen transformed agricultural robotics for sustainable farming with Amazon SageMaker AI
Aigen needed to scale machine learning pipelines across hundreds of distributed edge solar robots while managing data labeling and model training challenges in agricultural robotics.
Cloudflare Client-Side Security: smarter detection, now open to everyone
Detecting sophisticated client-side security threats like zero-day exploits while minimizing false positives in real-time across millions of requests.
AI Helping Build Better AI: How Agents Accelerate Model Experi...
Training and evaluating AI models is resource-intensive, requiring significant human effort to generate quality training data and assess model outputs.
Announcing Our LinkedIn-Cornell 2024 Grant Recipients
Advancing AI research requires collaboration between industry and academia, but funding and partnership models need structured programs.
Career stories: The math-music connection in data science
Data science teams need diverse skill sets that blend mathematical rigor with creative problem-solving to build effective ML systems.
Engineering the next generation of LinkedIn’s Feed
LinkedIn's Feed needed to evolve to handle increasing content diversity, real-time ranking signals, and personalization at massive scale.
Scaling LLM-Based ranking systems with SGLang at LinkedIn
LinkedIn's LLM-based ranking systems faced latency and throughput challenges when serving personalized results at scale.
The LinkedIn Generative AI Application Tech Stack: Personaliza...
Building personalized generative AI features at LinkedIn's scale required a robust and reliable application infrastructure that could serve millions of users.
AI for American-Produced Cement and Concrete
Designing high-quality, sustainable concrete mixes that are produced in the United States while optimizing for performance characteristics.
KernelEvolve: How Meta’s Ranking Engineer Agent Optimizes AI Infrastructure
Meta needed to automatically optimize low-level infrastructure and kernel-level parameters for AI ranking models to improve performance without manual tuning.
Meta Adaptive Ranking Model: Bending the Inference Scaling Curve to Serve LLM-Scale Models for Ads
Meta needed to scale their ads ranking models to LLM-scale complexity and size while maintaining inference latency requirements for real-time ad serving.