Reinventing data preparation on cloud to scale enterprise AINewsMike Murphy20 Sep 2024AIHybrid Cloud Infrastructure
A new chip architecture points to faster, more energy-efficient AINewsMike Murphy19 Oct 2023AIAI HardwareHybrid Cloud Infrastructure
Composable systems with OpenShiftTechnical noteChristian Pinto, Michele Gazzetti, and Michael Johnston28 Jul 2023Hybrid CloudHybrid Cloud Infrastructure
Why we built an AI supercomputer in the cloudResearchTalia Gershon, Seetharami Seelam, Jay Jubran, Eran Gampel, and Drew Thorstensen07 Feb 2023AIFoundation ModelsHybrid Cloud InfrastructureScaling AI
A Perspective on LLM Data Generation with Few-shot Examples: from Intent to Kubernetes ManifestAntonino AngiLiubov Nedoshivinaet al.2025ACL 2025
Towards Efficient Key-Value Cache Management for Prefix Prefilling in LLM InferenceYue ZhuHao Yuet al.2025CLOUD 2025
Mind the Memory Gap: Unveiling GPU Bottlenecks in Large-Batch LLM InferencePol G. RecasensFerran Agulloet al.2025CLOUD 2025
Rex: Closing the language-verifier gap with safe and usable kernel extensionsJinghao JiaRuowen Qinet al.2025USENIX ATC 2025