A
AI-CY25Q2- LLM Compression & Model Optimization
While there are incredibly capable open models available, inferencing and using them in production can be quite the challenge!How can you reduce hardware requirements and compute time needed to process data with LLMs?

AI-CY25Q2- LLM Compression & Model Optimization