EP2: ARCEE.AI small language models, open source and cost efficient AI | AWS for AI podcast

20/05/2025 1h 3min Episodio 2
EP2: ARCEE.AI small language models, open source and cost efficient AI | AWS for AI podcast

Listen "EP2: ARCEE.AI small language models, open source and cost efficient AI | AWS for AI podcast"

Episode Synopsis


Join us for an enlightening conversation with Julien Simon, VP and Chief Evangelist at ARCEE.AI , as he shares deep insights on building practical and cost-efficient AI solutions. From his extensive experience at AWS, Hugging Face, and now ARCEE.AI, Julien discusses why "small is beautiful" when it comes to language models, revealing how 10B parameter models can now match the performance of much larger 72B models from just months ago. Learn about innovative techniques like model merging, the importance of proper infrastructure choices, and practical advice for organizations starting their AI journey. This episode covers critical topics including:Why small language models are the future of enterprise AIHow to optimize costs while maintaining performanceThe role of CPU vs GPU inferenceEssential architecture considerations for AI workloadsBest practices for building production-ready AI systemsWhether you're a startup, enterprise, or public sector organization, this episode offers invaluable guidance on building scalable, efficient, and practical AI solutions in today's rapidly evolving landscape.Julien Simon Youtube channel : https://www.youtube.com/@juliensimonfrto learn more about ARCEE.AI : https://www.arcee.ai/

More episodes of the podcast AWS For AI