Listen "CausalML Book Ch10: Feature Engineering for Causal and Predictive Inference"
Episode Synopsis
This episode focuses on feature engineering, a technique that transforms complex data like text and images into numerical representations called embeddings for use in predictive and causal applications. It begins by explaining principal component analysis and autoencoders as methods for generating these embeddings. The text then specifically addresses text embeddings, detailing early methods like Word2Vec and later, more sophisticated sequence models such as ELMo and BERT, highlighting their architectural differences and advancements in capturing context. Finally, the chapter covers image embeddings through models like ResNet50 and illustrates their practical application in hedonic price modeling, demonstrating how these engineered features significantly improve prediction accuracy compared to traditional methods.DisclosureThe CausalML Book: Chernozhukov, V. & Hansen, C. & Kallus, N. & Spindler, M., & Syrgkanis, V. (2024): Applied Causal Inference Powered by ML and AI. CausalML-book.org; arXiv:2403.02467. Audio summary is generated by Google NotebookLM https://notebooklm.google/The episode art is generated by OpenAI ChatGPT
ZARZA We are Zarza, the prestigious firm behind major projects in information technology.