Skip to content

Mono repo for the PhD course AI for Business Research at Shanghai Jiao Tong University, Antai College of Economics and Management (Summer 2024)

Notifications You must be signed in to change notification settings

rphilipzhang/AI-PhD-Antai-Su2024

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

98 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Artificial Intelligence for Business Research @Antai (Summer 2024)

Teaching Team

  • Instructor: Renyu (Philip) Zhang, Associate Professor, Department of Decisions, Operations and Technology, CUHK Business School, [email protected]. Office Location: Antai 912.
  • Teaching Assistant: Zhenkang Peng, PhD Candidate, Management Science and Engineering, SJTU Antai College of Economics and Management, [email protected].

Basic Information

About

Welcome to the mono-repo of the PhD course AI for Business Research at SJTU Antai College of Economics and Management in Summer 2024. You may download the Syllabus of this course first. The purpose of this course is to learn the following:

  • Have a basic understanding of the fundamental concepts/methods in machine learning (ML) and artificial intelligence (AI) that are used (or potentially useful) in business research.
  • Understand how business researchers have utilized ML/AI and what managerial questions have been addressed by ML/AI in the recent decade.
  • Nurture a taste of what the state-of-the-art AI/ML technologies can do in the ML/AI community and, potentially, in your own research field.

We will meet at the time specified above in Xinshangyuan (新上院) S303. Please ask for my approval if you need to join us via the following Zoom Meeting link:

Most of the code in this course will be distributed through the Google CoLab cloud computing environment to avoid the incompatibility and version control issues on your local individual computer. On the other hand, you can always download the Jupyter Notebook from CoLab and run it your own computer.

  • The CoLab files of this course can be found at this folder.
  • The Google Sheet to sign up for groups and group tasks can be found here.

If you have any feedback on this course, please directly contact Philip at [email protected] and we will try our best to address it.

Brief Schedule

Session Date & Time Topic Key Words
1 May 8, 2:00pm-5:40pm AI/ML in a Nutshell Course Intro, ML Models, DL
2 May 10, 6:00pm-9:05pm Prediction and Traditional NLP Prediction in Biz Research, Pre-processing, $N$-gram, Naïve Bayes
3 May 11, 6:00pm-9:05pm NLP (II): Deep Learning Word2Vec, RNN, Seq2Seq, Attention, Transformer
4 May 12, 2:00pm-5:40pm NLP (III): LLM BERT, GPT, Emergent Abilities, Chain-of-Thought, In-context Learning, GenAI in Business Research
5 May 15, 2:00pm-5:40pm CV (I): Image Classification CNN, AlexNet, ResNet, ViT
6 May 16, 6:00pm-9:05pm CV (II): Image Segmentation and Video Analysis R-CNN, YOLO, 3D-CNN
7 May 17, 6:00pm-9:05pm Unsupervised Learning (I): Clustering & Topic Modeling GMM, EM Algorithm, LDA
8 May 18, 2:00pm-5:40pm Unsupervised Learning (II): Diffusion Models VAE, DDPM, LDM, DiT

Important Dates

Date Time Event Note
May 8 11:59pm Group Sign-Ups Each group has at most two students.
May 9 10:00am-noon Online Python Tutorial Given by Zhenkang Peng, Python Tutorial CoLab, Zoom Link
May 10 10:00am-noon Online PyTorch Tutorial Given by Zhenkang Peng, PyTorch Tutorial CoLab, Zoom Link
May 13 11:59pm Project 1 Choice Due
May 19 11:59pm Project 2 Choice Due
June 2 11:59pm All Problem Sets and Projects Due Project Report page limit: 5 each

Replication Projects

Choose one from the following to replicate for Project 1 (1st-Half):

Choose one from the following to replicate for Project 2 (2nd-Half):

Useful Resources

Find more on the Syllabus.

Detailed Schedule

The following schedule is tentative and subject to changes.

First-Half (May/08/2024-May/12/2024)

Session 1. Artificial Intelligence and Machine Learning in a Nutshell (May/08/2024, 2:00pm-5:40pm)

Session 2. Predictions in Business Research and Traditonal NLP (May/10/2024, 6:00pm-9:05pm)

  • Keywords: Prediction Problems in Business Research, Pre-processing and Word Representations in Traditional Natural Language Processing, N-Gram, Naïve Bayes, Traditional NLP Applied to Business/Econ Research
  • Slides: Prediction Problems in Business Research, NLP(I): Traditional NLP
  • CoLab Notebook Demos: NLP Pre-processing, N-Gram, Naïve Bayes
  • References:
    • Kleinberg, Jon, Jens Ludwig, Sendhil Mullainathan, and Ziad Obermeyer. 2015. Prediction policy problems. American Economic Review 105(5): 491-495.
    • Mullainathan, Sendhil, and Jann Spiess. 2017. Machine learning: an applied econometric approach. Journal of Economic Perspectives 31(2): 87-106.
    • Kleinberg, Jon, Himabindu Lakkaraju, Jure Leskovec, Jens Ludwig, and Sendhil Mullainathan. 2018. Human decisions and machine predictions. Quarterly Journal of Economics 133(1): 237-293.
    • Bajari, Patrick, Denis Nekipelov, Stephen P. Ryan, and Miaoyu Yang. 2015. Machine learning methods for demand estimation. American Economic Review, 105(5): 481-485.
    • Farias, Vivek F., and Andrew A. Li. 2019. Learning preferences with side information. Management Science 65(7): 3131-3149.
    • Cui, Ruomeng, Santiago Gallino, Antonio Moreno, and Dennis J. Zhang. 2018. The operational value of social media information. Production and Operations Management, 27(10): 1749-1769.
    • Gentzkow, Matthew, Bryan Kelly, and Matt Taddy. 2019. Text as data. Journal of Economic Literature, 57(3): 535-574.
    • Hansen, Stephen, Michael McMahon, and Andrea Prat. 2018. Transparency and deliberation within the FOMC: A computational linguistics approach. Quarterly Journal of Economics, 133(2): 801-870.
    • Tetlock, Paul. 2007. Giving content to investor sentiment: The role of media in the stock market. Journal of Finance, 62(3): 1139-1168.
    • Baker, Scott, Nicholas Bloom, and Steven Davis, 2016. Measuring economic policy uncertainty. Quarterly Journal of Economics, 131(4): 1593-1636.
    • Gentzkow, Matthew, and Jesse Shapiro. 2010. What drives media slant? Evidence from US daily newspapers. Econometrica, 78(1): 35-71.
    • Chapter 2, 12, & 13 Introduction to Information Retrieval, 2008, Cambridge University Press, by Christopher D. Manning, Prabhakar Raghavan and Hinrich Schutze. Link to the Book.
    • Chapter 2, 3, & 4, Speech and Language Processing (3rd ed. draft), 2023, by Dan Jurafsky and James H. Martin. Link to the Book.
    • Natural Language Tool Kit (NLTK) Documentation
    • Handwritten Notes (2)

Session 3. Deep-Learning-Based NLP: From Word2Vec to Transformers (May/11/2024, 6:00pm-9:05pm)

  • Keywords: Word2Vec: Continuous Bag of Words and Skip-Gram, RNN, LSTM, Seq2Seq, Attention Mechanism, Transformer
  • Slides: NLP(II): Deep Learning Based NLP
  • CoLab Notebook Demos: Word2Vec: CBOW, Word2Vec: Skip-Gram, RNN & LSTM, Attention Mechanism, Transformer
  • Homework: Problem Set 2 - Word2Vec & LSTM for Sentiment Analysis
  • References:
    • Ash, Elliot, and Stephen Hansen. 2023. Text algorithms in economics. Annual Review of Economics, 15: 659-688. Associated GitHub with Code Demonstrations.
    • Timoshenko, Artem, and John R. Hauser. 2019. Identifying customer needs from user-generated content. Marketing Science, 38(1): 1-20.
    • Mikolov, Tomas, Kai Chen, Greg Corrado, and Jeff Dean. 2013. Efficient estimation of word representations in vector space. ArXiv Preprint, arXiv:1301.3781.
    • Mikolov, Tomas, Ilya Sutskever, Kai Chen, Greg Corrado, and Jeff Dean. 2013. Distributed representations of words and phrases and their compositionality. Advances in Neural Information Processing Systems (NeurIPS) 26.
    • Li, Kai, Feng Mai, Rui Shen, and Xinyan Yan. 2021. Measuring corporate culture using machine learning. Review of Financial Studies, 34(7): 3265-3315.
    • Chen, Fanglin, Xiao Liu, Davide Proserpio, and Isamar Troncoso. 2022. Product2Vec: Leveraging representation learning to model consumer product choice in large assortments. Available at SSRN 3519358.
    • Qi, Meng, Yuanyuan Shi, Yongzhi Qi, Chenxin Ma, Rong Yuan, Di Wu, Zuo-Jun (Max) Shen. 2023. A Practical End-to-End Inventory Management Model with Deep Learning. Management Science, 69(2): 759-773.
    • Sarzynska-Wawer, Justyna, Aleksander Wawer, Aleksandra Pawlak, Julia Szymanowska, Izabela Stefaniak, Michal Jarkiewicz, and Lukasz Okruszek. 2021. Detecting formal thought disorder by deep contextualized word representations. Psychiatry Research, 304, 114135.
    • Sutskever, Ilya, Oriol Vinyals, and Quoc V. Le. 2014. Sequence to sequence learning with neural networks. Advances in neural information processing systems, 27.
    • Bahdanau, Dzmitry, Kyunghyun Cho, and Yoshua Bengio. 2015. Neural machine translation by jointly learning to align and translate. ICLR
    • Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., ... and Polosukhin, I. (2017). Attention is all you need. Advances in neural information processing systems, 30.
    • Parts 1, 2, 5, 6, and 8 Lecture Notes and Slides for CS224n: Natural Language Processing with Deep Learning, by Christopher D. Manning, Diyi Yang, and Tatsunori Hashimoto, https://web.stanford.edu/class/cs224n/.
    • Chapters 9, 10, and 11 Dive into Deep Learning (2nd Edition), 2023, by Aston Zhang, Zack Lipton, Mu Li, and Alex J. Smola, https://d2l.ai/.
    • Word Embeddings Trained on Google News Corpus
    • RNN and LSTM Visualizations
    • PyTorch's Tutorial of Seq2Seq for Machine Translation
    • Illustrated Transformer
    • Transformer from Scratch, with the Code on GitHub
    • Handwritten Notes (3)

Session 4. Deep-Learning-Based NLP: Pretraining and LLM (May/12/2024, 2:00pm-5:40pm)

  • Keywords: BERT (Bidirectional Encoder Representations from Transformers), GPT (Generative Pretrained Transformers), LLM (Large Language Model)
  • Slides: Deep Learning Computation, Pretraining and LLM
  • CoLab Notebook Demos: Crafting Intelligence: The Art of Deep Learning Modeling, BERT API @ Hugging Face
  • References:
    • Devlin, Jacob, Ming-Wei Chang, Kenton Lee, Kristina Toutanova. 2018. BERT: Pre-training of deep bidirectional transformers for language understanding. ArXiv preprint arXiv:1810.04805. GitHub Repo
    • Radford, Alec, Karthik Narasimhan, Tim Salimans, and Ilya Sutskever. 2018. Improving language understanding by generative pre-training, (GPT-1) PDF link, GitHub Repo
    • Radford, Alec, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever. 2019. Language models are unsupervised multitask learners. OpenAI blog, 1(8), 9. (GPT-2) PDF Link, GitHub Repo
    • Brown, Tom, et al. 2020. Language models are few-shot learners. Advances in neural information processing systems, 33, 1877-1901. (GPT-3) GitHub Repo
    • Huang, Allen H., Hui Wang, and Yi Yang. 2023. FinBERT: A large language model for extracting information from financial text. Contemporary Accounting Research, 40(2): 806-841. GitHub Repo
    • Wei, Jason, et al. 2021. Finetuned language models are zero-shot learners. ArXiv preprint arXiv:2109.01652, link to the paper.
    • Wei, Jason, et al. 2022. Emergent abilities of large language models. ArXiv preprint arXiv:2206.07682, link to the paper.
    • Ouyang, Long, et al. 2022. Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35, 27730-27744.
    • Wei, Jason, et al. 2022. Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems, 35, 24824-24837.
    • Kaplan, Jared. 2020. Scaling laws for neural language models. ArXiv preprint arXiv:2001.08361, link to the paper.
    • Hoffmann, Jordan, et al. 2022. Training compute-optimal large language models. ArXiv preprint arXiv:2203.15556, link to the paper.
    • Shinn, Noah, et al. 2023. Reflexion: Language agents with verbal reinforcement learning. ArXiv preprint arXiv:2303.11366, link to the paper.
    • Reisenbichler, Martin, Thomas Reutterer, David A. Schweidel, and Daniel Dan. 2022. Frontiers: Supporting content marketing with natural language generation. Marketing Science, 41(3): 441-452.
    • Romera-Paredes, B., Barekatain, M., Novikov, A. et al. 2023. Mathematical discoveries from program search with large language models. Nature, link to the paper.
    • Parts 9 & 10, Lecture Notes and Slides for CS 224N: Natural Language Processing with Deep Learning, by Christopher D. Manning, Diyi Yang, and Tatsunori Hashimoto. Link to CS 224N
    • COS 597G: Understanding Large Language Models, by Danqi Chen. Link to COS 597G
    • A Visual Guide to BERT, How GPT-3 Works
    • Andrej Karpathy's 1-hour Talk on LLM
    • CS224n, Hugging Face Tutorial
    • Andrej Karpathy's Talk to Build GPT from Scratch in Code Spelled Out
    • Handwritten Notes (4)

Second-Half (May/15/2024-May/18/2024)

Session 5. Deep-Learning-Based CV: Image Classification (May/15/2024, 2:00pm-5:40pm)

  • Keywords: Large Language Models Applications, Convolution Neural Nets (CNN), LeNet, AlexNet, VGG, ResNet, ViT
  • Slides: Image Classification
  • CoLab Notebook Demos: CNN, LeNet, & AlexNet, VGG, ResNet, ViT
  • Homework: Problem Set 3 - AlexNet and ResNet, due at 11:59pm, June 2, Sunday.
  • References:
    • Krizhevsky, Alex, Ilya Sutskever, and Geoffrey E. Hinton. 2012. Imagenet classification with deep convolutional neural networks. Advances in Neural Information Processing Systems, 25.
    • Simonyan, K., & Zisserman, A. (2014). Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556.
    • He, Kaiming, Xiangyu Zhang, Shaoqing Ren and Jian Sun. 2016. Deep residual learning for image recognition. Proceedings of the IEEE conference on computer vision and pattern recognition, 770-778.
    • Hu, J., Shen, L., & Sun, G. (2018). Squeeze-and-excitation networks. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 7132-7141).
    • Dosovitskiy, Alexey, et al. 2020. An image is worth 16x16 words: Transformers for image recognition at scale. ArXiv preprint, arXiv:2010.11929, link to the paper, link to the GitHub repo.
    • Jean, Neal, Marshall Burke, Michael Xie, Matthew W. Davis, David B. Lobell, and Stefand Ermon. 2016. Combining satellite imagery and machine learning to predict poverty. Science, 353(6301), 790-794.
    • Zhang, Mengxia and Lan Luo. 2023. Can consumer-posted photos serve as a leading indicator of restaurant survival? Evidence from Yelp. Management Science 69(1): 25-50.
    • Zhang, S., Lee, D., Singh, P. V., & Srinivasan, K. (2022). What makes a good image? Airbnb demand analytics leveraging interpretable image features. Management Science, 68(8), 5644-5666.
    • Course Notes (Lectures 5 & 6) for CS231n: Deep Learning for Computer Vision, by Fei-Fei Li, Ruohan Gao, & Yunzhu Li. Link to CS231n.
    • Chapters 7 and 8, Dive into Deep Learning (2nd Edition), 2023, by Aston Zhang, Zack Lipton, Mu Li, and Alex J. Smola. Link to the book.
    • Fine-Tune ViT for Image Classification with Hugging Face 🤗 Transformers
    • Hugging Face 🤗 ViT CoLab Tutorial

Session 6. Deep-Learning-Based CV (II): Object Detection & Video Analysis (May/16/2024, 6:00pm-9:05pm)

  • Keywords: Image Processing Applications, Localization, R-CNNs, YOLOs, Semantic Segmentation, 3D CNN, Video Analysis Applications
  • Slides: Object Detection and Video Analysis
  • CoLab Notebook Demos: Data Augmentation, Faster R-CNN & YOLO v5
  • References:
    • Girshick, R., Donahue, J., Darrell, T. and Malik, J., 2014. Rich feature hierarchies for accurate object detection and semantic segmentation. Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 580-587).
    • Girshick, R., Donahue, J., Darrell, T., & Malik, J. (2014). Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 580-587).
    • Girshick, R. (2015). Fast r-cnn. In Proceedings of the IEEE international conference on computer vision (pp. 1440-1448).
    • Ren, S., He, K., Girshick, R., & Sun, J. (2015). Faster r-cnn: Towards real-time object detection with region proposal networks. Advances in neural information processing systems, 28.
    • Redmon, Joseph, Santosh Divvala, Ross Girshick, and Ali Farhadi. 2016. You only look once: Unified, real-time object detection. Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 779-788).
    • Long, J., Shelhamer, E., & Darrell, T. (2015). Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 3431-3440).
    • Karpathy, A., Toderici, G., Shetty, S., Leung, T., Sukthankar, R. and Fei-Fei, L., 2014. Large-scale video classification with convolutional neural networks. Proceedings of the IEEE conference on Computer Vision and Pattern Recognition (pp. 1725-1732).
    • Glaeser, Edward L., Scott D. Kominers, Michael Luca, and Nikhil Naik. 2018. Big data and big cities: The promises and limitations of improved measures of urban life. Economic Inquiry, 56(1): 114-137.
    • Zhang, S., Xu, K. and Srinivasan, K., 2023. Frontiers: Unmasking Social Compliance Behavior During the Pandemic. Marketing Science, 42(3), pp.440-450.
    • Liu, Liu, Dzyabura, Daria, Mizik, Natalie. 2020. Visual listening in: Extracting brand image portrayed on social media. Marketing Science, 39(4): 669-686. Link to the Paper
    • Yang, Jeremy, Juanjuan Zhang, and Yuhan Zhang. 2023. Engagement that sells: Influencer video advertising on TikTok. Available at SSRN Link to the Paper
    • Course Notes (Lectures 10 & 11) for CS231n: Deep Learning for Computer Vision, by Fei-Fei Li, Ruohan Gao, & Yunzhu Li. Link to CS231n.
    • Chapter 14, Dive into Deep Learning (2nd Edition), 2023, by Aston Zhang, Zack Lipton, Mu Li, and Alex J. Smola. Link to the book.
    • Handwritten Notes

Session 7. Unsupervised Learning: Clustering, Topic Modeling & VAE (May/17/2024, 6:00pm-9:05pm)

  • Keywords: K-Means, Gaussian Mixture Models, EM-Algorithm, Latent Dirichlet Allocation, Variational Auto-Encoder
  • Slides: Clustering, Topic Modeling & VAE
  • CoLab Notebook Demos: K-Means, LDA, VAE
  • Homework: Problem Set 4 - Unsupervised Learning (EM & LDA), due at 11:59pm, June 2, Sunday.
  • References:
    • Blei, David M., Ng, Andrew Y., and Jordan, Michael I. 2003. Latent Dirichlet allocation. Journal of Machine Learning Research, 3(Jan): 993-1022.
    • Kingma, D.P. and Welling, M., 2013. Auto-encoding Variational Bayes. arXiv preprint arXiv:1312.6114.
    • Kingma, D.P. and Welling, M., 2019. An introduction to variational autoencoders. Foundations and Trends® in Machine Learning, 12(4), pp.307-392.
    • Bandiera, O., Prat, A., Hansen, S., & Sadun, R. 2020. CEO behavior and firm performance. Journal of Political Economy, 128(4), 1325-1369.
    • Liu, Jia and Olivier Toubia. 2018. A semantic approach for estimating consumer content preferences from online search queries. Marketing Science, 37(6): 930-952.
    • Tirunillai, S., & Tellis, G. J. (2014). Mining marketing meaning from online chatter: Strategic brand analysis of big data using latent dirichlet allocation. Journal of marketing research, 51(4), 463-479.
    • Mueller, Hannes, and Christopher Rauh. 2018. Reading between the lines: Prediction of political violence using newspaper text. American Political Science Review, 112(2): 358-375.
    • Tian, Z., Dew, R. and Iyengar, R., 2023. Mega or Micro? Influencer Selection Using Follower Elasticity. Journal of Marketing Research.
    • Chapters 8.5 and 14, The Elements of Statistical Learning (2nd Edition), 2009, by Trevor Hastie, Robert Tibshirani, Jerome Friedman, Link to Book.
    • Course Notes (Lectures 1 & 4) for CS294-158-SP24: Deep Unsupervised Learning, taught by Pieter Abbeel, Wilson Yan, Kevin Frans, Philipp Wu. Link to CS294-158-SP24.
    • Handwritten Notes

Session 8. Unsupervised Learning: Diffusion Models (May/18/2024, 2:00pm-5:40pm)

  • Keywords: Denoised Diffusion Probabilistic Models, Latent Diffusion Models, CLIP, Imagen, Diffusion Transformers, Course Summary
  • Slides: Diffusion Models, Course Summary
  • CoLab Notebook Demos: DDPM, DiT
  • References:
    • Kingma, D.P. and Welling, M., 2013. Auto-encoding Variational Bayes. arXiv preprint arXiv:1312.6114.
    • Kingma, D.P. and Welling, M., 2019. An introduction to variational autoencoders. Foundations and Trends® in Machine Learning, 12(4), pp.307-392.
    • Ho, J., Jain, A. and Abbeel, P., 2020. Denoising diffusion probabilistic models. Advances in neural information processing systems, 33, 6840-6851.
    • Chan, S.H., 2024. Tutorial on Diffusion Models for Imaging and Vision. arXiv preprint arXiv:2403.18103.
    • Peebles, W. and Xie, S., 2023. Scalable diffusion models with transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 4195-4205. Link to GitHub Repo.
    • Ludwig, J. and Mullainathan, S., 2024. Machine learning as a tool for hypothesis generation. Quarterly Journal of Economics, 139(2), 751-827.
    • Burnap, A., Hauser, J.R. and Timoshenko, A., 2023. Product aesthetic design: A machine learning augmentation. Marketing Science, 42(6), 1029-1056.
    • Dew, R., Ansari, A., & Toubia, O. (2022). Letting logos speak: Leveraging multiview representation learning for data-driven branding and logo design. Marketing Science, 41(2), 401-425.
    • Zhou, E., & Lee, D. (2024). Generative artificial intelligence, human creativity, and art. PNAS nexus, 3(3), pgae052.
    • Course Notes (Lecture 6) for CS294-158-SP24: Deep Unsupervised Learning, taught by Pieter Abbeel, Wilson Yan, Kevin Frans, Philipp Wu. Link to CS294-158-SP24.
    • CVPR 2022 Tutorial: Denoising Diffusion-based Generative Modeling: Foundations and Applications, by Karsten Kreis, Ruiqi Gao, and Arash Vahdat Link to the Tutorial
    • Lilian Weng (OpenAI)'s Blog on Diffusion Models
    • Lilian Weng (OpenAI)'s Blog on Diffusion Models for Video Generation
    • Hugging Face Diffusers 🤗 Library

About

Mono repo for the PhD course AI for Business Research at Shanghai Jiao Tong University, Antai College of Economics and Management (Summer 2024)

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published