
AI-PhD-S25
Mono-repo for the PhD course AI for Business Research (DOTE 6635, S25)
Stars: 52

AI-PhD-S25 is a mono-repo for the DOTE 6635 course on AI for Business Research at CUHK Business School. The course aims to provide a fundamental understanding of ML/AI concepts and methods relevant to business research, explore applications of ML/AI in business research, and discover cutting-edge AI/ML technologies. The course resources include Google CoLab for code distribution, Jupyter Notebooks, Google Sheets for group tasks, Overleaf template for lecture notes, replication projects, and access to HPC Server compute resource. The course covers topics like AI/ML in business research, deep learning basics, attention mechanisms, transformer models, LLM pretraining, posttraining, causal inference fundamentals, and more.
README:
Role | Name & Contact |
---|---|
Instructor |
Renyu (Philip) Zhang Associate Professor, Department of Decisions, Operations and Technology, CUHK Business School π§ [email protected] π @911 Cheng Yu Tung Building |
Teaching Assistant | Leo Cao Full-time TA, Department of Decisions, Operations and Technology, CUHK Business School π§ [email protected] |
Tutorial Instructor |
Xinyu Li PhD Candidate (Management Information Systems), CUHK Business School π§ [email protected] |
- π Website: https://github.com/rphilipzhang/AI-PhD-S25
- β° Time: Tuesday, 12:30pm-3:15pm (Jan 14 - Apr 15, 2025)
- Excluding: Jan 28 (Chinese New Year) and Mar 4 (Final Project Discussion)
- π Location: Wu Ho Man Yuen Building (WMY) 504
Welcome to the mono-repo of DOTE 6635: AI for Business Research at CUHK Business School!
- π§ Gain fundamental understanding of ML/AI concepts and methods relevant to business research.
- π‘ Explore applications of ML/AI in business research over the past decade.
- π Discover and nuture the taste of cutting-edge AI/ML technologies and their potential in your research domain.
Need to join remotely? Use our Zoom link (please seek approval from Philip):
- π₯ Join Meeting
- Meeting ID: 918 6344 5131
- Passcode: 459761
Most of the code in this course will be distributed through the Google CoLab cloud computing environment to avoid the incompatibility and version control issues on your local individual computer. On the other hand, you can always download the Jupyter Notebook from CoLab and run it your own computer.
- π The Literature References discussed in the slides can be found on this document.
- π The CoLab files of this course can be found at this folder.
- π The Google Sheet to sign up for groups and group tasks can be found here.
- π The overleaf template for scribing the lecture notes of this course can be found here.
- π¬ The replication projects can be found here.
- π₯οΈ The HPC Server compute resource of the CUHK DOT Department can be found here.
If you have any feedback on this course, please directly contact Philip at [email protected] and we will try our best to address it.
- ποΈ GitHub Repos: Spring 2024@CUHK, Summer 2024@SJTU Antai
- π₯ Video Recordings (You need to apply for access): Spring 2024@CUHK, Summer 2024@SJTU Antai
- π Scribed Notes: Spring 2024@CUHK
Subject to modifications. All classes start at 12:30pm and end at 3:15pm.
Session | Date | Topic | Key Words |
---|---|---|---|
1 | 1.14 | AI/ML in a Nutshell | Course Intro, Prediction in Biz Research |
2 | 1.21 | Intro to DL | ML Model Evaluations, DL Intro, Neural Nets |
3 | 2.04 | LLM (I) | DL Computations, Attention Mechanism |
4 | 2.11 | LLM (II) | Transformer, ViT, DiT |
5 | 2.18 | LLM (III) | BERT, GPT |
6 | 2.25 | LLM (IV) | LLM Pre-training, DL Computations |
7 | 3.04 | LLM (V) | Post-training, Fine-tuning, RLHF, Test-Time Scaling, Inference, Quantization |
8 | 3.11 | LLM (VI) | Agentic AI, AI as Human Simulators, Applications in Business Research |
9 | 3.18 | Causal (I) | Causal Inference Intro, RCT, IPW, AIPW |
10 | 3.25 | Causal (II) | Double Machine Learning, Neyman Orthogonality |
11 | 4.01 | Causal (III) | ML-Powered Causal Inference, Causal Trees and Forests |
12 | 4.08 | Causal (IV) | (Off-)Policy Evaluation, Policy Learning |
13 | 4.15 | Causal (V) | LLM x Causal Inference and Course Wrap-up |
All problem sets are due at 12:30pm right before class.
Date | Time | Event | Note |
---|---|---|---|
1.15 | 11:59pm | Group Sign-Ups | Each group has at most two students. |
1.17 | 7:00pm-9:00pm | Python Tutorial | Given by Xinyu Li, Python Tutorial CoLab |
1.24 | 7:00pm-9:00pm | PyTorch and DOT HPC Server Tutorial | Given by Xinyu Li, PyTorch Tutorial CoLab |
3.04 | 9:00am-6:00pm | Final Project Discussion | Please schedule a meeting with Philip. |
3.11 | 12:30pm | Final Project Proposal | 1-page maximum |
4.30 | 11:59pm | Scribed Lecture Notes | Overleaf link |
5.11 | 11:59pm | Project Paper, Slides, and Code | Paper page limit: 10 |
Find more on the Syllabus and the literature references discussed in the slides.
-
π Books:
-
π Courses:
- Foundations: ML Intro by Andrew Ng, DL Intro by Andrew Ng, Generative AI by Andrew Ng, Introduction to Causal Inference by Brady Neal
- Advanced Technologies: NLP (CS224N) by Chris Manning, CV (CS231N) by Fei-Fei Li, Deep Unsupervised Learning by Pieter Abbeel, DLR by Sergey Levine, DL Theory by Matus Telgarsky, LLM by Danqi Chen, LLM from Scratch (CS336) by Persy Liang, Efficient Deep Learning Computing by Song Han, Deep Generative Models by Kaiming He, LLM Agents by Dawn Song, Advanced LLM Agents by Dawn Song, Data, Learning, and Algorithms by Tengyuan Liang, Hands-on DRL by Weinan Zhang, Jian Shen, and Yu Yong (in Chinese), Understanding LLM: Foundations and Safety by Dawn Song
- Biz/Econ Applications of AI: Machine Learning and Big Data by Melissa Dell and Matthew Harding, Digital Economics and the Economics of AI by Martin Beraja, Chiara Farronato, Avi Goldfarb, and Catherine Tucker, Generative AI and Causal Inference with Texts, NLP for Computational Social Science by Diyi Yang
-
π‘ Tutorials and Blogs:
- GitHub of Andrej Karpathy, Blog of Lilian Weng, Double Machine Learning Package Documentation, Causality and Deep Learning (ICML 2022 Tutorial), Causal Inference and Machine Learning (KDD 2021 Tutorial), Online Causal Inference Seminar, Training a Chinese LLM from Scratch (in Chinese), Physics of Language Models (ICML 2024 Tutorial), Counterfactual Learning and Evaluation for Recommender Systems: Foundations, Implementations, and Recent Advances (RecSys 2021 Tutorial), Language Agents: Foundations, Prospects, and Risks (EMNLP 2024 Tutorial), GitHub Repo: Upgrading Cursor to Devin
The following schedule is tentative and subject to changes.
- π Keywords: Course Introduction, Prediction in Biz Research, Basic ML Models
- π Slides: Course Intro, Prediction, ML Intro
- π» CoLab Notebook Demos: Bootstrap, k-Nearest Neighbors, Decision Tree, Random Forest, Gradient Boosting Tree
- βοΈ Homework: Problem Set 1 - Housing Price Prediction, due at 12:30pm, Feb/4/2025
- π Online Python Tutorial: Python Tutorial CoLab, 7:00pm-9:00pm, Jan/17/2025 (Friday), given by Xinyu Li, [email protected]. Zoom Link, Meeting ID: 939 4486 4920, Passcode: 456911
- π References:
- The Elements of Statistical Learning (2nd Edition), 2009, by Trevor Hastie, Robert Tibshirani, Jerome Friedman, link to ESL.
- Probabilistic Machine Learning: An Introduction, 2022, by Kevin Murphy, link to PML.
- Mullainathan, Sendhil, and Jann Spiess. 2017. Machine learning: an applied econometric approach. Journal of Economic Perspectives 31(2): 87-106.
- Athey, Susan, and Guido W. Imbens. 2019. Machine learning methods that economists should know about. Annual Review of Economics 11: 685-725.
- Kleinberg, Jon, Jens Ludwig, Sendhil Mullainathan, and Ziad Obermeyer. 2015. Prediction policy problems. American Economic Review 105(5): 491-495.
- Hofman, Jake M., et al. 2021. Integrating explanation and prediction in computational social science. Nature 595.7866: 181-188.
- Bastani, Hamsa, Dennis Zhang, and Heng Zhang. 2022. Applied machine learning in operations management. Innovative Technology at the Interface of Finance and Operations. Springer: 189-222.
- Kelly, Brian, and Dacheng Xiu. 2023. Financial machine learning, SSRN, link to the paper.
- The Bitter Lesson, by Rich Sutton, which develops so far the most critical insight of AI: "The biggest lesson that can be read from 70 years of AI research is that general methods that leverage computation are ultimately the most effective, and by a large margin."
- Chatpers 1 & 3.2, Scribed Notes of Spring 2024 Course Offering.
- π Keywords: Bias-Variance Trade-off, Cross Validation, Bootstrap, Neural Nets, Computational Issues of Deep Learning
- π Slides: ML Intro, DL Intro
- π» CoLab Notebook Demos: Gradient Descent, Chain Rule, He Innitialization
- βοΈ Homework: Problem Set 2: Implementing Neural Nets, due at 12:30pm, Feb/11/2025
- π Online PyTorch and DOT HPC Server Tutorial: PyTorch Tutorial CoLab, 7:00pm-9:00pm, Jan/24/2025 (Friday), given by Xinyu Li, [email protected]. Zoom Link, Meeting ID: 939 4486 4920, Passcode: 456911
- π References:
- Deep Learning, 2016, by Ian Goodfellow, Yoshua Bengio and Aaron Courville, link to DL.
- Dive into Deep Learning (2nd Edition), 2023, by Aston Zhang, Zack Lipton, Mu Li, and Alex J. Smola, link to d2dl.
- Probabilistic Machine Learning: Advanced Topics, 2023, by Kevin Murphy, link to PML2.
- Deep Learning with PyTorch, 2020, by Eli Stevens, Luca Antiga, and Thomas Viehmann.
- Dell, Mellissa. 2024. Deep learning for economists. Journal of Economic Literature, forthcoming, link to the paper.
- Davies, A., VeliΔkoviΔ, P., Buesing, L., Blackwell, S., Zheng, D., TomaΕ‘ev, N., Tanburn, R., Battaglia, P., Blundell, C., JuhΓ‘sz, A. and Lackenby, M., 2021. Advancing mathematics by guiding human intuition with AI. Nature, 600(7887), pp.70-74.
- Ye, Z., Zhang, Z., Zhang, D., Zhang, H. and Zhang, R.P., 2023. Deep-Learning-Based Causal Inference for Large-Scale Combinatorial Experiments: Theory and Empirical Evidence. Available at SSRN 4375327, link to the paper.
- Luyang Chen, Markus Pelger, Jason Zhu (2023) Deep Learning in Asset Pricing. Management Science 70(2):714-750.
- Wang, Z., Gao, R. and Li, S. 2024. Neural-Network Mixed Logit Choice Model: Statistical and Optimality Guarantees. Working paper.
- Why Does Adam Work So Well? (in Chinese), Overview of gradient descent algorithms
- Chatpers 1 & 2, Scribed Notes of Spring 2024 Course Offering.
- π Keywords: Deep Learning Computations, Seq2Seq, Attention Mechanism, Transformer
- π Slides: What's New, DL Intro, Transformer
- π» CoLab Notebook Demos: Dropout, Micrograd, Attention Mechanism
- βοΈ Homework: Problem Set 2: Implementing Neural Nets, due at 12:30pm, Feb/11/2025
- π Presentation of Replication Project: By Jiaci Yi and Yachong Wang
- Gui, G. and Toubia, O., 2023. The challenge of using LLMs to simulate human behavior: A causal inference perspective. arXiv:2312.15524. Link to the paper. Replication Report, Code, and Slides.
- π References:
- Deep Learning, 2016, by Ian Goodfellow, Yoshua Bengio and Aaron Courville, link to DL.
- Dive into Deep Learning (2nd Edition), 2023, by Aston Zhang, Zack Lipton, Mu Li, and Alex J. Smola, link to d2dl.
- Dell, Mellissa. 2024. Deep learning for economists. Journal of Economic Literature, forthcoming, link to the paper.
- Sutskever, Ilya, Oriol Vinyals, and Quoc V. Le. 2014. Sequence to sequence learning with neural networks. Advances in neural information processing systems, 27.
- Bahdanau, Dzmitry, Kyunghyun Cho, and Yoshua Bengio. 2015. Neural machine translation by jointly learning to align and translate. ICLR
- Lecture Notes and Slides for CS224n: Natural Language Processing with Deep Learning, by Christopher D. Manning, Diyi Yang, and Tatsunori Hashimoto. Link to CS224n.
- Parameter Initialization and Batch Normalization (in Chinese), GPU Comparisons, GitHub Repo for Micrograd by Andrej Karpathy.
- RNN and LSTM Visualizations, PyTorch's Tutorial of Seq2Seq for Machine Translation.
- Chatpers 2 & 6, Scribed Notes of Spring 2024 Course Offering.
- Handwritten Notes
- π Keywords: Transformer, ViT, DiT, Decision Transformer
- π Slides: What's New, Transformer
- π» CoLab Notebook Demos: Attention Mechanism, Transformer
- βοΈ Homework: Problem Set 3: Sentiment Analysis with BERT, due at 12:30pm, Mar/4/2025
- π Presentation of Replication Project: By Xiqing Qin and Yuxin Chen
- Manning, B.S., Zhu, K. and Horton, J.J., 2024. Automated social science: Language models as scientists and subjects (No. w32381). National Bureau of Economic Research. Link to the paper, link to GitHub Repo. Replication Report, Code, and Slides.
- π References:
- Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., ... and Polosukhin, I. (2017). Attention is all you need. Advances in neural information processing systems, 30.
- Qi, Meng, Yuanyuan Shi, Yongzhi Qi, Chenxin Ma, Rong Yuan, Di Wu, Zuo-Jun (Max) Shen. 2023. A Practical End-to-End Inventory Management Model with Deep Learning. Management Science, 69(2): 759-773.
- Sarzynska-Wawer, Justyna, Aleksander Wawer, Aleksandra Pawlak, Julia Szymanowska, Izabela Stefaniak, Michal Jarkiewicz, and Lukasz Okruszek. 2021. Detecting formal thought disorder by deep contextualized word representations. Psychiatry Research, 304, 114135.
- Hansen, Stephen, Peter J. Lambert, Nicholas Bloom, Steven J. Davis, Raffaella Sadun, and Bledi Taska. 2023. Remote work across jobs, companies, and space (No. w31007). National Bureau of Economic Research.
- Chapter 11, Dive into Deep Learning (2nd Edition), 2023, by Aston Zhang, Zack Lipton, Mu Li, and Alex J. Smola, link to d2dl.
- Lecture Notes and Slides for CS224n: Natural Language Processing with Deep Learning, by Christopher D. Manning, Diyi Yang, and Tatsunori Hashimoto. Link to CS224n.
- Part 2, Slides for COS 597G: Understanding Large Language Models, by Danqi Chen. Link to COS 597G
- Illustrated Transformer, Transformer from Scratch with the Code on GitHub.
- Andrej Karpathy's Lecture: Deep Dive into LLM
- Chatpers 7 Scribed Notes of Spring 2024 Course Offering.
- Handwritten Notes
- π Keywords: Pretraining, Scaling Law, BERT, GPT, DeepSeek
- π Slides: What's New, Pretraining
- π» CoLab Notebook Demos: Attention Mechanism, Transformer
- βοΈ Homework: Problem Set 3: Sentiment Analysis with BERT, due at 12:30pm, Mar/4/2025
- π Presentation of Replication Project: By Guohao Li and Jin Wang
- Li, P., Castelo, N., Katona, Z. and Sarvary, M., 2024. Frontiers: Determining the validity of large language models for automated perceptual analysis. Marketing Science, 43(2), pp.254-266. Link to the paper. Link to the replication package. Replication Report, Code, and Slides.
- π References:
- Devlin, Jacob, Ming-Wei Chang, Kenton Lee, Kristina Toutanova. 2018. BERT: Pre-training of deep bidirectional transformers for language understanding. ArXiv preprint arXiv:1810.04805. GitHub Repo
- Radford, Alec, Karthik Narasimhan, Tim Salimans, and Ilya Sutskever. 2018. Improving language understanding by generative pre-training, (GPT-1) PDF link, GitHub Repo
- Radford, Alec, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever. 2019. Language models are unsupervised multitask learners. OpenAI blog, 1(8), 9. (GPT-2) PDF Link, GitHub Repo
- Brown, Tom, et al. 2020. Language models are few-shot learners. Advances in neural information processing systems, 33, 1877-1901. (GPT-3) GitHub Repo
- DeepSeek-AI, 2024. Deepseek-V3 Technical Report. arXiv:2412.19437. GitHub Repo
- Huang, Allen H., Hui Wang, and Yi Yang. 2023. FinBERT: A large language model for extracting information from financial text. Contemporary Accounting Research, 40(2): 806-841. (FinBERT) GitHub Repo
- Gorodnichenko, Y., Pham, T. and Talavera, O., 2023. The voice of monetary policy. American Economic Review, 113(2), pp.548-584.
- Reisenbichler, Martin, Thomas Reutterer, David A. Schweidel, and Daniel Dan. 2022. Frontiers: Supporting content marketing with natural language generation. Marketing Science, 41(3): 441-452.
- Books, Notes, and Courses: Chapter 11.9 of Dive into Deep Learning, Part 9 of CS224N: Natural Language Processing with Deep Learning, Part 2 & 4 of COS 597G: Understanding Large Language Models, Part 3 & 4 of CS336: Large Language Models from Scratch, Chatper 8 of Scribed Notes for Spring 2024 Course Offering, Handwritten Notes.
- Andrej Karpathy's Lectures: Build GPT-2 (124M) from Scratch GitHub Repo for NanoGPT, Deep Dive into LLM, Build the GPT Tokenizer
- Miscellaneous Resources: CS224n, Hugging Face π€ Tutorial, A Visual Guide to BERT, LLM Visualization, How GPT-3 Works, Video on DeepSeek MoE, TikTokenizer, Inference with Base LLM
- π Keywords: Posttraining, Instruct GPT, SFT, RLHF, DPO, Test-Time Scaling, Knowledge Distillation
- π Slides: What's New, Pretraining, Posttraining
- π» CoLab Notebook Demos: BERT API @ Hugging Face π€, BERT Finetuning
- βοΈ Homework: Problem Set 4: Finetuning LLM, due at 12:30pm, Mar/18/2025
- π Presentation of Replication Project: By Di Wu and Chuchu Sun
- Radford, Alec, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever. 2019. Language models are unsupervised multitask learners. OpenAI blog, 1(8), 9. (GPT-2) PDF Link, GitHub Repo, Replication Report, Code, and Slides.
- π References:
- Ouyang, Long, et al. 2022. Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35, 27730-27744.
- Chu, T., Zhai, Y., Yang, J., Tong, S., Xie, S., Schuurmans, D., Le, Q.V., Levine, S. and Ma, Y., 2025. Sft memorizes, rl generalizes: A comparative study of foundation model post-training. arXiv preprint arXiv:2501.17161.
- Wei, Jason, et al. 2022. Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems, 35, 24824-24837.
- Yao, S., Yu, D., Zhao, J., Shafran, I., Griffiths, T., Cao, Y. and Narasimhan, K., 2023. Tree of thoughts: Deliberate problem solving with large language models. Advances in neural information processing systems, 36, pp.11809-11822.
- Brynjolfsson, E., Li, D. and Raymond, L., 2025. Generative AI at work. The Quarterly Journal of Economics, p.qjae044.
- Hinton, G., Vinyals, O. and Dean, J., 2015. Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531.
- Books, Notes, and Courses: Part 10 of CS224N: Natural Language Processing with Deep Learning, Talk @ Stanford by Barret Zoph and John Schulman, Part 3, 4, 9 & 10 of CS336: Large Language Models from Scratch, Part 9 & 14 of MIT 6.5940: TinyML and Efficient Deep Learning Computing, Chatper 9 of Scribed Notes for Spring 2024 Course Offering, Handwritten Notes.
- Andrej Karpathy's Lectures: Build GPT-2 (124M) from Scratch GitHub Repo for NanoGPT, Deep Dive into LLM, Build the GPT Tokenizer
- Miscellaneous Resources: CS224n, Hugging Face π€ Tutorial, LLM Visualization, How GPT-3 Works, Video on DeepSeek MoE, Video on DeekSeep Native Sparse Attention, TikTokenizer, Inference with Base LLM
- π Keywords: Posttraining, SFT, PEFT, RLHF, DPO
- π Slides: What's New, What's Next, Posttraining
- π» CoLab Notebook Demos: LLM Finetuning, Quantization
- βοΈ Homework: Problem Set 4: Finetuning LLM, due at 12:30pm, Mar/18/2025
- π No Presentation of Replication Project in This Week.
- π References:
- Chu, T., Zhai, Y., Yang, J., Tong, S., Xie, S., Schuurmans, D., Le, Q.V., Levine, S. and Ma, Y., 2025. Sft memorizes, rl generalizes: A comparative study of foundation model post-training. arXiv preprint arXiv:2501.17161.
- Wei, Jason, et al. 2022. Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems, 35, 24824-24837.
- Yao, S., Yu, D., Zhao, J., Shafran, I., Griffiths, T., Cao, Y. and Narasimhan, K., 2023. Tree of thoughts: Deliberate problem solving with large language models. Advances in neural information processing systems, 36, pp.11809-11822.
- Brynjolfsson, E., Li, D. and Raymond, L., 2025. Generative AI at work. The Quarterly Journal of Economics, p.qjae044.
- Hinton, G., Vinyals, O. and Dean, J., 2015. Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531.
- Muennighoff, N., Yang, Z., Shi, W., Li, X.L., Fei-Fei, L., Hajishirzi, H., Zettlemoyer, L., Liang, P., Candès, E. and Hashimoto, T., 2025. s1: Simple test-time scaling. arXiv preprint arXiv:2501.19393.
- Books, Notes, and Courses: Part 10 & 11 of CS224N: Natural Language Processing with Deep Learning, Talk @ Stanford by Barret Zoph and John Schulman, Part 15 & 16 of CS336: Large Language Models from Scratch, Part 5, 6, 9 & 14 of MIT 6.5940: TinyML and Efficient Deep Learning Computing, Finetuning LLMs, Quantization Fundamentals, Chatper 9 of Scribed Notes for Spring 2024 Course Offering, Handwritten Note.
- Andrej Karpathy's Lectures: Build GPT-2 (124M) from Scratch GitHub Repo for NanoGPT, Deep Dive into LLM, How to Use LLM, Build the GPT Tokenizer
- Miscellaneous Resources: Video on DeepSeek R1, Video on DeekSeep MLA, Approximating KL Divergence, DeepSeek Open-Infra Repo
- π Keywords: Test-Time Scaling, Knowledge Distillation, Inference, Quantization, LLM Evaluations, LLM Agents
- π Slides: What's New, What's Next, Posttraining, Inference, Research Tools
- π» CoLab Notebook Demos: Quantization
- βοΈ Homework: Problem Set 4: Finetuning LLM, due at 12:30pm, Mar/18/2025
- π Presentation of Replication Project: By Tao Wang and Zhe Liu
- Jens Ludwig, Sendhil Mullainathan, Machine Learning as a Tool for Hypothesis Generation, The Quarterly Journal of Economics, Volume 139, Issue 2, May 2024, Pages 751β827, link to the paper, Replication Report, Code, and Slides.
- π References:
- Ouyang, Long, et al. 2022. Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35, 27730-27744.
- Chu, T., Zhai, Y., Yang, J., Tong, S., Xie, S., Schuurmans, D., Le, Q.V., Levine, S. and Ma, Y., 2025. Sft memorizes, rl generalizes: A comparative study of foundation model post-training. arXiv preprint arXiv:2501.17161.
- Wei, Jason, et al. 2022. Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems, 35, 24824-24837.
- Yao, S., Yu, D., Zhao, J., Shafran, I., Griffiths, T., Cao, Y. and Narasimhan, K., 2023. Tree of thoughts: Deliberate problem solving with large language models. Advances in neural information processing systems, 36, pp.11809-11822.
- Brynjolfsson, E., Li, D. and Raymond, L., 2025. Generative AI at work. The Quarterly Journal of Economics, p.qjae044.
- Hinton, G., Vinyals, O. and Dean, J., 2015. Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531.
- Muennighoff, N., Yang, Z., Shi, W., Li, X.L., Fei-Fei, L., Hajishirzi, H., Zettlemoyer, L., Liang, P., Candès, E. and Hashimoto, T., 2025. s1: Simple test-time scaling. arXiv preprint arXiv:2501.19393.
- Park, J.S., O'Brien, J., Cai, C.J., Morris, M.R., Liang, P. and Bernstein, M.S., 2023, October. Generative agents: Interactive simulacra of human behavior. In Proceedings of the 36th annual acm symposium on user interface software and technology (pp. 1-22).
- Ties de Kok (2025) ChatGPT for Textual Analysis? How to Use Generative LLMs in Accounting Research. Management Science forthcoming. GitHub Link
- Books, Notes, and Courses: Part 11 & 12 of CS224N: Natural Language Processing with Deep Learning, Talk @ Stanford by Barret Zoph and John Schulman, Part 16 & 17 of CS336: Large Language Models from Scratch, Part 5, 6, 9 & 14 of MIT 6.5940: TinyML and Efficient Deep Learning Computing, Berkeley LLM Agents, Finetuning LLMs, Quantization Fundamentals, Building and Evaluating Advanced RAG, RLHF Short Course, Chatper 9 of Scribed Notes for Spring 2024 Course Offering. Handwritten Note.
- Andrej Karpathy's Lectures: Build GPT-2 (124M) from Scratch GitHub Repo for NanoGPT, Deep Dive into LLM, How to Use LLM, Build the GPT Tokenizer
- Miscellaneous Resources: Video on DeepSeek R1, Video on DeekSeep MLA, Approximating KL Divergence, DeepSeek Open-Infra Repo, Language Agents: Foundations, Prospects, and Risks (EMNLP 2024 Tutorial)
- π Keywords: Potential Outcomes Model, RCT, Unconfoundedness, IPW, AIPW
- π Slides: What's New, What's Next, Causal Inference
- π» CoLab Notebook Demos: Causal Inference under Unconfoundedness
- βοΈ Homework: Problem Set 5: Bias and Variance with Mis-speficied Linear Regression, due at 12:30pm, Apr/2/2025
- π Presentation of Replication Project: By Keming Li and Qilin Huang
- Costello, T.H., Pennycook, G. and Rand, D.G., 2024. Durably reducing conspiracy beliefs through dialogues with AI. Science, 385(6714), p.eadq1814, link to the paper, Replication Report, Code, and Slides.
- π References:
- Rubin, D.B., 1974. Estimating causal effects of treatments in randomized and nonrandomized studies. Journal of educational Psychology, 66(5), p.688.
- Rosenbaum, P.R. and Rubin, D.B., 1983. The central role of the propensity score in observational studies for causal effects. Biometrika, 70(1), pp.41-55.
- Li, F., Morgan, K.L. and Zaslavsky, A.M., 2018. Balancing covariates via propensity score weighting. Journal of the American Statistical Association, 113(521), pp.390-400.
- Robins, J.M., Rotnitzky, A. and Zhao, L.P., 1994. Estimation of regression coefficients when some regressors are not always observed. Journal of the American statistical Association, 89(427), pp.846-866.
- Books, Notes, and Courses: Chapters 1, 2, & 3 of Causal Inference: A Statistical Learning Approach, Chapters 2 & 5 of Applied Causal Inference Powered by ML and AI, Chapters 2 & 3 of Duke STA 640: Causal Inference
- Miscellaneous Resources: Videos of Stanford ECON 293 Machine Learning and Causal Inference, AEA Continuing Education on Machine Learning and Econometrics
For Tasks:
Click tags to check more tools for each tasksFor Jobs:
Alternative AI tools for AI-PhD-S25
Similar Open Source Tools

AI-PhD-S25
AI-PhD-S25 is a mono-repo for the DOTE 6635 course on AI for Business Research at CUHK Business School. The course aims to provide a fundamental understanding of ML/AI concepts and methods relevant to business research, explore applications of ML/AI in business research, and discover cutting-edge AI/ML technologies. The course resources include Google CoLab for code distribution, Jupyter Notebooks, Google Sheets for group tasks, Overleaf template for lecture notes, replication projects, and access to HPC Server compute resource. The course covers topics like AI/ML in business research, deep learning basics, attention mechanisms, transformer models, LLM pretraining, posttraining, causal inference fundamentals, and more.

AI-PhD-S24
AI-PhD-S24 is a mono-repo for the PhD course 'AI for Business Research' at CUHK Business School in Spring 2024. The course aims to provide a basic understanding of machine learning and artificial intelligence concepts/methods used in business research, showcase how ML/AI is utilized in business research, and introduce state-of-the-art AI/ML technologies. The course includes scribed lecture notes, class recordings, and covers topics like AI/ML fundamentals, DL, NLP, CV, unsupervised learning, and diffusion models.

interpret
InterpretML is an open-source package that incorporates state-of-the-art machine learning interpretability techniques under one roof. With this package, you can train interpretable glassbox models and explain blackbox systems. InterpretML helps you understand your model's global behavior, or understand the reasons behind individual predictions. Interpretability is essential for: - Model debugging - Why did my model make this mistake? - Feature Engineering - How can I improve my model? - Detecting fairness issues - Does my model discriminate? - Human-AI cooperation - How can I understand and trust the model's decisions? - Regulatory compliance - Does my model satisfy legal requirements? - High-risk applications - Healthcare, finance, judicial, ...

LLMEvaluation
The LLMEvaluation repository is a comprehensive compendium of evaluation methods for Large Language Models (LLMs) and LLM-based systems. It aims to assist academics and industry professionals in creating effective evaluation suites tailored to their specific needs by reviewing industry practices for assessing LLMs and their applications. The repository covers a wide range of evaluation techniques, benchmarks, and studies related to LLMs, including areas such as embeddings, question answering, multi-turn dialogues, reasoning, multi-lingual tasks, ethical AI, biases, safe AI, code generation, summarization, software performance, agent LLM architectures, long text generation, graph understanding, and various unclassified tasks. It also includes evaluations for LLM systems in conversational systems, copilots, search and recommendation engines, task utility, and verticals like healthcare, law, science, financial, and others. The repository provides a wealth of resources for evaluating and understanding the capabilities of LLMs in different domains.

Awesome-LLM-Reasoning
**Curated collection of papers and resources on how to unlock the reasoning ability of LLMs and MLLMs.** **Description in less than 400 words, no line breaks and quotation marks.** Large Language Models (LLMs) have revolutionized the NLP landscape, showing improved performance and sample efficiency over smaller models. However, increasing model size alone has not proved sufficient for high performance on challenging reasoning tasks, such as solving arithmetic or commonsense problems. This curated collection of papers and resources presents the latest advancements in unlocking the reasoning abilities of LLMs and Multimodal LLMs (MLLMs). It covers various techniques, benchmarks, and applications, providing a comprehensive overview of the field. **5 jobs suitable for this tool, in lowercase letters.** - content writer - researcher - data analyst - software engineer - product manager **Keywords of the tool, in lowercase letters.** - llm - reasoning - multimodal - chain-of-thought - prompt engineering **5 specific tasks user can use this tool to do, in less than 3 words, Verb + noun form, in daily spoken language.** - write a story - answer a question - translate a language - generate code - summarize a document

LLM-PLSE-paper
LLM-PLSE-paper is a repository focused on the applications of Large Language Models (LLMs) in Programming Language and Software Engineering (PL/SE) domains. It covers a wide range of topics including bug detection, specification inference and verification, code generation, fuzzing and testing, code model and reasoning, code understanding, IDE technologies, prompting for reasoning tasks, and agent/tool usage and planning. The repository provides a comprehensive collection of research papers, benchmarks, empirical studies, and frameworks related to the capabilities of LLMs in various PL/SE tasks.

Time-LLM
Time-LLM is a reprogramming framework that repurposes large language models (LLMs) for time series forecasting. It allows users to treat time series analysis as a 'language task' and effectively leverage pre-trained LLMs for forecasting. The framework involves reprogramming time series data into text representations and providing declarative prompts to guide the LLM reasoning process. Time-LLM supports various backbone models such as Llama-7B, GPT-2, and BERT, offering flexibility in model selection. The tool provides a general framework for repurposing language models for time series forecasting tasks.

ChatLaw
ChatLaw is an open-source legal large language model tailored for Chinese legal scenarios. It aims to combine LLM and knowledge bases to provide solutions for legal scenarios. The models include ChatLaw-13B and ChatLaw-33B, trained on various legal texts to construct dialogue data. The project focuses on improving logical reasoning abilities and plans to train models with parameters exceeding 30B for better performance. The dataset consists of forum posts, news, legal texts, judicial interpretations, legal consultations, exam questions, and court judgments, cleaned and enhanced to create dialogue data. The tool is designed to assist in legal tasks requiring complex logical reasoning, with a focus on accuracy and reliability.

SLAM-LLM
SLAM-LLM is a deep learning toolkit for training custom multimodal large language models (MLLM) focusing on speech, language, audio, and music processing. It provides detailed recipes for training and high-performance checkpoints for inference. The toolkit supports various tasks such as automatic speech recognition (ASR), text-to-speech (TTS), visual speech recognition (VSR), automated audio captioning (AAC), spatial audio understanding, and music caption (MC). Users can easily extend to new models and tasks, utilize mixed precision training for faster training with less GPU memory, and perform multi-GPU training with data and model parallelism. Configuration is flexible based on Hydra and dataclass, allowing different configuration methods.

xgen
XGen is a research release for the family of XGen models (7B) by Salesforce AI Research. It includes models with support for different sequence lengths and tokenization using the OpenAI Tiktoken package. The models can be used for auto-regressive sampling in natural language generation tasks.

awesome-ml-gen-ai-elixir
A curated list of Machine Learning (ML) and Generative AI (GenAI) packages and resources for the Elixir programming language. It includes core tools for data exploration, traditional machine learning algorithms, deep learning models, computer vision libraries, generative AI tools, livebooks for interactive notebooks, and various resources such as books, videos, and articles. The repository aims to provide a comprehensive overview for experienced Elixir developers and ML/AI practitioners exploring different ecosystems.

Genesis
Genesis is a physics platform designed for general purpose Robotics/Embodied AI/Physical AI applications. It includes a universal physics engine, a lightweight, ultra-fast, pythonic, and user-friendly robotics simulation platform, a powerful and fast photo-realistic rendering system, and a generative data engine that transforms user-prompted natural language description into various modalities of data. It aims to lower the barrier to using physics simulations, unify state-of-the-art physics solvers, and minimize human effort in collecting and generating data for robotics and other domains.

llm-self-correction-papers
This repository contains a curated list of papers focusing on the self-correction of large language models (LLMs) during inference. It covers various frameworks for self-correction, including intrinsic self-correction, self-correction with external tools, self-correction with information retrieval, and self-correction with training designed specifically for self-correction. The list includes survey papers, negative results, and frameworks utilizing reinforcement learning and OpenAI o1-like approaches. Contributions are welcome through pull requests following a specific format.

intro_pharma_ai
This repository serves as an educational resource for pharmaceutical and chemistry students to learn the basics of Deep Learning through a collection of Jupyter Notebooks. The content covers various topics such as Introduction to Jupyter, Python, Cheminformatics & RDKit, Linear Regression, Data Science, Linear Algebra, Neural Networks, PyTorch, Convolutional Neural Networks, Transfer Learning, Recurrent Neural Networks, Autoencoders, Graph Neural Networks, and Summary. The notebooks aim to provide theoretical concepts to understand neural networks through code completion, but instructors are encouraged to supplement with their own lectures. The work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.

EDA-AI
EDA-AI is a repository containing implementations of cutting-edge research papers in the field of chip design. It includes DeepPlace, PRNet, HubRouter, and PreRoutGNN models for tasks such as placement, routing, timing prediction, and global routing. Researchers and practitioners can leverage these implementations to explore advanced techniques in chip design.
For similar tasks

Azure-Analytics-and-AI-Engagement
The Azure-Analytics-and-AI-Engagement repository provides packaged Industry Scenario DREAM Demos with ARM templates (Containing a demo web application, Power BI reports, Synapse resources, AML Notebooks etc.) that can be deployed in a customerβs subscription using the CAPE tool within a matter of few hours. Partners can also deploy DREAM Demos in their own subscriptions using DPoC.

sorrentum
Sorrentum is an open-source project that aims to combine open-source development, startups, and brilliant students to build machine learning, AI, and Web3 / DeFi protocols geared towards finance and economics. The project provides opportunities for internships, research assistantships, and development grants, as well as the chance to work on cutting-edge problems, learn about startups, write academic papers, and get internships and full-time positions at companies working on Sorrentum applications.

tidb
TiDB is an open-source distributed SQL database that supports Hybrid Transactional and Analytical Processing (HTAP) workloads. It is MySQL compatible and features horizontal scalability, strong consistency, and high availability.

zep-python
Zep is an open-source platform for building and deploying large language model (LLM) applications. It provides a suite of tools and services that make it easy to integrate LLMs into your applications, including chat history memory, embedding, vector search, and data enrichment. Zep is designed to be scalable, reliable, and easy to use, making it a great choice for developers who want to build LLM-powered applications quickly and easily.

telemetry-airflow
This repository codifies the Airflow cluster that is deployed at workflow.telemetry.mozilla.org (behind SSO) and commonly referred to as "WTMO" or simply "Airflow". Some links relevant to users and developers of WTMO: * The `dags` directory in this repository contains some custom DAG definitions * Many of the DAGs registered with WTMO don't live in this repository, but are instead generated from ETL task definitions in bigquery-etl * The Data SRE team maintains a WTMO Developer Guide (behind SSO)

mojo
Mojo is a new programming language that bridges the gap between research and production by combining Python syntax and ecosystem with systems programming and metaprogramming features. Mojo is still young, but it is designed to become a superset of Python over time.

pandas-ai
PandasAI is a Python library that makes it easy to ask questions to your data in natural language. It helps you to explore, clean, and analyze your data using generative AI.

databend
Databend is an open-source cloud data warehouse that serves as a cost-effective alternative to Snowflake. With its focus on fast query execution and data ingestion, it's designed for complex analysis of the world's largest datasets.
For similar jobs

NanoLLM
NanoLLM is a tool designed for optimized local inference for Large Language Models (LLMs) using HuggingFace-like APIs. It supports quantization, vision/language models, multimodal agents, speech, vector DB, and RAG. The tool aims to provide efficient and effective processing for LLMs on local devices, enhancing performance and usability for various AI applications.

mslearn-ai-fundamentals
This repository contains materials for the Microsoft Learn AI Fundamentals module. It covers the basics of artificial intelligence, machine learning, and data science. The content includes hands-on labs, interactive learning modules, and assessments to help learners understand key concepts and techniques in AI. Whether you are new to AI or looking to expand your knowledge, this module provides a comprehensive introduction to the fundamentals of AI.

awesome-ai-tools
Awesome AI Tools is a curated list of popular tools and resources for artificial intelligence enthusiasts. It includes a wide range of tools such as machine learning libraries, deep learning frameworks, data visualization tools, and natural language processing resources. Whether you are a beginner or an experienced AI practitioner, this repository aims to provide you with a comprehensive collection of tools to enhance your AI projects and research. Explore the list to discover new tools, stay updated with the latest advancements in AI technology, and find the right resources to support your AI endeavors.

go2coding.github.io
The go2coding.github.io repository is a collection of resources for AI enthusiasts, providing information on AI products, open-source projects, AI learning websites, and AI learning frameworks. It aims to help users stay updated on industry trends, learn from community projects, access learning resources, and understand and choose AI frameworks. The repository also includes instructions for local and external deployment of the project as a static website, with details on domain registration, hosting services, uploading static web pages, configuring domain resolution, and a visual guide to the AI tool navigation website. Additionally, it offers a platform for AI knowledge exchange through a QQ group and promotes AI tools through a WeChat public account.

AI-Notes
AI-Notes is a repository dedicated to practical applications of artificial intelligence and deep learning. It covers concepts such as data mining, machine learning, natural language processing, and AI. The repository contains Jupyter Notebook examples for hands-on learning and experimentation. It explores the development stages of AI, from narrow artificial intelligence to general artificial intelligence and superintelligence. The content delves into machine learning algorithms, deep learning techniques, and the impact of AI on various industries like autonomous driving and healthcare. The repository aims to provide a comprehensive understanding of AI technologies and their real-world applications.

promptpanel
Prompt Panel is a tool designed to accelerate the adoption of AI agents by providing a platform where users can run large language models across any inference provider, create custom agent plugins, and use their own data safely. The tool allows users to break free from walled-gardens and have full control over their models, conversations, and logic. With Prompt Panel, users can pair their data with any language model, online or offline, and customize the system to meet their unique business needs without any restrictions.

ai-demos
The 'ai-demos' repository is a collection of example code from presentations focusing on building with AI and LLMs. It serves as a resource for developers looking to explore practical applications of artificial intelligence in their projects. The code snippets showcase various techniques and approaches to leverage AI technologies effectively. The repository aims to inspire and educate developers on integrating AI solutions into their applications.

ai_summer
AI Summer is a repository focused on providing workshops and resources for developing foundational skills in generative AI models and transformer models. The repository offers practical applications for inferencing and training, with a specific emphasis on understanding and utilizing advanced AI chat models like BingGPT. Participants are encouraged to engage in interactive programming environments, decide on projects to work on, and actively participate in discussions and breakout rooms. The workshops cover topics such as generative AI models, retrieval-augmented generation, building AI solutions, and fine-tuning models. The goal is to equip individuals with the necessary skills to work with AI technologies effectively and securely, both locally and in the cloud.