Appendix A: Discipline-Specific AI Use Policy Matrix

OECD recommends discipline-specific AI policies that specify permitted, restricted, and prohibited uses at the assignment level rather than as a blanket course-level statement. The matrix below provides a starting framework for finance courses. Faculty should adapt it to their course context and publish it in the syllabus so students understand expectations before any assignment begins.

The progression from AI-Prohibited to AI-Transparent to AI-Required mirrors the framework in #4 of this guide and aligns with OECD’s recommendation that AI governance evolve alongside student and faculty competence.

Assignment Type AI-Prohibited AI-Transparent AI-Required
In-Class Exam / Quiz All AI use prohibited. Proctored conditions. Foundational knowledge verification. N/A for in-class exams. N/A for in-class exams.
Written Case Analysis / Financial Plan Draft must be written independently without AI. No AI grammar or idea assistance. AI permitted for revision feedback only. Full transcript required. Reflection memo required. Human instructor layer required. AI co-authoring permitted. Student grades the critique and process, not the AI output. Oral defense required.
Financial Modeling / Coding Assignment (#8) AI-generated final code submission without annotation. Copying without understanding the financial logic. AI may generate code. Student must annotate every section with financial logic explanation. Transcript submitted. Oral walkthrough required. Capstone / advanced courses: AI as primary co-pilot. Student architects requirements, audits logic, interprets outputs, and defends decisions.
Monte Carlo / Simulation (#10) Submitting AI-generated code or outputs without writing a specification document, annotating assumptions, or running sensitivity analysis. AI generates code from student specification. Student must run sensitivity analysis, interpret distribution outputs, and submit investment recommendation memo. Full transcript required. Risk Management / advanced courses: AI as simulation partner. Student designs the model logic, selects distributions, defends assumptions in Q&A, and connects results to a real investment decision.
Role-Play / Oral Simulation (#1) Submitting someone else’s transcript or completing the interaction without genuine engagement. No transcript submitted. AI persona used. Transcript + reflection memo required. Flags moments where the bot broke character. Oral defense adds human verification layer. Student designs the AI persona, writes the system prompt, runs the simulation, critiques the bot’s financial accuracy, and presents findings to the class.

“Prohibited” refers to the specific use described, not a blanket prohibition on AI in the course. Faculty are encouraged to share this matrix with students on the first day and revisit it at each major assignment. Policies should evolve as both student AI competence and course-level evidence accumulate.


Appendix B: OECD-Aligned Evaluation Framework

OECD’s Digital Education Outlook 2026 recommends that for each AI application, institutions specify: (1) the intended learning outcome, (2) the most likely unintended consequence, and (3) how both will be measured across five dimensions—performance, retention, engagement, equity, and instructor workload. The framework below applies this logic to each of the 10 finance applications.

Faculty are encouraged to select one or two applications to evaluate formally in the first semester of adoption and share findings with colleagues.

Application Intended Outcome Likely Unintended Consequence How to Measure Equity Check
#1 Role-Play Improved professional communication and analytical defense under pressure Surface-level interaction without genuine engagement; transcript padding Pre/post oral defense scores; rubric score distribution; reflection quality ratings Do students with speech anxiety or non-native English engage equitably? Offer written alternative.
#2 AI Feedback Faster feedback cycles; stronger revision quality; feedback literacy Over-reliance on AI critique; depersonalization of instructor relationship Compare pre/post rubric scores on revision quality; measure acceptance/rejection rationale quality Do ESL students or first-gen students trust AI feedback differently? Survey and compare.
#3 Socratic Tutor Reduced knowledge gaps on foundational concepts; increased office hours independence Students stop attempting problems independently before engaging AI AI-free quiz scores before and after adoption; verification log completion rates Do lower-GPA students benefit as much as higher-GPA students? Track by subgroup.
#4 Process Assessment Assessment validity in AI era; higher-order reasoning on display Student perception of subjectivity; oral check-in burden on faculty Compare oral check-in performance with written submission scores; track grade distribution equity Do oral assessments disadvantage non-native English speakers? Offer optional written equivalent.
#5 Data Analysis Data literacy without coding bottleneck; interpretation over syntax False confidence in AI-verified statistics; shallow interpretation memos Manual verification checklist accuracy; “so what?” memo quality ratings; exam performance on interpretation questions Do students without prior coding exposure benefit more? Track by major/background.
#6 Content Generation Faculty time savings; fresher, more varied assessment materials Curricular homogenization; AI fabricated financial data enters course materials Faculty time-savings log; student assessment score variance across variants; diversity audit of generated cases Are generated cases representative across firm size, geography, and industry? Run diversity check each semester.
#7 Case Updates Current-events integration; verification habits; analyst mindset Hallucinated market data; AI as citable source; paywalled data access gaps Source citation quality; verification accuracy; student performance on identifying material vs. noise Do students without Bloomberg/Refinitiv access face barriers? Ensure free-source alternatives are sufficient.
#8 Vibe Coding Quantitative finance access for non-CS students; code-as-communication skill Annotation without understanding; AI-generated financial logic errors go undetected Annotation quality scores; oral walkthrough performance; “break it” exercise accuracy Do students without prior programming background close the gap? Track by CS background.
#9 AI Literacy Professional skepticism; verification habits; AI evaluation methodology Cynicism about AI; failure to transfer audit methodology to new tools Error identification accuracy; cross-platform comparison depth; professional recommendation quality Do students from different majors/backgrounds show different AI literacy gaps? Survey by subgroup.
#10 Simulation Distributional thinking; decision-making under uncertainty; model-building intuition Code generation without engagement; assumption selection without theoretical grounding Specification document quality; sensitivity analysis depth; investment memo recommendations; oral Q&A scores Do students without strong quantitative backgrounds close the gap with AI assistance? Track pre/post.

Measurement Guidance

Performance
Rubric scores and exam results.
Retention
AI-free assessment scores 4–8 weeks after adoption.
Engagement
Dialogue log depth, reflection quality, participation rates.
Equity
Score distributions by subgroup (GPA, language background, major).
Workload
Faculty time-per-student on grading and feedback tasks.

Start with one application and one dimension. Build the evidence base incrementally.