Paper Discussion Round
The paper discussion round is often the most important interview in the research scientist loop. You will be asked to present your own papers, discuss papers the interviewer selects, and demonstrate that you can critically analyze research with depth and nuance. This lesson covers frameworks, common pitfalls, and practice strategies.
Part 1: Presenting Your Own Papers
When an interviewer says "Tell me about your best paper" or "Walk me through your most impactful work," they are evaluating multiple things simultaneously: your technical depth, your ability to communicate, your understanding of your contribution's place in the broader literature, and your intellectual honesty about limitations.
The 5-Minute Paper Presentation Framework
Structure your paper presentation using this framework. Practice until you can deliver it in exactly 5 minutes — interviewers will often interrupt with questions, so you need to be concise.
| Step | Time | What to Cover |
|---|---|---|
| 1. Problem & Motivation | 60 sec | What problem are you solving? Why does it matter? What was the state of the art before your work? Frame the problem so a researcher outside your subfield can understand why it is important. |
| 2. Key Insight | 60 sec | What is the core idea that makes your approach work? This is the most important part. Distill your contribution to one or two key insights that differentiate your work from prior approaches. |
| 3. Method Overview | 90 sec | High-level description of your method. Use a mental diagram. Do not get lost in implementation details — focus on the architecture decisions that follow from your key insight. |
| 4. Results & Evidence | 60 sec | What are your strongest results? Which experiments most convincingly support your claims? Mention specific numbers: "We improved BLEU by 3.2 points over the previous state of the art on WMT'23." |
| 5. Limitations & Future Work | 30 sec | Be honest about what your method does not do well. This demonstrates intellectual maturity. Mention specific limitations and how future work could address them. |
Anticipating Questions About Your Papers
Prepare answers for these questions about every paper you list on your CV:
Methodology Questions
"Why did you choose this approach over X?" "What happens if you remove component Y?" "How sensitive is your method to hyperparameter Z?" "Why not use a simpler baseline?" "What is the computational complexity?"
Results Questions
"Why does your method fail on dataset X?" "Is the improvement statistically significant?" "Did you run ablation studies?" "How does performance scale with data size?" "What is the variance across random seeds?"
Impact Questions
"Has anyone built on your work?" "How would you apply this to a different domain?" "What is the practical impact?" "If you could redo this paper, what would you change?" "What did you learn that surprised you?"
Adversarial Questions
"I think your baseline comparison is unfair because..." "Your theoretical assumptions do not hold when..." "This seems incremental over prior work X because..." Stay calm. Acknowledge valid points. Defend with evidence, not emotion.
Part 2: Discussing Others' Papers
Interviewers will often hand you a paper (sometimes one you have never seen) and ask you to discuss it. This tests whether you can quickly extract the core contribution, identify strengths and weaknesses, and propose improvements.
The Critical Analysis Framework
When given a paper to discuss, work through these layers systematically:
- Contribution: What is the paper claiming to contribute? Is this a new method, a theoretical result, an empirical finding, or a benchmark?
- Novelty: How does this differ from prior work? Is the novelty in the formulation, the architecture, the training procedure, or the evaluation?
- Soundness: Are the theoretical claims correct? Are the proofs rigorous? Are the assumptions reasonable for the target application?
- Empirical Rigor: Are the baselines fair and up-to-date? Are ablation studies included? Is there variance reporting? Are the datasets appropriate?
- Limitations: What does the paper not address? What assumptions could break in practice? What failure modes exist?
- Extensions: How could this work be improved or extended? What follow-up experiments would strengthen the claims?
5 Example Paper Discussion Frameworks
Practice these frameworks on recent papers from your subfield. Each represents a different type of paper you might encounter in an interview.
Framework 1: Discussing a Foundational Architecture Paper
Example: "Attention Is All You Need" (Vaswani et al., 2017)
Contribution: Replaces recurrence and convolution entirely with self-attention for sequence transduction. Key insight: attention alone can capture long-range dependencies more efficiently than RNNs.
Critical analysis: Strengths include O(1) sequential operations for self-attention (vs O(n) for RNNs), parallelizable training, and strong empirical results on WMT translation. Weaknesses include O(n²) memory complexity for sequence length n, no explicit positional inductive bias (sinusoidal encodings are a patch, not a principled solution), and the paper's theoretical justification for why attention works is thin — the success is largely empirical.
Extensions you should mention: Sparse attention (Longformer, BigBird), relative positional encodings (RoPE, ALiBi), flash attention for memory efficiency, and the open question of whether attention is truly sufficient or whether hybrid architectures (state space models) may be more principled.
Framework 2: Discussing a Training Methodology Paper
Example: "RLHF / InstructGPT" (Ouyang et al., 2022)
Contribution: Fine-tunes GPT-3 with human feedback via reinforcement learning from human feedback (RLHF) to align language model outputs with human preferences.
Critical analysis: Strengths include dramatic improvement in helpfulness and safety without sacrificing capability, demonstrates that alignment is tractable, and the reward model approach is elegant. Weaknesses include reliance on human annotator quality and consistency, reward hacking (model learns to exploit the reward model rather than genuinely improve), RLHF instability during training (PPO is notoriously finicky), and the philosophical question of whose preferences are being optimized.
Extensions: DPO (Direct Preference Optimization) as a simpler alternative to PPO, constitutional AI, debate as alignment, and RLAIF (using AI feedback instead of human feedback).
Framework 3: Discussing a Theoretical Paper
Example: "Neural Tangent Kernel" (Jacot et al., 2018)
Contribution: Shows that infinitely wide neural networks trained with gradient descent behave as kernel methods with a specific kernel (the neural tangent kernel). Provides a theoretical framework for understanding neural network training dynamics.
Critical analysis: Strengths include providing the first rigorous connection between neural networks and kernel methods, enabling closed-form predictions about training dynamics, and opening a new theoretical research direction. Weaknesses include the infinite-width assumption not holding for practical networks, NTK regime predicts lazy training (small parameter changes) which does not capture feature learning, and the gap between NTK predictions and actual network behavior grows with depth and practical hyperparameters.
Extensions: Mean-field theory as an alternative framework, feature learning regime analysis, tensor programs, and the ongoing debate about whether kernel perspectives or feature-learning perspectives better explain deep learning success.
Framework 4: Discussing an Empirical Scaling Paper
Example: "Scaling Laws for Neural Language Models" (Kaplan et al., 2020)
Contribution: Establishes power-law relationships between model performance and compute, dataset size, and parameter count. Provides predictive equations for optimal resource allocation.
Critical analysis: Strengths include actionable predictions for resource allocation, empirically validated across multiple orders of magnitude, and changed how the field thinks about scaling. Weaknesses include power laws not predicting emergent capabilities (phase transitions), the Chinchilla paper later showed the compute-optimal ratio was wrong (Kaplan undertrained relative to model size), and the laws may break at frontier scales.
Extensions: Chinchilla scaling laws (Hoffmann et al., 2022), emergent abilities debate, scaling laws for downstream tasks, and whether scaling laws hold for multimodal models.
Framework 5: Discussing a Safety/Alignment Paper
Example: "Constitutional AI" (Bai et al., 2022)
Contribution: Trains a harmless AI assistant using AI-generated self-critiques guided by a set of constitutional principles, reducing reliance on human feedback for safety.
Critical analysis: Strengths include scalability (less human labor), transparency (principles are explicit and auditable), and strong empirical results on harmlessness without sacrificing helpfulness. Weaknesses include the constitution itself may encode biases of its authors, self-critique quality depends on the base model's capability, circular dependency (using AI to align AI), and principles can conflict (helpfulness vs harmlessness) without clear resolution mechanisms.
Extensions: Debate as alignment, recursive reward modeling, mechanistic interpretability for validating alignment, and the broader question of whether behavioral alignment is sufficient or whether we need guaranteed alignment.
Common Paper Discussion Mistakes
| Mistake | Why It Hurts You | What to Do Instead |
|---|---|---|
| Only summarizing without critiquing | Shows you can read but not think critically | Always offer at least 2 strengths and 2 weaknesses |
| Criticizing without understanding | Shows arrogance and shallow reading | Demonstrate you understand the contribution before critiquing |
| Ignoring related work | Shows narrow reading habits | Connect the paper to 2–3 related works and explain the relationship |
| Being vague about limitations | "It could be better" is not a critique | Be specific: "The O(n²) attention complexity limits application to sequences beyond 4096 tokens" |
| Not proposing extensions | Misses the chance to show research thinking | Suggest concrete follow-up experiments or theoretical questions |
Key Takeaways
- Use the 5-minute framework to present your papers: Problem, Insight, Method, Results, Limitations
- Prepare for adversarial questions about every paper on your CV — especially methodology and baseline comparisons
- When discussing others' papers, systematically evaluate contribution, novelty, soundness, empirical rigor, limitations, and extensions
- Practice discussing papers from different categories: architecture, training methodology, theory, empirical scaling, and safety
- Always propose concrete extensions — this demonstrates research thinking, which is the primary skill being evaluated
Lilly Tech Systems