Skip to content

Claude/code review wg kt d#151

Closed
Sid3548 wants to merge 3 commits intoTHU-MAIC:mainfrom
Sid3548:claude/code-review-WGKtD
Closed

Claude/code review wg kt d#151
Sid3548 wants to merge 3 commits intoTHU-MAIC:mainfrom
Sid3548:claude/code-review-WGKtD

Conversation

@Sid3548
Copy link

@Sid3548 Sid3548 commented Mar 20, 2026

Summary

Related Issues

Changes

Type of Change

  • Bug fix (non-breaking change that fixes an issue)
  • New feature (non-breaking change that adds functionality)
  • Breaking change (fix or feature that would cause existing functionality to change)
  • Documentation update
  • Refactoring (no functional changes)
  • CI/CD or build changes

Verification

Steps to reproduce / test

What you personally verified

Evidence

  • CI passes (pnpm check && pnpm lint && npx tsc --noEmit)
  • Manually tested locally
  • Screenshots / recordings attached (if UI changes)

Checklist

  • My code follows the project's coding style
  • I have performed a self-review of my code
  • I have added/updated documentation as needed
  • My changes do not introduce new warnings

Sid3548 and others added 3 commits March 20, 2026 10:44
Implements POST /api/interview/turn using statelessGenerate() instead of
a direct callLLM() call, aligning the interview turn endpoint with the
same orchestration path used by /api/chat.

- Coach agent (turn 1): gives feedback on the candidate's latest answer
- Interviewer agent (turn 2): asks the next follow-up question
- Both agents are defined as inline agentConfigs (no registry required)
- Text deltas accumulated server-side; JSON response shape kept
  compatible with existing frontend: { feedback, nextQuestion,
  coachName, interviewerName }
- Provider config resolution (API key, baseUrl, SSRF guard) mirrors
  the chat route exactly

https://claude.ai/code/session_019MGKMBhcRnwEkGCRdN1m77
The coding track was always scoring full marks (e.g. 3/3) regardless of
solution correctness because:
- score was set to reviews.length (= number of problems, always = total)
- percentage was hardcoded to 100%
- all problem topics were unconditionally added as weakAreas
- the review-code route silently returned {} on parse failure

Changes:
- Add score (0-10) and verdict (pass/fail) fields to CodeReviewResult
- Update buildCodeReviewPrompt to request score + verdict from the LLM
- review-code route: clamp/normalise score, return 502 if fields missing
- quiz-runner: derive real score from passed.length, compute real %, only
  flag weak areas for problems with verdict=fail or score<5
- QuizResults display uses actual pass count instead of reviews.length
- CodeReviewPanel now shows per-problem score badge and missing points

https://claude.ai/code/session_019MGKMBhcRnwEkGCRdN1m77
@wyuc
Copy link
Contributor

wyuc commented Mar 21, 2026

Thanks for the contribution, Siddharth!

This PR covers a lot of ground (Interview module, Quiz module, Hindi locale — 86 files), which makes it hard to review as a single unit. We'd prefer smaller, focused PRs so each feature can be discussed and reviewed independently.

Suggested next steps:

  1. Open an issue or discussion for Interview / Quiz to align on scope and design before implementation
  2. For Hindi support, check out PR Fix classroom generation language normalization for non-English locales #176 which is taking a spec-based approach to multi-language generation — coordinating there would avoid conflicts

Happy to review focused follow-ups!

@wyuc wyuc closed this Mar 21, 2026
@wyuc wyuc mentioned this pull request Mar 21, 2026
13 tasks
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants