Spaces:
Paused
Paused
File size: 8,739 Bytes
b5d3943 6720344 6088c8f 6720344 b5d3943 f8bfc75 b5d3943 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 |
"""Utilities for assembling and exporting interview reports.
This module provides two primary helpers used by the recruiter dashboard:
``generate_llm_interview_report(application)``
Given a candidate's ``Application`` record, assemble a plain‑text report
summarising the interview. Because the interview process currently
executes entirely client‑side and does not persist questions or answers
to the database, this report focuses on the information available on
the server: the candidate's profile, the job requirements and a skills
match score. Should future iterations store richer interview data
server‑side, this function can be extended to include question/answer
transcripts, per‑question scores and LLM‑generated feedback.
``create_pdf_report(report_text)``
Convert a multi‑line string into a simple PDF. The implementation
leverages Matplotlib's PDF backend (available by default) to avoid
heavyweight dependencies such as ReportLab or WeasyPrint, which are
absent from the runtime environment. Text is wrapped and split
across multiple pages as necessary.
"""
from __future__ import annotations
import json
from io import BytesIO
import textwrap
from typing import List
import matplotlib.pyplot as plt
from matplotlib.backends.backend_pdf import PdfPages
def generate_llm_interview_report(application) -> str:
"""Generate a human‑readable interview report for a candidate.
The report includes the candidate's name and email, job details,
application date, a computed skills match summary and placeholder
sections for future enhancements. If server‑side storage of
question/answer pairs is added later, this function can be updated
to incorporate those details.
Parameters
----------
application : backend.models.database.Application
The SQLAlchemy Application instance representing the candidate's
job application. Assumed to have related ``job`` and
``date_applied`` attributes available.
Returns
-------
str
A multi‑line string containing the report contents.
"""
# Defensive imports to avoid circular dependencies at import time
try:
from datetime import datetime # noqa: F401
except Exception:
pass
# Extract candidate skills and job skills
try:
candidate_features = json.loads(application.extracted_features) if application.extracted_features else {}
except Exception:
candidate_features = {}
candidate_skills: List[str] = candidate_features.get('skills', []) or []
job_skills: List[str] = []
try:
job_skills = json.loads(application.job.skills) if application.job and application.job.skills else []
except Exception:
job_skills = []
# Compute skills match ratio and label. Normalise to lower case for
# comparison and avoid dividing by zero when ``job_skills`` is empty.
candidate_set = {s.strip().lower() for s in candidate_skills}
job_set = {s.strip().lower() for s in job_skills}
common = candidate_set & job_set
ratio = len(common) / len(job_set) if job_set else 0.0
if ratio >= 0.75:
score_label = 'Excellent'
elif ratio >= 0.5:
score_label = 'Good'
elif ratio >= 0.25:
score_label = 'Medium'
else:
score_label = 'Poor'
# Assemble report lines
lines: List[str] = []
lines.append('Interview Report')
lines.append('=================')
lines.append('')
lines.append(f'Candidate Name: {application.name}')
lines.append(f'Candidate Email: {application.email}')
if application.job:
lines.append(f'Job Applied: {application.job.role}')
lines.append(f'Company: {application.job.company}')
else:
lines.append('Job Applied: N/A')
lines.append('Company: N/A')
# Format date_applied if available
try:
date_str = application.date_applied.strftime('%Y-%m-%d') if application.date_applied else 'N/A'
except Exception:
date_str = 'N/A'
lines.append(f'Date Applied: {date_str}')
lines.append('')
lines.append('Skills Match Summary:')
# Represent required and candidate skills as comma‑separated lists. Use
# title‑case for presentation and handle empty lists gracefully.
formatted_job_skills = ', '.join(job_skills) if job_skills else 'N/A'
formatted_candidate_skills = ', '.join(candidate_skills) if candidate_skills else 'N/A'
formatted_common = ', '.join(sorted(common)) if common else 'None'
lines.append(f' Required Skills: {formatted_job_skills}')
lines.append(f' Candidate Skills: {formatted_candidate_skills}')
lines.append(f' Skills in Common: {formatted_common}')
lines.append(f' Match Ratio: {ratio * 100:.0f}%')
lines.append(f' Score: {score_label}')
lines.append('')
lines.append('Interview Transcript & Evaluation:')
try:
if application.interview_log:
try:
qa_log = json.loads(application.interview_log)
except Exception:
qa_log = []
if qa_log:
for idx, entry in enumerate(qa_log, 1):
q = entry.get("question", "N/A")
a = entry.get("answer", "N/A")
eval_score = entry.get("evaluation", {}).get("score", "N/A")
eval_feedback = entry.get("evaluation", {}).get("feedback", "N/A")
lines.append(f"\nQuestion {idx}: {q}")
lines.append(f"Answer: {a}")
lines.append(f"Score: {eval_score}")
lines.append(f"Feedback: {eval_feedback}")
else:
lines.append("No interview log data recorded.")
else:
lines.append("No interview log data recorded.")
except Exception as e:
lines.append(f"Error loading interview log: {e}")
return '\n'.join(lines)
def create_pdf_report(report_text: str) -> BytesIO:
"""
Alternative implementation using reportlab for better PDF generation.
Install with: pip install reportlab
"""
try:
from reportlab.lib.pagesizes import A4
from reportlab.lib.styles import getSampleStyleSheet, ParagraphStyle
from reportlab.lib.units import inch
from reportlab.platypus import SimpleDocTemplate, Paragraph, Spacer, PageBreak
from reportlab.lib.colors import HexColor
buffer = BytesIO()
doc = SimpleDocTemplate(
buffer,
pagesize=A4,
rightMargin=0.75*inch,
leftMargin=0.75*inch,
topMargin=1*inch,
bottomMargin=1*inch
)
# Create custom styles
styles = getSampleStyleSheet()
question_style = ParagraphStyle(
'Question',
parent=styles['Heading2'],
fontSize=12,
textColor=HexColor('#2C3E50'),
spaceAfter=6,
spaceBefore=12
)
answer_style = ParagraphStyle(
'Answer',
parent=styles['Normal'],
fontSize=10,
textColor=HexColor('#34495E'),
leftIndent=20,
spaceAfter=3
)
score_style = ParagraphStyle(
'Score',
parent=styles['Normal'],
fontSize=10,
textColor=HexColor('#27AE60'),
leftIndent=20,
fontName='Helvetica-Bold'
)
feedback_style = ParagraphStyle(
'Feedback',
parent=styles['Normal'],
fontSize=10,
textColor=HexColor('#E74C3C'),
leftIndent=20,
spaceAfter=6
)
# Build document content
story = []
lines = report_text.split('\n')
for line in lines:
stripped = line.strip()
if stripped.startswith('Question'):
story.append(Paragraph(stripped, question_style))
elif stripped.startswith('Answer:'):
story.append(Paragraph(stripped, answer_style))
elif stripped.startswith('Score:'):
story.append(Paragraph(stripped, score_style))
elif stripped.startswith('Feedback:'):
story.append(Paragraph(stripped, feedback_style))
elif stripped:
story.append(Paragraph(stripped, styles['Normal']))
else:
story.append(Spacer(1, 12))
# Build PDF
doc.build(story)
buffer.seek(0)
return buffer
__all__ = ['generate_llm_interview_report', 'create_pdf_report'] |