A Controlled Synthetic Benchmark for Educational Aspect-Based Sentiment Analysis

Yehudit Aperstein · Alexander Apartsin

Abstract

Educational aspect-based sentiment analysis (ABSA) can support course improvement, but public aspect-labeled student feedback remains scarce because educational reviews are private, institution-specific, and expensive to annotate. This study introduces a controlled synthetic benchmark for educational ABSA built from 10,000 synthetic course reviews with explicit train-validation-test splits and a 20-aspect pedagogical schema spanning instructional quality, assessment and course management, learning demand, learning environment, and engagement. The corpus is generated with sampled target labels, sampled nuance attributes, and a realism-tuned prompt refined through a three-cycle judge-editor procedure. On the resulting benchmark, local baselines with TF-IDF, two-step transformers, and joint encoders show that the task is nontrivial; the strongest untuned model, BERT, reaches a held-out detection micro-F1 of 0.2760, while a modest lower-rate BERT schedule improves this to 0.2930. Full-test GPT-based inference with gpt-5.2 reaches 0.2519 micro-F1 in zero-shot mode and 0.2501 with retrieval-based few-shot prompting, placing batch inference above the classical baseline and close to the compact joint encoders. A conservative external evaluation on 2,829 mapped student-feedback reviews from Herath et al. yields a micro-F1 of 0.4593 for BERT on a 9-aspect overlap, indicating partial synthetic-to-real transfer. Realism and faithfulness analyses are reported as generator diagnostics that clarify how the benchmark was stabilized and where label noise remains. The study therefore contributes a synthetic educational ABSA corpus, a documented generation procedure, and a reproducible benchmark setting for a domain in which public labeled data remain difficult to obtain.

1. Introduction

Student reviews contain actionable evidence about workload, clarity, support, fairness, materials, and overall learning experience, yet most institutional feedback analysis still depends on manual reading or coarse sentiment summaries. For pedagogy, this matters because interventions rarely target “sentiment” in the abstract; instructors and program leaders need to know whether learners are reacting to assessment design, instructional clarity, course relevance, staff support, or the overall classroom experience. Feedback research in higher education repeatedly shows that comments become educationally useful when they can be connected to actionable teaching conditions rather than treated as isolated satisfaction signals [20, 21, 22]. ABSA is therefore well matched to this setting because the same review can simultaneously praise one aspect and criticize another. The challenge is that high-quality educational ABSA datasets are scarce, expensive to annotate, and often limited to a single institution, course type, or annotation scheme.

That scarcity is not accidental. Real student-feedback data are difficult to release because they are tied to institutional processes, often contain identifying context, and require laborious annotation decisions about implicit aspects, mixed sentiment, and local pedagogical terminology. Existing educational text-mining studies show that open-ended feedback is useful for course improvement, but the underlying data are often confidential, institution-specific, or too small for straightforward public reuse [16, 23, 24]. Even when institutions collect large amounts of feedback, the resulting labels are rarely public, rarely harmonized across universities, and rarely aligned with the fine-grained aspect schemas needed for ABSA. The present study addresses that bottleneck by pairing synthetic data generation with downstream ABSA modeling. Rather than treating data generation and model training as separate tasks, it formulates them as one end-to-end system: a synthetic review generator produces multi-aspect course reviews in multiple student personas, and an analysis pipeline learns aspect detection and sentiment scoring from those labels.

The central contribution is therefore practical and scientific at the same time: a synthetic educational review corpus with controlled aspect labels can support internally consistent ABSA training and evaluation, while diversity and realism analyses help stabilize the generation process. Synthetic supervision is treated here as a practical way to study an educational NLP problem whose pedagogical value is high but whose public labeled data remain unusually scarce.

The remainder of the paper proceeds as follows. Section 2 situates the study in the ABSA, synthetic-data, and educational-feedback literatures. Section 3 describes the corpus design and generation protocol, Section 4 defines the benchmark task and evaluation setup, Section 5 reports generator-oriented validation analyses, and Section 6 presents the main benchmark and external-validation results before Section 7 discusses limitations and implications.

Contributions

  • Educational benchmark resource: a 10,000-review synthetic course-review corpus with explicit train, validation, and test partitions over a 20-aspect pedagogical schema that is substantially richer than the small educational sentiment resources typically available.
  • Controlled generation design: a generation procedure that separates supervision targets from nuance attributes, allowing the same aspect labels to appear under different course contexts, student backgrounds, and writing styles instead of collapsing into one prompt pattern.
  • Benchmark evidence: a full internal ABSA benchmark with TF-IDF and transformer baselines showing that the resulting task is nontrivial and supports reproducible model comparison under fixed splits.
  • Initial external relevance: a conservative synthetic-to-real evaluation on mapped Herath student feedback, together with realism and faithfulness analyses that clarify what the corpus supports and where its present limits remain.

2. Related Work

This study draws on three literatures that are usually discussed separately: general ABSA benchmarks and models, synthetic text generation for supervised NLP, and educational feedback analysis. The connection among them is central to the paper. General ABSA work provides the task formulation and evaluation vocabulary; synthetic-data research motivates controlled label generation when annotation is scarce; and educational feedback studies explain why the target problem matters pedagogically and why real labeled data are unusually difficult to assemble. SemEval task definitions helped establish ABSA as a standard fine-grained sentiment problem by formalizing aspect detection and aspect sentiment prediction with shared datasets and evaluation procedures [2, 3]. More recent ABSA work continues to show the value of pretrained transformers and sentiment-aware pretraining for fine-grained polarity modeling [4, 5].

On the data side, the synthetic-data component belongs within a broader NLP literature on augmentation, weak supervision, and LLM-based example generation. Data-augmentation surveys show that gains depend on whether generated instances preserve task semantics and enlarge useful coverage of the training distribution rather than merely adding lexical variation [25]. More recent work on LLM-generated supervision shows both promise and caution: synthetic examples can improve downstream classification when prompts, controls, and filtering are carefully designed, but task difficulty and label faithfulness remain decisive constraints [6, 26, 27]. This caution is especially relevant here because educational ABSA is a subjective and domain-specific task whose label space often reflects local pedagogical practice rather than a universal ontology.

In education, prior work has emphasized that student feedback is noisy, stylistically varied, and pedagogically important. A broader higher-education feedback literature also shows that formative feedback matters because it shapes student learning rather than merely administrative satisfaction tracking [1]. Welch and Mihalcea demonstrated the value of targeted sentiment analysis for student comments [7], while Chathuranga et al. released an annotated student course feedback corpus for opinion targets and polarity [8]. Herath et al. later reported a 3,000-instance student feedback corpus with annotations for aspects, opinion terms, and polarities, illustrating both the feasibility and the cost of building educational sentiment resources [11]. Nikolić et al. showed that ABSA can be informative for higher education reviews, but that source characteristics matter and performance varies across aspect frequencies and review sources [9]. Misuraca et al. repositioned opinion mining as an educational analytic rather than only a text-classification exercise, arguing that open-ended comments add value beyond numeric ratings [16]. More recent reviews synthesize the educational sentiment-analysis landscape and reiterate recurring bottlenecks around annotation cost, domain language, multi-polarity, and deployment in authentic educational settings [10, 17]. Together, this literature supports the relevance of the problem, clarifies why real labeled educational data are difficult to obtain at scale, and explains why a synthetic educational ABSA resource may be useful even when realism validation remains a separate concern.

Table 1. Educational feedback resources used to position the present study within prior work on student comments, annotated course feedback, and higher-education review analysis.
Study Data source and focus Label granularity Relation to this study
Welch and Mihalcea [7]Student commentsTargeted sentimentEstablishes pedagogical relevance, but not a public multi-aspect benchmark
Chathuranga et al. [8]Student course feedbackOpinion targets and polarityShows feasibility of real educational annotation, but with narrower task scope
Herath et al. [11]Student feedback corpusAspect, opinion-term, and polarity annotationsClosest educational benchmark reference and strongest evidence of annotation cost
Nikolić et al. [9]Higher-education reviewsAspect-based review analysisShows source sensitivity and motivates explicit realism validation
This studySynthetic course reviews with separate realism and mapped real-transfer validation sets20-aspect benchmark corpus, controllable prompt protocol, realism study, and conservative 9-aspect transfer checkPositions the contribution as synthetic supervision with initial external validation rather than a replacement for real annotated corpora

The dataset is intentionally more aspect-rich than narrow course-review setups that focus mainly on instructor praise or overall satisfaction. Its 20 aspects span workload, clarity, exam fairness, lecturer quality, relevance, interest, support, materials, overall experience, feedback quality, assessment design, pacing, organization, practical application, tooling usability, accessibility, grading transparency, peer interaction, and prerequisite fit. This wider aspect space matters pedagogically because the actionable question for instructors is rarely whether a course is “good” in the abstract; it is which design dimension is helping or harming student learning. It also matters methodologically because a broader aspect inventory creates a harder ABSA benchmark, especially for more ambiguous categories such as relevance, interest, overall experience, support, and prerequisite fit.

The novelty claim in this setting comes from the combination of pieces rather than from synthetic text alone. The study couples controlled synthetic generation with an explicitly educational 20-aspect schema, a train-validation-test benchmark protocol, and external validation on mapped student feedback. In other words, the contribution lies in the combination of task framing, controllable generation, educationally meaningful aspect design, and benchmark integration.

Synthetic augmentation and few-shot prompting also provide a natural methodological backdrop for this work. Prior augmentation research in ABSA shows that semantics-preserving edits can improve downstream learning without destroying polarity structure [12], while few-shot prompting methods demonstrate that label recovery can be possible even with very limited demonstrations [13, 18]. More recent instruction-tuned and weak-supervision studies further suggest that prompt-based ABSA can become competitive when label scarcity is severe, but that careful task formulation still matters [28, 29]. This makes it reasonable to compare trainable ABSA models with zero-shot and few-shot generative baselines on the same synthetic benchmark, even if the main paper contribution remains the dataset and the training pipeline rather than prompt engineering alone.

Generative ABSA work also supports this comparison more directly. Unified generative frameworks show that ABSA subtasks can be cast as structured text generation rather than only as classifier heads [14], and text-generation formulations of aspect category sentiment analysis report particular strength in zero-shot and few-shot settings [15]. At the same time, broader text-classification evidence suggests that in-context learning remains an informative but demanding baseline: smaller fine-tuned encoders often still outperform zero-shot and few-shot prompting when labels are structured and evaluation is strict [19], and broader sentiment-analysis reviews in the LLM era recommend caution against assuming prompt-based superiority by default [30]. This literature justifies including LLM-based baselines while keeping the central contribution focused on the synthetic educational dataset and its benchmark evaluation.

The most consequential contribution is therefore the benchmark resource itself, not synthetic text generation in isolation. What makes that resource scientifically useful is the combination of three properties. First, the aspect schema is pedagogically motivated: it separates actionable dimensions of teaching and course design rather than collapsing them into overall satisfaction. Second, the generator is controlled rather than generic: target aspect sentiments and sampled nuance attributes are manipulated independently so that labels can be realized under varied educational situations and writing styles. Third, the corpus is paired with explicit ABSA evaluation, so the study contributes not only text generation, but a reproducible experimental setting for a domain where public aspect-labeled data remain scarce.

3. Data Resource and Generation Protocol

The methodology has two linked components: a synthetic data generation protocol and a downstream ABSA analysis pipeline. The first produces educational reviews with controlled aspect-level sentiment labels and varied review conditions. The second evaluates whether those reviews function as useful ABSA training and evaluation data. Diversity and realism analyses are treated as quality analyses of the generator, whereas the main empirical results concern model behavior on the synthetic corpus.

3.1 Synthetic Data Generation Pipeline

The generation process is organized as a sequence of four decisions. First, the system samples the supervision target itself: one, two, or three aspect-sentiment pairs drawn from the 20-aspect pedagogical schema. Second, it samples a separate nuance state from the attribute families that define course context, student background, assessment conditions, writing style, and realism controls. This separation is intentional. The target labels specify what the downstream ABSA model should recover, whereas the nuance state specifies how those labels are realized in the surface review. The same aspect combination can therefore appear under different course names, student trajectories, and rhetorical styles instead of collapsing into one repeated prompt pattern.

After those two states are sampled, they are merged only at prompt construction time. A realism-constrained instruction is appended to the sampled labels and attributes to form the generation request. The generator then produces a draft review, a refinement stage removes recurrent synthetic cues while attempting to preserve the declared labels, and the exported record retains both its aspect labels and its sampled nuance attributes. Large-scale generation is therefore not driven by a single free-form prompt, but by a fixed template whose row-level variation comes from independently sampled targets and contextual controls.

Figure 2 condenses the generation protocol to its main structural idea. One input stream defines the supervision target, namely one to three aspect-sentiment labels. A second input stream defines how those labels are realized in text through sampled course context, student state, pedagogical circumstances, and style controls. These streams meet only at prompt construction time, after which generation yields a review-level benchmark record that preserves both the text and the sampled control state. The dashed feedback path is shown as an inter-cycle revision rather than a per-row operation because realism validation updates the stabilized instruction between complete prompt states.

Synthetic data generation pipeline
Figure 2. Supervision targets and contextual attributes are sampled separately, merged through a stabilized prompt, and exported as review-level benchmark records.

3.2 ABSA Analysis Pipeline

The benchmark pipeline begins from the assembled review-level corpus and applies one shared train-validation-test split to every reported method. The train partition is used for parameter estimation, the validation partition is used for early stopping, threshold calibration, and prompt selection, and the test partition is reserved for final reporting only. This split discipline is especially important here because the paper also reports realism validation and mapped real-data transfer; those additional analyses remain outside the internal benchmark and are never merged into the synthetic split.

Within that split, the primary task formulation is two-step ABSA. A first stage predicts which aspects are present in a review, and a second stage assigns sentiment only to the aspects selected by the detector. This decomposition keeps omission errors and polarity errors analytically separate and makes the evaluation easier to interpret than a single opaque score. Reported results therefore combine multi-label detection metrics such as precision, recall, and F1 with a detected-aspect sentiment MSE that measures how well polarity is recovered once an aspect has been predicted. Prompt-based methods and mapped real-data transfer are evaluated under the same output contract, but they are reported as supporting analyses rather than as the central benchmark family.

Figure 3 summarizes the evaluation side of the study. Every reported approach begins from the same review-level corpus and the same 8,000/1,000/1,000 split, so differences in outcome come from the modeling approach rather than from incompatible data conditions. The diagram therefore emphasizes a single shared evaluation contract rather than separate model-specific pipelines: one corpus, one split, one validation stage, and one held-out reporting stage. The side annotations mark the two most important boundaries around that core flow, namely that validation is reserved for calibration and model choice, while mapped real-data transfer and generator-validation analyses are treated as complementary checks rather than replacements for the main synthetic benchmark.

ABSA analysis pipeline
Figure 3. All reported methods share one corpus, one split policy, one validation stage, and one held-out test report.

3.3 Aspect Inventory and Pedagogical Scope

The benchmark uses a 20-aspect inventory so that distinct student concerns can be represented directly instead of being collapsed into broad categories such as overall_experience or relevance. For readability, the inventory is organized into five pedagogical blocks: instructional quality (clarity, lecturer_quality, materials, feedback_quality), assessment and course management (exam_fairness, assessment_design, grading_transparency, organization, tooling_usability), learning demand and readiness (difficulty, workload, pacing, prerequisite_fit), learning environment (support, accessibility, peer_interaction), and engagement and value (relevance, interest, practical_application, overall_experience). Appendix A.1 lists the full inventory with these group assignments.

These additions are motivated by pedagogy rather than by label inflation. Students frequently distinguish between a course that is difficult and one that is poorly paced, between a fair exam and a badly aligned assessment structure, and between strong content and weak tooling or feedback loops. Making those distinctions explicit is one of the main novelty claims: the goal is not only more labels, but a more educationally actionable aspect space than the small public student-feedback corpora described in prior work.

3.4 Prompting Protocol for Diversity

Diversity does not refer to surface paraphrasing alone. The generator varies course name, lecturer, grade, writing style, and one to three aspect labels, but that space would still be too narrow if reviews differed only superficially. The protocol therefore samples from a richer attribute schema that captures situational, pedagogical, and stylistic variation together. The schema contains four groups of controls: core context, assessment and teaching context, linguistic diversity, and realism controls.

Core-context attributes include course name, course level, semester stage, student background, motivation for taking the course, attendance pattern, study context, and grade band. Assessment-context attributes include assessment profile, instructional delivery, support-channel experience, administrative friction, feedback timing, prerequisite fit, collaboration structure, and platform or tooling experience. Linguistic-diversity attributes include writing style, emotional temperature, hedging level, specificity level, review length band, formality level, and recommendation stance. Realism-oriented attributes include review shape, contradiction pattern, time-pressure context, natural-noise style, comparison frame, and memory anchors. Each attribute is parameterized with a compact set of 4 to 6 values so the prompt remains controllable while still widening the review space. Each synthetic review samples five core-context attributes with course_name always required, four assessment-and-teaching attributes, three linguistic-diversity attributes, and three realism-control attributes. The result is one prompt template with a different contextual state for each review rather than a fixed bundle repeated at scale.

3.5 Prompt Stabilization and Data Generation Process

The generation process can be read as a separation of supervision targets from contextual realization. First, target attributes are sampled as one to three aspect-sentiment pairs; these are the labels the downstream ABSA system must recover. Second, nuance attributes are sampled independently from the four attribute groups described above so that the same target labels can appear in different pedagogical and stylistic contexts. Third, the prompt combines the sampled labels, the sampled nuance state, and the realism-constrained instruction used throughout corpus generation. The model then produces a draft review, a refinement step removes obvious synthetic cues while trying to preserve label faithfulness, and the stored item retains both its target labels and its sampled nuance attributes.

$$K \sim \mathrm{Categorical}(0.30, 0.40, 0.30), \qquad K \in \{1,2,3\}$$ $$A = \{(a_i, s_i)\}_{i=1}^{K}, \qquad a_i \in \mathcal{A}, \quad s_i \in \{\text{negative}, \text{neutral}, \text{positive}\}$$ $$N = \bigcup_{g \in \mathcal{G}} \{(n_{g,j}, v_{g,j})\}_{j=1}^{m_g}, \qquad v_{g,j} \sim P_g(\cdot)$$ $$p_c = T(A, N, I_c), \qquad x^{(0)} \sim p_{\theta}(\cdot \mid p_c), \qquad x = R_{\phi}(x^{(0)}, A, N, I_c)$$

In this formalization, \(K\) is the sampled number of aspects, \(A\) is the set of target aspect-sentiment labels, \(N\) is the sampled nuance state drawn from the attribute groups \(\mathcal{G}\), \(I_c\) is the realism-constrained instruction, \(T\) is the prompt-construction function, \(p_{\theta}\) is the base review generator, and \(R_{\phi}\) is the refinement step that attempts to preserve labels while removing cues that repeatedly trigger synthetic detection. The notation highlights the main design choice of the paper: labels and contextual variation are sampled separately and only later combined into one review-generation prompt.

The full dataset is generated from one prompt specification rather than from mixed prompt states. That specification is rendered into OpenAI Batch requests with gpt-5-nano as the generator, minimal reasoning effort, and low text verbosity. Length control is enforced both textually and operationally: every prompt includes a sampled review_length_band attribute, the prompt injects hard length guidance for that band, and the request builder sets band-specific max_output_tokens budgets that further depend on whether the review has one, two, or three target aspects. The resulting corpus is therefore not produced by an unconstrained free-form prompt, but by a fixed template with per-row target labels, nuance states, and output-budget controls.

Realism is treated as more than surface messiness. The stabilized instruction explicitly discourages sentence-by-sentence checklist coverage of aspect labels, over-balanced tradeoff language, stock recommendation summaries, and generic domain-term stacking. Instead, it asks for asymmetric detail, incidental mention of some aspects, and limited but grounded specificity. The number of labeled aspects is also constrained. The synthetic corpus contains an almost uniform distribution over one-, two-, and three-aspect reviews, with counts of 2,008, 1,969, and 2,007 respectively. This empirical pattern justifies restricting generation to one to three aspects only; allowing more aspects would likely create unnaturally dense reviews and increase synthetic detectability. The protocol therefore uses either the empirical distribution directly or a rounded practical policy of 0.30 / 0.40 / 0.30 for one, two, and three aspects respectively.

3.6 Synthetic Training Data and Real Validation Data

The final corpus contains 10,000 synthetic records generated from the realism-tuned prompt package. All rows contain review text and at least one valid labeled aspect, although 841 rows were marked incomplete by the API because they hit the output-token cap. The assembled corpus spans 8 course names, 6 observed writing-style labels, 5 grade bands, and the full 20-aspect inventory. The mean review length is 117.2 words, the median is 108 words, and each review contains an average of 2.0 labeled aspects. In practical terms, the corpus is large enough for controlled internal benchmarking, while realism validation and synthetic-to-real transfer remain separate supporting analyses.

Real data are used in two different but still narrower roles than the main synthetic benchmark. The realism-validation pool contains 32 public OMSCS reviews drawn from four course pages: CS-6200, CS-6250, CS-6400, and CS-7641. These reviews are not used for training, validation, or test evaluation in the main ABSA benchmark; they are used only as blinded reference items in the real-versus-synthetic judge loop so that prompt changes can be checked against naturally occurring educational-review language. Separately, the external transfer evaluation uses the annotated student-feedback corpus of Herath et al. [11], conservatively mapped to a 9-aspect overlap with our schema. That mapped real set is used only for out-of-domain evaluation after synthetic training and validation. This distinction is important for interpretation: synthetic data define the main ABSA benchmark, while real data are reserved for generator analysis and external evaluation.

The richness of the generator comes from the combination of target attributes and nuance attributes. Target attributes define the supervision target, while nuance attributes introduce course- and student-level variation that helps the same aspect labels appear in different contexts. In this protocol, the nuance space is grouped into four functional blocks so the prompt remains interpretable: one block anchors the course situation, one changes the pedagogical events being described, one changes linguistic realization, and one suppresses templated synthetic regularity. Table 2 lists representative variables from each block rather than exhausting the full schema, because the scientific point is the role each block plays in controlled diversity; the aspect blocks themselves are listed separately in Appendix A.1.

Table 2. Representative nuance-attribute groups used to diversify synthetic reviews while keeping the target aspect labels fixed. The table emphasizes the functional role of each block; individual reviews sample only a subset of these variables.
Group Representative attributes Function in the generation protocol
Core contextcourse_name, course_level, semester_stage, student_backgroundPlaces the review in a recognizable educational situation so the same labels can appear under different student positions and course settings.
Assessment and teachingassessment_profile, instruction_delivery, support_channel_experience, feedback_timingChanges the pedagogical substance of the review by varying what the student is reacting to, not only how the review sounds.
Linguistic diversitywriting_style, emotional_temperature, hedging_level, review_length_bandChanges tone, compression, and rhetorical texture while preserving the declared aspect-sentiment targets.
Realism controlsreview_shape, contradiction_pattern, memory_anchor, natural_noiseDiscourages checklist-like coverage and overbalanced summaries by introducing the small irregularities common in human reviews.

The benchmark uses the full 20-aspect inventory. The five aspect groups span instructional quality, assessment and course management, learning demand and readiness, learning environment, and engagement and value. Those dimensions are pedagogically important because they capture implementation details that often determine whether a student experience feels coherent or frustrating, whether the social learning environment is supportive, and whether course expectations match student preparation. In other words, Table 2 is not a style table; it is a control map for how pedagogical content, student position, and writing surface are disentangled during generation.

4. Benchmark Task and Evaluation Setup

4.1 Task Definition and Split Protocol

The ABSA benchmark is defined entirely on the synthetic corpus. Reviews are divided into train, validation, and test partitions using a strict three-way split. Training updates model parameters only on the training split; the validation split is reserved for early stopping, threshold calibration, and prompt-variant selection; and the test split is held out for final reporting only. This separation is central to the experimental design because the real-review pool is not part of the ABSA benchmark at all.

The experiments use the 10,000-review 20-aspect corpus with an 8,000/1,000/1,000 train-validation-test split. Metrics include per-aspect precision, recall, and F1 for aspect detection, together with sentiment mean-squared error on detected aspects. Because this is a sparse multilabel setting with many more negative than positive label decisions, the study also tracks macro balanced accuracy, macro specificity, and macro Matthews correlation as complementary diagnostics derived from the per-aspect confusion counts. Thresholds for discriminative detectors are chosen on the validation split only.

The split itself is deterministic. The benchmark harness applies a seeded permutation with seed 42, then assigns rows to train, validation, and test according to the fixed 0.80 / 0.10 / 0.10 proportions. No real-data rows are ever merged into this split. This point is worth stating explicitly because the paper contains both synthetic and real evaluations: the synthetic split is the main benchmark, the OMSCS pool is used only for realism validation, and the mapped Herath corpus is used only after training as an external evaluation set.

4.2 Evaluation Plan for the ABSA Pipeline

The evaluation plan compares two families of ABSA approaches under one shared task definition. The first family contains trainable encoders that learn aspect detection and per-aspect sentiment from synthetic supervision. The second contains GPT-based inference methods that recover the same outputs without task-specific fine-tuning. Both families are executed and reported in the study. The difference is not whether they are tested, but how they are deployed: the local supervised models are trained on the synthetic split, whereas the GPT-based methods perform direct batch inference on the held-out test split under the same output schema.

Formally, let \(x\) denote a review, let \(\mathcal{A}=\{a_1,\dots,a_K\}\) denote the aspect inventory, let \(z \in \{0,1\}^K\) denote aspect presence indicators, and let \(s \in \{-1,0,1\}^K\) denote aspect sentiments for the aspects that are present. The benchmark can then be viewed as learning or inferring a mapping \(x \mapsto (\hat{z}, \hat{s})\), with detection quality evaluated against \(z\) and sentiment quality evaluated only on detected or gold-present aspects depending on the reporting view. The main supervised decomposition treats ABSA as

$$\hat{z} = f_\theta(x), \qquad \hat{s} = g_\phi(x, \hat{z})$$

where \(f_\theta\) is a multi-label detector and \(g_\phi\) is a sentiment predictor applied only to the aspects selected by the first stage. This two-step formulation remains the main discriminative benchmark because it cleanly separates omission errors from polarity errors. The reported benchmark includes the TF-IDF two-step baseline together with transformer two-step models built on distilbert-base-uncased, bert-base-uncased, albert-base-v2, and roberta-base, all evaluated on the same 10K / 20-aspect train-validation-test split.

The training protocol is identical across the transformer baselines except for the encoder backbone. Reviews are truncated or padded to 192 tokens, optimization uses AdamW with learning rate \(3\times 10^{-5}\), and both the detection and sentiment stages are trained for up to three epochs with patience-based early stopping after two non-improving validation checks. The detection head is trained with sigmoid logits and a per-aspect weighted binary cross-entropy loss whose positive weights are estimated from the training split and clipped to the range \([1, 50]\) to avoid extreme imbalance. The sentiment head predicts one bounded score per aspect via a \(\tanh\) output layer and is optimized with masked mean-squared error so that loss is accumulated only on gold-present aspects. This explicit protocol matters because it explains why the benchmark emphasizes reproducibility and interpretable stage-wise errors rather than aggressive architecture tuning.

Threshold calibration is also separated cleanly from test evaluation. After training the detection stage, each aspect receives its own decision threshold chosen on the validation split by a grid search from 0.05 to 0.95 in steps of 0.05, maximizing per-aspect F1 on validation only. Those fixed thresholds are then applied once on the test split. Reported micro- and macro-F1 therefore reflect calibrated binary decisions, whereas the sentiment MSE is measured only on aspects that the model actually predicts as present. The balanced-accuracy and specificity diagnostics are useful here because they remain interpretable when the label space is sparse and class prevalence differs strongly across aspects. That detected-aspect view intentionally couples polarity quality to detection quality, which makes the sentiment metric more operational but also more conservative than evaluating sentiment only on gold-present aspects.

We also retain a single-stage joint family as an explicit decomposition baseline. In that setting, one model predicts both outputs simultaneously,

$$ (\hat{z}, \hat{s}) = h_\psi(x), $$

so that the benchmark can test whether jointly coupling aspect presence and polarity is competitive with the more interpretable two-step design. In the evaluation framework these joint variants are represented by compact bert_joint and distilbert_joint baselines, and they are reported as executed local comparisons alongside the stronger two-step models.

The prompt-based branch casts the same output space as structured generation rather than parameter updates. Given an instruction template \(\pi\), an optional demonstration set \(D\), and possibly a retrieval operator \(R(\cdot)\), a prompted model produces

$$ (\hat{z}, \hat{s}) = G_{\text{LLM}}(x; \pi, D), \qquad D = \varnothing \text{ for zero-shot}, \qquad D = R(x) \text{ for retrieval-based few-shot}. $$

This family includes six variants. The zero-shot model uses constrained structured decoding without demonstrations. Fixed few-shot prompting uses a static set of labeled training examples, while diverse few-shot prompting chooses examples that deliberately vary aspect count, tone, and review style. Retrieval-based few-shot prompting replaces static demonstrations with lexical nearest neighbors from the training split so that the context is review-dependent. The batch evaluation path uses a strict schema-constrained output contract, and the reported GPT results cover zero-shot, fixed few-shot, diverse few-shot, and retrieval-based few-shot inference under that shared contract.

Two additional prompt decompositions are included because they parallel common ABSA design choices. The two-pass prompted variant first predicts aspects and then conditions sentiment on those detected aspects,

$$\hat{z} = G_{\text{det}}(x; \pi_{\text{det}}, D_{\text{det}}), \qquad \hat{s} = G_{\text{sent}}(x, \hat{z}; \pi_{\text{sent}}, D_{\text{sent}}),$$

which makes its structure directly comparable to the two-step discriminative pipeline. The aspect-by-aspect prompted variant instead factors inference over the aspect inventory,

$$\hat{z}_k = G^{(k)}_{\text{pres}}(x), \qquad \hat{s}_k = G^{(k)}_{\text{pol}}(x) \text{ only if } \hat{z}_k = 1,$$

so it asks a separate binary presence question for each \(a_k\) and only then queries sentiment where needed. This formulation is more expensive but useful for diagnosing whether per-aspect prompting improves recall control or sentiment calibration.

These GPT-based variants are part of the ABSA evaluation plan, not the realism protocol. The realism protocol remains a separate judge-based validation loop used to improve the generator prompt, whereas the evaluation plan asks whether different ABSA approaches can recover the synthetic aspect labels once the dataset has been fixed. Across all planned approaches, model selection, threshold calibration, and prompt-choice decisions are confined to the validation split, and final benchmark numbers are reported only on held-out data.

Table 3 functions as a compact map of the executed experiment families. Rather than repeating implementation details, it keeps the benchmark logic visible in one place: which approaches learn from synthetic supervision, which recover labels through inference-time prompting, and which analyses are used to examine robustness and transfer.

Table 3. Benchmark matrix for the ABSA evaluation plan. The table shows the executed model families and the role each one plays in the reported evidence.
Family Modeling idea Representative variants Role in the study
Classical baselineTwo-step sparse lexical detector plus sentiment regressortfidf_two_stepMain benchmark
Low-cost reference point under the same validation and test contract.
Transformer encodersTwo-step supervised detection and sentiment modelsdistilbert-base-uncased, bert-base-uncased, albert-base-v2, roberta-baseMain benchmark
Primary discriminative comparison family on the full 10K / 20-aspect split.
Joint predictionSingle model predicts aspect presence and polarity togetherbert_joint, distilbert_jointExecuted comparison
Used to test whether a compact joint formulation can match the stronger two-step decomposition.
GPT-based inferenceStructured generation with demonstrations or retrieval contextgpt-5.2 zero-shot, few-shot, few-shot-diverse, retrieval-few-shotExecuted comparison
Inference-only ABSA baselines under the same structured aspect-sentiment contract and the full held-out test split.

This matrix is intentionally compact. The later results section returns to exact scores, robustness analyses, and per-aspect behavior, but Table 3 keeps the benchmark logic visible in one place: what is learned from synthetic supervision, what is inferred at test time only, and how those families complement the transfer and generator-validation analyses.

Table 4. Exact corpus-level counts and ranges for the synthetic benchmark and the two real-data pools used in realism analysis and external transfer. The table provides the precise summary values that complement the broader visual distribution shown later in Figure 5.
Measure Value Interpretation
Clean reviews10,000Usable synthetic records in the final corpus
Real validation reviews32Public OMSCS reviews used only for the realism-validation loop
Mapped external test reviews2,829Annotated student-feedback reviews from Herath et al. used only for synthetic-to-real evaluation on a 9-aspect overlap
Course names8Multiple academic contexts rather than a single course
Observed writing styles6Style-conditioned prompting remains present in the assembled dataset
Aspect inventory20The benchmark runs directly on the 20-aspect pedagogical schema listed in Appendix A.1
Mean / median words117.2 / 108The corpus mixes short comments with longer reflective reviews
Mean aspects per review2.0Many reviews are genuinely multi-aspect
Min-max words34-273Substantial variation in compression and detail remains visible even after length controls

5. Generator Validation Analyses

5.1 Realism-Validation Protocol

To assess generator quality without redefining the paper's main task, we built a small-scale real-versus-synthetic discrimination protocol. The reference pool is a public OMSCS review source from which 32 reviews were prepared across four computer science courses. This source is useful for a first realism stress test because it provides naturally written educational reviews with concrete course experiences, but it is not representative enough to serve as the only real benchmark. It is graduate-level, computer-science-heavy, English-only, and drawn from a single public review website. Accordingly, OMSCS is treated as a constrained validation source rather than a final external evaluation set, and these reviews are never mixed into the ABSA train, validation, or test splits.

The realism-validation procedure is iterative. In this protocol, one cycle means a complete pass over 60 judge questions: 30 real reviews and 30 synthetic reviews, each judged independently as a binary REAL-versus-SYNTHETIC decision with confidence, cue tags, and justification. Synthetic reviews are generated from sampled aspect sentiments together with a sampled subset of nuance attributes rather than the full schema, so the prompt remains controllable without looking templated. Each cycle records judge labels, confidence, entropy, a chance-confusion percentage, and exact-versus-chance significance statistics. When the judge correctly identifies synthetic cues, those explanations are fed into an editor step that rewrites the instruction before the next cycle, so each cycle is evaluated under one fixed prompt specification rather than an ad hoc prompt mixture. These checks are positioned as generator validation steps that support the main synthetic-data ABSA study, not as a substitute for the paper's core training-and-evaluation experiments. They are diagnostic, single-source, and single-judge evidence rather than a broad realism claim.

$$\hat{y}_{c,i} \in \{\text{real}, \text{synthetic}\}, \qquad q_{c,i} \in [0,1]$$ $$\mathrm{Acc}_c = \frac{1}{M}\sum_{i=1}^{M} \mathbf{1}[\hat{y}_{c,i} = y_i], \qquad M = 60$$ $$\mathrm{Confusion}_c = 100\left(1 - \frac{|\mathrm{Acc}_c - 0.5|}{0.5}\right)$$ $$H_c = -\frac{1}{M}\sum_{i=1}^{M}\left(q_{c,i}\log q_{c,i} + (1-q_{c,i})\log(1-q_{c,i})\right)$$ $$I_{c+1} = E\left(I_c, \{(\hat{y}_{c,i}, q_{c,i}, r_{c,i}) : \hat{y}_{c,i}=\text{synthetic} \land y_i=\text{synthetic}\}\right)$$

Here \(y_i\) is the hidden ground-truth source label, \(\hat{y}_{c,i}\) is the judge decision for item \(i\) in cycle \(c\), \(q_{c,i}\) is the judge confidence, \(r_{c,i}\) denotes the textual cue explanation, \(\mathrm{Acc}_c\) is the judge accuracy, \(\mathrm{Confusion}_c\) is the paper-facing chance-confusion statistic, \(H_c\) is mean decision entropy, and \(E\) is the editor function that produces the next stable realism instruction only from cues that correctly exposed synthetic items.

Realism improvement curve
Figure 1. Realism-validation outcomes across three judge-editor cycles. The three panels summarize judge accuracy, chance-confusion, and decision entropy over complete 60-question cycles. The figure documents how judge uncertainty changes across cycles rather than claiming statistical indistinguishability from real reviews.

5.2 Complementary Generator Diagnostics

We also ran compact model-assisted auxiliary audits with gpt-5.2. On 30 synthetic reviews generated directly from the corpus-generation prompt, all declared aspects were judged to be textually supported (\(1.0000\) support rates at both the aspect and review level), while exact sentiment agreement was lower (\(0.7377\) at the aspect level and \(0.5333\) at the full-review level). This prompt-level result should be read together with the harsher full-dataset audit reported later in Section 6.7B: the combination suggests that target aspect presence can be preserved under the generation prompt, but large-scale generation still introduces substantial label-faithfulness drift, especially for sentiment polarity. Together these diagnostics justify treating realism and faithfulness checks as generator audits rather than as evidence that the 10K corpus is a synthetic gold standard.

6. Results

The results are presented in the same order as the paper's claim structure. We begin with the corpus itself, then report the main local benchmark, then the full-test GPT and real-data comparisons, and only afterward return to generator-quality diagnostics such as output-control drift, label faithfulness, and realism validation. This ordering is intentional: the central claim concerns the value of the synthetic corpus as an educational ABSA benchmark resource, whereas realism and faithfulness analyses explain the strengths and limits of the generation process.

Aspect sentiment heatmap
Figure 4. Aspect-level label support across the 10K / 20-aspect corpus. The heatmap shows polarity composition for each aspect, ordered by total support, and the companion bar panel shows the total number of labeled mentions per aspect.
Corpus profile
Figure 5. Corpus profile and diversity cues. The panels visualize the distributional shape of review length, aspect-count mixture, course coverage, and observed style cues, while Table 4 provides the corresponding exact summary values.

6.1 Corpus Profile and Representative Examples

Figure 4 and Figure 5 establish the corpus-level structure before any model results are introduced. Figure 4 shows how label support is distributed across aspects and sentiment polarities, which makes it easier to distinguish heavily represented pedagogical dimensions from thinner parts of the schema. Figure 5 complements that view with review length, aspect-count mixture, course coverage, and style variation. Table 5 then gives a qualitative sense of how these distributions appear in actual review text. These examples are included to help the reader interpret the benchmark qualitatively, not as a substitute for aggregate model evidence.

Table 5. Representative synthetic review snippets showing how aspect labels appear under different writing styles and pedagogical situations.
Example type Style Aspect labels Excerpt
Tradeoff-heavy Neutral everyday prose difficulty=positive, materials=positive "I took Computer Networks mainly to fit my schedule ... the material felt accessible at times ... Dr. Chen’s delivery varies week to week ... the LMS was smooth, but feedback came too late to be useful."
Compressed complaint Analytic but Simple interest=negative "I stuck with Cyber Security Basics only when I needed it ... the lead-in theory never clicked for me ... the few but heavy deadlines left me frustrated but still fair about the effort."
Pedagogical frustration Analytic but Simple assessment_design=negative, relevance=negative "Data Structures felt like a required hurdle more than a doorway to real understanding ... the assessments were exhausting and overbearing ... even when I passed a task, I was not convinced it reflected useful learning."

6.2 Scope of Reported Analyses

Table 6. Evidence map for the reported analyses. Unlike Table 3, which organizes the method families, this table aligns each analysis block with its data source, split, and interpretive role in the study.
Analysis block Data and split Primary outputs Interpretive role
Synthetic benchmark10,000 generated reviews with an 8,000 / 1,000 / 1,000 train-validation-test splitDetection F1, sentiment MSE, runtime, per-aspect summaries, and robustness analysesPrincipal evidence for learnability and model comparison on the 20-aspect benchmark.
GPT-based inferenceThe same 1,000-review held-out synthetic test split, evaluated through batch promptingStructured sparse JSON predictions and full-test detection/sentiment metricsInference-time comparison under the same label contract as the learned local models.
External mapped evaluation2,829 mapped Herath student-feedback reviews over the shared 9-aspect overlapTransfer metrics, overlap support counts, and overlap-matched synthetic-versus-real comparisonsConservative real-data check on whether synthetic supervision connects to one annotated educational feedback space.
Realism studyThree 60-item cycles built from public OMSCS reviews and matched synthetic reviewsJudge accuracy, confusion, entropy, cue tags, and prompt-state transitionsGenerator-side validation showing how realism cues were diagnosed and reduced across cycles.
Faithfulness auditA 250-review full-corpus sample plus the 25-review pilot subsetAspect support rates and sentiment-match rates between declared labels and textual evidenceClarifies how the benchmark should be interpreted as a useful but imperfect supervised resource.

Table 6 therefore complements rather than repeats Table 3. Table 3 says which method families are executed, while Table 6 says what each analysis block is meant to establish, which split it uses, and how it should be interpreted alongside the main benchmark.

6.3 Internal Benchmark on the 10K / 20-Aspect Corpus

The main benchmark uses the 10,000-review corpus with a strict 8,000/1,000/1,000 train-validation-test split. Seven local approaches are reported: TF-IDF, four two-step transformer encoders, and two joint encoders that predict aspect presence and polarity together. Among the untuned models, bert-base-uncased gives the strongest held-out micro-F1 at 0.2760 with a detected-aspect sentiment MSE of 0.4959, followed by distilbert-base-uncased at 0.2691. The joint variants remain competitive but lower, with distilbert_joint at 0.2524 and bert_joint at 0.2447, which supports the two-step decomposition as the clearest baseline family in this benchmark.

Additional robustness analyses sharpen that picture. Across three seeds for TF-IDF, DistilBERT, and BERT, the highest mean detection score comes from BERT at \(0.2791 \pm 0.0140\) micro-F1, while DistilBERT remains the most stable model at \(0.2694 \pm 0.0005\). A modest lower-rate four-epoch BERT schedule further raises the held-out score to 0.2930 and lowers sentiment MSE to 0.4728, whereas the same change does not improve DistilBERT. Appendix A.9 reports the full seed-stability, joint-versus-two-step, and tuning tables and figures that support these observations.

Absolute scores remain moderate, which is itself informative. The benchmark covers 20 aspects across five pedagogical groups, includes subtle categories such as feedback_quality, peer_interaction, and prerequisite_fit, and allows one to three aspects per review across multiple writing styles and course contexts. The ranking also has a concrete error-profile interpretation. BERT and DistilBERT improve primarily by lifting recall above 0.43 while retaining enough precision to prevent indiscriminate overprediction. ALBERT and RoBERTa, by contrast, drift toward recall-heavy detectors whose false positives then inflate downstream sentiment error. In this sense, Table 7 shows that the benchmark rewards models that recover many aspect mentions without collapsing into broad positive prediction.

The additional diagnostics tell a fuller story than F1 alone. On the principal local comparison, BERT and DistilBERT are nearly tied in macro balanced accuracy at 0.6229 and 0.6207 respectively, while BERT now also holds the strongest untuned sentiment MSE and macro MCC. This is useful because the 20-aspect benchmark contains many negative label decisions per review, so balanced accuracy and specificity clarify whether a model is improving by recovering minority positives or merely by exploiting the dominant negatives. Appendix A.10 reports these complementary diagnostics for the principal synthetic-benchmark, GPT-based, and mapped real-data comparisons.

Table 7 provides the exact values for the executed local benchmark. Figure 6 serves a different role: instead of repeating the table cell-by-cell, it visualizes the main trade-offs among detection quality, recall, runtime, and sentiment error. This separation is intentional so that the table remains the precise archival record while the figure highlights the comparative structure of the result space.

Table 7. Main held-out test results on the 10K / 20-aspect synthetic benchmark, ranked by detection micro-F1.
Rank Approach Micro-F1 Macro-F1 Micro-recall Sentiment MSE Runtime (min)
1bert-base-uncased0.27600.33640.43960.495921.86
2distilbert-base-uncased0.26910.33760.45310.504415.95
3distilbert_joint0.25240.32480.47190.54286.29
4bert_joint0.24470.32080.51220.528811.58
5tfidf_two_step0.23260.28670.45950.68300.10
6albert-base-v20.18290.18281.00000.577322.33
7roberta-base0.18290.18281.00000.683822.19
Benchmark comparison
Figure 6. Local-benchmark trade-offs on the 10K / 20-aspect corpus. The left panel visualizes the cost-quality frontier using runtime, micro-F1, and sentiment MSE, while the right panel contrasts micro-F1 with micro-recall to show how some models gain coverage at the cost of more false positives.

6.4 GPT-Based Inference Methods

GPT-based ABSA inference is also evaluated as a full-test executed method family. Using OpenAI Batch mode with gpt-5.2, we ran zero-shot, fixed few-shot, diverse few-shot, and retrieval-based few-shot prompting over the entire 1,000-review held-out test split under the same structured output contract used throughout the benchmark. The strongest configuration is zero-shot structured prompting with a micro-F1 of 0.2519, followed closely by retrieval-based few-shot prompting at 0.2501, fixed few-shot at 0.2450, and diverse few-shot at 0.2374.

These GPT runs use the same gold label schema as the local benchmark but a different inference path. The batch job uses an exact-key sparse JSON contract so that outputs can only contain recognized aspect names and ternary sentiments, and all four variants achieved a parse success rate of 1.0 on the full test split. The zero-shot variant uses no demonstrations, the fixed few-shot variant uses three static labeled demonstrations from the synthetic training split, the diverse few-shot variant uses five demonstrations deliberately chosen to vary aspect count and tone, and retrieval-based few-shot prompting replaces static demonstrations with nearest-neighbor examples from the training split.

Table 7A. Full-test GPT-based ABSA inference results under the same structured aspect-sentiment output contract used in the batch evaluation pipeline.
Approach Micro-F1 Macro-F1 Micro-recall Sentiment MSE
gpt-5.2 zero-shot0.25190.24170.31350.7179
gpt-5.2 retrieval-few-shot0.25010.23950.31000.7244
gpt-5.2 few-shot0.24500.23390.30450.7325
gpt-5.2 few-shot-diverse0.23740.22610.29460.7386

The ranking is informative in its own right. Under this strict sparse-schema contract, zero-shot prompting and retrieval-based few-shot prompting are the strongest GPT variants, which suggests that example selection matters but that a highly constrained zero-shot formulation is already competitive. Relative to the local benchmark, the full-test GPT runs outperform TF-IDF and both joint encoders while remaining below the strongest two-step transformers. This places batch GPT inference in a meaningful middle band of the benchmark rather than at the edges of the comparison space. Appendix A.6 records the configuration choices that distinguish the four GPT-based inference variants.

The confusion-based diagnostics reinforce that interpretation. The strongest GPT variants reach macro balanced accuracy near 0.59 with macro specificity around 0.87, which indicates conservative but well-controlled prediction behavior: the batch prompts avoid broad overprediction, but they still miss many positive aspects that the stronger local encoders recover. In other words, GPT-based inference is competitive partly because it remains selective rather than because it dominates recall. Appendix A.10 includes these complementary values alongside the main F1-based ranking.

6.5 Per-Aspect Behavior of the Best Model

The per-aspect profile of the best untuned model, BERT, is uneven in a way that matches the pedagogical complexity of the label space. Stronger aspect categories concentrate in the assessment-and-management and learning-demand blocks, especially workload, grading_transparency, exam_fairness, pacing, and tooling_usability. Harder categories cluster in the instructional-quality, learning-environment, and engagement blocks, including peer_interaction, support, interest, feedback_quality, and clarity, which likely require more implicit reasoning and weaker lexical cues. This pattern shows that the benchmark is not uniformly easy even when labels are known by construction.

Best-model per-aspect benchmark
Figure 7. Per-aspect detection F1 and sentiment MSE for the best untuned local model, BERT, on the held-out test split. The figure separates the strongest and weakest aspect groups, making it easier to see where the benchmark remains lexically explicit versus pedagogically latent.
Table 8. Strongest and weakest aspect categories for the best local model, showing where the benchmark is comparatively explicit or comparatively latent.
Model Group Aspect F1 Sentiment MSE Precision Recall
bert-base-uncasedtop-5workload0.58640.37270.64370.5385
bert-base-uncasedtop-5grading_transparency0.55130.58380.89580.3981
bert-base-uncasedtop-5exam_fairness0.53160.37350.75000.4118
bert-base-uncasedtop-5pacing0.50930.31990.50000.5189
bert-base-uncasedtop-5tooling_usability0.48730.48920.53330.4486
bert-base-uncasedbottom-5clarity0.20060.59260.11660.7172
bert-base-uncasedbottom-5feedback_quality0.19610.59170.11170.8049
bert-base-uncasedbottom-5interest0.18140.53220.11140.4889
bert-base-uncasedbottom-5support0.18050.69770.10490.6458
bert-base-uncasedbottom-5peer_interaction0.13850.84160.39130.0841

6.6 External Validation on a Mapped Real Student-Feedback Dataset

The study also includes one external validation on the annotated student-feedback dataset of Herath et al. [11]. We conservatively mapped the original annotation scheme to the subset of our pedagogical schema with defensible correspondence, yielding a 9-aspect overlap consisting of accessibility, assessment_design, exam_fairness, grading_transparency, lecturer_quality, materials, organization, overall_experience, and workload. This produced 2,829 mapped real reviews, which were used purely as an external evaluation set: no real reviews were used during synthetic training, validation-threshold calibration, or prompt refinement.

The transfer result is encouraging and should be read as a focused external validation. On this overlap benchmark, bert-base-uncased achieved the strongest detection micro-F1 at 0.4593, followed by distilbert-base-uncased at 0.4156 and tfidf_two_step at 0.3740. DistilBERT produced the lowest detected-aspect sentiment MSE at 0.3888, slightly better than BERT's 0.3990. These numbers are materially stronger than the internal 20-aspect benchmark scores, which is plausible because the external test covers fewer aspects, includes a dominant lecturer_quality signal, and uses a conservative overlap mapping rather than the full pedagogical schema. The result therefore supports partial synthetic-to-real transfer on one mapped educational annotation space.

To interpret that result more carefully, we also ran an overlap-matched internal comparison using the same 9 aspects on the synthetic corpus. On that overlap slice, BERT rises from 0.3869 micro-F1 internally to 0.4811 on the mapped real set, DistilBERT rises from 0.3809 to 0.4156, and TF-IDF stays roughly flat at 0.3811 versus 0.3740. This pattern indicates that the mapped real benchmark is shaped by the conservative overlap definition and by strong support for externally visible categories such as lecturer_quality, rather than by a uniform domain-shift penalty. Accordingly, the real-data result is best interpreted as a constrained overlap evaluation that demonstrates compatibility between the synthetic supervision scheme and one real educational annotation space. Table 8B makes this comparison explicit side by side, while Appendix A.4 lists the overlap counts and label balance and Appendix A.5 expands the result into a per-aspect transfer view.

The same comparison becomes more interpretable when balanced metrics are included. On the mapped real benchmark, BERT achieves the strongest macro balanced accuracy among the compared transfer models at 0.5925, slightly above DistilBERT at 0.5778 and clearly above TF-IDF at 0.5403. This supports the main transfer conclusion from another angle: the best synthetic-trained encoder is not merely matching the real overlap through dominant negative decisions, but is retaining a stronger balance between sensitivity and specificity across the nine mapped aspects. Appendix A.10 reports these values explicitly.

Table 8A. Exact synthetic-to-real transfer scores on the mapped Herath benchmark over the conservative nine-aspect overlap. The table gives the precise model values, while Figure 8 adds the overlap-support context needed to interpret them.
Rank Approach Micro-F1 Macro-F1 Micro-recall Sentiment MSE Real reviews Overlap aspects
1bert-base-uncased0.45930.30590.62110.399028299
2distilbert-base-uncased0.41560.35150.67640.388828299
3tfidf_two_step0.37400.23030.40170.701928299
External validation on mapped real student feedback
Figure 8. External validation on mapped real student feedback. The left panel provides overlap-support context from the mapped Herath corpus, while the center and right panels visualize the transfer scores reported exactly in Table 8A. The figure supports a partial-transfer claim only for the conservative 9-aspect overlap.
Overlap-matched internal versus external comparison
Figure 9. Overlap-matched internal versus external comparison on the same nine aspects. The absence of a uniform external drop indicates that the mapped real benchmark is partly shaped by overlap support and aspect composition rather than only by domain shift.
Table 8B. Side-by-side comparison on the shared nine-aspect space, contrasting the same models on the synthetic overlap test split and the mapped real-data benchmark.
Approach Synthetic overlap micro-F1 Mapped real micro-F1 Δ real minus synthetic Synthetic overlap sentiment MSE Mapped real sentiment MSE
bert-base-uncased0.38690.4811+0.09420.51690.3914
distilbert-base-uncased0.38090.4156+0.03460.48110.3888
tfidf_two_step0.38110.3740-0.00710.65190.7019
Table 8C. Prior real-data educational sentiment results included for context; the final column states how directly each study compares to the present benchmark.
Study Setting Reported metric Task Comparability to this study
Welch and Mihalcea [7]Real student comments with automatically extracted entitiesF1 = 0.586Targeted sentiment toward courses and instructorsLow; entity-targeted sentiment is narrower than review-level 20-aspect ABSA
Welch and Mihalcea [7]Real student comments with ground-truth entitiesF1 = 0.695Targeted sentiment with gold entitiesLow; useful upper-bound reference, but not a directly matched benchmark
Herath et al. [11]Real student feedback baselineF1 = 0.750Aspect-level sentiment analysisMedium; closest real educational sentiment benchmark, but different label unit and task structure
Herath et al. [11]Real student feedback best tree-based ALSCF1 = 0.710Aspect-level sentiment classification with FT-RoBERTa induced RGATMedium; stronger educational baseline under their own annotation protocol
This studyMapped Herath overlap after synthetic trainingmicro-F1 = 0.4593Synthetic-to-real transfer on a 9-aspect overlapMedium; the closest comparison here because it uses real educational reviews, but it still depends on conservative aspect mapping

Table 8C is included to prevent overinterpretation of the transfer result. Our mapped-overlap score should not be read as “worse than Herath” or “better than Welch” in any simple leaderboard sense, because the task definitions differ substantially. Welch and Mihalcea study targeted sentiment toward extracted course or instructor entities, while Herath et al. evaluate aspect-level sentiment under their own annotation structure. The most defensible reading is narrower: the reported result shows that models trained only on the synthetic corpus recover some real educational signal on one mapped real benchmark, but the paper does not yet claim parity with prior real-data educational sentiment systems under their native tasks.

6.7 Corpus-Scale Generation Diagnostics

The next three subsections return from downstream model behavior to generator quality. They document how the 10K corpus behaves as generated data: first in terms of output control at corpus scale, then in terms of label faithfulness, and finally in terms of the three-cycle realism procedure that shaped the final prompt specification.

6.7A Full-Corpus Output Control

The full dataset is large enough to support a realistic internal benchmark, and it also reveals how output control behaves at scale. All 10,000 rows returned usable review text and no duplicates were detected. At the same time, 841 rows were marked incomplete because they reached the output-token cap, and overall length-band adherence settled at 0.6819 over the full corpus. The resulting dataset is therefore broad and usable while still leaving room for stronger long-form control in later releases.

This behavior does not invalidate the benchmark, because every row still contains the target labels and usable text, and the train-validation-test pipeline runs cleanly on the assembled corpus. It does, however, matter for interpretation: stronger long-form length control would further improve the controllability of future releases. Figure 5 makes this visible directly by combining length, aspect-count, course, and style evidence in one place.

6.7B Model-Assisted Label-Faithfulness Audit

A complementary label-faithfulness analysis asks whether generated reviews visibly express the aspect-sentiment labels they were assigned, not only whether the text looks realistic. We therefore ran a conservative model-assisted audit with gpt-5.2 as a strict checker over two samples: 250 reviews from the 10K corpus and all 25 reviews from the pilot sample. For each review, the auditor received the raw text and declared aspect sentiments, then judged whether each declared aspect was supported in the text and whether the polarity matched.

Table 8D. Model-assisted label-faithfulness audit showing how often declared aspects and polarities are visibly supported in generated text.
Split Audit model Reviews Declared aspects Aspect support rate Aspect sentiment-match rate Full-row support rate Full-row sentiment-match rate
Full 10K samplegpt-5.22505010.77050.42320.59200.2120
Pilot samplegpt-5.225440.77270.31820.60000.2800

These results make the character of the corpus clearer. Many declared aspects are visibly present in the text, but polarity faithfulness is weaker, especially at full-corpus scale. At the same time, the larger 250-review audit is more favorable than the earlier smaller probe, which suggests that the benchmark preserves declared aspect presence more consistently than the harsher initial sample implied, even though polarity drift remains substantial. The benchmark is therefore strongest as a controllable and useful educational ABSA resource whose labels remain informative but imperfect. This interpretation supports benchmarking and method comparison while also motivating future filtering or post-generation verification. Appendix A.8 retains the detailed audit table for direct comparison with later regenerated versions of the corpus.

6.7C Three-Cycle Realism-Validation Procedure

The realism study remains diagnostic, and it is reported here as a complete three-cycle procedure rather than as a single pass. Each cycle used 60 independent judge questions: 30 real OMSCS reviews and 30 synthetic reviews generated from the richer attribute prompt, with binary REAL-versus-SYNTHETIC decisions, confidence scores, cue tags, and textual justification. Across cycles 0, 1, and 2, judge item accuracy was 0.5167, 0.4333, and 0.4333; mean confusion was 0.5737, 0.6417, and 0.6430; mean entropy was 0.6912, 0.7020, and 0.7068 bits; and the number of correctly detected synthetic reviews fell from 5 to 1 to 0. These diagnostics indicate that the stabilized prompt reduced several synthetic cues, even though the accuracy-based chance-confusion statistic was not monotonic across cycles.

The final cycle does not support an equivalence claim. In cycle 2, the judge accuracy remained 0.4333, the exact binomial test against chance yielded p = 0.366294, and the Wilson 95% interval was [0.3157, 0.5590], which is too wide for the predefined ±0.10 equivalence margin. The value of the procedure is therefore diagnostic rather than confirmatory. Judge explanations repeatedly highlighted generic specificity, overbalanced structure, overpolished prose, stacked motifs, and occasional persona inconsistencies; those cues were then used by the editor step to stabilize the next-cycle prompt. The realism analysis therefore contributes a cycle-level improvement curve, cue summaries, prompt states, and side-by-side appendix examples rather than only a final binary verdict.

Table 9. Realism-study summary across the three complete judge-editor cycles, including accuracy, confusion, entropy, and whether prompt rewriting was triggered.
Prompt state Cycle Judge accuracy Mean confusion Mean entropy (bits) Binomial p-value Editor triggered
rich_attributes_baseline00.51670.57370.69120.8974yes
reduce_synthetic_signatures10.43330.64170.70200.3663yes
messier_realism20.43330.64300.70680.3663no

To make the realism analysis auditable, the appendix includes side-by-side real and synthetic review examples together with the synthetic review's sampled aspects and nuance attributes. The paper therefore reports not only the final cycle summary, but also the prompt instructions and qualitative evidence used to justify revision between cycles; the paired examples appear in Appendix A.2.

Realism validation and prompt refinement pipeline
Figure 10. The realism-validation loop. A judge evaluates one blended pool of real and synthetic reviews, the resulting metrics and cue evidence are summarized, and prompt revision affects the next cycle.

6.8 Overall Interpretation

Taken together, the study presents a coherent account of controlled synthetic supervision for educational ABSA: corpus characterization, qualitative examples, a documented ABSA modeling pipeline, calibrated 20-aspect discriminative baselines, a mapped 9-aspect synthetic-to-real evaluation on annotated student feedback, and a three-cycle realism study. The evidence supports internal learnability on synthetic reviews, informative differences among modeling choices, and partial transfer to one real educational corpus, while the label-faithfulness audit clarifies how the benchmark can be interpreted productively as a useful but noisy resource.

7. Discussion

The results support interpreting the study as a synthetic educational ABSA benchmark together with a strong internal modeling analysis and one conservative external transfer check. The strongest evidence lies where those pieces connect directly: controlled generation produces aspect-labeled educational reviews, and a downstream ABSA pipeline tests whether those labels are learnable under a strict train-validation-test regime.

The study is therefore strongest as an investigation of dataset construction and internal learnability under synthetic supervision. Its value lies in offering a controllable benchmark resource for a domain where public aspect-labeled data remain scarce, together with a transparent account of how that benchmark connects to one real educational annotation space.

7.1 Limitations

Five limitations remain central. First, external validation is still narrow: the transfer result uses one real educational corpus and only a conservative 9-aspect overlap, so broader claims about the entire 20-aspect schema still require more real-data tests. Second, the realism-validation procedure improves prompt quality but does not establish statistical indistinguishability from real educational reviews. Third, the 10K dataset is usable but not perfectly controlled, because a nontrivial share of rows reached the output-token cap and weakened length-band adherence at scale. Fourth, the label-faithfulness audit indicates that many declared aspect polarities are expressed only approximately in the generated text, so the corpus is best treated as a useful noisy benchmark rather than as a gold-standard annotation set. Fifth, the GPT-based comparison now covers a full-test gpt-5.2 benchmark, but it still represents one provider model and one structured prompting family rather than the full space of proprietary LLM baselines.

These limitations help define the next stage of work rather than undercut the contribution of the study. The reported evidence supports a resource-and-benchmark study with meaningful generator validation, informative internal comparisons, and a first mapped real-data check.

The most important extensions are broader synthetic-to-real tests across more real educational datasets and more of the 20-aspect schema, improved generation or filtering to raise label faithfulness and sentiment correctness on generated reviews, and broader prompt-based comparisons that add decomposition variants and cross-provider LLM baselines on the same fixed benchmark split.

7.2 Educational Implications

Although this study is framed as a benchmark-and-corpus contribution, the underlying motivation is practical: institutions accumulate large volumes of course feedback, and aspect-level summaries are pedagogically more useful than overall sentiment. A controllable synthetic resource has three complementary roles in this setting. First, it lowers the cost of model bootstrapping. Programs that cannot fund the per-comment annotation required for a private ABSA model can pretrain on synthetic data calibrated to their own aspect schema and fine-tune on a small adjudicated slice of real reviews. Second, it supports longitudinal monitoring. Course-improvement cycles benefit from comparable aspect distributions across terms, instructors, and cohorts; a documented synthetic benchmark gives the modeling layer a stable yardstick against which a department-specific model can be tracked. Third, it allows controlled stress-testing. By construction the synthetic corpus exposes minority aspects (e.g. accessibility, grading_transparency, peer_interaction) that are rarely well represented in any single institution's archive, so it can act as a probing set for known coverage gaps. None of these uses presume that synthetic supervision replaces human annotation; rather, the corpus is a scaffold that makes the model-development stage cheaper and the evaluation stage more reproducible.

The label-faithfulness audit constrains those use cases. With overall aspect-sentiment match at 0.42 on a 250-review sample, the corpus should be treated as a noisy training resource rather than as a calibrated ground truth, and any deployment built on top of it should be checked against locally adjudicated examples before consequential decisions, especially for sentiment polarity. The faithfulness-aware filtering extension flagged in Section 7.1 is the natural next step for institutional users who plan to base course-improvement signals on the model's outputs.

7.3 Ethics Statement

The synthetic corpus released with this study contains no identifiable student data: instructor names, course codes, and personas are sampled or invented as part of the controlled-generation protocol, and the corpus is generated rather than scraped. The mapped external evaluation uses the publicly released Herath et al. 2022 student-feedback corpus under its MIT license; that release was prepared by the authors of the original study with their institutional approvals, and our use is limited to evaluation, citation, and a conservative schema mapping documented at paper/real_transfer/herath_mapping.json. Preliminary realism-validation work also drew on the public OMSCS course-review pages; only public text was used and no individual reviewer is identified in any downstream artifact. Any institutional deployment of the resulting models should re-consent or re-anonymize student-authored text per local research-ethics policy and should make clear to students and instructors that an automated aspect summary is produced from feedback they submit. We do not recommend using model outputs as inputs to high-stakes personnel decisions about instructors without human review.

8. Conclusion

This study contributes a 10K synthetic educational ABSA corpus over a 20-aspect pedagogical schema, a documented generation protocol, and a benchmark setting that makes internal train-validation-test evaluation possible in a domain where public labeled data remain difficult to obtain. It also includes a conservative synthetic-to-real evaluation on mapped annotated student feedback, which supports partial transfer on a 9-aspect overlap, and generator validations that clarify how realism control and label faithfulness behave at corpus scale. Taken together, the evidence supports controlled synthetic supervision as a productive way to study educational ABSA, compare model families, and establish an openly reusable benchmark in a data-scarce domain.

Appendix

A.1 Twenty-Aspect Inventory

Table A1 lists the 20-aspect inventory used for generation, realism validation, and benchmarking, grouped into the five pedagogical blocks discussed in Sections 3 and 6.

Table A1. Twenty-aspect pedagogical inventory used throughout generation, validation, and benchmarking, grouped into five pedagogical blocks.
Group Aspect Description
Instructional qualityclarityHow understandable the teaching and explanations feel.
Instructional qualitylecturer_qualityPerceived quality of the lecturer or lead instructor.
Instructional qualitymaterialsUsefulness of slides, notes, readings, and resources.
Instructional qualityfeedback_qualityUsefulness and timeliness of feedback on student work.
Assessment and course managementexam_fairnessWhether exams feel aligned and fair.
Assessment and course managementassessment_designAlignment and structure of assignments, projects, and exams.
Assessment and course managementgrading_transparencyHow clearly grading criteria, rubrics, and score interpretation are communicated.
Assessment and course managementorganizationAdministrative clarity, course structure, and coordination.
Assessment and course managementtooling_usabilityFriction or support created by LMS, submission systems, and required software.
Learning demand and readinessdifficultyConceptual or technical challenge of the course.
Learning demand and readinessworkloadAmount of sustained effort required across the term.
Learning demand and readinesspacingWhether the course tempo and weekly rhythm are manageable.
Learning demand and readinessprerequisite_fitHow well the course matches the advertised prerequisite level and student preparation.
Learning environmentsupportQuality of help from instructor, TAs, or forums.
Learning environmentaccessibilityPerceived accessibility and inclusiveness of materials, pace, and course participation.
Learning environmentpeer_interactionWhether peer discussion, teamwork, and class community help or hinder learning.
Engagement and valuerelevancePerceived usefulness to the program or future goals.
Engagement and valueinterestLevel of engagement or curiosity the course creates.
Engagement and valuepractical_applicationConnection to real-world practice or authentic tasks.
Engagement and valueoverall_experienceGlobal student impression after tradeoffs.

A.2 Side-by-Side Real and Synthetic Examples

Table A2 shows representative real and synthetic review pairs from the realism study. For the synthetic reviews, the table includes both the target aspect labels and the sampled nuance attributes used to construct the prompt.

Table A2. Representative real and synthetic review pairs from the realism study. Each pair is grouped into a short real row and a matched synthetic row so the review text remains readable while the synthetic example still exposes its label and prompt context.
Source Review text Annotation context
Pair 0
Real
I took this class my first semester in the program, alongside networks (6250). I almost feel like I shouldn’t review this – I didn’t take advantage of the resources it offered. The instructor and TAs all seemed great and response time to questions and concerns was near immediate. But I acted like they didn’t exist except the few times I had a specific question, and I suspect my learning suffered for it. The lectures comprise the bulk of the learning material, along with journal articles assigned for most lectures.
Public real review used only as a blinded realism-validation item.
Synthetic
CS-6200 / Graduate Introduction to Operating Systems was useful for me, but honestly not a fun class. I took it my last semester before graduation and it definitely felt like one of those mid-program core courses that everyone should probably do anyway. What I remember most is one systems assignment week where I was up way too late debugging something dumb, and the peer chat helped more than the actual course materials.
Target aspects
{"lecturer_quality": "negative", "relevance": "positive"}
Sampled nuance attributes
{"course_name": "Graduate Introduction to Operating Systems", "lecturer_name": "Prof. Alvarez", "course_level": "mid-program core course", "semester_stage": "last semester before graduation", "attendance_pattern": "used the materials selectively when needed", "collaboration_structure": "peer discussion helps more than the official materials", "platform_and_tooling": "the LMS and submission flow are smooth", "prerequisite_fit": "starts accessible but ramps up sharply", "support_channel_experience": "email support is slow when deadlines hit"}
Pair 1
Real
TLDR: This was a great course, got a B overall but thoroughly enjoyed it. This course requires a lot of time and prerequisite knowledge of C and C++. Exams are disconnected from the project work. 25+ hours/week. This was my first class in OMSCS. I wanted a challenge and I definitely got it.
Public real review used only as a blinded realism-validation item.
Synthetic
Honestly the platform stuff in CS-6200 was a pain and caused way more stress than it needed to. I took Graduate Introduction to Operating Systems mostly because it fit my schedule, not because OS was my thing, and that probably showed. I have a pretty strong coding background, but the theory side was rough for me, so I kept falling behind and then basically crammed around project deadlines.
Target aspects
{"interest": "positive", "accessibility": "negative"}
Sampled nuance attributes
{"course_name": "Graduate Introduction to Operating Systems", "student_background": "strong coding background but weak theory", "attendance_pattern": "fell behind and caught up around deadlines", "course_level": "advanced elective", "motivation_for_taking_course": "fit the schedule more than the interests", "platform_and_tooling": "platform quirks create avoidable stress", "feedback_timing": "scores arrive fast but explanations are thin", "collaboration_structure": "collaboration expectations feel unclear"}

A.3 Prompt Instructions Across Realism-Validation Cycles

The realism-validation procedure rewrites the prompt between complete 60-question cycles. The prompt used in the next cycle is the editor-refined output from the previous cycle when correctly detected synthetic reviews provide sufficient cue evidence.

Table A3. Prompt instructions across realism-study cycles, showing how the generation instruction changed as synthetic cues were identified and reduced.
Stage Prompt instruction
Baseline cycle-0 instruction
Write a realistic first-person course review with uneven detail. Avoid textbook sentiment wording, avoid obvious label leakage, and keep the tone consistent with the sampled student persona.
Cycle-1 stable instruction
Write a realistic first-person course review with uneven detail and a mildly informal voice. Include 1-2 concrete, course-plausible specifics (for example a project, tool, rubric quirk, deadline pattern, exam format, or memorable incident), but do not force every aspect to appear. Avoid textbook sentiment wording, explicit label leakage, neat pros/cons symmetry, and generic praise/complaint lists. Let the review sound a little partial or messy, like recalled experience rather than a polished summary, and end naturally with a complete final thought. Keep all details consistent with the sampled student persona and the actual course/instructor context.
Cycle-2 stable instruction
Write a realistic first-person course review in a mildly informal voice. Make it feel like recalled experience, not a balanced evaluation: focus on 1-2 things the student would actually remember, with uneven detail and some partialness. Include at most 1-2 concrete, course-plausible specifics (such as a project, tool, grading quirk, deadline pattern, exam format, or small incident), and make at least one of them individualized rather than just subject jargon. Do not try to cover every aspect or balance praise and criticism; avoid tidy contrast patterns, stacked common review motifs, generic domain-term lists, and polished summary phrasing. Keep the sentiment and details consistent with the sampled student persona and the real course/instructor context, and end with a natural complete final sentence.
Final generation instruction
You are writing one realistic student course review for research validation.
The review must feel like a naturally written student comment rather than a labeled synthetic sample.

Target aspect sentiments:
{aspect_block}

Target attributes:
{attribute_block}

Requirements:
- Keep the review first-person and specific.
- Do not mention aspect labels or sentiment labels explicitly.
- Do not force a tidy conclusion.
- Do not cover every aspect with the same level of detail.
- Let at least one point feel incidental rather than checklist-driven.
- Preserve mixed feelings when the attributes imply them.
- Additional stable realism instruction: Write a realistic first-person course review in a mildly informal voice. Make it feel like recalled experience, not a balanced evaluation: focus on 1-2 things the student would actually remember, with uneven detail and some partialness. Include at most 1-2 concrete, course-plausible specifics (such as a project, tool, grading quirk, deadline pattern, exam format, or small incident), and make at least one of them individualized rather than just subject jargon. Do not try to cover every aspect or balance praise and criticism; avoid tidy contrast patterns, stacked common review motifs, generic domain-term lists, and polished summary phrasing. Keep the sentiment and details consistent with the sampled student persona and the real course/instructor context, and end with a natural complete final sentence.

Return only the review text.

A.4 Mapped Real-Data Overlap and Label Balance

Table A4 documents the conservative overlap used for external validation against the Herath student-feedback corpus. The mapped real benchmark is intentionally narrower than the full 20-aspect schema, because only defensible correspondences were retained.

Table A4. Exact support and polarity balance for the nine overlap aspects retained in the mapped Herath benchmark. This appendix table preserves the full counts behind the lighter overlap context summarized in Figure 8.
Aspect Reviews Positive Neutral Negative
lecturer_quality21901402398390
overall_experience55739561101
organization47726615556
materials39013719459
assessment_design2358310052
exam_fairness184299758
grading_transparency146496631
workload75122142
accessibility35111212

A.5 Per-Aspect Synthetic-to-Real Diagnostics

Figure A1 expands the external-validation result into an aspect-by-model matrix. It is included in the appendix because it is diagnostically valuable, but too detailed for the main body.

Per-aspect F1 heatmap on the mapped real benchmark
Figure A1. Per-aspect detection F1 for each synthetic-trained model on the mapped real benchmark. The heatmap shows that cross-dataset transfer is highly uneven: lecturer_quality and exam_fairness transfer more cleanly than sparse or weakly aligned categories such as accessibility.

A.6 GPT-Based Inference Configuration Details

Table A5 complements Table 7A by documenting the configuration differences behind the four GPT-based inference variants rather than repeating the same held-out scores. All four methods used the same exact-key sparse JSON contract and achieved parse success of 1.00 on the full 1,000-review test split.

Table A5. Configuration details for the GPT-based inference methods reported in Table 7A.
Approach Demonstration policy Context selection Test reviews Parse success Intended comparison
gpt-5.2 zero-shotNo demonstrationsSchema instructions only10001.00Measures the strongest constrained inference baseline without example conditioning.
gpt-5.2 retrieval-few-shotNearest-neighbor labeled demonstrationsExamples retrieved from the synthetic training split per test review10001.00Tests whether instance-level example selection improves over fixed prompting.
gpt-5.2 few-shotThree static labeled demonstrationsFixed examples drawn from the synthetic training split10001.00Reference few-shot condition with stable prompt context across the full test set.
gpt-5.2 few-shot-diverseFive static demonstrations with varied tone and aspect countCurated fixed examples from the synthetic training split10001.00Tests whether broader example diversity changes the precision-recall trade-off.

A.7 Pilot Subset Procedure-Validation Details

The pilot subset is included to document the small-scale validation used before full-scale generation, even though it is far too small for substantive model comparison. Table A6 records the acceptance criteria and Figure A2 shows the sampled aspect-count mix used in that validation subset.

Table A6. Pilot-subset validation criteria used before full-scale generation.
Criterion Value Interpretation
Pilot subset25 reviewsSmall-scale generation used for end-to-end validation of the data contract
Split sizes21 / 2 / 2Train / validation / test
Completed rate1.00All pilot responses completed
Text success rate1.00All rows returned parsable review text
Duplicate rate0.00No duplicate reviews detected
Length-band match rate0.80Length control cleared the predeclared pilot threshold
Mean review length125.6 wordsObserved average output length under the validated generation settings
Aspect count distribution in the pilot validation subset
Figure A2. Distribution of sampled target-aspect counts in the pilot validation subset. This appendix figure is retained because it documents the pilot mixture without duplicating the acceptance criteria already summarized in Table A6.

A.8 Label-Faithfulness Audit Details

Table A7 records the strict model-assisted label-faithfulness audit used to check whether declared aspect sentiments are visibly supported in the generated text. The main body reports the headline implication; this appendix keeps the exact audited rates visible for future comparison after regeneration or filtering.

Table A7. Detailed label-faithfulness audit rates for the full-corpus sample and the pilot sample.
Split Audit model Reviews Declared aspects Aspect support rate Aspect sentiment-match rate Full-row support rate Full-row sentiment-match rate
Full 10K samplegpt-5.22505010.77050.42320.59200.2120
Pilot samplegpt-5.225440.77270.31820.60000.2800

A.9 Additional Local-Benchmark Robustness Analyses

Tables A8-A10 and Figures A3-A4 collect the robustness analyses cited in Section 6.3. They show how the two-step and joint formulations compare directly, how stable the strongest local models are across three seeds, and how a modest training-budget change affects BERT and DistilBERT.

Table A8. Joint-versus-two-step comparison on the held-out synthetic benchmark.
Family Approach Micro-F1 Macro-F1 Micro-recall Sentiment MSE Runtime (min)
Two-stepbert-base-uncased0.27600.33640.43960.495921.86
Two-stepdistilbert-base-uncased0.26910.33760.45310.504415.95
Jointdistilbert_joint0.25240.32480.47190.54286.29
Jointbert_joint0.24470.32080.51220.528811.58
Seed stability across the strongest local models
Figure A3. Micro-F1 across three random seeds for TF-IDF, DistilBERT, and BERT. The figure highlights the strong stability of DistilBERT and the larger upside variance of BERT.
Table A9. Three-seed stability summary for the strongest local benchmark models.
Approach Seeds Micro-F1 mean Micro-F1 std Sentiment MSE mean Sentiment MSE std
bert-base-uncased30.27910.01400.50040.0296
distilbert-base-uncased30.26940.00050.50970.0188
tfidf_two_step30.23530.00580.66670.0158
Training-budget tuning comparison for BERT and DistilBERT
Figure A4. Effect of a modest training-budget change on the two strongest transformer baselines. BERT benefits noticeably from the longer lower-rate schedule, while DistilBERT remains essentially flat.
Table A10. Targeted training-budget comparison for the two strongest transformer baselines.
Approach Baseline micro-F1 Tuned micro-F1 Δ micro-F1 Baseline sentiment MSE Tuned sentiment MSE Δ sentiment MSE
bert-base-uncased0.27600.2930+0.01700.49590.4728-0.0231
distilbert-base-uncased0.26910.2683-0.00090.50440.5239+0.0195

A.10 Additional Detection Diagnostics

Table A11 complements the headline F1 metrics with three confusion-based summaries that are less sensitive to label sparsity: macro balanced accuracy, macro specificity, and macro Matthews correlation. These values are especially useful in the present 20-aspect multilabel setting because they show whether a model is improving through better minority-positive recovery, through conservative rejection of negatives, or through a more balanced combination of both.

Table A11. Additional confusion-based detection diagnostics for the principal synthetic-benchmark, GPT-based, and mapped real-data comparisons.
Family Approach Micro-F1 Macro-F1 Macro balanced accuracy Macro specificity Macro MCC Sentiment MSE
Local synthetic benchmarkbert-base-uncased0.27600.33640.62290.80500.27660.4959
Local synthetic benchmarkdistilbert-base-uncased0.26910.33760.62070.78630.27130.5044
Local synthetic benchmarktfidf_two_step0.23260.28670.59550.72250.19200.6830
GPT batch inferencegpt-5.2 zero-shot0.25190.24170.58990.86860.17990.7179
GPT batch inferencegpt-5.2 retrieval-few-shot0.25010.23950.58830.86930.18230.7244
GPT batch inferencegpt-5.2 few-shot0.24500.23390.58480.86790.17980.7325
GPT batch inferencegpt-5.2 few-shot-diverse0.23740.22610.58000.86730.16530.7386
Mapped real-data transferbert-base-uncased0.45930.30590.59250.83270.18740.3990
Mapped real-data transferdistilbert-base-uncased0.41560.35150.57780.79760.21820.3888
Mapped real-data transfertfidf_two_step0.37400.23030.54030.79920.11620.7019

Bibliography

  1. J.W. Gikandi, D. Morrow, and N.E. Davis. 2011. Online formative assessment in higher education: A review of the literature. Computers & Education, 57(4), 2333-2351.
  2. Maria Pontiki, Haris Papageorgiou, Dimitrios Galanis, Ion Androutsopoulos, John Pavlopoulos, and Suresh Manandhar. 2014. SemEval-2014 Task 4: Aspect Based Sentiment Analysis. Proceedings of the 8th International Workshop on Semantic Evaluation, pages 27-35.
  3. Maria Pontiki, Dimitris Galanis, Haris Papageorgiou, Ion Androutsopoulos, Suresh Manandhar, Mohammad AL-Smadi, Mahmoud Al-Ayyoub, Yanyan Zhao, Bing Qin, Orphée De Clercq, Véronique Hoste, Marianna Apidianaki, Xavier Tannier, Natalia Loukachevitch, Evgeniy Kotelnikov, Nuria Bel, Salud María Jiménez-Zafra, and Gülşen Eryiğit. 2016. SemEval-2016 Task 5: Aspect Based Sentiment Analysis. Proceedings of SemEval-2016, pages 19-30.
  4. Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. Proceedings of NAACL-HLT 2019, pages 4171-4186.
  5. Yice Zhang, Yifan Yang, Bin Liang, Shiwei Chen, Bing Qin, and Ruifeng Xu. 2023. An Empirical Study of Sentiment-Enhanced Pre-Training for Aspect-Based Sentiment Analysis. Findings of ACL 2023, pages 9633-9651.
  6. Zhuoyan Li, Hangxiao Zhu, Zhuoran Lu, and Ming Yin. 2023. Synthetic Data Generation with Large Language Models for Text Classification: Potential and Limitations. EMNLP 2023, pages 10443-10461.
  7. Charles Welch and Rada Mihalcea. 2016. Targeted Sentiment to Understand Student Comments. COLING 2016, pages 2471-2481.
  8. Janaka Chathuranga, Shanika Ediriweera, Ravindu Hasantha, Pranidhith Munasinghe, and Surangika Ranathunga. 2018. Annotating Opinions and Opinion Targets in Student Course Feedback. LREC 2018 Workshop paper.
  9. Nikola Nikolić, Olivera Grljević, and Aleksandar Kovačević. 2020. Aspect-based sentiment analysis of reviews in the domain of higher education. The Electronic Library, 38(1), 44-64.
  10. Thanveer Shaik, Xiaohui Tao, Yan Li, Christopher Dann, Jacquie McDonald, Petrea Redmond, and Linda Galligan. 2022. A Review of the Trends and Challenges in Adopting Natural Language Processing Methods for Education Feedback Analysis. IEEE Access, 10, 56720-56739.
  11. Missaka Herath, Kushan Chamindu, Hashan Maduwantha, and Surangika Ranathunga. 2022. Dataset and Baseline for Automatic Student Feedback Analysis. Proceedings of the 13th Conference on Language Resources and Evaluation, pages 2042-2049.
  12. Ting-Wei Hsu, Chung-Chi Chen, Hen-Hsen Huang, and Hsin-Hsi Chen. 2021. Semantics-Preserved Data Augmentation for Aspect-Based Sentiment Analysis. Proceedings of EMNLP 2021, pages 4067-4079.
  13. Timo Schick and Hinrich Schütze. 2021. Exploiting Cloze-Questions for Few-Shot Text Classification and Natural Language Inference. Proceedings of EACL 2021, pages 255-269.
  14. Hang Yan, Junqi Dai, Tuo Ji, Xipeng Qiu, and Zheng Zhang. 2021. A Unified Generative Framework for Aspect-Based Sentiment Analysis. Proceedings of ACL-IJCNLP 2021, pages 2416-2429.
  15. Jian Liu, Zhiyang Teng, Leyang Cui, Hanmeng Liu, and Yue Zhang. 2021. Solving Aspect Category Sentiment Analysis as a Text Generation Task. Proceedings of EMNLP 2021, pages 4406-4416.
  16. Michelangelo Misuraca, Germana Scepi, and Maria Spano. 2021. Using Opinion Mining as an educational analytic: An integrated strategy for the analysis of students’ feedback. Studies in Educational Evaluation, 68, 100979.
  17. Thanveer Shaik, Xiaohui Tao, Christopher Dann, Haoran Xie, Yan Li, and Linda Galligan. 2023. Sentiment analysis and opinion mining on educational data: A survey. Natural Language Processing Journal, 2, 100003.
  18. Charalampos Dervenis, Giannis Kanakis, and Panos Fitsilis. 2024. Sentiment analysis of student feedback: A comparative study employing lexicon and machine learning techniques. Studies in Educational Evaluation, 83, 101406.
  19. Aleksandra Edwards and Jose Camacho-Collados. 2024. Language Models for Text Classification: Is In-Context Learning Enough?. Proceedings of LREC-COLING 2024, pages 10448-10461.
  20. John Hattie and Helen Timperley. 2007. The Power of Feedback. Review of Educational Research, 77(1), 81-112.
  21. David Carless and David Boud. 2018. The development of student feedback literacy: enabling uptake of feedback. Assessment & Evaluation in Higher Education, 43(8), 1315-1325.
  22. Michael Henderson, Michael Phillips, Tracii Ryan, David Boud, Phillip Dawson, Elizabeth Molloy, and Paige Mahoney. 2019. Conditions that enable effective feedback. Higher Education Research & Development, 38(7), 1401-1416.
  23. Swapna Gottipati, Venky Shankararaman, and Jeff Rongsheng Lin. 2018. Text analytics approach to extract course improvement suggestions from students’ feedback. Research and Practice in Technology Enhanced Learning, 13, 6.
  24. Maija Hujala, Antti Knutas, Timo Hynninen, and Heli Arminen. 2020. Improving the quality of teaching by utilising written student feedback: A streamlined process. Computers & Education, 157, 103965.
  25. Steven Y. Feng, Varun Gangal, Jason Wei, Sarath Chandar, Soroush Vosoughi, Teruko Mitamura, and Eduard Hovy. 2021. A Survey of Data Augmentation Approaches for NLP. Findings of ACL 2021, pages 968-988.
  26. Yu Meng, Jiaxin Huang, Yu Zhang, and Jiawei Han. 2022. Generating Training Data with Language Models: Towards Zero-Shot Language Understanding. Advances in Neural Information Processing Systems 35.
  27. Yizhong Wang, Yeganeh Kordi, Swaroop Mishra, Alisa Liu, Noah A. Smith, Daniel Khashabi, and Hannaneh Hajishirzi. 2023. Self-Instruct: Aligning Language Models with Self-Generated Instructions. Proceedings of ACL 2023, pages 13484-13508.
  28. Siddharth Varia, Shuai Wang, Kishaloy Halder, Robert Vacareanu, Miguel Ballesteros, Yassine Benajiba, Neha Anna John, Rishita Anubhai, Smaranda Muresan, and Dan Roth. 2023. Instruction Tuning for Few-Shot Aspect-Based Sentiment Analysis. Proceedings of WASSA 2023, pages 28-37.
  29. Robert Vacareanu, Siddharth Varia, Kishaloy Halder, Shuai Wang, Giovanni Paolini, Neha Anna John, Miguel Ballesteros, and Smaranda Muresan. 2024. A Weak Supervision Approach for Few-Shot Aspect Based Sentiment Analysis. Proceedings of EACL 2024, pages 2381-2395.
  30. Wenxuan Zhang, Yue Deng, Bing Liu, Sinno Jialin Pan, and Lidong Bing. 2024. Sentiment Analysis in the Era of Large Language Models: A Reality Check. Findings of NAACL 2024, pages 3892-3910.

Reproducibility note: generation artifacts, local benchmark outputs, and realism-study files are kept distinct throughout the text.