Giant language fashions (LLMs) now drive essentially the most superior conversational brokers, artistic instruments, and decision-support programs. Nonetheless, their uncooked output typically incorporates inaccuracies, coverage misalignments, or unhelpful phrasing—points that undermine belief and restrict real-world utility. Reinforcement Fantastic‑Tuning (RFT) has emerged as the popular technique to align these fashions effectively, utilizing automated reward alerts to switch expensive handbook labeling.
On the coronary heart of recent RFT is reward capabilities. They’re constructed for every area by verifiable reward capabilities that may rating LLM generations by a bit of code (Reinforcement Studying with Verifiable Rewards or RLVR) or with LLM-as-a-judge, the place a separate language mannequin evaluates candidate responses to information alignment (Reinforcement Studying with AI Suggestions or RLAIF). Each these strategies present scores to the RL algorithm to nudge the mannequin to resolve the issue at hand. On this submit, we take a deeper have a look at how RLAIF or RL with LLM-as-a-judge works with Amazon Nova fashions successfully.
Why RFT with LLM‑as‑a-judge in comparison with generic RFT?
Reinforcement Fantastic-Tuning can use any reward sign, easy hand‑crafted guidelines (RLVR), or an LLM that evaluates mannequin outputs (LLM-as-a-judge or RLAIF). RLAIF makes alignment way more versatile and highly effective, particularly when reward alerts are obscure and arduous to craft manually. Not like generic RFT rewards that depend on blunt numeric scoring like substring matching, an LLM decide causes throughout a number of dimensions—correctness, tone, security, relevance—offering context-aware suggestions that captures subtleties and domain-specific nuances with out task-specific retraining. Moreover, LLM judges supply built-in explainability by rationales (for instance, “Response A cites peer-reviewed research”), offering diagnostics that speed up iteration, pinpoint failure modes instantly, and scale back hidden misalignments, one thing static reward capabilities can’t do.
Implementing LLM-as-a-judge: Six essential steps
This part covers the important thing steps concerned in designing and deploying LLM-as-a-judge reward capabilities.
Choose the decide structure
The primary essential choice is deciding on your decide structure. LLM-as-a-judge provides two main analysis modes: Rubric-based (point- based mostly) judging and Desire-based judging, every suited to totally different alignment situations.
Standards
Rubric-based judging
Desire-based judging
Analysis technique
Assigns a numeric rating to a single response utilizing predefined standards
Compares two candidate responses side-by-side and selects the superior one
High quality measurement
Absolute high quality measurements
Relative high quality by direct comparability
Most well-liked used when
Clear, quantifiable analysis dimensions exist (accuracy, completeness, security compliance)
Coverage mannequin ought to discover freely with out reference knowledge restrictions
Knowledge necessities
Solely requires cautious immediate engineering to align the mannequin to reward specs
Requires at the least one response pattern for choice comparability
Generalizability
Higher for out-of-distribution knowledge, avoids knowledge bias
Depends upon high quality of reference responses
Analysis fashion
Mirrors absolute scoring programs
Mirrors pure human analysis by comparability
Beneficial place to begin
Begin right here if choice knowledge is unavailable and RLVR unsuitable
Use when comparative knowledge is offered
Outline your analysis standards
After you’ve chosen your decide kind, articulate the particular dimensions that you simply wish to enhance. Clear analysis standards are the muse of efficient RLAIF coaching.
For Desire-based judges:
Write clear prompts explaining what makes one response higher than one other. Be express about high quality preferences with concrete examples. Instance: “Desire responses that cite authoritative sources, use accessible language, and instantly tackle the person’s query.”
For Rubric-based judges:
We advocate utilizing Boolean (cross/fail) scoring for rubric-based judges. Boolean scoring is extra dependable and reduces decide variability in comparison with fine-grained 1–10 scales. Outline clear cross/fail standards for every analysis dimension with particular, observable traits.
Choose and configure your decide mannequin
Select an LLM with ample reasoning functionality to guage your goal area, configured by Amazon Bedrock and referred to as utilizing a reward AWS Lambda operate. For widespread domains like math, coding, and conversational capabilities, smaller fashions can work effectively with cautious immediate engineering.
Mannequin tier
Most well-liked for
Value
Reliability
Amazon Bedrock mannequin
Giant/Heavyweight
Complicated reasoning, nuanced analysis, multi-dimensional scoring
Excessive
Very Excessive
Amazon Nova Professional, Claude Opus, Claude Sonnet
Medium/Light-weight
Basic domains like math or coding, balanced cost-performance
Low-Medium
Average-Excessive
Amazon Nova 2 Lite, Claude Haiku
Refine your decide mannequin immediate
Your decide immediate is the muse of alignment high quality. Design it to supply structured, parseable outputs with clear scoring dimensions:
- Structured output format – Specify JSON or parseable format for easy extraction
- Clear scoring guidelines – Outline precisely how every dimension needs to be calculated
- Edge case dealing with – Deal with ambiguous situations (for instance, “If response is empty, assign rating 0”)
- Desired behaviors – Explicitly state behaviors to encourage or discourage
Align decide standards with manufacturing analysis metrics
Your reward operate ought to mirror the metrics that you’ll use to guage the ultimate mannequin in manufacturing. Align your reward operate with manufacturing success standards to allow fashions designed for the proper aims.
Alignment workflow:
- Outline manufacturing success standards (for instance, accuracy, security) with acceptable thresholds
- Map every criterion to particular decide scoring dimensions
- Validate that decide scores correlate along with your analysis metrics
- Check the decide on consultant samples and edge instances
Constructing a sturdy reward Lambda operate
Manufacturing RFT programs course of hundreds of reward evaluations per coaching step. Construct a resilient reward Lambda operate to assist present coaching stability, environment friendly compute utilization, and dependable mannequin conduct. This part covers the right way to construct a reward Lambda operate that’s resilient, environment friendly, and manufacturing prepared.
Composite reward rating structuring
Don’t rely solely on LLM judges. Mix them with quick, deterministic reward parts that catch apparent failures earlier than costly decide evals:
Core parts
Element
Goal
When to make use of
Format correctness
Confirm JSON construction, required fields, schema compliance
All the time – catches malformed outputs instantly. Low cost and instantaneous suggestions.
Size penalties
Discourage overly verbose or terse responses
When output size issues (for instance, summaries)
Language consistency
Confirm responses match enter language
Vital for multilingual functions
Security filters
Rule-based checks for prohibited content material
All the time – prevents unsafe content material from reaching manufacturing
Infrastructure readiness
- Implement exponential backoff: Handles Amazon Bedrock API charge limits and transient failures gracefully
- Parallelization technique: Use ThreadPoolExecutor or async patterns to parallelize decide calls throughout rollouts to cut back latency
- Keep away from Lambda chilly begin delays: Set an acceptable Lambda timeout (quarter-hour advisable) and provisioned concurrency (~100 for typical setups)
- Error dealing with: Add complete error dealing with that returns impartial/noisy rewards (0.5) moderately than failing all the coaching step
Check your reward Lambda operate for resilience
Validate decide consistency and calibration:
- Consistency: Check decide on the identical samples a number of instances to measure rating variance (needs to be low for deterministic analysis)
- Cross-judge comparability: Evaluate scores throughout totally different decide fashions to determine analysis blind spots
- Human calibration: Periodically pattern rollouts for human evaluate to catch decide drift or systematic errors
- Regression testing: Create a “decide check suite” with identified good/unhealthy examples to regression check decide conduct
RFT with LLM-as-a-judge – Coaching workflow
The next diagram illustrates the whole end-to-end coaching course of, from baseline analysis by decide validation to manufacturing deployment. Every step builds upon the earlier one, making a resilient pipeline that balances alignment high quality with computational effectivity whereas actively stopping reward hacking and supporting production-ready mannequin conduct.
Actual-world case research: Automating authorized contract evaluate
On this part, we consult with a real-world use case with a number one authorized trade associate. The duty is to generate feedback on dangers, assessments, and actions on authorized documentation with respect to the insurance policies and former contracts as reference paperwork.
Problem
Accomplice was occupied with fixing the issue of automating the method of reviewing, assessing, and flagging dangers in authorized contract paperwork. Particularly, they wished to guage potential new contracts towards inner pointers and rules, previous contracts, and legal guidelines of the nation pertaining to the contract.
Answer
We formulated this downside as one the place we’re offering a goal doc (the “contract” that wants analysis), and a reference doc (the grounding doc and context) and anticipate the LLM to generate a JSON with a number of feedback, remark varieties, and advisable actions to take based mostly on the evaluation. The unique dataset accessible for this use case was comparatively small that included full contracts together with annotations and feedback from authorized consultants. We used LLM as a decide utilizing GPT OSS 120b mannequin because the decide and a customized system immediate throughout RFT.
RFT workflow
Within the following part we cowl particulars of the important thing elements within the RFT workflow for this use case.
Reward Lambda operate for LLM-as-a-judge
The next code snippets current the important thing parts of the reward Lambda operate.
Be aware: title of Lambda operate ought to have “SageMaker”, for instance, “arn:aws:lambda:us-east-1:123456789012:operate:MyRewardFunctionSageMaker“
a) Begin with defining a high-level goal
# Contract Assessment Analysis – Unweighted Scoring
You might be an knowledgeable contract reviewer evaluating AI-generated feedback. Your PRIMARY goal is to evaluate how effectively every predicted remark identifies points within the TargetDocument contract clauses and whether or not these points are justified by the Reference pointers.
b) Outline the analysis strategy
## Analysis Method
For every pattern, you obtain:
– **TargetDocument**: The contract textual content being reviewed (the doc beneath analysis)
– **Reference**: Reference pointers/requirements used for the evaluate (the analysis standards)
– **Prediction**: A number of feedback from the AI mannequin
**Essential**: The SystemPrompt exhibits what directions the mannequin acquired. Contemplate whether or not the mannequin adopted these directions when evaluating the prediction high quality.
**CRITICAL**: Every remark should determine a selected concern, hole, or concern IN THE TARGETDOCUMENT CONTRACT TEXT ITSELF. The remark’s text_excerpt area ought to quote problematic contract language from the TargetDocument, NOT quote textual content from the Reference pointers. The Reference justifies WHY the contract clause is problematic, however the concern should exist IN the contract.
Consider EACH predicted remark independently. Feedback ought to flag issues within the contract clauses, not merely cite Reference necessities.
c) Describe the scoring dimensions with clear specs on how a selected rating needs to be calculated
## Scoring Dimensions (Per Remark)
**EVALUATION ORDER**: Consider on this sequence: (1) TargetDocument_Grounding, (2) Reference_Consistency, (3) Actionability
### 1. TargetDocument_Grounding
**Evaluates**: (a) Whether or not text_excerpt quotes from TargetDocument contract textual content, and (b) Whether or not the remark is related to the quoted text_excerpt
**MANDATORY**: text_excerpt should quote from TargetDocument contract textual content. If text_excerpt quotes from Reference as a substitute, rating MUST be 1.
– **5**: text_excerpt appropriately quotes TargetDocument contract textual content AND remark identifies a extremely related, legitimate, and notable concern in that quoted textual content
– **4**: text_excerpt appropriately quotes TargetDocument contract textual content AND remark identifies a legitimate and related concern in that quoted textual content
– **3**: text_excerpt appropriately quotes TargetDocument contract textual content AND remark is considerably related to that quoted textual content, however concern has reasonable validity
– **2**: text_excerpt appropriately quotes TargetDocument contract textual content BUT remark has weak relevance to that quoted textual content, or concern is questionable
– **1**: text_excerpt does NOT quote TargetDocument contract textual content (quotes Reference as a substitute, or no precise quote), OR remark is irrelevant to the quoted textual content
### 2. Reference_Consistency
…
…
d) Clearly outline the ultimate output format to parse
## Scoring Calculation
**Comment_Score** = Easy common of the three dimensions:
– Comment_Score = (TargetDocument_Grounding + Reference_Consistency + Actionability) / 3
**Aggregate_Score** = Common of all Comment_Score values for the pattern
## Output Format
For every pattern, consider ALL predicted feedback and supply:
“`json
{ “feedback”: [
{ “comment_id”: “…”,
“TargetDocument_Grounding”: {“score”: X, “justification”: “…”, “supporting_evidence”: “Verify text_excerpt quotes actual TargetDocument contract text and comment is relevant to it”},
“Reference_Consistency”: {“score”: X, “justification”: “…”, “supporting_reference”: “Quote from Reference that justifies the concern OR explain meaningful reasoning”},
“Actionability”: {“score”: X, “justification”: “Assess if action is clear, grounded in TargetDocument and Reference, and relevant to comment”},
“Comment_Score”: X.XX
} ],
“Aggregate_Score”: {
“rating”: X.XX,
“total_comments”: N,
“rationale”: “…”
}
}
“`
e) Create a high-level Lambda handler, offering ample multithreading for quicker inference
def lambda_handler(occasion, context):
scores: Record[RewardOutput] = []
samples = occasion
max_workers = len(samples)
print(f”Evaluating {len(samples)} gadgets with {max_workers} threads…”)
with ThreadPoolExecutor(max_workers=max_workers) as executor:
futures = [executor.submit(judge_answer, sample) for sample in samples]
scores = [future.result() for future in futures]
print(f”Accomplished {len(scores)} evaluations”)
return [asdict(score) for score in scores]
Deployment of the Lambda operate
We used the next AWS Identification and Entry Administration (IAM) permissions and settings within the Lambda operate. The next configurations are required for reward Lambda capabilities. RFT coaching can fail if any of them are lacking.
a) Permissions for Amazon SageMaker AI execution position
Your Amazon SageMaker AI execution position will need to have permission to invoke your Lambda operate. Add this coverage to your Amazon SageMaker AI execution position:
{
“Model”: “2012-10-17”,
“Assertion”: [
{
“Effect”: “Allow”,
“Action”: [
“lambda:InvokeFunction”
],
“Useful resource”: “arn:aws:lambda:area:account-id:operate:function-name”
}
]
}
b) Permissions for Lambda operate execution position
Your Lambda operate’s execution position wants primary Lambda execution permissions and the permissions to Invoke the decide Amazon Bedrock mannequin.
Be aware: This resolution follows the AWS shared duty mannequin. AWS is accountable for securing the infrastructure that runs AWS companies within the cloud. You might be accountable for securing your Lambda operate code, configuring IAM permissions, implementing encryption and entry controls, managing knowledge safety and privateness, configuring monitoring and logging, and verifying compliance with relevant rules. Observe the precept of least privilege by scoping permissions to particular useful resource ARNs. For extra data, see Safety in AWS Lambda and Amazon SageMaker AI Safety within the AWS documentation.
c) Add provisioned concurrency
Publish a model of the Lambda and to allow the operate to scale with out fluctuations in latency, we added some provisioned concurrency. 100 was ample on this case, nonetheless, there’s extra room for value enhancements right here.
d) Set Lambda timeout to fifteen minutes
Customizing the coaching configuration
We launched Nova Forge SDK that can be utilized for all the mannequin customization lifecycle—from knowledge preparation to deployment and monitoring. Nova Forge SDK removes the necessity to seek for the suitable recipes or container URI for particular methods.
You need to use the Nova Forge SDK to customise coaching parameters in two methods: present a full recipe YAML utilizing recipe_path or cross particular fields utilizing overrides for selective adjustments. For this use case, we use overrides to tune the rollout and coach settings as proven within the following part.
# Launch coaching with recipe overrides
outcome = customizer.prepare(
job_name=”my-rft-run”,
rft_lambda_arn=””,
overrides={
# Coaching config
“max_length”: 64000,
“global_batch_size”: 64,
“reasoning_effort”: None,
# Knowledge
“shuffle”: False,
# Rollout
“kind”: “off_policy_async”,
“age_tolerance”: 2,
“proc_num”: 6,
“number_generation”: 8,
“max_new_tokens”: 16000,
“set_random_seed”: True,
“temperature”: 1,
“top_k”: 0,
“lambda_concurrency_limit”: 100,
# Coach
“max_steps”: 516,
“save_steps”: 32,
“save_top_k”: 17,
“refit_freq”: 4,
“clip_ratio_high”: 0.28,
“ent_coeff”: 0.0,
“loss_scale”: 1,
},
)
Outcomes
RFT with Amazon Nova 2 Lite achieved a 4.33 mixture rating—the best efficiency throughout all evaluated fashions—whereas sustaining excellent JSON schema validation. This represents a big enchancment, demonstrating that RFT can produce production-ready, specialised fashions that outperform bigger general-purpose alternate options.
We evaluated fashions utilizing a “better of okay” single-comment setting, the place every mannequin generated a number of feedback per pattern and we scored the highest-quality output. This strategy establishes an higher certain on efficiency and permits a good comparability between fashions that produce single versus a number of outputs.
Determine 1 — JSON Schema Validation Scores (0–1 scale, greater is best)
Determine 2 — Mixture LLM decide scores (1–5 scale, greater is best)
Key takeaways:
- RFT achieved the best efficiency amongst evaluated fashions on this research.
Amazon Nova 2 Lite with RFT achieved a 4.33 mixture rating, outperforming each Claude Sonnet 4.5 and Claude Haiku 4.5, whereas additionally attaining excellent JSON schema validation.
- Removes pointless coaching artifacts
Throughout SFT iterations, we noticed problematic behaviors together with repetitive remark era and unnatural Unicode character predictions. These points, probably attributable to overfitting or dataset imbalances, didn’t seem in RFT checkpoints. RFT’s reward-based enhancements naturally discourages such artifacts, producing extra sturdy and dependable outputs.
- Robust generalization to new decide standards
After we evaluated RFT fashions utilizing a modified decide immediate (aligned however not equivalent to the coaching reward operate), efficiency remained sturdy. This demonstrates that RFT learns generalizable high quality patterns moderately than overfitting particular analysis standards. It is a essential benefit for real-world deployment the place necessities evolve.
- Compute concerns
RFT required 4–8 rollouts per coaching pattern, growing compute prices in comparison with SFT. This overhead is amplified when utilizing non-zero reasoning effort settings. Nonetheless, for mission-critical functions the place alignment high quality instantly impacts enterprise outcomes—comparable to authorized contract evaluate, monetary compliance, or healthcare documentation, the efficiency good points justify the extra compute prices.
Conclusion
Reinforcement Fantastic-Tuning (RFT) with LLM-as-a-judge represents a robust strategy to aligning LLMs for domain-specific functions. As demonstrated in our authorized contract evaluate case research, this technique delivers vital enhancements over each base fashions and conventional supervised fine-tuning (SFT) approaches, with RFT attaining the best mixture scores throughout all analysis dimensions. For groups constructing mission-critical AI programs the place alignment high quality instantly impacts enterprise outcomes, RFT with LLM-as-a-judge provides a compelling path ahead. The methodology’s explainability, flexibility, and superior efficiency make it significantly invaluable for advanced domains like authorized evaluate (or Monetary Providers or Healthcare) the place delicate nuances matter.
Organizations contemplating this strategy ought to begin small—validate their decide design on curated benchmarks, confirm infrastructure resilience, and scale step by step whereas monitoring for reward hacking. With correct implementation, RFT can remodel succesful base fashions into extremely specialised, production-ready programs that persistently ship aligned, reliable outputs.
References:
- Amazon Nova Developer Information for Amazon Nova 2
- Nova Forge SDK- GitHub
- Reinforcement Fantastic-Tuning (RFT) with Amazon Nova fashions
Disclaimer:
The authorized contract evaluate use case described on this submit is for technical demonstration functions solely. AI-generated contract evaluation is just not an alternative to skilled authorized recommendation. Seek the advice of certified authorized counsel for authorized issues.
In regards to the authors
Hemanth Kumar Jayakumar is an Utilized Scientist at Amazon AGI, the place he works on reinforcement studying and basis fashions. He interprets the most recent ML analysis into scalable options, unlocking area specialization of basis fashions for patrons. Outdoors of labor, Hemanth enjoys touring and climbing.
Daniel Suarez Souto is a Options Architect at Amazon Net Providers, specializing in Synthetic Intelligence. He helps prospects speed up their AI adoption and construct safe, scalable AI programs end-to-end, turning real-world edge instances into reusable patterns that assist prospects transfer quicker. In his free time, Daniel enjoys taking part in soccer, operating, and climbing.
Ajit Kumar Ok.P. is a Senior Generative AI Accomplice Options Architect at AWS, the place he works with enterprise prospects and companions deploying AI options within the cloud. He brings deep experience bridging the hole between platform engineering and enterprise-scale AI, having constructed Laptop Imaginative and prescient options on the Edge, and AIML and Generative AI options within the Cloud. Ajit enjoys studying biographies and taking part in sports activities in his free time.
Bharathan Balaji is a Senior Utilized Scientist at Amazon Net Providers, engaged on reinforcement studying and basis mannequin companies. His work focuses on constructing AI capabilities that assist prospects remodel their companies.

