Enhanced fine-tuning is coming! How to make AI truly "understand" human needs

Written by
Audrey Miles
Updated on:June-25th-2025
Recommendation

To explore new ways for AI to understand human needs, reinforcement fine-tuning technology deserves attention.

Core content:
1. The role and core elements of reinforcement learning in AI
2. The difference and advantages between reinforcement learning and supervised learning
3. The concept, process and significance of reinforcement fine-tuning to the development of AI

Yang Fangxian
Founder of 53A/Most Valuable Expert of Tencent Cloud (TVP)

In today's rapidly developing AI field, how to make models better understand human needs and provide truly valuable services has always been the direction of developers' efforts. The emergence of reinforcement finetuning has undoubtedly brought new hope to this goal. It combines human feedback and reinforcement learning to enable the model to continuously adjust its behavior to better meet human values ​​and expectations. Today, let's take a deep look at the mystery of reinforcement finetuning and see how it plays an important role in modern AI development.

1. Reinforcement Learning: The Cornerstone of Reinforcement Fine-tuning

Before we delve into reinforcement fine-tuning, we first need to understand its core principle - reinforcement learning. Unlike traditional supervised learning, reinforcement learning does not rely on clear correct answers, but guides AI system learning through rewards and penalties. In this process, the AI ​​system is called an "agent", which generates actions through interaction with the environment, and adjusts its own strategy based on the rewards fed back by the environment to maximize the cumulative rewards.

The four core elements of reinforcement learning are as follows:

  • Agent : A learning system, such as our language model.
  • Environment : The context in which the agent is located. For language models, this includes input prompts and task specifications.
  • Actions : Responses or outputs produced by an agent.
  • Rewards : Feedback signals used to indicate whether a behavior is good or bad.

By continuously interacting with the environment and receiving reward signals, the agent gradually learns a policy, which is a method of selecting behaviors to maximize expected rewards.

2. Reinforcement Learning and Supervised Learning: A Paradigm Shift

To better understand the value of reinforcement fine-tuning, let’s first compare the characteristics of reinforcement learning and supervised learning:

Features
Supervised Learning
Reinforcement Learning
Learning Signal
Correct label/answer
Quality-based rewards
Feedback timing
Immediately and clearly
Delayed, sometimes sparse
Target
Minimize prediction error
Maximize cumulative rewards
Data requirements
Annotation example
Reward Signal
Training process
One-time optimization
Interactive, iterative exploration

While supervised learning relies on a clear correct answer for each input, reinforcement learning guides learning through a more flexible reward signal. This flexibility makes reinforcement fine-tuning particularly important when optimizing language models, where “correctness” is often subjective and context-dependent.

3. What is enhanced fine-tuning?

Reinforcement fine-tuning refers to improving the pre-trained language model through reinforcement learning technology to make it better conform to human preferences and values. Unlike traditional training methods, reinforcement fine-tuning does not only focus on the accuracy of predictions, but optimizes the model to produce outputs that humans consider useful, harmless, and honest. This method solves the problem that traditional training objectives are difficult to specify clearly.

Human feedback plays a central role in reinforcement fine-tuning. Humans evaluate the quality of the model output, such as whether it is helpful, accurate, safe, and natural in tone. These evaluation results generate reward signals that guide the model in the direction of human preferences. The typical reinforcement fine-tuning workflow is as follows:

  1. Start with a pretrained language model : Choose a model that has been pretrained and fine-tuned with supervision.
  2. Generate responses : The model generates multiple responses to various prompts.
  3. Collecting human preferences : Human evaluators rank or score these responses.
  4. Train reward model : Use these evaluation results to train a reward model that can predict human preferences.
  5. Reinforcement Learning Fine-tuning : Use reinforcement learning to optimize the original model to maximize the prediction reward.
  6. Validation : Test the improved model on the retained samples to ensure its generalization ability.

4. How Enhanced Fine-tuning Works

Reinforcement fine-tuning improves model performance by generating responses, collecting feedback, training a reward model, and optimizing the original model. The following are the detailed steps of the reinforcement fine-tuning workflow:

1. Preparing the dataset

First, a diverse set of prompts covering the target domain needs to be carefully curated and an assessment benchmark created.

2. Response Generation

The model generates multiple responses to each prompt, which are used for subsequent human evaluation.

3. Human Evaluation

Human raters rank or score these responses based on quality criteria—for example, assessing whether a response is more helpful, more accurate, or safer.

4. Reward model training

The role of the reward model is to act as a proxy for human judgment. It receives the prompt and response as input and outputs a scalar value representing the predicted human preference. The following is a simplified pseudocode for reward model training:

def train_reward_model (preference_data, model_params) : 
    for  epoch  in  range(EPOCHS):
        for  prompt, better_response, worse_response  in  preference_data:
            # Get the reward prediction values ​​for the two responses
            better_score = reward_model(prompt, better_response, model_params)
            worse_score = reward_model(prompt, worse_response, model_params)
            
            # Calculate the log probability of the correct preference
            log_prob = log_sigmoid(better_score - worse_score)
            
            # Update the model to increase the probability of the correct preference
            loss = -log_prob
            model_params = update_params(model_params, loss)
    
    return  model_params

5. Applied Reinforcement Learning

Reinforcement fine-tuning can be achieved using a variety of algorithms, such as:

  • Proximal Policy Optimization (PPO) : OpenAI used PPO when fine-tuning the GPT model. It optimizes the policy by limiting the update amplitude to prevent destructive changes to the model.
  • Direct Preference Optimization (DPO) : This method optimizes directly from the preference data without the need for a separate reward model and is more efficient.
  • Reinforcement Learning from AI Feedback (RLAIF) : Using another AI system to provide training feedback can reduce the costs and scale limitations of human feedback.

During optimization, one needs to improve the reward signal while preventing the model from “forgetting” its pre-training knowledge or finding exploitative behaviors that maximize the reward without actually improving upon it.

5. Why is enhanced fine-tuning superior when data is scarce?

When labeled data is limited, reinforcement fine-tuning shows many advantages:

  • Learning from preferences : Reinforcement fine-tuning can learn from judgments about the output, not just from what the ideal output is.
  • Efficient use of feedback : Through the generalization ability of reward models, a single feedback can guide many related behaviors.
  • Policy Exploration : Reinforcement fine-tuning can discover novel response patterns that are not present in the training examples.
  • Dealing with ambiguity : When there are multiple valid responses, reinforcement fine-tuning can maintain diversity rather than averaging to a safe but bland middle ground.

Therefore, even without a fully labeled dataset, enhanced fine-tuning can produce more helpful and natural models.

6. Strengthen the key advantages of fine-tuning

Enhanced fine-tuning brings many significant advantages to AI models, making them more valuable in practical applications.

1. Better conformity with human values

Through iterative feedback, the model is able to learn the subtleties of human preferences that are difficult to specify explicitly through programming. Reinforcement fine-tuning enables the model to better understand:

  • Appropriate tone and style
  • Moral and ethical considerations
  • Cultural sensitivity
  • Helpful and manipulative responses

This alignment process makes the model a more trustworthy and helpful partner, rather than just a powerful prediction engine.

2. Adaptability to Specific Tasks

While retaining general capabilities, the model after reinforcement fine-tuning can focus on a specific domain by incorporating domain-specific feedback. This allows the model to:

  • Implementing customized assistant behaviors
  • Demonstrate expertise in fields such as medicine, law or education
  • Provide customized responses for specific user groups

The flexibility of reinforcement fine-tuning makes it ideal for creating purpose-built AI systems without having to start from scratch.

3. Long-term performance improvement

Models trained with enhanced fine-tuning tend to maintain performance better across scenarios because they optimize for essential qualities rather than superficial patterns. This leads to the following benefits:

  • Generalizes better to new topics
  • More consistent quality across different inputs
  • More robust to cue changes

4. Reducing hallucinations and harmful outputs

By explicitly punishing undesirable outputs, reinforcement nudging significantly reduces problem behaviors:

  • Fabricating information is negatively rewarded
  • Harmful, offensive or misleading content is suppressed
  • Honest uncertainty is reinforced, not confident misrepresentation

5. More helpful and detailed responses

Most importantly, reinforcement nudges produced responses that users actually found more valuable:

  • Better understanding of implicit requirements
  • Further reasoning
  • Appropriate level of detail
  • A balanced view of complex issues

These improvements make the reinforced fine-tuned models even more useful as assistants and sources of information.

VII. Variants of Enhanced Fine-tuning and Related Techniques

There are many different ways to implement reinforcement fine-tuning, each with its own unique advantages and application scenarios.

RLHF (Reinforcement Learning from Human Feedback)

RLHF is a classic implementation of reinforcement fine-tuning, with preference signals provided by human evaluators. Its workflow is usually as follows:

  1. Humans compare the model outputs and choose the superior response.
  2. Use these preferences to train a reward model.
  3. The language model is optimized via PPO (Proximal Policy Optimization) to maximize the expected reward.

The following is a simplified code implementation of RLHF:

def train_rihf (model, reward_model, dataset, optimizer, ppo_params) : 
   # PPO hyperparameters
   kl_coef = ppo_params[ 'kl_coef' ]
   epochs = ppo_params[ 'epochs' ]

   for  prompt  in  dataset:
       # Generate a response using the current strategy
       responses = model.generate_responses(prompt, n= 4 )
      
       # Get rewards from the reward model
       rewards = [reward_model(prompt, response)  for  response  in  responses]
      
       # Calculate the logarithmic probability of the response under the current strategy
       log_probs = [model.log_prob(response, prompt)  for  response  in  responses]
      
       for  _  in  range(epochs):
           # Update the policy to increase the probability of high reward responses
           # While staying close to the original strategy
           new_log_probs = [model.log_prob(response, prompt)  for  response  in  responses]
          
           # Strategy ratio
           ratios = [torch.exp(new - old)  for  new, old  in  zip(new_log_probs, log_probs)]
          
           # PPO clipping target and KL penalty
           kl_penalties = [kl_coef * (new - old)  for  new, old  in  zip(new_log_probs, log_probs)]
          
           # Strategy loss
           policy_loss = -torch.mean(torch.stack([
               ratio * reward - kl_penalty
               for  ratio, reward, kl_penalty  in  zip(ratios, rewards, kl_penalties)
           ]))
          
           # Update the model
           optimizer.zero_grad()
           policy_loss.backward()
           optimizer.step()   
   return  model

RLHF has made breakthrough progress in aligning language models with human values, but its scalability faces challenges due to the bottleneck of human labeling.

2. DPO (Direct Preference Optimization)

DPO simplifies the process of reinforcement fine-tuning by eliminating separate reward models and PPO optimizations. The following is the code implementation of DPO:

import  torch
import  torch.nn.functional  as  F


def dpo_loss (model, prompt, preferred_response, rejected_response, beta) : 
   # Calculate the log probability of two responses
   preferred_logprob = model.log_prob(preferred_response, prompt)
   rejected_logprob = model.log_prob(rejected_response, prompt)
  
   # Calculate loss, encouraging preferred responses > rejected responses
   loss = -F.logsigmoid(beta * (preferred_logprob - rejected_logprob))
  
   return  loss

Benefits of DPO include:

  • Easier implementation, fewer components
  • Training dynamics are more stable
  • Usually more sample efficient

3. RLAIF (Reinforcement Learning from AI Feedback)

RLAIF replaces human raters with another AI system that is trained to mimic human preferences. This approach:

  • Significantly reduced feedback collection costs
  • Scalable to larger datasets
  • Maintaining consistency in assessment criteria

The following is the code implementation of RLAIF:

import  torch


def train_with_rlaif (model, evaluator_model, dataset, optimizer, config) : 
   """
   Fine-tune the model using RLAIF (Reinforcement Learning from AI Feedback)
  
   parameter:
   - model: the language model being fine-tuned
   - evaluator_model: The AI ​​model trained to evaluate the response
   - dataset: a collection of prompts for generating responses
   - optimizer: optimizer for model update
   - config: dictionary containing 'batch_size' and 'epochs'
   """

   batch_size = config[ 'batch_size' ]
   epochs = config[ 'epochs' ]

   for  epoch  in  range(epochs):
       for  batch  in  dataset.batch(batch_size):
           # Generate multiple candidate responses for each prompt
           all_responses = []
           for  prompt  in  batch:
               responses = model.generate_candidate_responses(prompt, n= 4 )
               all_responses.append(responses)
          
           # Let the estimator model score each response
           all_scores = []
           for  prompt_idx, prompt  in  enumerate(batch):
               scores = []
               for  response  in  all_responses[prompt_idx]:
                   # AI evaluator provides quality scores based on defined criteria
                   score = evaluator_model.evaluate(
                       prompt,
                       response,
                       criteria=[ "helpfulness""accuracy""harmlessness" ]
                   )
                   scores.append(score)
               all_scores.append(scores)
          
           # Optimize the model to increase the probability of high scoring responses
           loss =  0
           for  prompt_idx, prompt  in  enumerate(batch):
               responses = all_responses[prompt_idx]
               scores = all_scores[prompt_idx]
              
               # Find the best response according to the estimator
               best_idx = scores.index(max(scores))
               best_response = responses[best_idx]
              
               # Increase the probability of the best response
               loss -= model.log_prob(best_response, prompt)
          
           # Update the model
           optimizer.zero_grad()
           loss.backward()
           optimizer.step()

   return  model

Despite the potential introduction of bias in the estimator model, RLAIF shows promising results when the estimator is well calibrated.

4. Constitutional AI

Constitutional AI adds another layer to reinforcement fine-tuning by introducing explicit principles or “constitutions” to guide the feedback process. This approach:

  • Providing more consistent guidance
  • Make value judgments more transparent
  • Reduce reliance on individual annotator bias

The following is a simplified code implementation of Constitution AI:

def train_constitutional_ai (model, constitution, dataset, optimizer, config) : 
   """
   Fine-tuning the model using the Constitutional AI approach


   - model: the language model being fine-tuned
   - constitution: the set of principles used to evaluate responses
   - dataset: a collection of prompts for generating responses
   """

   principles = constitution[ 'principles' ]
   batch_size = config[ 'batch_size' ]


   for  batch  in  dataset.batch(batch_size):
       for  prompt  in  batch:
           # Generate initial response
           initial_response = model.generate(prompt)


           # Self-criticism phase: the model evaluates its response according to the constitution
           critiques = []
           for  principle  in  principles:
               critique_prompt =  f"""
               Principle:  {principle[ 'description' ]}


               Your response:  {initial_response}


               Does this response violate the principle? If so, explain how:
               """

               critique = model.generate(critique_prompt)
               critiques.append(critique)


           # Modification phase: the model improves its response based on criticism
           revision_prompt =  f"""
           Original prompt:  {prompt}


           Your initial response:  {initial_response}


           Critiques of your response:
           { ' ' .join(critiques)}


           Please provide an improved response that addresses these critiques:
           """

           improved_response = model.generate(revision_prompt)


           # Training the model directly produces improved responses
           loss = -model.log_prob(improved_response | prompt)


           # Update the model
           optimizer.zero_grad()
           loss.backward()
           optimizer.step()


   return  model

Anthropic pioneered this approach when it developed its Claude model, focusing on principles such as being helpful, doing no harm, and being honest.

8. Fine-tune the LLM practice using enhanced fine-tuning

Implementing reinforcement fine-tuning requires choosing between different algorithmic approaches (RLHF/RLAIF vs. DPO), reward model types, and appropriate optimization procedures (such as PPO).

RLHF/RLAIF vs. DPO

When implementing reinforcement fine-tuning, practitioners need to choose between different algorithmic approaches:

Features
RLHF/RLAIF
DPO
Components
Separate Reward Model + RL Optimization
Single-stage optimization
Implementation complexity
Higher (multi-stage training)
Lower (direct optimization)
Computing requirements
Higher (PPO required)
Lower (single loss function)
Sample efficiency
Lower
Higher
Control over training dynamics
More clear
Less clear

Organizations should choose between these methods based on their specific constraints and goals. OpenAI has historically used RLHF for reinforcement fine-tuning of its models, and recent research shows the effectiveness of DPO with less computational overhead.

2. Categories of Human Preference Reward Models

The reward model for reinforcement fine-tuning can be trained on various types of human preference data:

  • Binary comparison : humans choose between two model outputs (A vs B).
  • Likert scale rating : Humans rate responses numerically.
  • Multi-attribute evaluation : separate ratings are given for different qualities (e.g. helpfulness, accuracy, safety).
  • Free-form feedback : Converting qualitative comments into quantitative signals.

Different feedback types have a trade-off between annotation efficiency and signal richness. Many enhanced fine-tuning systems use a combination of multiple feedback types to capture different aspects of quality.

3. Using PPO for enhanced fine-tuning

PPO (Proximal Policy Optimization) is a popular algorithm for reinforcement fine-tuning due to its stability. The process involves:

  1. Initial sampling : Generates responses using the current strategy.
  2. Reward calculation : Responses are scored using a reward model.
  3. Advantage estimation : Compare rewards to a baseline to determine which behaviors perform better than average.
  4. Policy update : Optimize the policy to increase the probability of high reward output.
  5. KL divergence constraint : Prevents the model from deviating too much from the initial version, avoiding catastrophic forgetting or degradation.

Through this balancing mechanism, PPO improves model performance while ensuring that the model does not lose its original knowledge and capabilities due to over-optimization.

9. Strengthening fine-tuning practice in mainstream LLM

Today, enhanced fine-tuning has become a key step in the training process of many mainstream large language models (LLMs). Here are some typical application cases:

1. OpenAI’s GPT series

OpenAI was one of the first companies to apply reinforcement fine-tuning at scale. Their GPT model implements reinforcement fine-tuning in the following way:

  1. Collect a large amount of human preference data : Obtain human evaluation of model output through crowdsourcing and other means.
  2. Iteratively optimize the reward model : Continuously improve the accuracy of the reward model based on human feedback.
  3. Multi-stage training : Enhanced fine-tuning is used as the final alignment step to ensure that the model can match human values ​​after large-scale pre-training.

For example, both GPT-3.5 and GPT-4 have undergone extensive hardening fine-tuning, significantly improving the usefulness and safety of the models while reducing harmful outputs.

2. Anthropic Claude Model

Anthropic introduces explicit principles into the reinforcement fine-tuning process through its unique constitutional AI approach. The training process of the Claude model is as follows:

  1. Human Preference-Based Initial RLHF : Training a reward model with feedback from human evaluators.
  2. Constitutional reinforcement learning : Use explicit principles to guide the feedback process, ensuring that model behavior conforms to a specific ethical framework.
  3. Multiple rounds of improvement : Iteratively refine the model, focusing on principles such as helpfulness, non-harm, and honesty.

This approach makes the Claude model perform well within a specific ethical framework, demonstrating the great potential of enhanced fine-tuning in achieving specific value alignment.

3. Google DeepMind’s Gemini model

Google's Gemini model extends reinforcement fine-tuning to the multimodal domain. Its training process includes:

  1. Multimodal preference learning : Combine feedback from multiple modalities such as text and images to optimize the overall performance of the model.
  2. Security-focused fine-tuning : The reward model is specifically designed to improve the safety and reliability of the model.
  3. Reward models for different capabilities : Customize the reward model for different functions of the model to ensure that each aspect is optimal.

The practice of the Gemini model shows that reinforcement fine-tuning can not only be applied to text generation, but also play an important role in multimodal scenarios.

4. Meta’s LLaMA series

Meta also introduced enhanced fine-tuning technology in its open source LLaMA model. Their practice shows that:

  1. Enhanced fine-tuning can significantly improve the performance of open source models : by applying RLHF to models of different sizes, the alignment effect of the models is significantly improved.
  2. Public documentation and community extension : Meta attracts extensive community participation and further optimization by publicly disclosing the implementation details of the enhancement fine-tuning.

The practice of the LLaMA series provides a valuable reference for the open source community, demonstrating the great potential of reinforcement fine-tuning in improving the performance of open source models.

5. Mistral and Mixtral variants

Mistral AI introduces enhanced fine-tuning in its model development, focusing on achieving efficient alignment in resource-constrained environments. Their practices include:

  1. Lightweight reward model : An efficient reward model is designed for smaller architectures.
  2. Efficient reinforcement fine-tuning implementation : By optimizing algorithms and processes, the computing cost is reduced.
  3. Open variants : By open sourcing parts of the implementation, the community is encouraged to conduct more extensive experiments and optimizations.

The practices of Mistral and Mixtral show that enhanced fine-tuning can adapt to different resource environments, providing more developers with the opportunity to apply this technology.

10. Challenges and limitations of strengthening fine-tuning

Although reinforcement fine-tuning brings many advantages, it also faces some challenges and limitations in practical applications:

1. The cost and speed of human feedback

  1. Collecting high-quality human preferences requires a lot of resources : annotation work is time-consuming and labor-intensive, and requires professional annotators.
  2. Complex annotator training and quality control : Different annotators may have inconsistent standards, resulting in uneven feedback quality.
  3. Feedback collection becomes an iteration bottleneck : Frequent human feedback requirements limit the rapid iteration speed of the model.
  4. Human judgment can be biased : the subjectivity of annotators can cause the model to learn incorrect preferences.

These issues have motivated researchers to explore synthetic feedback and more efficient preference acquisition methods.

2. Reward hijacking and alignment issues

  1. Models may optimize superficial patterns rather than true preferences : certain behaviors may obtain high rewards by exploiting loopholes in the reward function without actually improving quality.
  2. Complex goals are difficult to express with reward signals : For example, goals such as "realism" are difficult to measure with simple reward functions.
  3. Reward signals can inadvertently reinforce manipulative behavior : If rewards are not designed well, the model may learn to obtain rewards by misleading users.

Researchers are continually improving techniques to detect and prevent this kind of reward hijacking.

3. Explainability and Control

  1. The optimization process is like a “black box” : it is difficult to understand which behaviors are being reinforced by the model, and the changes are scattered throughout the parameters.
  2. Difficulty isolating and modifying specific behaviors : Once a model has been fine-tuned through reinforcement learning, it is difficult to adjust specific aspects.
  3. Difficulty in providing assurances about model behavior : Due to the lack of transparency, it is difficult to ensure that the model behaves as expected in all scenarios.

These explainability challenges pose difficulties in strengthening the governance and oversight of fine-tuning systems.

XI. Latest developments and trends in strengthening fine-tuning

As technology continues to advance, reinforcement fine-tuning is also evolving, and here are some trends to watch:

1. The rise of open source tools and libraries

The implementation of reinforcement fine-tuning increasingly relies on open source tools and libraries, which greatly reduce the entry barrier:

  1. **Transformer Reinforcement Learning (TRL)**: provides a ready-made reinforcement fine-tuning component.
  2. PEFT tool by Hugging Face : supports efficient fine-tuning process.
  3. Community benchmarks : Help standardize model evaluation and facilitate fair comparisons.

These tools and resources make reinforcement tuning more accessible, allowing more developers to apply and improve this technology.

2. The rise of synthetic feedback

In order to break through the limitations of human feedback, synthetic feedback has become an important research direction:

  1. Model-generated criticism and evaluation : Leverage feedback generated by the model itself to guide training.
  2. Guided feedback : Let stronger models evaluate weaker models to achieve "self-improvement".
  3. Hybrid feedback : Combining human feedback and synthetic feedback to balance efficiency and quality.

Widespread use of synthetic feedback is expected to significantly reduce the cost of reinforcement fine-tuning and improve its scalability.

3. Enhanced fine-tuning in multimodal models

As AI models gradually expand from plain text to multimodal fields, reinforcement fine-tuning is also constantly adapting to new application scenarios:

  1. Image Generation : Optimizing image generation models based on human aesthetic preferences.
  2. Video-model alignment : Optimizing the behavior of video generation models via feedback.
  3. Cross-modal alignment : Achieve better consistency between text and other modalities.

These applications demonstrate the powerful flexibility of reinforcement fine-tuning as a general alignment method.

12. Future Prospects for Strengthening Fine-tuning

Reinforcement fine-tuning has played an important role in AI development. It solves the alignment problem that is difficult to solve with traditional methods by directly incorporating human preferences into the optimization process. Looking ahead, reinforcement fine-tuning is expected to achieve greater breakthroughs in the following aspects:

  1. Breaking the human annotation bottleneck : Reducing reliance on human annotation through synthetic feedback and more efficient preference acquisition methods.
  2. Improve model interpretability : Develop a more transparent optimization process so that developers can better understand and control model behavior.
  3. Deepening of multimodal scenarios : In multimodal fields such as images, videos, and voice, enhanced fine-tuning will play a greater role and promote the comprehensive development of AI systems.
  4. Wider application scenarios : From language generation to intelligent decision-making, reinforcement fine-tuning will help AI systems better adapt to various complex scenarios and provide more valuable services to humans.

As technology continues to advance, reinforcement fine-tuning will continue to guide the development of AI models, ensuring that they remain aligned with human values ​​and creating more trustworthy intelligent assistants for humans.

In the world of AI, reinforcement fine-tuning is not only a technical means, but also a concept - to make machines truly understand human needs and become our reliable partners. This is a profound change and a journey full of hope. Let us wait and see how reinforcement fine-tuning will shape the future of AI!