Format Training Dataset#

Learn how to format a training dataset to work with the model type you want to train, such as a chat or completion model.

  • Chat Model — Requires data that adheres to the messages schema.

  • Completion Model — Requires data that adheres to the prompt-completion schema.

Customizer expects all datasets to use JSONL format, where each line in the dataset is a training example serialized in JSON.

Chat Models#

Chat and instruction models require additional structure in their data to capture concepts such as multi-turn conversations. We support the the OpenAI messages format.

Format a Conversation Dataset#

Train a chat model to optimize for generating responses using multiple messages as context.

Basic Schema#

A conversational dataset contains a sequence of messages that represent interactions between users and assistants. Each message has:

  • A role field to categorize the message text. Options include system, user, and assistant.

  • A content field for the actual body of information communicated by that role.

Tip

For best training results, the assistant role should be the last message in each training example.

Example entry formatted for JSONL dataset file:

{"messages": [{"role": "system","content": "<system message>"}, {"role": "user","content": "<user message>"}, {"role": "assistant","content": "<assistant message>"}]}
Expanded JSON example

For illustrative purposes only, we show an example entry as multi-line JSON.

{
  "messages": [
    {
      "role": "system",
      "content": "<system message>"
    }, {
      "role": "user",
      "content": "<user message>"
    }, {
      "role": "assistant",
      "content": "<assistant message>"
    }
  ]
}

Reasoning Considerations#

Some models (such as Llama Nemotron) support a detailed thinking mode, which you can toggle in the system message. This setting controls whether the model is encouraged to show step-by-step reasoning in its responses.

  • Training data without reasoning: Use detailed thinking off in the system message.

  • Training data with reasoning: Use detailed thinking on in the system message.

Note

If you have an existing system message that must be preserved, prepend detailed thinking on or detailed thinking off to the beginning of your system message.

For example, if your original system message is You are a helpful assistant., you should use detailed thinking on\nYou are a helpful assistant. or detailed thinking off\nYou are a helpful assistant.

{"messages": [
  {"role": "system", "content": "detailed thinking off"},
  {"role": "user", "content": "What is 2 + 2?"},
  {"role": "assistant", "content": "4"}
]}
{"messages": [
  {"role": "system", "content": "detailed thinking on"},
  {"role": "user", "content": "What is 2 + 2?"},
  {"role": "assistant", "content": "<think>To solve 2 + 2, add 2 and 2 together.</think> The answer is 4."}
]}

You can adjust the system message for each training example to match the style of your data and the behavior you want the model to learn.

Schema with Tool Calling#

Tool calling (also known as function calling) allows the model to directly interact with external systems based on user inputs. By integrating the model with external applications and APIs, you can significantly expand its capabilities to include:

  • Data retrieval from databases or services

  • Execution of specific actions in connected systems

  • Performance of computation tasks

  • Implementation of complex business logic

Inline Tools#

To train a model with tool calling capabilities, use the conversational dataset format with additional fields. Beyond the standard messages structure, you’ll need to define the list of tools the model can access along with their function parameters. Within the messages, include tool_calls when you want the model to invoke a tool with specific arguments.

Every sample must be a single line as in this example below.

{"messages": [{"role": "user","content": ""},{"role": "assistant","content": "","tool_calls": [{"type": "function","function": {"name": "fibonacci","arguments": {"n": 20}}}]}],"tools": [{"type": "function","function": {"name": "fibonacci","description": "Calculates the nth Fibonacci number.","parameters": {"type": "object","properties": {"n": {"description": "The position of the Fibonacci number.","type": "integer"}}}}}]}
Expanded JSON example

For illustrative purposes only, we show an example entry as multi-line JSON.

{
  "messages": [
    {
      "role": "user",
      "content": ""
    },
    {
      "role": "assistant",
      "content": "",
      "tool_calls": [{
        "type": "function",
        "function": {
          "name": "fibonacci",
          "arguments": {"n": 20}
        }
      }]
    }
  ],
  "tools": [{
    "type": "function",
    "function": {
      "name": "fibonacci",
      "description": "Calculates the nth Fibonacci number.",
      "parameters": {
        "type": "object",
        "properties": {
          "n": {
            "description": "The position of the Fibonacci number.",
            "type": "integer"
          }
        }
      }
    }
  }]
}
Shared Tools#

When your dataset uses the same set of tools across all examples, you can streamline your configuration by omitting the tools field from individual dataset entries. Instead, specify these tools once in the dataset_parameters section during job creation.

Important

The config field must include a version, for example: meta/llama-3.2-1b-instruct@v1.0.0+A100. Omitting the version will result in an error like:

{ "detail": "Version is not specified in the config URN: meta/llama-3.2-1b-instruct" }

You can find the correct config URN (with version) by inspecting the output of the /v1/customization/configs endpoint. Use the name and version fields to construct the URN as name@version.

curl -X "POST" \
    "${CUSTOMIZER_SERVICE_URL}/v1/customization/jobs" \
    -H 'accept: application/json' \
    -H 'Content-Type: application/json' \
    -d '{
  "config": "meta/llama-3.2-1b-instruct@v1.0.0+A100",
  "dataset": "namespace/test-dataset",
  "dataset_parameters": {
    "tools": [{
      "type": "function",
      "function": {
          "name": "fibonacci",
          "description": "Calculates the nth Fibonacci number.",
          "parameters": {
            "type": "object",
            "properties": {
            "n": {
              "description": "The position of the Fibonacci number.",
              "type": "integer"
            }
          }
        }
      }
    }]
  },
  "hyperparameters": {
    "training_type": "sft",
    "finetuning_type": "lora",
    "epochs": 10,
    "batch_size": 16,
    "learning_rate": 0.0001,
    "lora": {
      "adapter_dim": 8,
      "adapter_dropout": 0.1
    }
  }
}'

Find Chat Models#

You can perform a GET request to Entity Store’s model registry to discover chat models, which have spec.is_chat set to true.

curl http://nemo.test/v1/models/my-namespace/model-name@version | jq
{
  "name": "model-name@version",
  "namespace": "my-namespace",
  "spec": {
    "num_parameters": 8000000000,
    "context_size": 4096,
    "num_virtual_tokens": 0,
    "is_chat": true
  },
  "artifact": {
    "gpu_arch": "Ampere",
    "precision": "bf16",
    "tensor_parallelism": 1,
    "backend_engine": "nemo",
    "status": "upload_completed",
    "files_url": "hf://my-namespace/model-name@version"
  },
  "base_model": "meta/llama-3.1-8b-instruct",
  "peft": {
    "finetuning_type": "lora"
  }
}

Chat with the Model#

To run inference with a chat model, use the /v1/chat/completions endpoint for the model.

curl -X POST http://nim.test/v1/chat/completions \
  -H 'Accept: application/json' \
  -H 'Content-Type: application/json' \
  -d '{
    "model": "my-namespace/model-name@version",
    "messages": [
      {
        "role":"system",
        "content":""
    },
    {
      "role":"user",
      "content":""
    }
   ],
   "temperature": 0.5,
   "top_p": 1,
   "max_tokens": 1024
  }'

Completion Models#

Train a model using the completion dataset format for tasks like text summarization, information extraction, question answering, text classification, reasoning, or story writing.

Format a Prompt-Completion Dataset#

Prompt completion datasets have a simple schema. Each datum has:

  • A prompt field for the body of information provided by the user.

  • A completion field for output of the model.

{"prompt": "Hello", "completion": " world."}

Prompt the Model#

To run inference with a completion model, use the /v1/completions endpoint for the model.

curl -X POST http://nim.test/v1/completions \
  -H 'Accept: application/json' \
  -H 'Content-Type: application/json' \
  -d '{
    "model": "my-namespace/model-name@version",
    "prompt": "",
   "temperature": 0.5,
   "top_p": 1,
   "max_tokens": 1024
  }'