🚀 This repository contains the NVIDIA Triton Inference Server configuration for serving the AgriBERT model (recobo/agriculture-bert-uncased) using the PyTorch LibTorch backend 🧠⚙️, focusing on the config.pbtxt that defines the model’s inputs, outputs, and execution settings required for deployment on Triton 🖥️🔥.
This repository contains the NVIDIA Triton Inference Server configuration for serving the AgriBERT model (recobo/agriculture-bert-uncased) using the PyTorch LibTorch backend. The repository currently focuses on the model configuration (config.pbtxt), which defines inputs, outputs, and execution settings required to deploy the model on Triton.
undefinedAgriBERT is a BERT-based language model trained specifically for the agriculture domain.
The model was trained on a large agriculture-focused corpus consisting of:
This balanced dataset includes both scientific and general agricultural content, enabling strong domain-specific language understanding.
This allows AgriBERT to learn deep contextual representations relevant to agricultural text.
While AgriBERT is commonly used for fill-mask tasks, this Triton configuration exposes the model as a general-purpose inference endpoint, suitable for:
The output shape indicates a 7-class FP32 output, suggesting use in a classification or scoring setup.
The config.pbtxt defines how Triton serves the model.
agribertpytorch_libtorch| Name | Data Type | Shape | Description |
|---|---|---|---|
input_ids |
INT32 | [batch_size, seq_len] |
Token IDs from the tokenizer |
token_type_ids |
INT32 | [batch_size, seq_len] |
Segment IDs for sentence pairs |
attention_mask |
INT32 | [batch_size, seq_len] |
Attention mask for padding tokens |
All inputs support dynamic batch size and dynamic sequence length.
| Name | Data Type | Shape | Description |
|---|---|---|---|
OUTPUT__0 |
FP32 | [batch_size, 7] |
Model logits or class scores |
KIND_GPU.
├── agribert/
│ ├── config.pbtxt
│ └── 1/
│ └── model.pt (expected)
└── README.md
⚠️ Note: The actual serialized TorchScript model (
model.pt) must be placed under a versioned directory (e.g.,1/) for Triton to load the model successfully.
torch.jit.trace or torch.jit.script)recobo/agriculture-bert-uncasedundefined📄 License & Attribution: All credit for the model architecture and pretraining goes to Recobo.ai and the original AgriBERT authors. This repository only contains deployment configuration for NVIDIA Triton Inference Server.