🚀 This repository contains the NVIDIA Triton Inference Server configuration for serving the AgriBERT model (recobo/agriculture-bert-uncased) using the PyTorch LibTorch backend 🧠⚙️, focusing on the config.pbtxt that defines the model’s inputs, outputs, and execution settings required for deployment on Triton 🖥️🔥.
This repository contains the NVIDIA Triton Inference Server configuration for serving the AgriBERT model (recobo/agriculture-bert-uncased) using the PyTorch LibTorch backend. The repository currently focuses on the model configuration (config.pbtxt), which defines inputs, outputs, and execution settings required to deploy the model on Triton.
undefinedAgriBERT is a BERT-based language model trained specifically for the agriculture domain.
The model was trained on a large agriculture-focused corpus consisting of:
This balanced dataset includes both scientific and general agricultural content, enabling strong domain-specific language understanding.
This allows AgriBERT to learn deep contextual representations relevant to agricultural text.
While AgriBERT is commonly used for fill-mask tasks, this Triton configuration exposes the model as a general-purpose inference endpoint, suitable for:
The output shape indicates a 7-class FP32 output, suggesting use in a classification or scoring setup.
The config.pbtxt defines how Triton serves the model.
agribertpytorch_libtorch| Name | Data Type | Shape | Description |
|---|---|---|---|
input_ids |
INT32 | [batch_size, seq_len] |
Token IDs from the tokenizer |
token_type_ids |
INT32 | [batch_size, seq_len] |
Segment IDs for sentence pairs |
attention_mask |
INT32 | [batch_size, seq_len] |
Attention mask for padding tokens |
All inputs support dynamic batch size and dynamic sequence length.
| Name | Data Type | Shape | Description |
|---|---|---|---|
OUTPUT__0 |
FP32 | [batch_size, 7] |
Model logits or class scores |
KIND_GPU.
├── agribert/
│ ├── config.pbtxt
│ └── 1/
│ └── model.pt (expected)
└── README.md
⚠️ Note: The actual serialized TorchScript model (
model.pt) must be placed under a versioned directory (e.g.,1/) for Triton to load the model successfully.
torch.jit.trace or torch.jit.script)recobo/agriculture-bert-uncasedundefined📄 License & Attribution: All credit for the model architecture and pretraining goes to Recobo.ai and the original AgriBERT authors. This repository only contains deployment configuration for NVIDIA Triton Inference Server.
We use cookies
We use cookies to analyze traffic and improve your experience. You can accept or reject analytics cookies.