Skip to content

Config reference: training

  • Enterprise tuning surface


    Defaults + constraints are rendered directly from Pydantic.

  • Env keys when available


    Many fields have an env-style alias (from TriBridConfig.to_flat_dict()).

  • Tooltip-level guidance


    If a matching glossary entry exists, you’ll see deeper tuning notes.

Config reference Config API & workflow Glossary

Total parameters: 36

Group index
  • (root)

(root)

JSON key Env key(s) Type Default Constraints Summary
training.learning_reranker_backend LEARNING_RERANKER_BACKEND Literal["auto", "mlx_qwen3"] "auto" allowed="auto", "mlx_qwen3" Learning reranker backend: auto (prefer MLX Qwen3 on Apple Silicon), mlx_qwen3 (force). Legacy values 'transformers'/'hf' normalize to 'auto'.
training.learning_reranker_base_model LEARNING_RERANKER_BASE_MODEL str "Qwen/Qwen3-Reranker-0.6B" Base model to fine-tune for MLX Qwen3 learning reranker
training.learning_reranker_grad_accum_steps LEARNING_RERANKER_GRAD_ACCUM_STEPS int 8 ≥ 1, ≤ 128 Gradient accumulation steps per optimizer update for MLX Qwen3 learning reranker training
training.learning_reranker_lora_alpha LEARNING_RERANKER_LORA_ALPHA float 32.0 > 0.0, ≤ 512.0 LoRA alpha for MLX Qwen3 learning reranker
training.learning_reranker_lora_dropout LEARNING_RERANKER_LORA_DROPOUT float 0.05 ≥ 0.0, ≤ 0.5 LoRA dropout for MLX Qwen3 learning reranker
training.learning_reranker_lora_rank LEARNING_RERANKER_LORA_RANK int 16 ≥ 1, ≤ 128 LoRA rank for MLX Qwen3 learning reranker
training.learning_reranker_lora_target_modules list[str] ["q_proj", "k_proj", "v_proj", "o_proj"] min_length=1 Module name suffixes to apply LoRA to (MLX Qwen3)
training.learning_reranker_negative_ratio LEARNING_RERANKER_NEGATIVE_RATIO int 5 ≥ 1, ≤ 20 Negative pairs per positive during learning reranker training
training.learning_reranker_promote_epsilon LEARNING_RERANKER_PROMOTE_EPSILON float 0.0 ≥ 0.0, ≤ 1.0 Minimum improvement required to auto-promote (primary metric delta)
training.learning_reranker_promote_if_improves LEARNING_RERANKER_PROMOTE_IF_IMPROVES int 1 ≥ 0, ≤ 1 Promote trained learning artifact to active path only if primary metric improves
training.learning_reranker_telemetry_interval_steps LEARNING_RERANKER_TELEMETRY_INTERVAL_STEPS int 2 ≥ 1, ≤ 20 Emit trainer telemetry every N optimizer steps (plus first/final)
training.learning_reranker_unload_after_sec LEARNING_RERANKER_UNLOAD_AFTER_SEC int 0 ≥ 0, ≤ 86400 Unload MLX learning reranker model after idle seconds (0 = never)
training.ragweld_agent_backend RAGWELD_AGENT_BACKEND str "mlx_qwen3" Ragweld agent backend (in-process chat model). Currently: mlx_qwen3
training.ragweld_agent_base_model RAGWELD_AGENT_BASE_MODEL str "mlx-community/Qwen3-1.7B-4bit" Shipped base model for the ragweld agent (MLX).
training.ragweld_agent_grad_accum_steps RAGWELD_AGENT_GRAD_ACCUM_STEPS int 8 ≥ 1, ≤ 128 Gradient accumulation steps per optimizer update for ragweld agent training.
training.ragweld_agent_lora_alpha RAGWELD_AGENT_LORA_ALPHA float 32.0 > 0.0, ≤ 512.0 LoRA alpha for ragweld agent MLX fine-tuning.
training.ragweld_agent_lora_dropout RAGWELD_AGENT_LORA_DROPOUT float 0.05 ≥ 0.0, ≤ 0.5 LoRA dropout for ragweld agent MLX fine-tuning.
training.ragweld_agent_lora_rank RAGWELD_AGENT_LORA_RANK int 16 ≥ 1, ≤ 128 LoRA rank for ragweld agent MLX fine-tuning.
training.ragweld_agent_lora_target_modules list[str] ["q_proj", "k_proj", "v_proj", "o_proj"] min_length=1 Module name suffixes to apply LoRA to (ragweld agent; MLX Qwen3).
training.ragweld_agent_model_path RAGWELD_AGENT_MODEL_PATH str "models/learning-agent-epstein-files-1" Active ragweld agent adapter artifact path (directory containing adapter.npz + adapter_config.json).
training.ragweld_agent_promote_epsilon RAGWELD_AGENT_PROMOTE_EPSILON float 0.0 ≥ 0.0, ≤ 10.0 Minimum eval_loss improvement required to auto-promote (baseline_loss - new_loss >= epsilon).
training.ragweld_agent_promote_if_improves RAGWELD_AGENT_PROMOTE_IF_IMPROVES int 1 ≥ 0, ≤ 1 Auto-promote trained ragweld agent adapter only if eval_loss improves.
training.ragweld_agent_reload_period_sec RAGWELD_AGENT_RELOAD_PERIOD_SEC int 60 ≥ 0, ≤ 600 Adapter reload check period (seconds). 0 = check every request.
training.ragweld_agent_telemetry_interval_steps RAGWELD_AGENT_TELEMETRY_INTERVAL_STEPS int 2 ≥ 1, ≤ 20 Emit ragweld agent trainer telemetry every N optimizer steps (plus first/final).
training.ragweld_agent_train_dataset_path RAGWELD_AGENT_TRAIN_DATASET_PATH str "" Training dataset path for the ragweld agent (empty = use evaluation.eval_dataset_path).
training.ragweld_agent_unload_after_sec RAGWELD_AGENT_UNLOAD_AFTER_SEC int 0 ≥ 0, ≤ 86400 Unload ragweld agent model after idle seconds (0 = never).
training.reranker_train_batch RERANKER_TRAIN_BATCH int 16 ≥ 1, ≤ 128 Training batch size
training.reranker_train_epochs RERANKER_TRAIN_EPOCHS int 2 ≥ 1, ≤ 20 Training epochs for reranker
training.reranker_train_lr RERANKER_TRAIN_LR float 2e-05 ≥ 1e-06, ≤ 0.001 Learning rate
training.reranker_warmup_ratio RERANKER_WARMUP_RATIO float 0.1 ≥ 0.0, ≤ 0.5 Warmup steps ratio
training.tribrid_reranker_mine_mode TRIBRID_RERANKER_MINE_MODE str "replace" pattern=^(replace|append)$ Triplet mining mode
training.tribrid_reranker_mine_reset TRIBRID_RERANKER_MINE_RESET int 0 ≥ 0, ≤ 1 Reset triplets file before mining
training.tribrid_reranker_model_path TRIBRID_RERANKER_MODEL_PATH str "models/learning-reranker-epstein-files-1" Active learning reranker artifact path (MLX adapter directory).
training.tribrid_triplets_path TRIBRID_TRIPLETS_PATH str "data/training/triplets__epstein-files-1.jsonl" Training triplets file path
training.triplets_min_count TRIPLETS_MIN_COUNT int 100 ≥ 10, ≤ 10000 Min triplets for training
training.triplets_mine_mode TRIPLETS_MINE_MODE str "replace" pattern=^(replace|append)$ Triplet mining mode

Details (glossary)

training.learning_reranker_backend (LEARNING_RERANKER_BACKEND) — Learning Reranker Backend

Category: reranking

Selects the execution stack used to train and serve the learning-based reranker. In this project, backend choice determines hardware assumptions, supported model formats, and how adapters are loaded during inference, so it directly affects throughput, reproducibility, and operational complexity. Keep the backend consistent across training and deployment environments whenever possible, or validate compatibility boundaries before shipping adapters. If performance or stability regresses, backend mismatch is one of the first places to investigate.

Badges: - Backend selection

Links: - Qwen3 Embeddings and Rerankers (arXiv) - MLX-LM Repository - Qwen3-Reranker-0.6B - PEFT LoRA Guide

training.learning_reranker_base_model (LEARNING_RERANKER_BASE_MODEL) — Learning Reranker Base Model

Category: reranking

Base checkpoint that LoRA adapters are trained against and later mounted on during reranking inference. Adapter weights are architecture-specific, so changing the base model after training usually invalidates existing adapters and can silently degrade ranking quality if not caught. Pin this value explicitly, record it in experiment metadata, and keep train/infer parity to make evaluation deltas trustworthy. In practice, base-model drift is a common root cause of non-reproducible reranker performance.

Badges: - Model compatibility

Links: - Qwen3 Embeddings and Rerankers (arXiv) - Qwen3-Reranker-0.6B - MLX-LM Repository - PEFT LoRA Guide

training.learning_reranker_grad_accum_steps (LEARNING_RERANKER_GRAD_ACCUM_STEPS) — Learning Reranker Grad Accum Steps

Category: reranking

Number of micro-batches whose gradients are accumulated before each optimizer update during reranker training. Increasing this value raises effective batch size without requiring equivalent VRAM, which can stabilize ranking-objective learning but also slows update frequency and may require learning-rate retuning. For RAG rerankers, this parameter is most useful when negatives are hard and memory is constrained, because larger effective batches improve signal diversity per update. Tune it jointly with global batch size, learning rate, and training time budget rather than in isolation.

Badges: - Training dynamics

Links: - ERank: Efficient Learning-to-Rank for RAG (arXiv) - PyTorch Gradient Accumulation - MLX-LM Repository - PEFT LoRA Guide

training.learning_reranker_lora_alpha (LEARNING_RERANKER_LORA_ALPHA) — Learning Reranker LoRA Alpha

Category: reranking

LoRA scaling factor that determines how strongly adapter updates influence the frozen base reranker during training and inference. The effective adaptation strength is tied to alpha relative to rank, so increasing alpha without considering rank can over-amplify updates and destabilize relevance calibration. In ranking-focused fine-tuning, this value is best tuned with validation metrics that emphasize ordering quality, not just loss reduction. Use conservative increments and track metric movement on hard negatives to avoid overfitting narrow retrieval patterns.

Badges: - LoRA scaling

Links: - How Relevance Emerges in Fine-tuned Rerankers (arXiv) - PEFT LoRA Guide - MLX-LM Repository - Qwen3-Reranker-0.6B

training.learning_reranker_lora_dropout (LEARNING_RERANKER_LORA_DROPOUT) — Learning Reranker LoRA Dropout

Category: reranking

Dropout probability applied inside LoRA adapter paths during reranker fine-tuning. This acts as regularization against overfitting on narrow training pairs, especially when mined positives and negatives are repetitive or domain-skewed. Too little dropout can produce brittle rankers that fail on unseen queries, while too much can underfit and flatten relevance separation. Tune with validation sets that include both in-domain and near-domain queries so improvements generalize beyond the training distribution.

Badges: - LoRA regularization

Links: - How Relevance Emerges in Fine-tuned Rerankers (arXiv) - PEFT LoRA Guide - MLX-LM Repository - Qwen3-Reranker-0.6B

training.learning_reranker_lora_rank (LEARNING_RERANKER_LORA_RANK) — Learning Reranker LoRA Rank

Category: reranking

Adapter rank determines the capacity of LoRA updates layered onto the base reranker. Higher rank can capture richer relevance transformations and improve difficult ranking tasks, but it increases memory, training cost, and overfitting risk if data volume is limited. Lower rank is cheaper and often sufficient for moderate domain adaptation, especially when base model quality is already high. Select rank by balancing retrieval-quality gains against training budget and inference latency targets, then confirm with held-out ranking benchmarks.

Badges: - LoRA capacity

Links: - How Relevance Emerges in Fine-tuned Rerankers (arXiv) - PEFT LoRA Guide - MLX-LM Repository - Qwen3-Reranker-0.6B

training.learning_reranker_negative_ratio (LEARNING_RERANKER_NEGATIVE_RATIO) — Learning Reranker Negative Ratio

Category: reranking

LEARNING_RERANKER_NEGATIVE_RATIO controls how many negative (query, document) pairs are generated per positive pair during learning-reranker training (default 5, range 1-20). Higher ratios usually improve separation between relevant and non-relevant candidates, but they also increase training time, GPU memory usage, and the risk of overfitting to easy negatives if sampling quality is poor. Lower ratios train faster and can be sufficient when negatives are already hard and diverse, but may leave the reranker under-discriminative on near-miss results. Treat this as a quality-versus-cost dial and tune it alongside hard-negative mining strategy and dev-set ranking metrics.

Badges: - Quality vs cost

Links: - Reranker Optimization via Geodesic Distances on k-NN Manifolds (arXiv) - Sentence Transformers: Train Cross-Encoder for Reranking - Sentence Transformers Loss Functions - PyTorch MarginRankingLoss

training.learning_reranker_promote_epsilon (LEARNING_RERANKER_PROMOTE_EPSILON) — Learning Reranker Promotion Epsilon

Category: reranking

LEARNING_RERANKER_PROMOTE_EPSILON sets the minimum dev-metric delta required before a newly trained reranker is allowed to replace the active baseline (range 0.0-1.0, default 0.0). This threshold is a noise guard: if metric gains are smaller than epsilon, the run is treated as statistically or operationally insignificant and promotion should be blocked. Small nonzero values (for example around 0.001-0.005 depending on metric stability) reduce churn from random variation, label noise, and temporary data drift. Calibrate epsilon from repeated baseline evaluations so promotion decisions reflect durable quality changes rather than measurement jitter.

Badges: - Prevents noise promotions

Links: - Scaling Search Relevance: Augmenting App Store Ranking with LLM-Generated Judgments (arXiv) - MLflow Model Registry Tutorial - Amazon SageMaker Model Registry - SciPy Paired t-test (ttest_rel)

training.learning_reranker_promote_if_improves (LEARNING_RERANKER_PROMOTE_IF_IMPROVES) — Learning Reranker Promotion Gate

Category: reranking

LEARNING_RERANKER_PROMOTE_IF_IMPROVES is the hard promotion gate for learning-reranker training. When set to 1 (default), a successful training job promotes the candidate artifact only if the primary dev metric exceeds the current baseline by at least LEARNING_RERANKER_PROMOTE_EPSILON; when set to 0, every successful run can overwrite the active path. Keeping this enabled is safer in continuous-training loops because it preserves model stability during noisy data windows and imperfect labeling periods. Disable it only for controlled experiments with manual review and explicit rollback procedures.

Badges: - Safety

Links: - Scaling Search Relevance: Augmenting App Store Ranking with LLM-Generated Judgments (arXiv) - MLflow Model Registry Tutorial - Amazon SageMaker Model Registry - SciPy Paired t-test (ttest_rel)

training.learning_reranker_telemetry_interval_steps (LEARNING_RERANKER_TELEMETRY_INTERVAL_STEPS) — Learning Reranker Telemetry Interval Steps

Category: reranking

Defines how often trainer telemetry is emitted in optimizer-step units. Lower values (for example 1-2) produce smoother live curves and faster anomaly detection, but increase event traffic, UI render pressure, and log volume. Higher values reduce overhead and can stabilize weak machines, but hide short-lived instability such as gradient spikes or transient loss explosions. In practice, use tighter intervals while tuning a new objective or dataset, then relax interval size once behavior is stable. This parameter directly shapes observability quality, so tune it with both monitoring fidelity and system cost in mind.

Links: - Query-focused and Memory-aware Reranker for Unrestricted Context (arXiv, 2026) - Weights & Biases: Log Data During Experiments - PyTorch Lightning Logging - REARANK: Reinforcement Learning for RAG Reranking (arXiv, 2025)

training.learning_reranker_unload_after_sec (LEARNING_RERANKER_UNLOAD_AFTER_SEC) — Learning Reranker Idle Unload

Category: reranking

Idle-time model eviction threshold for the local MLX reranker. When this timer is greater than zero, the reranker is unloaded after inactivity so RAM/VRAM can be reclaimed for other work; when set to zero, the model stays resident and avoids cold-start reload latency. This is a classic memory-latency tradeoff: aggressive unloading helps constrained laptops, while persistent residency is better for frequent back-to-back reranks. Tune this using actual interaction cadence, not defaults: if people pause briefly between trials, use a longer window to avoid repeated thrash from unload/reload cycles.

Badges: - MLX only - Affects latency

Links: - Query-focused and Memory-aware Reranker for Unrestricted Context (arXiv, 2026) - Ollama API Reference - MLX Documentation - REARANK: Reinforcement Learning for RAG Reranking (arXiv, 2025)

training.ragweld_agent_backend (RAGWELD_AGENT_BACKEND) — RAGWELD_AGENT_BACKEND

Category: general

Chooses the runtime stack used to train and evaluate the agent model (for example, a Transformers/TRL pipeline with DeepSpeed, or another backend with different distributed semantics). This setting controls how batches are scheduled, how optimizer state is sharded, how mixed precision is handled, and what checkpoint artifacts are produced. In practice, backend choice affects throughput, memory headroom, resume reliability, and reproducibility more than most single hyperparameters. Keep backend, precision mode, and checkpoint format aligned so promoted adapters can be reloaded without silent drift.

Links: - LoRAFusion: Advancing Efficient Fine-Tuning in Production-Scale Language Models (arXiv 2025) - Hugging Face TRL: PEFT Integration - Transformers Trainer API - DeepSpeed Config: Batch and Training Parameters

training.ragweld_agent_base_model (RAGWELD_AGENT_BASE_MODEL) — RAGWELD_AGENT_BASE_MODEL

Category: general

Specifies the pretrained foundation model that LoRA adapters are attached to. Tokenizer vocabulary, context length behavior, architecture names, and module layout all come from this base model, so changing it after training usually invalidates existing adapters. Treat this as an ABI contract for fine-tuning: adapter weights, target modules, and optimizer state are only portable when the base model family and revision are compatible. Pin exact model revisions to make evaluation and rollback deterministic.

Links: - Linearization of Language Models under Parameter-Efficient Fine-Tuning (arXiv 2026) - Transformers PreTrainedModel Reference - Transformers AutoModel Classes - Hugging Face Model Cards

training.ragweld_agent_grad_accum_steps (RAGWELD_AGENT_GRAD_ACCUM_STEPS) — RAGWELD_AGENT_GRAD_ACCUM_STEPS

Category: general

Defines how many micro-batches are accumulated before one optimizer update. Effective batch size is roughly per_device_batch_size * grad_accum_steps * world_size, so increasing this value lets you emulate larger batches under limited VRAM. The tradeoff is fewer optimizer steps per wall-clock minute and slightly staler gradients, which can change convergence behavior. Tune this together with learning rate and scheduler warmup, not in isolation, because accumulation directly changes update frequency.

Links: - PROMA: Continual Pretraining and RL Fine-Tuning Framework (arXiv 2026) - Transformers TrainingArguments: gradient_accumulation_steps - DeepSpeed Batch Size and Gradient Accumulation Settings - PyTorch Automatic Mixed Precision

training.ragweld_agent_lora_alpha (RAGWELD_AGENT_LORA_ALPHA) — RAGWELD_AGENT_LORA_ALPHA

Category: general

Controls LoRA adapter scaling (commonly applied as alpha / rank). Higher alpha increases the effective strength of adapter updates and can speed adaptation, but it also raises the risk of overshooting or overfitting on narrow datasets. Lower alpha makes updates conservative and may underfit unless training runs longer. Treat alpha as a stability/capacity dial coupled to rank and learning rate; when rank changes, revisit alpha instead of keeping a fixed absolute value.

Links: - Linearization of Language Models under Parameter-Efficient Fine-Tuning (arXiv 2026) - PEFT LoRA Configuration Reference - PEFT LoRA Developer Guide - TRL PEFT Integration

training.ragweld_agent_lora_dropout (RAGWELD_AGENT_LORA_DROPOUT) — RAGWELD_AGENT_LORA_DROPOUT

Category: general

Applies dropout on the adapter path during training to reduce co-adaptation and improve generalization when data is limited or noisy. Setting this to 0.0 maximizes determinism and can help on very large, clean corpora, while moderate values often improve robustness on mixed-quality data. Too much dropout suppresses useful signal and slows learning. Tune it alongside alpha and rank because stronger regularization usually needs either more steps or slightly higher adapter capacity.

Links: - FedSA-LoRA: Bayesian Sparse and Adaptive Low-Rank Adaptation (arXiv 2026) - PEFT LoRA Configuration Reference - PyTorch Dropout Layer Reference - TRL PEFT Integration

training.ragweld_agent_lora_rank (RAGWELD_AGENT_LORA_RANK) — RAGWELD_AGENT_LORA_RANK

Category: general

Sets the low-rank adapter dimension (r), which is the main capacity knob for LoRA. Higher rank increases representational power and typically improves fit on complex tasks, but also raises VRAM, compute, and risk of overfitting. Lower rank is cheaper and often adequate for style or narrow-domain adaptation. Choose rank using validation metrics under fixed budget constraints, and retune alpha when rank changes because update scaling depends on both.

Links: - Adaptive LoRA Exploration in Federated Fine-Tuning (arXiv 2026) - PEFT LoRA Configuration Reference - PEFT LoRA Developer Guide - TRL PEFT Integration

training.ragweld_agent_model_path (RAGWELD_AGENT_MODEL_PATH) — RAGWELD_AGENT_MODEL_PATH

Category: general

Points to the model artifact location used by the agent pipeline (local directory, checkpoint file, or hub identifier). This path determines what tokenizer/config/weights are loaded and where resumed training continues from. Use immutable, versioned paths for reproducibility and avoid ambiguous symlinks in production pipelines. For LoRA workflows, clearly separate base-model path from adapter output path so promotion and rollback do not accidentally mix incompatible artifacts.

Links: - LLMTailor: Fine-Tuning LLMs by Checking Components and LoRA (arXiv 2026) - Transformers PreTrainedModel Loading and Saving - Hugging Face Hub Download Guide - Safetensors Documentation

training.ragweld_agent_promote_epsilon (RAGWELD_AGENT_PROMOTE_EPSILON) — RAGWELD_AGENT_PROMOTE_EPSILON

Category: general

Sets the minimum metric gain required before a new checkpoint is promoted over the current best. This prevents noisy, statistically insignificant fluctuations from constantly replacing promoted models. Use epsilon in the same unit as your monitored metric (for example absolute NDCG gain or loss decrease), and calibrate it from historical run variance. If epsilon is too small, promotion churn increases; if too large, meaningful improvements are ignored and iteration slows.

Links: - UGCS: Better Checkpoint Selection for LLM Optimization (arXiv 2025) - Transformers TrainingArguments: load_best_model_at_end - PyTorch Lightning EarlyStopping (min_delta) - PyTorch Lightning ModelCheckpoint

training.ragweld_agent_promote_if_improves (RAGWELD_AGENT_PROMOTE_IF_IMPROVES) — RAGWELD_AGENT_PROMOTE_IF_IMPROVES

Category: general

Boolean gate for checkpoint promotion based on validation outcomes. When enabled, a candidate checkpoint is promoted only if the tracked metric improves versus the current promoted model (typically with PROMOTE_EPSILON as the noise threshold). This keeps production candidates aligned to measured quality instead of recency. When disabled, every completed run can overwrite promoted state, which is useful for exploratory debugging but risky for stable deployments.

Links: - UGCS: Better Checkpoint Selection for LLM Optimization (arXiv 2025) - Transformers Trainer and Best-Checkpoint Selection - PyTorch Lightning ModelCheckpoint - PyTorch Lightning EarlyStopping

training.ragweld_agent_telemetry_interval_steps (RAGWELD_AGENT_TELEMETRY_INTERVAL_STEPS) — RAGWELD_AGENT_TELEMETRY_INTERVAL_STEPS

Category: general

Controls how often the training/runtime loop emits telemetry snapshots, measured in optimizer or agent steps. Lower values improve observability granularity (you see loss drift, action-quality regressions, and tool-failure spikes sooner) but increase logging overhead, storage volume, and dashboard cost. Higher values reduce overhead but can hide short-lived failures and make root-cause analysis harder because fewer intermediate states are preserved. Tune this together with batch size and run duration: keep intervals small during experiments and incident triage, then increase once behavior is stable.

Links: - AgentSight: Visualizing and Monitoring Foundation Agent Dynamics (arXiv 2025) - OpenTelemetry Documentation - Prometheus Instrumentation Best Practices - LangSmith Observability Quickstart

training.ragweld_agent_train_dataset_path (RAGWELD_AGENT_TRAIN_DATASET_PATH) — RAGWELD_AGENT_TRAIN_DATASET_PATH

Category: general

Filesystem path to the training dataset used by the agent/reranker training pipeline. This path determines what examples are loaded, so a wrong mount or stale directory silently changes training behavior and can invalidate evaluation comparisons. Prefer explicit absolute paths and versioned artifacts, and keep schema/format checks near load time (for example JSONL field validation) so bad rows fail early. In multi-environment setups (local, CI, container), treat this as an environment-specific input that must be pinned per run.

Links: - PIPES: Programmatic Pipeline Search for Scalable Data Synthesis and Curation (arXiv 2025) - Hugging Face Datasets: Load - JSON Lines Format - DVC Data Management

training.reranker_train_batch (RERANKER_TRAIN_BATCH) — Training Batch Size

Category: embedding

Number of training examples processed per optimization step during learning-reranker fine-tuning. Larger batches can improve gradient stability and throughput on strong hardware, but they increase memory pressure and can destabilize local/containerized setups if oversized. Smaller batches are safer for constrained environments and can be paired with gradient accumulation to emulate larger effective batch sizes. Tune batch size together with learning rate and sequence length, since all three interact with convergence speed and overfitting risk.

Badges: - Lower = safer on Colima

Links: - GraLoRA: Gradient-Driven Low-Rank Adaptation (arXiv 2025) - GoRA: Gradient-guided LoRA (arXiv 2025) - PyTorch Optimizer Documentation - Hugging Face PEFT Documentation

training.reranker_train_epochs (RERANKER_TRAIN_EPOCHS) — Training Epochs

Category: reranking

Defines how many full passes over the reranker training dataset are executed. More epochs can improve fit on stable, representative triplets, but excessive epochs on small or noisy data usually reduce generalization and hurt real query performance. Use held-out validation queries and early stopping signals rather than only training loss to choose this value. As your mined data grows or distribution shifts, retune epochs because the optimal point moves with dataset size and difficulty.

Badges: - Quality vs overfit

Links: - Sensitivity-LoRA: Hyperparameter Sensitivity (arXiv 2025) - GraLoRA: Training Stability for LoRA (arXiv 2025) - Transformers Optimizer & Schedule Docs - Jina MLX Retrieval Repository

training.reranker_train_lr (RERANKER_TRAIN_LR) — Training Learning Rate

Category: reranking

Learning rate for reranker fine-tuning updates. It controls update magnitude and is often the highest-impact training hyperparameter: too high causes unstable loss and catastrophic drift, too low undertrains and wastes epochs. Choose LR jointly with batch size, adapter rank, and warmup schedule, and validate using ranking metrics rather than loss alone. For reranker adaptation, conservative starting values with short sweeps are usually safer than aggressive defaults.

Badges: - Advanced ML training - Requires tuning

Links: - Sensitivity-LoRA: Learning-Rate Sensitivity (arXiv 2025) - GoRA: Gradient-guided LoRA Optimization (arXiv 2025) - PyTorch Optimizer Documentation - Transformers Optimizer & Schedule Docs

training.reranker_warmup_ratio (RERANKER_WARMUP_RATIO) — Warmup Ratio

Category: reranking

RERANKER_WARMUP_RATIO defines what fraction of total optimization steps uses a gradual learning-rate ramp before entering the main scheduler phase. In this training stack, warmup protects early updates when the reranker head and backbone are still unstable, reducing gradient spikes and divergence risk that can otherwise corrupt the first checkpoints. Operationally, this value interacts with total step count: short runs need a smaller warmup fraction so useful learning starts early, while longer runs can tolerate a larger warmup to improve stability. Tune it together with batch size and base LR, because warmup that is too short can destabilize training, while warmup that is too long can waste compute on underpowered updates.

Badges: - Advanced ML training - Stabilizes training

Links: - Warmup-Stable-Decay Learning Rates in Language Model Pre-Training (arXiv) - Hugging Face Optimizer Schedules (get_linear_schedule_with_warmup) - Hugging Face TrainingArguments.warmup_ratio - PyTorch LinearLR Scheduler

training.tribrid_reranker_mine_mode (TRIBRID_RERANKER_MINE_MODE) — Triplet Mining Mode

Category: general

Negative-sampling policy used when generating triplets for reranker training. Random negatives are stable but often weak; semi-hard negatives improve discrimination without overwhelming optimization; hard negatives are highest signal but can inject false negatives and noise if mining quality is low. The right mode depends on corpus ambiguity and label fidelity, so teams typically stage mining as a curriculum (random to semi-hard to hard) with periodic audit sets. Treat this as a data-quality lever first and a model-quality lever second.

Badges: - Advanced

Links: - BiCA: Dense Retrieval with Citation-Aware Hard Negatives (arXiv 2025) - RRRA: Resampling and Reranking through a Retriever Adapter (arXiv 2025) - SentenceTransformers Utility Functions (including hard-negative mining) - SentenceTransformers Losses (Triplet and ranking losses)

training.tribrid_reranker_mine_reset (TRIBRID_RERANKER_MINE_RESET) — Reset Triplets Before Mining

Category: general

Whether to clear previously mined triplets before a new mining run. Enabling reset gives a clean dataset snapshot and avoids mixing stale and fresh negatives, which is useful for controlled experiments. Disabling reset preserves historical data and can improve coverage, but it also increases the risk of drift and duplicate/noisy samples. Use this setting with explicit dataset versioning so you can reproduce training results and roll back when mining quality drops.

Badges: - Destructive

Links: - BiCA: Dense Retrieval with Citation-Aware Hard Negatives (arXiv 2025) - Hugging Face Datasets: Loading Local JSON/JSONL - JSON Lines Format Specification - MLflow Tracking (experiment and artifact reproducibility)

training.tribrid_reranker_model_path (TRIBRID_RERANKER_MODEL_PATH) — Reranker Model Path

Category: general

Filesystem location of the active reranker checkpoint or adapter bundle used at inference time. This path is effectively a deployment control: whichever artifact is loaded here defines live ranking behavior. Use stable, versioned directories and atomically swap symlinks or folder names to avoid partial reads during update windows. In adapter-based pipelines, keep base model version and adapter metadata aligned so path changes do not silently load incompatible weights.

Links: - RRRA: Resampling and Reranking through a Retriever Adapter (arXiv 2025) - PEFT Checkpoint Format and Loading - Transformers from_pretrained() Model Loading - Python pathlib (path handling and portability)

training.tribrid_triplets_path (TRIBRID_TRIPLETS_PATH) — Triplets Dataset Path

Category: general

Location of the JSONL triplets corpus used to mine and train the reranker. Because this dataset defines supervision quality, the path should point to a durable, versioned artifact rather than an ad hoc local file. Maintain a consistent schema (query, positive, negative, metadata) and track generation provenance so model regressions can be traced back to specific triplet revisions. In practice, good triplet hygiene often improves ranking quality more than additional training steps.

Links: - BiCA: Dense Retrieval with Citation-Aware Hard Negatives (arXiv 2025) - SentenceTransformers Losses (Triplet and ranking losses) - Hugging Face Datasets: Loading Local JSON/JSONL - JSON Lines Format Specification

training.triplets_min_count (TRIPLETS_MIN_COUNT) — Triplets Min Count

Category: general

Minimum mined triplets required before training starts. Default: 100. Range: 10-10000. If training skips for insufficient data, mine more triplets or lower this for experimentation.

Badges: - Data quality gate - Production needs 500+

Links: - Triplet Loss for Ranking - Hard Negative Mining - Triplet Mining in RAG (ACL 2025) - Learning to Rank

training.triplets_mine_mode (TRIPLETS_MINE_MODE) — Triplets Mine Mode

Category: general

Controls how newly mined triplets are persisted to disk: replace creates a clean dataset for a reproducible training run, while append extends an existing corpus for incremental hard-negative collection. Use replace when you want strict experiment comparability, fixed train/validation splits, and clear provenance. Use append when your retrieval index, query set, or domain vocabulary is evolving and you intentionally want longitudinal data accumulation. In production retraining pipelines, pair this setting with dataset versioning and a run manifest so you can trace exactly which mined triplets entered each reranker checkpoint.

Badges: - Advanced training control - Use semi-hard for production

Links: - Reranker Optimization via Geodesic Distances on k-NN Manifolds (arXiv 2026) - PyTorch TripletMarginLoss - Sentence-Transformers Loss Functions - Sentence-Transformers MS MARCO Training Example