Research that makes faith-facing AI inspectable.
Fide AI studies how AI systems behave when people ask questions of faith, morality, doctrine, formation, and care. The current benchmark name and paper titles may change, but the public-standard work is already underway.
Featured research
Faith & Moral Guidance Benchmark v1
A benchmark for evaluating faith, moral guidance, and pastoral-boundary behavior in large language models and AI systems.
- Status
- Pre-publication; academic submission and public release preparation
- Artifacts
- Corpus, manifests, scoring rubric, runner, figures, analysis scripts
- Next validity step
- Human calibration panel and public release package
Research questions
Current benchmark program
Faith & Moral Guidance Benchmark v1 evaluates whether model and system responses preserve theological triage, represent disagreement accurately, avoid fabricated grounding, follow user or tradition preferences, and respect pastoral referral boundaries.
What exists now
The full local benchmark corpus and public release package are in the Fide AI benchmark tree.
The standalone runner, model configuration, scoring code, and tests are separated from product-specific application code.
Whitepaper drafts, LaTeX targets, figures, analysis scripts, and submission plans are staged under the publications directory.
Private raw outputs, production logs, and repeated run artifacts are governed by storage rules instead of being treated as public web copy.
Reports and release artifacts
Public benchmark outputs should eventually live here as reports, dataset references, reproducibility links, and versioned release notes. Until publication, this section records the intended release structure.
Publication pipeline
The near-term publication strategy prioritizes AI ethics, NLP, and responsible-evaluation venues, with longer-term plans for faith and theology audiences.
Research areas
Fide AI's research areas include faith-sensitive evaluation, theological grounding, comparative tradition representation, pastoral-adjacent safety, institutional readiness, and claims/public trust.
View research areasInterpretation limits
Benchmark scores are not theological authority, pastoral authority, or universal product approval. They are evidence about behavior under named versions, prompts, conditions, rubrics, and evaluation procedures. Human calibration remains necessary before making strong claims about judge validity or pastoral adequacy.
Why Fide AI publishes this work
This benchmark is larger than any one product. Fide AI is the correct home because the work is a public research and accountability layer for any AI system that gives faith, moral, or pastoral-adjacent guidance. Product companies belong as external or related participants under the same published rules.