Archives

Seekr Debuts AI Evaluation Tool for AI Action Plan Compliance

Seekr

Seekr, a leader in explainable and trustworthy artificial intelligence designed to power mission‑critical decisions in enterprises and government, announced SeekrGuard, to evaluate and certify AI models. SeekrGuard moves beyond generic benchmarks by delivering model evaluation and interrogation capabilities that measure bias, accuracy, and reliability with transparent risk scoring, flexible testing, custom evaluators and audit‑ready governance on their own data, policies, and operational requirements. SeekrGuardmodel penetration testing is a critical advancement in detecting adverse model behavior.

“When the President released America’s AI Action plan it was made very clear that an evaluation ecosystem was needed to prevent National Security risks and ensure America remains at the very forefront of AI. Seekr answers this call with SeekrGuard.” said Rob Clark, President of Seekr.

The risk of Adverse AI Models

The rapid spread of unvetted AI models is exposing U.S. systems and global enterprises to adversarial manipulation, embedded bias, and strategic vulnerabilities at scale. According to McKinsey’s 2024 State of AI survey, roughly two‑thirds of organizations now report regular use of generative AI in at least one business function, a sharp increase from the prior year and a clear sign that adoption is outpacing governance. When AI models and large language models are deployed without rigorous evaluation, they can jeopardize critical decisions and core systems, introducing bias, weak oversight and openings for manipulation that erode trust in both public institutions and private companies.

Also Read: Vijil Secures $17 Million to Bolster Resilience of AI Agents 

How Seekr is Redefining AI risk assessment

Unlike static public leaderboards that rely on fixed, generic datasets, SeekrGuard gives control back to the organizations that deploy AI, so they can continuously re‑evaluate models as threats, policies and business conditions change.

SeekrGuard is designed to fix the key gaps in traditional AI risk assessment by using the organization’s own context as the benchmark:

  • Clear scoring. Transparent benchmarking produces side‑by‑side scorecards across real‑world scenarios for every model under evaluation.
  • Quantified model risk. Custom risk profiles let teams define their own risk frameworks and convert them into mission‑ or business‑specific risk scores.
  • Flexible testing. Users can mix and match datasets, evaluators and both open‑weight and proprietary models to run targeted, domain‑specific tests at scale.
  • Custom evaluators and data. Teams can quickly build custom evaluators for edge cases and use Seekr‘s AI‑Ready Data Engine in SeekrFlow to turn their own documents into model test datasets on any topic.

Source: PRNewswire