Isn’t there a risk of misuse if the AI can act “agentically”? How does Owkin ensure responsible use of agentic capabilities in AI?
Agentic behavior is strictly managed by role based permissions and safety checks. K Navigator is not designed to self-initiate actions beyond its assigned domain and cannot operate without authenticated users and guardrails.
At Owkin, we recognize the importance of managing AI systems that can act “agentically”—that is, systems capable of taking independent actions within defined boundaries. To ensure responsible use, K Navigator’s agentic behavior is strictly governed by robust role-based permissions and multiple layers of safety checks.
K Navigator is deliberately not designed to self-initiate actions outside its clearly assigned domain. Every function it performs is subject to explicit user authentication and operates within carefully established limits. All actions are traceable to authenticated users, and comprehensive guardrails are in place to prevent unauthorized or unintended operations.
In addition, we regularly review and update these controls to keep pace with emerging best practices in AI safety and governance. By combining technical safeguards with clear operational policies, we ensure that agentic capabilities are used solely to support and empower qualified users—never to replace human oversight or operate without accountability.
This approach allows us to harness the power of advanced AI while maintaining trust, transparency, and full control over its actions at all times.