Ensuring AI systems behave as intended — safely, reliably, and in alignment with human values — even as they become more capable.
AI Safety is the field of research and practice concerned with ensuring that artificial intelligence systems behave as their designers intend — and that they remain beneficial as they become more capable.
Unlike traditional software bugs that produce obvious errors, AI systems can fail in subtle, unpredictable ways. A system optimizing for a narrowly defined objective might achieve it in ways that are technically correct but deeply harmful. Safety research tries to prevent this — before it happens at scale.
The challenge is particularly acute for advanced AI systems that can learn, adapt, and act autonomously across diverse environments.
Sri Lanka is adopting AI in public services, financial systems, and healthcare. These are high-stakes domains where AI failures can harm people directly. Building awareness of safety principles now — before large-scale deployment — is far more effective than retrofitting safety after harm has occurred.
Sector Example: In Agriculture, we can deploy "Red-Teaming" to prevent data poisoning in crop-yield prediction models, ensuring that food security decisions are based on robust, unmanipulated data.
Additionally, as a community connected to a global movement, Sri Lanka practitioners can contribute to international safety research and standards — shaping global norms that will inevitably affect us.