• IEEE.org
  • IEEE CS Standards
  • Career Center
  • About Us
  • Subscribe to Newsletter

0

IEEEFairness and Non-discrimination: AI systems should be built and used in a way that treats everyone fairly and avoids discrimination based on factors like race, gender, or religion. This requires careful consideration of the data used to train AI models and ongoing monitoring to identify and mitigate any potential biases. Biases embedded in training data can lead to discriminatory outcomes in AI systems. This can have serious ramifications, perpetuating social inequalities in areas like loan approvals or criminal justice.AI professionals must be vigilant in detecting and mitigating bias during data collection, model development, and deployment. (Link: Gebru, Timnit et al. Can large language models provide useful feedback on research papers? A large-scale empirical analysis: https://arxiv.org/abs/2310.01783)
  • Transparency and Explainability: AI systems often learn intricate patterns from vast datasets, making their decision-making processes opaque. Understanding how an AI system reaches a conclusion is crucial. This allows for debugging and improvement of the system and helps to build trust with users. In some cases, it may also be necessary to explain the reasoning behind an AI decision to ensure fairness and accountability. AI systems should be designed and operated in a way that allows users to understand how decisions are made. This includes clear documentation of algorithms and limitations.AI professionals should strive to develop and deploy models that are interpretable and auditable. (Link: Amodei, Dario et al. Concrete Problems in AI Safety: https://arxiv.org/abs/1606.06565)
  • Safety and Reliability: AI systems, particularly those used in safety-critical applications, need to be rigorously tested and validated to ensure they are reliable and safe. This involves identifying and mitigating potential risks associated with the system and having a plan in place for how to handle unexpected situations.AI professionals have a responsibility to anticipate unexpected situations and design fail-safe mechanisms to minimize harm. In essence, they must build robust systems that function as intended, even in unforeseen circumstances.
  • Privacy and Security: AI systems often handle sensitive data, so it's crucial to protect user privacy and security. This includes secure data practices and user control over personal information.AI professionals have a responsibility to implement robust security measures and follow established privacy regulations. (Link: European Commission. General Data Protection Regulation (GDPR): https://gdpr.eu/)
  • Accountability: There should be clear lines of accountability for the development, deployment, and use of AI systems. This includes identifying who is responsible for decisions made by the system, and having a mechanism in place to address any issues that may arise.Developers and deployers of AI systems must be held responsible for their impact. This involves mechanisms for identifying and mitigating bias, as well as clear ownership of unintended consequences.
  • Accountability and Responsibility: The development and deployment of AI systems carries inherent risks. Clear lines of accountability must be established, particularly when dealing with critical applications like healthcare or autonomous vehicles.AI professionals need to actively share responsibility for the consequences of their work. (Link: Hutson, Matthew. AI Fairness 360: An extensible toolkit for detecting, understanding, and mitigating unwanted algorithmic bias: https://github.com/Trusted-AI/AIF360)
  • These principles provide a basic framework for ethical AI development but applying them in real-world scenarios requires ongoing vigilance from practitioners. Building a culture of professional integrity within the AI field requires a collective effort. By adhering to these principles, professionals can ensure that AI is used for good and contributes to a more just and equitable future.

    Putting Professional Integrity into Practice


    Upholding these principles requires ongoing effort from AI professionals. Here are some ways to put professional integrity into action:

    • Be aware of the potential risks and biases associated with AI systems.
    • Advocate for ethical AI development and deployment within your organization.
    • Stay informed about the latest developments in AI ethics and best practices.
    • Report any unethical or unsafe practices you encounter.

    By following these practices, AI professionals can play a vital role in ensuring the responsible development and use of AI technology. As AI continues to evolve, adhering to these principles will be crucial for building trust in AI and ensuring its positive impact on society.

    Looking Forward


    The future of AI is bright, but it hinges on responsible development practices. By prioritizing professional integrity, we can unlock the immense potential of AI while mitigating its risks. As the field progresses, continuous dialogue and collaboration amongst researchers, developers, policymakers, and the public will be crucial in shaping a future where AI serves all of humanity.

    Here are some additional resources that you may find helpful:

    • The IEEE Ethically Aligned Design: A Vision for Prioritizing Human Values in Artificial Intelligence: https://ieeexplore.ieee.org/document/8058187
    • DARPA Explainable Artificial Intelligence (XAI) Program: https://www.darpa.mil/program/explainable-artificial-intelligence
    • Ben Shneiderman. 2021. Responsible AI: bridging from ethics to practice. Commun. ACM 64, 8 (August 2021), 32–35. https://doi.org/10.1145/3445973

    About the Author


    Dr. Nihad Bassis is a Global Expert in Management of Innovation and Technology leading Business and Solution Architecture Projects for over 20 years in the fields of Digital Transformation, Smart Mobility, Smart Homes, IoT, UAV and Artificial Intelligence (NLP, RPA, Quality, Compliance & Regulations). During his professional career, Dr. Bassis held positions at organizations such as Desjardins Bank (Canada), Ministry of Justice (Canada), Alten Inc. (France), United Nations, UNESCO, UNODC, IFX Corporation, Cofomo Development Inc. (Canada), Ministry of Foreign Affairs (Brazil). His deep well of knowledge and experience earned him a singular distinction: participation in international committees shaping international standards for Software Engineering, Technological Innovation, Project Management and Artificial Intelligence. He lent his expertise to renowned institutions like ISO, IEEC, IEEE, SCC, and ABNT.

    Disclaimer: The author is completely responsible for the content of this article. The opinions expressed are their own and do not represent IEEE's position nor that of the Computer Society nor its Leadership.

    LATEST NEWS