The National Institute of Standards and Technology (NIST), as very white mountain, is operating under the federal government’s auspices, and the primary responsibility is to design a new policy that will address the data security, privacy, and public responsibility concerns that are linked to the increasing success of artificial intelligence (AI).
Escalating AI-driven cyber risks
Though the appointment of NIST by the U.S. Secretary of Commerce last year to pursue the introduction of new security standard by AI came with the opportunity, only the allocation of funding to the level beyond the resources and implementation of the strategies for that development will reassure as to its readiness and ability to lead such an ambitious undertaking effectively.
AI’s introduction into cyber-threat scenery made it unbelievably faster for hackers. Analytical reports predict turnover will skyrocket to $10.5 quadrillion by 2025 as AI systems become more powerful and integral.
AI empowers the hackers to fast-track their jobs, devising increasingly advanced virus strains and digitally manipulating video and voice content, in a scary perspective of the so-called ‘deepfake’, worldwide-united disinformation campaigns to be orchestrated. Achieving AI’s ethical implementation and keeping the mentioned risks at a minimum is today’s economic and national security priority.
As part of the new directive from NIST, safe AI design would provide a detached standard that will enable rigorous testing programs targeting responsible commercial use of AI controllers and high-risk systems.
Strategic collaborations
NIST services are being made run power-loned in the building where the R&D institution is currently housed because maintenance funds have not been budgeted, which has led to documented leakages and mold. The federal budget for the future year to be considered proposes a 10% more cut to NIST funds that are frozen already.
Meeting the financial needs to handle the AI risks responsibly will sustain NIST’s work. The Department may have to research alternative sources of funding and forge partnerships to achieve this. AI programs remain a big cause of concern for the privacy of most major tech leaders, like Google and Amazon, who have, therefore, signed up for their own AI security initiatives, which provide an opportunity for collective knowledge-sharing.
Collaboration with private sector innovators through public-private partnerships could provide NIST with preferential access to leading AI technical expertise, computing resources, and funding so that they are not distracted from core company priorities.
Source:securityintelligence.