# Introduction n the contemporary era, many governments worldwide shifted its' policy research agenda to understand and assess the uses of social media, eservices, digital transformation, smart cities, open government data, robotics, deep learning, big data, machine learning blockchain, and artificial intelligence. The idiom "artificial intelligence" was first coined by John McCarthy at a famously held workshop at Dartmouth College, Hanover, USA, during 1956. According to John McCarthy, the father of Artificial Intelligence (AI) defined as "the science and engineering of making intelligent machines," and researchers define AI as the aim to "mimic human cognitive functions." For healthcare, AI is bringing a paradigm shift, powered by increasing healthcare data availability and the rapid progress of analytics techniques. AI generally encompasses of various activities such as machine learning, robotics, and deep learning. For the context of this perspective review, deep learning is where there are artificial neural networks. Secondly, machine learning is making machines that learn from data, such as Automatic Teller Machine cheque readers. And finally, robotics is creating devices and machines that move, such as autonomous vehicles. AI has become the new edge for digital transformation. Many factors support and drive the fast and powerful evolution of Artificial Intelligence across industries. Most common amongst these are: Access to sophisticated, fast, and cost-effective computing (processing) tools, hardware, and software and applications, Availability of large (big) and longitudinal data sets generated by digital efforts worldwide and technologies like IoT. Availability of open-source coding resources, online communities, users (coders and managers) sharing know-how. However, many companies are still struggling with real business value, and many Governments are still toying with the idea. In a nutshell, everyone wishes to weigh the risk and reward before committing to such an expensive effort. The AI business risks can around [1,17,18] Hence, with the growing market potential and interest in AI, it is imperative to develop a thoughtthrough regulatory and legal framework on the adoption and use of AI. Several hypotheses are set forward to design a policy framework for AI technologies; the authors will discuss them. This review also suggests a framework that we think is a better case involving "responsible AI" and "permission less innovation." # Size of the problem As per the Grand View Research report, The global artificial intelligence market size was valued at USD 62.3 billion in 2020 and is expected to grow at a compound annual growth rate (CAGR) of 42.2% from 2020 to 2027. [1] AI decision-making applications that use algorithmic, neural networks, deep learning, expert and learning systems are used in education, digital imaging, healthcare, manufacturing, robotics, government, supply chain, manufacturing, and production can replace humans for a variety of processes and tasks. This dependency on automated AI-centric systems has raised enormous concern about over-allocating resources towards mitigating AI's most extreme impacts. # II. Underlying Factors Regulations: There is an ongoing global debate on opaque AI systems, data protection regulations, and the lack of transparency on automated data processing. Regulatory approvals and interventions must have access and understanding of concrete definitions; however, the consensus around AI has been broadly worded, an elusive feat, especially in policy discussions. The United Kingdom and the European Union have already implemented AI policies that promote trustworthy AI. Europe has some stringent digital rules that are more strict than HIPAA rules in the US. For example, Article 22 stipulates that citizens cannot be submitted to medical decisions generated by an automated source. [2]NIST's revised data standards have become central to AI policy under the US's Trump order. [3] Policy versus Practice: AI advocates and researchers define AI that highlights its usability, functionality, and process. On the other hand, while designing Policy frameworks, policymakers recognize AI as a tool that should have caution, sensitivity, and prudence like human beings compared to human behavior. Hence, sometimes policies tend to over accentuate concern on the future use of these technologies, ignoring current usability and present-day issues. [4] impact. There is consideration required to understand the destructive power of AI as well. As suggested by Taddeo & Floridi (2018), there is a pertinent risk that the AI arms race [5] can trigger inadvertent development and AI use. Hence, in addition to Fairness, Accountability, Transparency, and Ethics, human rights serve as a complementary framework for guiding and governing AI and machine learning research and development. [6] Governance: Here, we are taking the example of healthcare as an industry to understand governancerelated challenges. Healthcare, as an industry, has established processes and frameworks. The fast pace development and roll-out of AI-related projects may hamper such frameworks. Hence, to maintain such processes and frameworks, an overarching framework must assess and establish potential areas of impact and how regulations may view these changes. Innovation in processes, analysis, and research needs to be developed in the light of maintaining transparency, accountability, and social impact/public interest, as stated in the problem statement above. In addition to these frameworks, it is also essential to develop skill sets amongst the subject matter experts and the user community to plan, assess, and evaluate the best use case of AI for their respective industries. # III. # Policy Recommendations & Implications Fig. 1: Recommendations on the AI policy framework Policy suggestions below would impact multiple stakeholders in the value chain. This is because efficient and responsible use of Artificial intelligence tools would mean culture, data management, technology shifts in the industry, and required up-grading and training professionals for better coordination. To achieve the promise AI technology brings in and its efficient use, these policy suggestions will form the policy framework upon which key stakeholders collaborate. The key factors and elements crucial for informing policy with sufficient evidence include collaboration, facilitation, oversight management, quality structure, education, benchmarking and best practices, ethics and accountability and 'responsible AI. ' Given the risk imposed with the advancement and uptake of AI amongst industries, here are seven high-level recommendations summed up in Figure 1: Collaboration: AI development and implementation should involve multi-stakeholders to collaborate for social, economic, ethical, and legal implications of AI. Public funding should be provided wherever possible to drive mandates for such collaborations nationally and internationally. Collaborations and partnerships should promote knowledge sharing, building access to information, and innovation. Hence, policymakers need to collaborate with AI experts and researchers to design and implement frameworks that facilitate research initiatives and are aligned with the technical practice of AI gaping the divide between policy and practice. Facilitation: Involvement of experts and relevant stakeholders in discussing challenges and possible safeguards against threats. Both Public and Private sectors should pool inappropriate funding for the R&D efforts pertaining to AI. All parties (regulatory and industry stakeholders) should come together to provide access to resources that help facilitate digitization, building data access, and encouraging incentives like tax credits for both profit and non-profit research that prioritizes transparency and evidence-based validation. Policy frameworks should enable data access by creating a cooperation culture among policymakers, experts, technology users, and the general public. Oversight management: Safety and efficacy of AI are contingent upon well-thought-out risk management approaches and processes to align standards and drive compliance. "What the eye doesn't see, and the mind doesn't know, doesn't exist" [7]. Hence, awareness of possible misuse, abuse, and bias is necessary amongst both researchers and policymakers alike to influence norms, design, and applications, proactively analyzing and flagging potential misuse. The policy framework should highlight all actors-roles, process risks, liabilities, and incentives to highlight opacity, bias, discrimination, inefficiency, and any other negative impact (responsible disclosure). Quality structure: It is vital that stakeholders understand AI risk distribution and liability while using AI tools. An ideal AI structure/ technological framework should support: Guiding principles of being explainable, transparent (auditable), and fair (unbiased) And augment human capabilities and maintain human well-being by being safe, ethical, and equitable (human-centric). # Design # Structure Developme nt Education # Facilitation Collaborati on # Ethics & Accounta bility # Oversight management # Best Practices Hence quality assurance should be taken into perspective while designing, developing, and deploying AI tools. Policy frameworks also need to match realworld workflows, usability principles, and end-user needs. These AI-driven systems should also solve the redundant, disjointed, and dysfunctions of the technology/ operational systems. opportunities of AI, it is also important to realize that an uneven distribution of technology and resources can hamper equitable access to AI resources. Hence, policymakers should influence investments in building AI infrastructure, training personnel, and building an engaging community of users and researchers that help demonstrate AI value leading to voluntary adoption and standards compliance. The education interventions with stakeholder involvement should also encourage keeping these frameworks up-to-date and perceptive to upcoming challenges. Ethics and Accountability: AI adoption will only progress and reach its potential if it is used ethically to protect its users (that is, humans). A digital economic policy has been adopted by almost 40 countries, including the US and the European Union. For private organizations, the personal data protection commission (PDPC), Singapore, proposed a model that guides how ethical principles can be converted into implementable practices as per the World Economic Forum regulations. In 2018, the UK also mandated five principles that could become the basis for a shared ethical AI framework. These include [8]: Development for common good # Act with fairness and clarity Preserve data rights and privacy of communities. AI to help improve citizens' cognitive intelligence alongside artificial intelligence. Should not be used to destroy or deceive human beings autonomously. IV. Education: In addition to understanding the risk and # Responsible AI Responsible AI is a framework that emphasizes ethical, accountable, and transparent use of AI technologies congruous with human rights, societal norms, user expectations, and organizational values. The overarching eight principles of AI ethics and reliability as adapted from the Responsible AI framework by IT tech law is mentioned in Figure 2. [9] Independent non-profit bodies like AI-Global [10] [15], submitted a report on the implication of AI implementation from the angle of safety and liability. As two-thirds of the value creation by AI contributes to the B2B segment, it is a call for us all researchers, academicians, business owners, governments, and industry leaders to come together and provide due consideration to ethical automation with the use of such technologies. Neural fuzzing can be used to test large amounts of random input data within the software to identify its vulnerabilities. [18] Compliance to data privacy regulations (European Union's General Data Protection Regulation (GDPR) and the California Consumer Privacy Act (CCPA). V. # Conclusion The review can influence policymakers and stakeholders to develop AI and data privacy policies and guidelines across countries globally in healthcare facilities, especially during the current drive towards the future of AI. Future research could investigate the effect of specific variables on healthcare facility users' perceptions that might influence AI use and data privacy. # VI. Statements ![and regulatory governance, and Outcomes leading to mental or physical harm.](image-2.png "") Compliance-Does the design comply with industryregulations?Governments, private businesses, and non-governmental organizations across the Middle Eastregion are recognizing the shift globally towards AI andadvanced technologies. PWC [13] estimates that theMiddle East is expected to ensue 2% of the total globalbenefits of AI in 2030, which is equal to US$320 billion.The UAE's national program on ArtificialIntelligenceaimsatenhancingGovernmentperformance and efficiency. Recently, the Governmentof Dubai, Smart Dubai, published Dubai's Ethical AIToolkit. The toolkit has been created to provide hands-on support across a metropolitan ecosystem. Itsupports academia, industry, and citizens inunderstanding how AI systems can be utilizedresponsibly. It comprises principles and guidelines anda self-assessment tool for developers to assess theirplatforms. [14] Europe's Communication on ArtificialIntelligence, 2018Explainability-what was predicted and how "x" waspredicted?Fairness-Is it ethical or unfair to a particular group?Robustness-Can the model be fooled? How robust isthe model? 1DrawbacksEthical challengeImplementation considerations1 AI Black boxUnexplained PredictionsBuild Transparencya) Transparentdesign,Interpretable output-Developdecision or prediction modelThough AI algorithms can learnwith its explanation.from massive amounts of data and internalize them to make decisions, these algorithms could be a blackPredictions and decisions without reasonsb) Model Inspection-> Model explanation->Outcome explanationbox to even their creators. [16]c) Use what can be explained.Treatself-learningneuralnetworks and solutions withcare.2 Algorithmic ComplexityDifficult to understand and comprehend the "how?"Provide adequate training & Validate modelsTraining is required for the endThere is more emphasis on models to give smart decisions than ethical ones.Technical secrecy and complexity can be deceptionLittle understanding or skills around comprehending the algorithm, its functional elements, modus operandi, and relationship across system may blind decision makingprofessional to interpret and explicably understand the AI models and the application. Enough test cases (vertical domain) should be run to validatethe results of AI. * By End Use, By Region, And Segment Forecasts Trends Analysis Report By Solution (Hardware, Software, Services) 2020. 2020 Rep. No. GVR-1-68038-955-5 By Technology (Deep Learning * Differences between Europe and the United States on AI/Digital Policy PGourraud doi:10. 1177/2470289720907103 Comment Response to Roundtable Discussion on AI. Gender and the Genome 4 247028972090710 2020 * Lawyers say NIST's revised data standards should be central to AI policy under trump order. Inside Cybersecurity RWeber 2019 * Defining AI in Policy versus Practice PMKrafft MYoung MKatell KHuang GBugingo Proceedings of the AAAI the AAAI ACM 2020 * doi:10. 1145/3375627.3375835 Conference on AI, Ethics, and Society * Regulate artificial intelligence to avert cyber arms race Taddeo LucianoFloridi 296-298.10.1038/d41586-018-04602-6 Nature 556 2018 * Leap of FATE: human rights as a complementary framework for AI policy and practice CorinneCath MarkLatonero VidushiMarda RoyaPakzad 10.1145/3351095.3375665 Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency (FAT*' 20) the 2020 Conference on Fairness, Accountability, and Transparency (FAT*' 20)New York, NY, USA Association for Computing Machinery 2020 702 * What the eye doesn't see and the mind doesn't know, doesn't exist DHLawrence * AI in the UK: Ready, willing and able? (Rep. No. HL Paper 2019 100 Retrieved 2020, from Authority of the House of Lords website * Digital economy is next priority: The digital readiness of government ensures business continuity, regardless of the circumstances August 17, 2020. 2020 Responsible AI Policy Framework (Tech. * Cortex Certifai: Evaluate AI models for robustness, fairness, and explainability PWC. 2020. 2020. 2020 The Institute for Ethical AI & Machine Learning. The potential impact of AI in the Middle East * The accompanying Staff Working Document 2018 137 provided a first mapping of liability challenges that occur in the context of emerging digital technologies * Can We Open the Black Box of AI? DavideCastelvecchi NATURE Oct. 5, 2016 characterizing "opening up the black box" as the "equivalent of neuroscience to understand the networks inside" the brain * FRA 2018a * Data quality and artificial intelligence -mitigating bias and error to protect fundamental rights SBarocas ADSelbst 2016 * Healthcare Professional and User Perceptions of e Health Data and Record Privacy in Dubai. Information FMAlmarzooqi IAMoonesar R&alqutob 10.3390/info11090415 2020 11 415