{"id":1019,"date":"2020-06-06T19:11:22","date_gmt":"2020-06-06T17:11:22","guid":{"rendered":"http:\/\/blogs.uned.es\/workshopadvancingtowards\/?page_id=1019"},"modified":"2020-06-07T14:38:21","modified_gmt":"2020-06-07T12:38:21","slug":"contents-8","status":"publish","type":"page","link":"https:\/\/blogs.uned.es\/workshopadvancingtowards\/contents-8\/","title":{"rendered":"Contents 8"},"content":{"rendered":"\t\t<div data-elementor-type=\"wp-page\" data-elementor-id=\"1019\" class=\"elementor elementor-1019\">\n\t\t\t\t\t\t<section class=\"elementor-section elementor-top-section elementor-element elementor-element-c158b9b elementor-section-boxed elementor-section-height-default elementor-section-height-default\" data-id=\"c158b9b\" data-element_type=\"section\" data-e-type=\"section\">\n\t\t\t\t\t\t<div class=\"elementor-container elementor-column-gap-default\">\n\t\t\t\t\t<div class=\"elementor-column elementor-col-100 elementor-top-column elementor-element elementor-element-bcdc733\" data-id=\"bcdc733\" data-element_type=\"column\" data-e-type=\"column\">\n\t\t\t<div class=\"elementor-widget-wrap elementor-element-populated\">\n\t\t\t\t\t\t<div class=\"elementor-element elementor-element-5592a8c elementor-widget elementor-widget-text-editor\" data-id=\"5592a8c\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<h4>8. IMPORTANT RISKS NOT MENTIONED IN WHITE PAPER<\/h4><p dir=\"ltr\">\u25cf In all the forums where the ethical risks of AI are discussed, a set of common themes are mentioned: beneficence and non-maleficence, autonomy, justice and explicability. To our understanding the White Paper does not cover all the dimensions of the principles of Autonomy and Justice.<\/p><p dir=\"ltr\">\u25cf Although the optimistic view of technology also predicts a generally positive contribution of AI to the ability to live a happy and fulfilling life, there are also important risks in this area for individuals, their values and their personal development. It is important to emphasize the warning of possible psychological damage, of how intelligent technology could negatively influence mental health and the development of personal potential, personal fulfillment, human fulfillment, and living a life with meaning and purpose.<\/p><p dir=\"ltr\">No mention is made of important problems associated with the use of this technology such as the alteration of the concept of identity and the nature of human interactions, the difficulty of distinguishing between the real and the virtual, escapism towards virtual worlds, the replacement and deterioration of human bonds, cognitive overload, the loss of meaning and purpose due to being replaced by intelligent machines, etc. Nor is there any discussion of the potential loss of human values: wisdom, creativity, empathy, affection, social skills&#8230; It does not warn about the possibilities of control, manipulation, attack on autonomy, etc. that the field of affective computing opens up, given the susceptibility of humans to emotional influence. Nor does it warn of the important effects on mental and physical health that interactive immersive virtual reality applications (with intensive application of Artificial Intelligence techniques) could have. Experimental work with social science experts should be promoted with a view to assessing these risks.<\/p><p dir=\"ltr\">\u25cf To ensure that AI really contributes to \u00abhuman welfare\u00bb requires a prudent multidisciplinary and eco-centric approach (in harmony with an authentic anthropocentrism that recognizes the essence of the human being and his or her interests) to AI research, as opposed to an excessively techno-optimistic and technocratic approach. Given its disruptive power, we cannot assume out of hand that AI facilitates more efficient exploitation of resources; to simply assume that the market will regulate\u00a0 good uses of AI is to abdicate responsability.<\/p><p dir=\"ltr\">\u25cf While implementing intelligent technologies with a purely economic or technocratic perspective could contribute to economic growth, as a counterpart, it could have an environmental and inequality cost.\u00a0 Some studies estimate that of the sub-targets into which the SDGs are decomposed, AI could contribute positively to 134 (79%) and act as an inhibitor to 59 (21%).\u00a0 The potential of AI to increase productivity could, in turn, increase the overexploitation of resources if economic, social and environmental variables are not integrated, which is not always the case in private sector studies. In addition, while AI can increase efficiency in energy production, advanced AI technology requires massive computing resources only available in large computer centers that require a lot of energy and have a high ecological and carbon footprint (this aspect is addressed in the White Paper).<\/p><p dir=\"ltr\">\u25cf The recently published UN report by the Special Rapporteur on extreme poverty and human rights warns of the \u00abrisk of tripping over like zombies in a digital welfare dystopia\u00bb where \u00abBig Tech has been a driver of growing inequality and has facilitated the creation of a vast digital underclass\u201d. The report provides many well-documented examples in different countries of how dehumanized smart technologies are creating barriers to access to a range of social rights for those without Internet access and digital skills.<\/p><p dir=\"ltr\">\u25cf One of the most important ethical requirements of AI is explainability and transparency, since algorithmic decisions can affect the most sensitive areas in people&#8217;s lives (health, civil and social rights, criminal law, credit).\u00a0 With machine learning, and particularly deep learning applications, however, explaining the decision process is very difficult. As an example of the effect of such lack of transparency, we cite a real life case in which a claimant was informed that he or she did not qualify for a government subsidy. When an explanation was demanded, the only one given was that an algorithm had made the decision. A request to see the algorithm was refused on intellectual property grounds since the decision-making task had been subcontracted to a private company.<\/p><p dir=\"ltr\">\u25cf Other relevant risks are commercial and political manipulation, and intensive coercion and surveillance by governments and large corporations, which can damage social cohesion and contravene democratic principles and human rights.<\/p><p dir=\"ltr\">\u25cf Privacy is another well-known risk, and though it is treated in the White Paper, in our view, important aspects are not addressed. Privacy regulations must protect people but should also offer solutions to the social and public use of data, since the \u00abnon-use\u00bb of data in circumstances in which there is a clear public interest in its use is a social disadvantage. So far privacy has mainly served to protect the interests of private companies while what is needed is a move towards models of responsible data sharing. There is a need for a clear public policy on the use of data oriented towards the common good, especially data generated by public institutions (for example open access to results of publicly-funded research).<\/p><p dir=\"ltr\">\u25cf The list of high-risk areas mentioned in the White Paper (health, transport, energy and parts of the public sector such as asylum, migration, border control and justice, social security and employment services), is not very complete. Applications that may seem harmless a priori (in marketing or financial or insurance services, or in social assistance provided by NGOs, etc.), may bring about threats to rights if they produce discriminatory, biased results, etc. It seems risky to leave this list open to future revisions and amendments depending on relevant developments in practice, instead of carrying out a deeper analysis now.<\/p><p dir=\"ltr\">\u25cf In general AI requires a proactive approach to risk management, involving continual risk identification and handling.<\/p><h4>\u00a0<\/h4>\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<section class=\"elementor-section elementor-inner-section elementor-element elementor-element-e23da47 elementor-section-boxed elementor-section-height-default elementor-section-height-default\" data-id=\"e23da47\" data-element_type=\"section\" data-e-type=\"section\">\n\t\t\t\t\t\t<div class=\"elementor-container elementor-column-gap-default\">\n\t\t\t\t\t<div class=\"elementor-column elementor-col-50 elementor-inner-column elementor-element elementor-element-fc93974\" data-id=\"fc93974\" data-element_type=\"column\" data-e-type=\"column\">\n\t\t\t<div class=\"elementor-widget-wrap elementor-element-populated\">\n\t\t\t\t\t\t<div class=\"elementor-element elementor-element-0d5cf09 elementor-align-center elementor-widget elementor-widget-button\" data-id=\"0d5cf09\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"button.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<div class=\"elementor-button-wrapper\">\n\t\t\t\t\t<a class=\"elementor-button elementor-button-link elementor-size-md\" href=\"http:\/\/blogs.uned.es\/workshopadvancingtowards\/contents-general\/\">\n\t\t\t\t\t\t<span class=\"elementor-button-content-wrapper\">\n\t\t\t\t\t\t\t\t\t<span class=\"elementor-button-text\">Back to contents<\/span>\n\t\t\t\t\t<\/span>\n\t\t\t\t\t<\/a>\n\t\t\t\t<\/div>\n\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t<\/div>\n\t\t\t\t<div class=\"elementor-column elementor-col-50 elementor-inner-column elementor-element elementor-element-dd879cc\" data-id=\"dd879cc\" data-element_type=\"column\" data-e-type=\"column\">\n\t\t\t<div class=\"elementor-widget-wrap elementor-element-populated\">\n\t\t\t\t\t\t<div class=\"elementor-element elementor-element-fa2914d elementor-align-center elementor-widget elementor-widget-button\" data-id=\"fa2914d\" data-element_type=\"widget\" data-e-type=\"widget\" data-widget_type=\"button.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t<div class=\"elementor-button-wrapper\">\n\t\t\t\t\t<a class=\"elementor-button elementor-button-link elementor-size-md\" href=\"http:\/\/blogs.uned.es\/workshopadvancingtowards\/news\">\n\t\t\t\t\t\t<span class=\"elementor-button-content-wrapper\">\n\t\t\t\t\t\t\t\t\t<span class=\"elementor-button-text\">Back to response summary<\/span>\n\t\t\t\t\t<\/span>\n\t\t\t\t\t<\/a>\n\t\t\t\t<\/div>\n\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t<\/section>\n\t\t\t\t\t<\/div>\n\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t<\/section>\n\t\t\t\t<\/div>\n\t\t","protected":false},"excerpt":{"rendered":"<p>8. IMPORTANT RISKS NOT MENTIONED IN WHITE PAPER \u25cf In all the forums where the ethical risks of AI are discussed, a set of common themes are mentioned: beneficence and non-maleficence, autonomy, justice and explicability. To our understanding the White Paper does not cover all the dimensions of the principles of Autonomy and Justice. \u25cf Although the optimistic view of technology also predicts a generally positive contribution of AI to the ability to live a happy and fulfilling life, there are also important risks in this area for individuals, their values and their personal development. It is important to emphasize the warning of possible psychological damage, of how intelligent technology could negatively influence mental health and the development of personal potential, personal fulfillment, human fulfillment, and living a life with meaning and purpose. No mention is made of important problems associated with the use of this technology such as the&hellip;<\/p>\n<p> <a class=\"more-link\" href=\"https:\/\/blogs.uned.es\/workshopadvancingtowards\/contents-8\/\">Leer m\u00e1s<\/a><\/p>\n","protected":false},"author":7201,"featured_media":0,"parent":0,"menu_order":0,"comment_status":"closed","ping_status":"closed","template":"","meta":{"footnotes":""},"class_list":["post-1019","page","type-page","status-publish"],"_links":{"self":[{"href":"https:\/\/blogs.uned.es\/workshopadvancingtowards\/wp-json\/wp\/v2\/pages\/1019","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/blogs.uned.es\/workshopadvancingtowards\/wp-json\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/blogs.uned.es\/workshopadvancingtowards\/wp-json\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/blogs.uned.es\/workshopadvancingtowards\/wp-json\/wp\/v2\/users\/7201"}],"replies":[{"embeddable":true,"href":"https:\/\/blogs.uned.es\/workshopadvancingtowards\/wp-json\/wp\/v2\/comments?post=1019"}],"version-history":[{"count":4,"href":"https:\/\/blogs.uned.es\/workshopadvancingtowards\/wp-json\/wp\/v2\/pages\/1019\/revisions"}],"predecessor-version":[{"id":1301,"href":"https:\/\/blogs.uned.es\/workshopadvancingtowards\/wp-json\/wp\/v2\/pages\/1019\/revisions\/1301"}],"wp:attachment":[{"href":"https:\/\/blogs.uned.es\/workshopadvancingtowards\/wp-json\/wp\/v2\/media?parent=1019"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}