{"id":80209,"date":"2024-06-05T16:30:34","date_gmt":"2024-06-05T16:30:34","guid":{"rendered":"https:\/\/neclink.com\/index.php\/2024\/06\/05\/ai-workers-demand-stronger-whistleblower-protections-in-open-letter\/"},"modified":"2024-06-05T16:30:34","modified_gmt":"2024-06-05T16:30:34","slug":"ai-workers-demand-stronger-whistleblower-protections-in-open-letter","status":"publish","type":"post","link":"https:\/\/neclink.com\/index.php\/2024\/06\/05\/ai-workers-demand-stronger-whistleblower-protections-in-open-letter\/","title":{"rendered":"AI workers demand stronger whistleblower protections in open letter"},"content":{"rendered":"<p> <br \/>\n<\/p>\n<div>\n<p>A group of current and former employees from leading AI companies like OpenAI, Google DeepMind and Anthropic has signed an <a data-i13n=\"elm:context_link;elmt:doNotAffiliate;cpos:1;pos:1\" class=\"link \" href=\"https:\/\/righttowarn.ai\/\" rel=\"nofollow noopener\" target=\"_blank\" data-ylk=\"slk:open letter;elm:context_link;elmt:doNotAffiliate;cpos:1;pos:1;itc:0;sec:content-canvas\">open letter<\/a> asking for greater transparency and protection from retaliation for those who speak out about the potential concerns of AI. \u201cSo long as there is no effective government oversight of these corporations, current and former employees are among the few people who can hold them accountable to the public,\u201d the letter, which was published on Tuesday, says. \u201cYet broad confidentiality agreements block us from voicing our concerns, except to the very companies that may be failing to address these issues.\u201d<\/p>\n<p>The letter comes just a couple of weeks after a <em>Vox<\/em> <a data-i13n=\"elm:context_link;elmt:doNotAffiliate;cpos:2;pos:1\" class=\"link \" href=\"https:\/\/www.vox.com\/future-perfect\/351132\/openai-vested-equity-nda-sam-altman-documents-employees\" rel=\"nofollow noopener\" target=\"_blank\" data-ylk=\"slk:investigation;elm:context_link;elmt:doNotAffiliate;cpos:2;pos:1;itc:0;sec:content-canvas\">investigation<\/a> revealed OpenAI had attempted to muzzle recently departing employees by forcing them to chose between signing an aggressive non-disparagement agreement, or risk losing their vested equity in the company. After the report, OpenAI CEO Sam Altman <a data-i13n=\"elm:context_link;elmt:doNotAffiliate;cpos:3;pos:1\" class=\"link \" href=\"https:\/\/x.com\/sama\/status\/1791936857594581428\" rel=\"nofollow noopener\" target=\"_blank\" data-ylk=\"slk:said;elm:context_link;elmt:doNotAffiliate;cpos:3;pos:1;itc:0;sec:content-canvas\">said<\/a> that he had been genuinely embarrassed&#8221; by the provision and claimed it has been removed from recent exit documentation, though it&#8217;s unclear if it remains in force for some employees. After this story was published, nn OpenAI spokesperson told Engadget that the company had removed a non-disparagement clause from its standard departure paperwork and released all former employees from their non-disparagement agreements.<\/p>\n<p>The 13 signatories include former OpenAI employees Jacob Hinton, William Saunders and Daniel Kokotajlo. Kokotajlo <a data-i13n=\"cpos:4;pos:1\" href=\"https:\/\/x.com\/DKokotajlo67142\/status\/1797994238468407380\" rel=\"nofollow noopener\" target=\"_blank\" data-ylk=\"slk:said;cpos:4;pos:1;elm:context_link;itc:0;sec:content-canvas\" class=\"link \">said<\/a> that he resigned from the company after losing confidence that it would responsibly build artificial general intelligence, a term for AI systems that is as smart or smarter than humans. The letter \u2014 which was endorsed by prominent AI experts Geoffrey Hinton, Yoshua Bengio and Stuart Russell \u2014 expresses grave concerns over the lack of effective government oversight for AI and the financial incentives driving tech giants to invest in the technology. The authors warn that the unchecked pursuit of powerful AI systems could lead to the spread of misinformation, exacerbation of inequality and even the loss of human control over autonomous systems, potentially resulting in human extinction.<\/p>\n<p>\u201cThere is a lot we don\u2019t understand about how these systems work and whether they will remain aligned to human interests as they get smarter and possibly surpass human-level intelligence in all areas,\u201d <a data-i13n=\"elm:context_link;elmt:doNotAffiliate;cpos:5;pos:1\" class=\"link \" href=\"https:\/\/x.com\/DKokotajlo67142\/status\/1797994251185471762\" rel=\"nofollow noopener\" target=\"_blank\" data-ylk=\"slk:wrote;elm:context_link;elmt:doNotAffiliate;cpos:5;pos:1;itc:0;sec:content-canvas\">wrote<\/a> Kokotajlo on X. \u201cMeanwhile, there is little to no oversight over this technology. Instead, we rely on the companies building them to self-govern, even as profit motives and excitement about the technology push them to \u2018move fast and break things.\u2019 Silencing researchers and making them afraid of retaliation is dangerous when we are currently some of the only people in a position to warn the public.\u201d<\/p>\n<p>In a statement shared with Engadget, an OpenAI spokesperson said: \u201cWe\u2019re proud of our track record providing the most capable and safest AI systems and believe in our scientific approach to addressing risk. We agree that rigorous debate is crucial given the significance of this technology and we&#8217;ll continue to engage with governments, civil society and other communities around the world.\u201d They added: \u201cThis is also why we have avenues for employees to express their concerns including an anonymous <a data-i13n=\"cpos:6;pos:1\" href=\"https:\/\/urldefense.com\/v3\/__https:\/\/url.avanan.click\/v2\/___https:\/openai.integrityline.com\/___.YXAzOm91dGNhc3Q6YTpvOjViNjQ2MTEyYzZiYzhhYTRmNjdjMTY5MGEyYWQ3OWIyOjY6ZmQ2NTpkNjhhMDNiZWI4ZGFjODAwODJjNzIwYWY5ZDUwMjkzNjU0MGZkNzIwOGY3MjgyYTQ2YTQxYjFmNjA1YzFiOTViOmg6Rg__;!!Op6eflyXZCqGR5I!HS_i2ditgvlRlsw2aZt9V4vuZYgAaxBV057twmRcgGV15L_nzhv3Fzlcer9hnlNw8t5D-d3AMMVbXiPSPpchXfSxL_Wplw$\" rel=\"nofollow noopener\" target=\"_blank\" data-ylk=\"slk:integrity hotline;cpos:6;pos:1;elm:context_link;itc:0;sec:content-canvas\" class=\"link \">integrity hotline<\/a> and a Safety and Security Committee led by members of our board and safety leaders from the company.\u201d<\/p>\n<p>Google and Anthropic did not respond to request for comment from Engadget. In a <a data-i13n=\"cpos:7;pos:1\" href=\"https:\/\/www.bloomberg.com\/news\/articles\/2024-06-04\/openai-employees-call-for-protections-to-speak-out-on-ai-risks?accessToken=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJzb3VyY2UiOiJTdWJzY3JpYmVyR2lmdGVkQXJ0aWNsZSIsImlhdCI6MTcxNzUxNjM0OCwiZXhwIjoxNzE4MTIxMTQ4LCJhcnRpY2xlSWQiOiJTRUtBMTNEV1gyUFMwMCIsImJjb25uZWN0SWQiOiJFODA3NUYyRkZGMjA0NUI2QTlEQzA5M0EyQTdEQTE4NiJ9.wFpVycx4Gun0EF4R47ffHeOt1f9W2VdyR7pb5A8jF5E&amp;sref=10lNAhZ9\" rel=\"nofollow noopener\" target=\"_blank\" data-ylk=\"slk:statement;cpos:7;pos:1;elm:context_link;itc:0;sec:content-canvas\" class=\"link \">statement<\/a> sent to <em>Bloomberg<\/em>, an OpenAI spokesperson said the company is proud of its \u201ctrack record providing the most capable and safest AI systems&#8221; and it believes in its &#8220;scientific approach to addressing risk.\u201d It added: \u201cWe agree that rigorous debate is crucial given the significance of this technology and we&#8217;ll continue to engage with governments, civil society and other communities around the world.\u201d<\/p>\n<p>The signatories are calling on AI companies to commit to four key principles:<\/p>\n<ul class=\"caas-list caas-list-bullet\">\n<li>\n<p>Refraining from retaliating against employees who voice safety concerns<\/p>\n<\/li>\n<li>\n<p>Supporting an anonymous system for whistleblowers to alert the public and regulators about risks<\/p>\n<\/li>\n<li>\n<p>Allowing a culture of open criticism<\/p>\n<\/li>\n<li>\n<p>And avoiding non-disparagement or non-disclosure agreements that restrict employees from speaking out<\/p>\n<\/li>\n<\/ul>\n<p>The letter comes amid growing scrutiny of OpenAI&#8217;s practices, including the <a data-i13n=\"elm:context_link;elmt:doNotAffiliate;cpos:8;pos:1\" class=\"link \" href=\"https:\/\/www.engadget.com\/the-openai-team-tasked-with-protecting-humanity-is-no-more-183433377.html\" data-ylk=\"slk:disbandment;elm:context_link;elmt:doNotAffiliate;cpos:8;pos:1;itc:0;sec:content-canvas\">disbandment<\/a> of its &#8220;superalignment&#8221; safety team and the <a data-i13n=\"elm:context_link;elmt:doNotAffiliate;cpos:9;pos:1\" class=\"link \" href=\"https:\/\/www.engadget.com\/openai-co-founder-and-chief-scientist-ilya-sutskever-is-leaving-the-company-054650964.html\" data-ylk=\"slk:departure;elm:context_link;elmt:doNotAffiliate;cpos:9;pos:1;itc:0;sec:content-canvas\">departure<\/a> of key figures like co-founder Ilya Sutskever and Jan Leike, who <a data-i13n=\"elm:context_link;elmt:doNotAffiliate;cpos:10;pos:1\" class=\"link \" href=\"https:\/\/x.com\/janleike\/status\/1791498184671605209\" rel=\"nofollow noopener\" target=\"_blank\" data-ylk=\"slk:criticized;elm:context_link;elmt:doNotAffiliate;cpos:10;pos:1;itc:0;sec:content-canvas\">criticized<\/a> the company&#8217;s prioritization of &#8220;shiny products&#8221; over safety.<\/p>\n<p><strong>Update, June 05 2024, 11:51AM ET:<\/strong> This story has been updated to include statements from OpenAI.<\/p>\n<\/div>\n<p><br \/>\n<br \/><a href=\"https:\/\/www.engadget.com\/former-openai-google-and-anthropic-workers-are-asking-ai-companies-for-more-whistleblower-protections-175916744.html?src=rss\">Source link <\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>A group of current and former employees from leading AI companies like OpenAI, Google DeepMind and Anthropic has signed an open letter asking for greater<\/p>\n","protected":false},"author":1,"featured_media":80210,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_monsterinsights_skip_tracking":false,"_monsterinsights_sitenote_active":false,"_monsterinsights_sitenote_note":"","_monsterinsights_sitenote_category":0,"footnotes":""},"categories":[157],"tags":[],"class_list":["post-80209","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-gadget"],"aioseo_notices":[],"_links":{"self":[{"href":"https:\/\/neclink.com\/index.php\/wp-json\/wp\/v2\/posts\/80209","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/neclink.com\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/neclink.com\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/neclink.com\/index.php\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/neclink.com\/index.php\/wp-json\/wp\/v2\/comments?post=80209"}],"version-history":[{"count":0,"href":"https:\/\/neclink.com\/index.php\/wp-json\/wp\/v2\/posts\/80209\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/neclink.com\/index.php\/wp-json\/wp\/v2\/media\/80210"}],"wp:attachment":[{"href":"https:\/\/neclink.com\/index.php\/wp-json\/wp\/v2\/media?parent=80209"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/neclink.com\/index.php\/wp-json\/wp\/v2\/categories?post=80209"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/neclink.com\/index.php\/wp-json\/wp\/v2\/tags?post=80209"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}