post by [deleted] · · ? · GW · 0 comments

This is a link post for

0 comments

Comments sorted by top scores.

comment by M. Y. Zuo · 2023-05-04T12:47:16.225Z · LW(p) · GW(p)

Public assessments of existing generative AI systems. The Administration is announcing an independent commitment from leading AI developers, including Anthropic, Google, Hugging Face, Microsoft, NVIDIA, OpenAI, and Stability AI, to participate in a public evaluation of AI systems, consistent with responsible disclosure principles—on an evaluation platform developed by Scale AI—at the AI Village at DEFCON 31. This will allow these models to be evaluated thoroughly by thousands of community partners and AI experts to explore how the models align with the principles and practices outlined in the Biden-Harris Administration’s Blueprint for an AI Bill of Rights and AI Risk Management Framework.

I don't know anything about the 'evaluation platform developed by Scale AI—at the AI Village at DEFCON 31'.

Does anyone know if this is a credible method?

Replies from: StellaAthena, gworley, ESRogs
comment by StellaAthena · 2023-05-04T23:48:59.006Z · LW(p) · GW(p)

Hi, I’m helping support the event. I think that some mistranslation happened by a non-AI person. The event is about having humans get together and do prompt hacking and similar on a variety of models side-by-side. ScaleAI built the app that’s orchestrating the routing of info, model querying, and human interaction. Scale’s platform isn’t doing the evaluation itself. That’s being done by users on-site and then by ML and security researchers analyzing the data after the fact.

comment by Gordon Seidoh Worley (gworley) · 2023-05-04T16:04:07.566Z · LW(p) · GW(p)

My guess would be that it'll be on the level of evals done internally by these companies today to make sure generative AI models don't say racist things or hand out bomb making instructions, etc.

comment by ESRogs · 2023-05-04T20:09:23.105Z · LW(p) · GW(p)

I don't know anything about the 'evaluation platform developed by Scale AI—at the AI Village at DEFCON 31'.

Looks like it's this.

comment by Jayson_Virissimo · 2023-05-04T16:41:19.215Z · LW(p) · GW(p)

I noticed that Meta (Facebook) isn't mentioned as being participants. Is that because they weren't asked to or because they were asked but declined?

Replies from: RomanS
comment by RomanS · 2023-05-04T19:19:42.246Z · LW(p) · GW(p)

Also no Tesla, it spite of:

  • Tesla's ~4 million of AI-powered wheeled robots on the road
  • Elon being one of the most prominent people pushing for AI regulations
  • Elon himself claiming that the Tesla AI is among the smartest AIs (which makes sense, given the complexity of the task, and how Teslas are solving it) 

Maybe Meta and Tesla will join later. If not, perhaps there is some political conflict in the play. 

Replies from: ESRogs, StellaAthena
comment by ESRogs · 2023-05-04T20:12:30.451Z · LW(p) · GW(p)

I would imagine one of the major factors explaining Tesla's absence is that people are most worried about LLMs at the moment, and Tesla is not a leader in LLMs.

(I agree that people often seem to overlook Tesla as a leader in AI in general.)

comment by StellaAthena · 2023-05-05T00:21:01.779Z · LW(p) · GW(p)

What deployed LLM system does Tesla make that you think should be evaluated alongside ChatGPT, Bard, etc?

Replies from: RomanS
comment by RomanS · 2023-05-05T09:02:16.185Z · LW(p) · GW(p)

I'm not aware of any LLM systems by Tesla. 

But their self-driving AI is definitely worth evaluating. The task of self-driving on a busy city road is extremely hard to solve (if not AGI-complete), yet their AI is surprisingly good at that. It still fails in many circumstances, but is surprisingly good overall. Tesla could be closer to AGI than most people realize.

comment by Mitchell_Porter · 2023-05-04T13:30:14.199Z · LW(p) · GW(p)

This is a big day. Up to this point, the future of AI in the US has mostly been in the hands of the tech companies and "the market". Presumably IT people in the military and in intelligence were keeping up too... But now, the US government is getting involved in a serious way, in managing the evolution of AI and its impact on society. AI is now, very officially, a subject of public policy. The NSF's seven new institutes are listed here. 

Replies from: gworley
comment by Gordon Seidoh Worley (gworley) · 2023-05-04T16:02:22.208Z · LW(p) · GW(p)

Unclear to me how "serious" this really is. The US government has its hands in lots of things and spends money on lots of stuff. It's more serious than it was before, but to me this seems pretty close to the least they could be doing and not be seen as ignoring AI in ways that would be used against them in the next election cycle.

Replies from: nathan-helm-burger, None
comment by Nathan Helm-Burger (nathan-helm-burger) · 2023-05-05T22:43:59.993Z · LW(p) · GW(p)

Here's the details on the NSF institutes...  Sound mostly irrelevant to AInotkilleveryoneism. Some seem likely to produce minor good things for the world, like perhaps the education and agriculture focused programs. Others seem potentially harmfully accelerationist, like the Neural & Cognitive science program. Cybersecurity might be good, we certainly could do with better cybersecurity. The Trustworthy AI just sounds like Social Justice AI concerns not relevant to AInotkilleveryoneism.

Trustworthy AI

NSF Institute for Trustworthy AI in Law & Society (TRAILS)

Led by the University of Maryland, TRAILS aims to transform the practice of AI from one driven primarily by technological innovation to one driven with attention to ethics, human rights, and support for communities whose voices have been marginalized into mainstream AI. TRAILS will be the first Institute of its kind to integrate participatory design, technology, and governance of AI systems and technologies and will focus on investigating what trust in AI looks like, whether current technical solutions for AI can be trusted, and which policy models can effectively sustain AI trustworthiness. TRAILS is funded by a partnership between NSF and NIST.

Intelligent Agents for Next-Generation Cybersecurity

AI Institute for Agent-based Cyber Threat Intelligence and Operation (ACTION)

Led by the University of California, Santa Barbara, this Institute will develop novel approaches that leverage AI to anticipate and take corrective actions against cyberthreats that target the security and privacy of computer networks and their users. The team of researchers will work with experts in security operations to develop a revolutionary approach to cybersecurity, in which AI-enabled intelligent security agents cooperate with humans across the cyber-defense life cycle to jointly improve the resilience of security of computer systems over time. ACTION is funded by a partnership between NSF, DHS S&T, and IBM.

Climate Smart Agriculture and Forestry

AI Institute for Climate-Land Interactions, Mitigation, Adaptation, Tradeoffs and Economy (AI-CLIMATE)

Led by the University of Minnesota Twin Cities, this Institute aims to advance foundational AI by incorporating knowledge from agriculture and forestry sciences and leveraging these unique, new AI methods to curb climate effects while lifting rural economies. By creating a new scientific discipline and innovation ecosystem intersecting AI and climate-smart agriculture and forestry, our researchers and practitioners will discover and invent compelling AI-powered knowledge and solutions. Examples include AI-enhanced estimation methods of greenhouse gases and specialized field-to-market decision support tools. A key goal is to lower the cost of and improve accounting for carbon in farms and forests to empower carbon markets and inform decision-making. The Institute will also expand and diversify rural and urban AI workforces. AI-CLIMATE is funded by USDA-NIFA.

Neural and Cognitive Foundations of Artificial Intelligence

AI Institute for Artificial and Natural Intelligence (ARNI)

Led by Columbia University, this Institute will draw together top researchers across the country to focus on a national priority: connecting the major progress made in AI systems to the revolution in our understanding of the brain. ARNI will meet the urgent need for new paradigms of interdisciplinary research between neuroscience, cognitive science, and AI. This will accelerate progress in all three fields and broaden the transformative impact on society in the next decade. ARNI is funded by a partnership between NSF and OUSD (R&E).

AI for Decision Making

AI-Institute for Societal Decision Making (AI-SDM)

Led by Carnegie Mellon University, this Institute seeks to create human-centric AI for decision making to bolster effective response in uncertain, dynamic, and resource-constrained scenarios like disaster management and public health. By bringing together an interdisciplinary team of AI and social science researchers, AI-SDM will enable emergency managers, public health officials, first responders, community workers, and the public to make decisions that are data driven, robust, agile, resource efficient, and trustworthy. The vision of AI-SDM will be realized via development of AI theory and methods, translational research, training, and outreach, enabled by partnerships with diverse universities, government organizations, corporate partners, community colleges, public libraries, and high schools.

AI-Augmented Learning to Expand Education Opportunities and Improve Outcomes

AI Institute for Inclusive Intelligent Technologies for Education (INVITE)

Led by the University of Illinois, Urbana-Champaign, this Institute seeks to fundamentally reframe how educational technologies interact with learners by developing AI tools and approaches to support three crucial noncognitive skills known to underlie effective learning: persistence, academic resilience, and collaboration. The Institute’s use-inspired research will focus on how children communicate STEM content, how they learn to persist through challenging work, and how teachers support and promote noncognitive skill development. The resultant AI-based tools will be integrated into classrooms to empower teachers to support learners in more developmentally appropriate ways.

AI Institute for Exceptional Education (AI4ExceptionalEd)

Led by the University at Buffalo, this Institute will work toward universal speech and language screening for children. The framework, the AI screener, will analyze video and audio streams of children during classroom interactions and assess the need for evidence-based interventions tailored to individual needs of students. The institute will serve children in need of ability-based speech and language services, advance foundational AI technologies and enhance understanding of childhood speech and language development. The AI Institute for Exceptional Education was previously announced in January 2023. The INVITE and AI4ExceptionalEd Institutes are funded by a partnership between NSF and ED-IES.

comment by [deleted] · 2023-05-05T10:00:07.980Z · LW(p) · GW(p)

What would be a reasonable standard of action by you? Genuinely asking

Replies from: nathan-helm-burger
comment by Nathan Helm-Burger (nathan-helm-burger) · 2023-05-05T22:46:48.737Z · LW(p) · GW(p)

From my point of view, I'd love to have the US and UK govs classify cutting edge AI tech as dangerous weapons technology, and start applying military rules of discipline around these like they do for high-tech weapons R&D. Security clearances, export controls, significant government oversight and cybersecurity requirements, etc. I think that's a reasonable step at this point.

comment by VojtaKovarik · 2023-05-08T19:19:10.943Z · LW(p) · GW(p)

Regarding the "assesment platform with ScaleAI": Does anybody here plan to take some academia-related actions based on this? Or do you know somebody considering this? (I am considering this, but see below.)

 

Context: This seems like an opportunity to legitimise red-teaming as an academically respectable thing. Also, an opportunity to have a more serious (incl. academic) discussion on mechanism/incentive design regarding red-teaming (cf. Legitimising AI Red-Teaming by Pulic [LW · GW]). I am considering taking some actions in this respect --- for example, getting together a white-paper or call for proposals. However, it seems that other people might be interested in this as well, and it seems important to coordinate (etc etc).

Replies from: StellaAthena
comment by StellaAthena · 2023-05-09T06:30:51.670Z · LW(p) · GW(p)

Red teaming has always been a legitimate academic thing? I don’t know what background you’re coming from but… you’re very far off.

But yes, the event organizers will be writing a paper about it and publishing the data (after it’s been anonymized).

Replies from: VojtaKovarik, VojtaKovarik
comment by VojtaKovarik · 2023-05-09T18:33:39.508Z · LW(p) · GW(p)

But yes, the event organizers will be writing a paper about it and publishing the data (after it’s been anonymized).

I imagine this would primarily be a report from the competition? What I was thinking about was more about how this sort of assessment should be done in general, what are the similarities and differences between cybersecurity, and how to squeeze more utility out of it. For example, a (naive version of) one low-hanging fruit is to withhold 10% of the obtained data (from the AI companies, then test those jailbreak strategies later). This would give us some insight into whether the current "alignment" methods generalise, or whether we are closer to playing whack-a-mole. Similarly to how we use test data in ML.

There are many more considerations, and many more things you can do. And I don't claim to have all the answers, nor to be the optimal person to be writing about them. Just that it would be good if somebody was doing that (and wondering whether that is happening :-) ).

comment by VojtaKovarik · 2023-05-09T18:15:56.731Z · LW(p) · GW(p)

Red teaming has always been a legitimate academic thing? I don’t know what background you’re coming from but… you’re very far off.

Theoretical CS/AI/game theory, rather than cybersecurity. Given the lack of cybersec background, I acknowledge I might be very far off.

To me, it seems that the perception in cybersecurity might be different from the perception outside of it. Also, red teaming in the context of AI models might have important differences from cybersecurity context. Also, red teaming by public seems, to me, different from internal red-teaming or bounties. (Though this might be one of the things where I am far off.)