{"id":15390,"date":"2023-12-07T14:49:25","date_gmt":"2023-12-07T14:49:25","guid":{"rendered":"http:\/\/scannn.com\/introducing-purple-llama-for-safe-and-responsible-ai-development\/"},"modified":"2023-12-07T14:49:25","modified_gmt":"2023-12-07T14:49:25","slug":"introducing-purple-llama-for-safe-and-responsible-ai-development","status":"publish","type":"post","link":"https:\/\/scannn.com\/lv\/introducing-purple-llama-for-safe-and-responsible-ai-development\/","title":{"rendered":"Introducing Purple Llama for Safe and Responsible AI Development"},"content":{"rendered":"<p> <br \/>\n<\/p>\n<div>\n<p><span style=\"font-weight: 400;\">With over 100 million downloads of <\/span><a href=\"https:\/\/about.fb.com\/news\/2023\/07\/llama-2\/\"><span style=\"font-weight: 400;\">Llama<\/span><\/a><span style=\"font-weight: 400;\"> models to date, a lot of this innovation is being fueled by open models. In order to build trust in the developers driving this new wave of innovation, we\u2019re launching <\/span><a href=\"https:\/\/ai.meta.com\/blog\/purple-llama-open-trust-safety-generative-ai\"><span style=\"font-weight: 400;\">Purple Llama<\/span><\/a><span style=\"font-weight: 400;\">, an umbrella project that will bring together tools and evaluations to help developers build responsibly with open generative AI models.\u00a0<\/span><\/p>\n<p><b>Why purple? <\/b><span style=\"font-weight: 400;\">Borrowing a <\/span><a href=\"https:\/\/danielmiessler.com\/p\/red-blue-purple-teams\/\"><span style=\"font-weight: 400;\">concept<\/span><\/a><span style=\"font-weight: 400;\"> from the cybersecurity world, we believe that to truly mitigate the challenges that generative AI presents, we need to take both attack (red team) and defensive (blue team) postures. Purple teaming, composed of both red and blue team responsibilities, is a collaborative approach to evaluating and mitigating potential risks.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">To start, Purple Llama will include tools and evaluations for cybersecurity and input\/output safeguards, with more to come in the near future. Components within the Purple Llama project will be licensed permissively enabling both research and commercial use. We believe this is a major step in enabling collaboration among developers and standardizing trust and safety tools for generative AI.<\/span><\/p>\n<h2><span style=\"font-weight: 400;\">Cybersecurity<\/span><\/h2>\n<p><span style=\"font-weight: 400;\">We are sharing what we believe is the first industry-wide set of cyber security safety evaluations for Large Language Models (LLMs). These benchmarks are based on industry guidance and standards and are built in collaboration with our security experts. With this initial release, we aim to provide tools that will help address risks outlined in the <\/span><a href=\"https:\/\/www.whitehouse.gov\/briefing-room\/statements-releases\/2023\/07\/21\/fact-sheet-biden-harris-administration-secures-voluntary-commitments-from-leading-artificial-intelligence-companies-to-manage-the-risks-posed-by-ai\/\"><span style=\"font-weight: 400;\">White House commitments<\/span><\/a><span style=\"font-weight: 400;\"> including:\u00a0<\/span><\/p>\n<ul>\n<li style=\"font-weight: 400;\" aria-level=\"1\"><span style=\"font-weight: 400;\">Metrics for quantifying LLM cybersecurity risk<\/span><\/li>\n<li style=\"font-weight: 400;\" aria-level=\"1\"><span style=\"font-weight: 400;\">Tools to evaluate the frequency of insecure code suggestions\u00a0<\/span><\/li>\n<li style=\"font-weight: 400;\" aria-level=\"1\"><span style=\"font-weight: 400;\">Tools to evaluate LLMs to make it harder to generate malicious code or aid in carrying out cyber attacks<\/span><\/li>\n<\/ul>\n<p><span style=\"font-weight: 400;\">We believe these tools will reduce the frequency of insecure AI-generated code suggested by LLMs and reduce the helpfulness of LLMs to cyber adversaries.\u00a0<\/span><\/p>\n<h2><span style=\"font-weight: 400;\">Input\/Output Safeguards\u00a0<\/span><\/h2>\n<p><span style=\"font-weight: 400;\">As we outlined in <\/span><a href=\"https:\/\/ai.meta.com\/llama\/responsible-use-guide\/\"><span style=\"font-weight: 400;\">Llama 2\u2019s Responsible Use Guide<\/span><\/a><span style=\"font-weight: 400;\">, we recommend that all inputs and outputs to the LLM be checked and filtered in accordance with <\/span><span style=\"font-weight: 400;\">content guidelines appropriate to the application<\/span><span style=\"font-weight: 400;\">.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">To support this, we are releasing Llama Guard, an openly available foundational model to help developers avoid generating potentially risky outputs. As part of our ongoing commitment to open and transparent science, we are releasing our methodology and an extended discussion of our results in our paper. This model has been trained on a mix of publicly available datasets to enable detection of common types of potentially risky or violating content. Ultimately, our vision is to enable developers to customize future versions to support relevant use cases based on their own requirements and make it easier to adopt best practices and improve the open ecosystem.<\/span><\/p>\n<h2><span style=\"font-weight: 400;\">An Open Ecosystem<\/span><\/h2>\n<p><span style=\"font-weight: 400;\">Taking an open approach to AI is not new for Meta. Exploratory research, open science and cross-collaboration are foundational to our AI efforts <\/span><span style=\"font-weight: 400;\">and we believe there\u2019s an important opportunity to create an open ecosystem. <\/span><span style=\"font-weight: 400;\">This collaborative mindset was at the forefront when Llama 2 launched in July with over 100 partners and we\u2019re excited to share that many of those same partners are working with us on open trust and safety including: AI Alliance, AMD, Anyscale, AWS, Bain, CloudFlare, Databricks, Dell Technologies, Dropbox, Google Cloud, Hugging Face, IBM, Intel, Microsoft, MLCommons, Nvidia, Oracle, Orange, Scale AI, Together.AI and many more to come.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">We\u2019re excited to collaborate with our partners and others who share the same vision of an open ecosystem of responsibly-developed generative AI.\u00a0<\/span><\/p>\n<\/p><\/div>\n<p><script async defer crossorigin=\"anonymous\" src=\"https:\/\/connect.facebook.net\/en_US\/sdk.js#xfbml=1&#038;version=v5.0\"><\/script><br \/>\n<br \/><br \/>\n<br \/><a href=\"https:\/\/about.fb.com\/news\/2023\/12\/purple-llama-safe-responsible-ai-development\/\">Source link <\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>With over 100 million downloads of Llama models to date, a lot of this innovation is being fueled by open models. In order to build trust in the developers driving this new wave of innovation, we\u2019re launching Purple Llama, an umbrella project that will bring together tools and evaluations to help developers build responsibly with [&hellip;]<\/p>\n","protected":false},"author":16,"featured_media":15391,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[123],"tags":[],"class_list":["post-15390","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-facebook"],"_links":{"self":[{"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/posts\/15390","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/users\/16"}],"replies":[{"embeddable":true,"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/comments?post=15390"}],"version-history":[{"count":0,"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/posts\/15390\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/media\/15391"}],"wp:attachment":[{"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/media?parent=15390"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/categories?post=15390"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/tags?post=15390"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}