{"id":130692,"date":"2022-12-20T16:55:44","date_gmt":"2022-12-20T16:55:44","guid":{"rendered":"https:\/\/news.microsoft.com\/?p=447965"},"modified":"2022-12-20T16:55:44","modified_gmt":"2022-12-20T16:55:44","slug":"tapping-into-large-language-models-with-microsofts-turing-academic-program","status":"publish","type":"post","link":"https:\/\/sickgaming.net\/blog\/2022\/12\/20\/tapping-into-large-language-models-with-microsofts-turing-academic-program\/","title":{"rendered":"Tapping into Large Language Models with Microsoft\u2019s Turing Academic Program"},"content":{"rendered":"<div><img decoding=\"async\" src=\"https:\/\/blogs.bing.com\/CMSPages\/GetFile.aspx?guid=ece82b01-416d-46a3-b761-00fa21ffc601\" class=\"ff-og-image-inserted\"><\/div>\n<h5>&#8211; By <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/horvitz\/\">Eric Horvitz<\/a>, Chief Scientific Officer, and <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/satiwary\/\">Saurabh Tiwary<\/a>, Corporate VP and Technical Fellow, Microsoft Turing&nbsp;<\/h5>\n<p>The AI research community is focusing attention on large language models (LLMs) given their impressive performance with difficult tasks and value in useful applications. The capabilities of the models have stimulated many research questions: How do these models work and what do they learn? How can we leverage their potential for new innovations and discoveries? How can we mitigate potentially harmful behaviors? These and many other questions are inspiring new and challenging directions for AI research and underscoring the need for different perspectives and proficiencies.&nbsp;<br \/>&nbsp;<br \/>To explore what\u2019s ahead for LLMs and discuss the ways industry, academia, and government could work together to advance understanding of these models, we organized a panel titled, \u201cTowards a Healthy Research Ecosystem for Large Language Models.\u201d We were joined by <a href=\"https:\/\/nam06.safelinks.protection.outlook.com\/?url=https%3A%2F%2Fwww.microsoft.com%2Fen-us%2Fresearch%2Fpeople%2Fhassanam%2F&amp;data=05%7C01%7Csarmc%40microsoft.com%7C18988197d88843ad0ef008dab60ec697%7C72f988bf86f141af91ab2d7cd011db47%7C1%7C0%7C638022471832288302%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&amp;sdata=Pq7B3UrGdaVE%2BeXubQeGnLjLgyl4bSwuoObli1VsfcA%3D&amp;reserved=0\" target=\"_blank\" rel=\"noopener\">Ahmed Awadallah<\/a> from Microsoft Research, <a href=\"https:\/\/nam06.safelinks.protection.outlook.com\/?url=https%3A%2F%2Fbeta.nsf.gov%2Ftip%2Fleadership&amp;data=05%7C01%7Csarmc%40microsoft.com%7C18988197d88843ad0ef008dab60ec697%7C72f988bf86f141af91ab2d7cd011db47%7C1%7C0%7C638022471832288302%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&amp;sdata=p%2FbUal2YgTPzMHMJVefL2H27BgimuNcn%2FI1Au2FSU8A%3D&amp;reserved=0\" target=\"_blank\" rel=\"noopener\">Erwin Gianchandani<\/a> from the National Science Foundation, and <a href=\"https:\/\/cs.stanford.edu\/~pliang\/\">Percy Liang<\/a> from Stanford University who each brought fascinating insights and ideas for consideration. <\/p>\n<div>[embedded content]<br \/>&nbsp;<\/div>\n<p>A key part of the discussion is on the need to expand access to large language models. Building and experimenting with the largest models requires a great deal of data and computing resources, which are often beyond the reach of university-based teams. This challenge was our motivation for launching the <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/collaboration\/microsoft-turing-academic-program\/\">Microsoft Turing Academic Program<\/a> (MS-TAP). Since 2021, we have sought to provide leading academic teams with access to some of the world&#8217;s largest language models. The program reflects our belief in the importance of having diverse and talented teams from academia working with these models.<\/p>\n<p>MS-TAP has supported multiple in-depth collaborations with partner universities. With deep engagement from researchers and domain experts in Microsoft Research, Microsoft Turing, and the Office of the Chief Scientific Officer, we work to better understand model behavior, identify novel applications, explore potential risks, develop mitigations, and improve future models. Participants receive unprecedented access to our 530B parameter <a href=\"https:\/\/arxiv.org\/abs\/2201.11990\">Natural Language Generation model (T-NLGv2)<\/a>, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/blog\/efficiently-and-effectively-scaling-up-language-model-pretraining-for-best-language-representation-model-on-glue-and-superglue\/\">Natural Language Representation model (T-NLRv5)<\/a>, and 2.5B parameter <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/blog\/microsoft-turing-universal-language-representation-model-t-ulrv2-tops-xtreme-leaderboard\/#:~:text=The%20Turing%20Universal%20Language%20Representation,in%20the%20same%20vector%20space.\">Universal Image Language Representation model (T-UILRv2)<\/a> and Azure compute resources to run experiments and evaluations.<br \/>&nbsp;<br \/>MS-TAP Phase 1 involved collaborations with six universities on five projects (see below) and included numerous highlights like these papers: <a href=\"https:\/\/aclanthology.org\/2021.emnlp-main.790.pdf\"><em>Was it \u201csaid\u201d or was it \u201cclaimed\u201d? How linguistic bias affects generative language models<\/em><\/a> (Brown University) <a href=\"https:\/\/arxiv.org\/pdf\/2110.08413.pdf\">and <em>Invariant Language Modeling<\/em><\/a> (\u00c9cole Polytechnique F\u00e9d\u00e9rale de Lausanne) published at the <a href=\"https:\/\/2022.emnlp.org\/\">EMNLP 2022<\/a> conference. More information including contributors and ongoing updates is available on the <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/collaboration\/microsoft-turing-academic-program\/phase-1-collaborations\/\">MS-TAP Phase 1 Collaborations<\/a> page.<br \/>&nbsp;\n<\/p>\n<ul>\n<li><strong>University of California, Berkeley<\/strong> and <strong>University of California, San Francisco<\/strong>: Leveraging large language models for transfer learning in medical notes<\/li>\n<li><strong>Brown Unive<\/strong><strong>rsity<\/strong>: The extent to which large language models exacerbate bias when given different types of biased and unbiased inputs<\/li>\n<li><strong>\u00c9cole Polytechnique F\u00e9d\u00e9rale de Lausanne (EPFL)<\/strong>: Enhancing the robustness of massive language models via invariant representation learning<\/li>\n<li><strong>Georgia Tech<\/strong>: Analyzing and using large pretrained language models for societal good<\/li>\n<li><strong>University <\/strong><strong>of Washington<\/strong>: Analyzing toxicity, factuality and memory<\/li>\n<\/ul>\n<p>The second phase is currently underway and focuses on larger and more complex models. We are collaborating with seven universities on nine projects. Additional details about each effort including contributors are available on the <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/collaboration\/microsoft-turing-academic-program\/phase-2-collaborations\/\">MS-TAP Phase 2 Collaborations<\/a> page. We will share links to forthcoming papers at the close of Phase 2.<br \/>&nbsp;<\/p>\n<ul>\n<li><strong>Carnegie Mellon University<\/strong>: Large language models for dialog evaluation<\/li>\n<li><strong>Carnegie Mellon University<\/strong>: Learning instructible visuo-motor agents through multimodal interactive teaching<\/li>\n<li><strong>\u00c9cole Polytechnique F\u00e9d\u00e9rale de Lausanne (EPFL): <\/strong>Impact of Decoding Strategies for LLMs<\/li>\n<li><strong>Harvard University<\/strong>: Transferring word representations to the electronic health records with disparity<\/li>\n<li><strong>Harvard University<\/strong>: Improving the reasoning ability of large pretrained models by instructional scaffolding<\/li>\n<li><strong>Massachusetts Institute of Technology<\/strong>: Speeding up training and fine-tuning for large-scale NLP models (Paper: <a href=\"https:\/\/arxiv.org\/pdf\/2211.10438.pdf\"><em>SmoothQuant: Accurate and Efficient Post Training Quantization for Large Language Models<\/em><\/a>)<\/li>\n<li><strong>Mila \u2013 Quebec Artificial Intelligence Institute: <\/strong>Reducing the impact of Summaries generated by LLMs<\/li>\n<li><strong>University of <\/strong><strong>Michigan<\/strong>: Enabling transparency and interpretability in Turing natural language representation models<\/li>\n<li><strong>Stanford University<\/strong>: A multi-faceted benchmark for large language models (Paper: <a href=\"https:\/\/arxiv.org\/pdf\/2211.09110.pdf\"><em>Holistic Evaluation of Language Models<\/em><\/a>; Resource: <a href=\"https:\/\/hai.stanford.edu\/news\/language-models-are-changing-ai-we-need-understand-them\"><em>Language Models are Changing AI. We Need to Understand Them<\/em><\/a><em>)<\/em><\/li>\n<\/ul>\n<p>We look forward to continuing our collaborations with our academic partners and to welcoming new ones in future phases.<\/p>\n<p>Academic labs interested in participating in future phases of the program should send email to <a href=\"mailto:turingacademic@microsoft.com\">turingacademic@microsoft.com<\/a><\/p>\n<p><strong>Resources<\/strong><br \/><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/collaboration\/microsoft-turing-academic-program\/phase-2-collaborations\/\">Microsoft Turing Academic Program<\/a><br \/><a href=\"https:\/\/turing.microsoft.com\/\">Microsoft Turing<\/a><br \/><a href=\"https:\/\/www.microsoft.com\/en-us\/ai\/responsible-ai?activetab=pivot1:primaryr6\">Microsoft Responsible AI<\/a><br \/><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/event\/microsoft-research-summit-2022\/\">Microsoft Research Summit 2022<\/a> <\/p>\n","protected":false},"excerpt":{"rendered":"<p>&#8211; By Eric Horvitz, Chief Scientific Officer, and Saurabh Tiwary, Corporate VP and Technical Fellow, Microsoft Turing&nbsp; The AI research community is focusing attention on large language models (LLMs) given their impressive performance with difficult tasks and value in useful applications. The capabilities of the models have stimulated many research questions: How do these models [&hellip;]<\/p>\n","protected":false},"author":2,"featured_media":0,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[49],"tags":[166,50],"class_list":["post-130692","post","type-post","status-publish","format-standard","hentry","category-microsoft-news","tag-bing","tag-recent-news"],"_links":{"self":[{"href":"https:\/\/sickgaming.net\/blog\/wp-json\/wp\/v2\/posts\/130692","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/sickgaming.net\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/sickgaming.net\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/sickgaming.net\/blog\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/sickgaming.net\/blog\/wp-json\/wp\/v2\/comments?post=130692"}],"version-history":[{"count":0,"href":"https:\/\/sickgaming.net\/blog\/wp-json\/wp\/v2\/posts\/130692\/revisions"}],"wp:attachment":[{"href":"https:\/\/sickgaming.net\/blog\/wp-json\/wp\/v2\/media?parent=130692"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/sickgaming.net\/blog\/wp-json\/wp\/v2\/categories?post=130692"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/sickgaming.net\/blog\/wp-json\/wp\/v2\/tags?post=130692"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}