{"id":2697,"date":"2024-10-23T15:21:40","date_gmt":"2024-10-23T13:21:40","guid":{"rendered":"https:\/\/wordpress-test.app.u-pariscite.fr\/diip\/?p=2697"},"modified":"2024-10-23T15:22:15","modified_gmt":"2024-10-23T13:22:15","slug":"laurent-daudet-promises-and-challenges-of-massive-scale-ai-the-case-of-large-language-models","status":"publish","type":"post","link":"https:\/\/wordpress-test.app.u-pariscite.fr\/diip\/laurent-daudet-promises-and-challenges-of-massive-scale-ai-the-case-of-large-language-models\/","title":{"rendered":"Laurent Daudet &#8211; Promises and challenges of massive-scale AI \u2013 the case of large language models"},"content":{"rendered":"<p>[et_pb_section fb_built=&#8221;1&#8243; admin_label=&#8221;section&#8221; _builder_version=&#8221;3.22&#8243;][et_pb_row _builder_version=&#8221;3.22.1&#8243; border_color_all=&#8221;#3255c9&#8243; border_style_all=&#8221;groove&#8221;][et_pb_column type=&#8221;1_2&#8243; _builder_version=&#8221;3.22.1&#8243;][et_pb_text _builder_version=&#8221;3.22.1&#8243; min_height=&#8221;11px&#8221; custom_margin=&#8221;||-25px|||&#8221; custom_padding=&#8221;||0px|||&#8221;]<\/p>\n<p class=\"et_pb_module_header\"><span style=\"color: #3255c9\">Laurent Daudet<\/span><\/p>\n<p><span style=\"color: #3255c9\">November 3, 4pm<\/span><\/p>\n<p><span style=\"color: #3255c9\">Room Turing Conseil, 45 rue des Saints P\u00e8res 75006 Paris &amp; Online (Zoom)<\/span><\/p>\n<p>&nbsp;<\/p>\n<p>[\/et_pb_text][\/et_pb_column][et_pb_column type=&#8221;1_2&#8243; _builder_version=&#8221;3.22.1&#8243;][\/et_pb_column][\/et_pb_row][et_pb_row custom_padding=&#8221;44px|||||&#8221; custom_margin=&#8221;80px||80px&#8221; _builder_version=&#8221;3.22.1&#8243;][et_pb_column type=&#8221;2_3&#8243; _builder_version=&#8221;3.22.1&#8243;][et_pb_text _builder_version=&#8221;3.22.1&#8243; custom_margin=&#8221;-51px|||||&#8221; custom_padding=&#8221;0px|||||&#8221;]<\/p>\n<h3><strong>Abstract<\/strong><\/h3>\n<p>&nbsp;<\/p>\n<p>OpenAi\u2019s GPT-3 language model has triggered a new generation of Machine Learning models. Leveraging Transformers architectures at billion-size parameters trained on massive unlabeled datasets, these language models achieve new capabilities such as text generation, question answering, or even zero-shot learning \u2013 tasks the model has not been explicitly trained for. However, training these models represent massive computing tasks, now done on dedicated supercomputers. Scaling up these models will require new hardware and optimized training algorithms.<\/p>\n<p>At LightOn \u2013 a spinoff of university research -, we develop a set of technologies to address these challenges. The Optical Processing Unit (OPU) technology makes some matrix-vector multiplications in a massively parallel fashion, at record-low power consumption. Now accessible on-premises or through the cloud, the OPU technology has been used by engineers and researchers worldwide in a variety of applications, for Machine Learning and scientific computing. We also train in an efficient manner large language models, such as PAGnol (demo at https:\/\/pagnol.lighton.ai ), the largest language model in French, that can be used for various<\/p>\n<p>[\/et_pb_text][\/et_pb_column][et_pb_column type=&#8221;1_3&#8243; _builder_version=&#8221;3.22.1&#8243;][et_pb_video_slider _builder_version=&#8221;3.22.1&#8243;][et_pb_video_slider_item src=&#8221;https:\/\/www.youtube.com\/watch?v=Les4CT5KCIM&#8221; src_webm=&#8221;https:\/\/www.youtube.com\/watch?v=Les4CT5KCIM&#8221; _builder_version=&#8221;3.22.1&#8243; show_image_overlay=&#8221;off&#8221;][\/et_pb_video_slider_item][\/et_pb_video_slider][et_pb_text admin_label=&#8221;Text&#8221; _builder_version=&#8221;3.22.1&#8243;]<\/p>\n<p><em>Click the image below to see slides<\/em><\/p>\n<p>[\/et_pb_text][et_pb_image src=&#8221;https:\/\/wordpress-test.app.u-pariscite.fr\/diip\/wp-content\/uploads\/sites\/27\/2024\/10\/Screenshot-2024-10-23-152145.png&#8221; url=&#8221;https:\/\/wordpress-test.app.u-pariscite.fr\/diip\/wp-content\/uploads\/sites\/27\/2021\/11\/LightOn_diiP_Nov21_slides-min.pdf&#8221; _builder_version=&#8221;3.22.1&#8243; transform_styles__hover_enabled=&#8221;on&#8221; transform_scale__hover_enabled=&#8221;on&#8221; transform_translate__hover_enabled=&#8221;on&#8221; transform_rotate__hover_enabled=&#8221;on&#8221; transform_skew__hover_enabled=&#8221;on&#8221; transform_origin__hover_enabled=&#8221;on&#8221; transform_translate__hover=&#8221;-4px|-4px&#8221;][\/et_pb_image][\/et_pb_column][\/et_pb_row][\/et_pb_section][et_pb_section fb_built=&#8221;1&#8243; _builder_version=&#8221;3.22.1&#8243;][et_pb_row _builder_version=&#8221;3.22.1&#8243;][et_pb_column type=&#8221;1_3&#8243; _builder_version=&#8221;3.22.1&#8243;][et_pb_image src=&#8221;https:\/\/wordpress-test.app.u-pariscite.fr\/diip\/wp-content\/uploads\/sites\/27\/2021\/10\/portrait.jpeg&#8221; align=&#8221;center&#8221; _builder_version=&#8221;3.22.1&#8243; custom_margin=&#8221;20px|125px||||&#8221;][\/et_pb_image][\/et_pb_column][et_pb_column type=&#8221;2_3&#8243; _builder_version=&#8221;3.22.1&#8243;][et_pb_text _builder_version=&#8221;3.22.1&#8243; custom_padding=&#8221;0px|||||&#8221;]<\/p>\n<h3><strong>Laurent Daudet<br \/><\/strong>CTO and co-founder at LightOn, Professor (on leave) of physics (Universit\u00e9 Paris Cit\u00e9)<\/p>\n<p><strong><br \/><\/strong><\/h3>\n<p>Laurent Daudet is currently employed as CTO at LightOn, a startup he co-founded in 2016, where he manages cross-disciplinary R&amp;D projects, involving machine learning, optics, signal processing, electronics, and software engineering. Laurent is a recognized expert in signal processing and wave physics, and is currently on leave from his position of Professor of Physics at the Universit\u00e9 Paris Cit\u00e9. Prior to that or in parallel, he has held various academic positions: fellow of the Institut Universitaire de France, associate professor at Universit\u00e9 Pierre et Marie Curie, Visiting Senior Lecturer at Queen Mary University of London, UK, Visiting Professor at the National Institute for Informatics in Tokyo, Japan. Laurent has authored or co-authored more than 200 scientific publications, has been a consultant to various small and large companies, and is a co-inventor in several patents. He is a graduate in physics from Ecole Normale Sup\u00e9rieure in Paris, and holds a PhD in Applied Mathematics from Marseille University.<\/p>\n<p>[\/et_pb_text][\/et_pb_column][\/et_pb_row][et_pb_row custom_margin=&#8221;120px||&#8221; _builder_version=&#8221;3.22.1&#8243; locked=&#8221;off&#8221;][et_pb_column type=&#8221;4_4&#8243; _builder_version=&#8221;3.0.47&#8243;][et_pb_divider _builder_version=&#8221;3.22.1&#8243;][\/et_pb_divider][et_pb_text admin_label=&#8221;\u00c0 lire aussi&#8221; _builder_version=&#8221;3.22.1&#8243; z_index_tablet=&#8221;500&#8243;]<\/p>\n<h2><span class=\"st\">Other distinguished lectures<br \/><\/span><\/h2>\n<p>[\/et_pb_text][et_pb_blog posts_number=&#8221;4&#8243; include_categories=&#8221;65&#8243; show_author=&#8221;off&#8221; show_date=&#8221;off&#8221; show_pagination=&#8221;off&#8221; module_id=&#8221;page_type_blog&#8221; _builder_version=&#8221;3.22.1&#8243; header_level=&#8221;h4&#8243; border_width_bottom_fullwidth=&#8221;1px&#8221; border_color_bottom_fullwidth=&#8221;rgba(51,51,51,0.18)&#8221; custom_padding=&#8221;||50px|&#8221; z_index_tablet=&#8221;500&#8243;][\/et_pb_blog][\/et_pb_column][\/et_pb_row][\/et_pb_section]<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Laurent Daudet November 3, 4pm Room Turing Conseil, 45 rue des Saints P\u00e8res 75006 Paris &amp; Online (Zoom) &nbsp;Abstract &nbsp; OpenAi\u2019s GPT-3 language model has triggered a new generation of Machine Learning models. Leveraging Transformers architectures at billion-size parameters trained on massive unlabeled datasets, these language models achieve new capabilities such as text generation, question&hellip; <a class=\"continue\" href=\"https:\/\/wordpress-test.app.u-pariscite.fr\/diip\/laurent-daudet-promises-and-challenges-of-massive-scale-ai-the-case-of-large-language-models\/\">Lire la suite<span> Laurent Daudet &#8211; Promises and challenges of massive-scale AI \u2013 the case of large language models<\/span><\/a><\/p>\n","protected":false},"author":560,"featured_media":674,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"_et_pb_use_builder":"on","_et_pb_old_content":"","_et_gb_content_width":"","footnotes":""},"categories":[57,1,65],"tags":[],"class_list":["post-2697","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-57","category-diip","category-distinguished-lectures"],"_links":{"self":[{"href":"https:\/\/wordpress-test.app.u-pariscite.fr\/diip\/wp-json\/wp\/v2\/posts\/2697","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/wordpress-test.app.u-pariscite.fr\/diip\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/wordpress-test.app.u-pariscite.fr\/diip\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/wordpress-test.app.u-pariscite.fr\/diip\/wp-json\/wp\/v2\/users\/560"}],"replies":[{"embeddable":true,"href":"https:\/\/wordpress-test.app.u-pariscite.fr\/diip\/wp-json\/wp\/v2\/comments?post=2697"}],"version-history":[{"count":3,"href":"https:\/\/wordpress-test.app.u-pariscite.fr\/diip\/wp-json\/wp\/v2\/posts\/2697\/revisions"}],"predecessor-version":[{"id":2702,"href":"https:\/\/wordpress-test.app.u-pariscite.fr\/diip\/wp-json\/wp\/v2\/posts\/2697\/revisions\/2702"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/wordpress-test.app.u-pariscite.fr\/diip\/wp-json\/wp\/v2\/media\/674"}],"wp:attachment":[{"href":"https:\/\/wordpress-test.app.u-pariscite.fr\/diip\/wp-json\/wp\/v2\/media?parent=2697"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/wordpress-test.app.u-pariscite.fr\/diip\/wp-json\/wp\/v2\/categories?post=2697"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/wordpress-test.app.u-pariscite.fr\/diip\/wp-json\/wp\/v2\/tags?post=2697"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}