From ed15c6c3ab6fbca37d1076355f50573836aaf204 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Ezequiel=20Leonardo=20Casta=C3=B1o?= <14986783+ELC@users.noreply.github.com> Date: Tue, 17 Jun 2025 22:05:10 -0300 Subject: [PATCH 1/8] Scraped pytorchconf-2024 Fixes #xxx Event config: ~~~yaml repo_dir: W:\Repositories\pyvideo-data # Copy the event template here and adapt to the event parameters # Only repo_dir: and events: are loaded # ============================================================================= events: # - title: SciPy 2024 # dir: scipy-2024 # youtube_list: # - https://www.youtube.com/playlist?list=PL1PbeFStIOoO7rDLs431H-rn0h24Wr80S # related_urls: # - label: Conference Website # url: https://www.scipy2024.scipy.org/ # language: eng # dates: # begin: 2024-07-08 # end: 2024-07-14 # default: 2024-07-08 # minimal_download: false # issue: xxx # overwrite: # # all: true # takes precedence over add_new_files and existing_files_fields # add_new_files: true # existing_files_fields: # - duration # - thumbnail_url # - videos # - description # - language # - recorded # - related_urls # - speakers # - tags # - title # tags: - title: JupyterCon 2020 dir: jupytercon-2020 youtube_list: - https://www.youtube.com/playlist?list=PL_1BH3ug7n1KCiM-g0x9ZoWuDNhwuM1kr related_urls: - label: Conference Website url: https://web.archive.org/web/20201030085456/https://jupytercon.github.io/jupytercon2020-website/ language: eng dates: begin: 2020-10-05 end: 2020-10-17 default: 2020-10-05 minimal_download: false issue: xxx overwrite: # all: true # takes precedence over add_new_files and existing_files_fields add_new_files: true existing_files_fields: - duration - thumbnail_url - videos - description - language - recorded - related_urls - speakers - tags - title tags: - title: JupyterCon 2023 dir: jupytercon-2023 youtube_list: - https://www.youtube.com/playlist?list=PL_1BH3ug7n1Ih_Yy2TmM7MZ2zogSLZvzE related_urls: - label: Conference Website url: https://web.archive.org/web/20230531110007/https://www.jupytercon.com/ language: eng dates: begin: 2023-05-10 end: 2023-05-12 default: 2023-05-10 minimal_download: false issue: xxx overwrite: # all: true # takes precedence over add_new_files and existing_files_fields add_new_files: true existing_files_fields: - duration - thumbnail_url - videos - description - language - recorded - related_urls - speakers - tags - title tags: - title: Py4AI 2024 dir: py4ai-2024 youtube_list: - https://www.youtube.com/playlist?list=PL0RwQVm3YPu5k9iIaQUehwgh2M1DgKWaT related_urls: - label: Conference Website url: https://web.archive.org/web/20240511071059/https://www.py4ai.com/ language: eng dates: begin: 2024-03-16 end: 2024-03-16 default: 2024-03-16 minimal_download: false issue: xxx overwrite: # all: true # takes precedence over add_new_files and existing_files_fields add_new_files: true existing_files_fields: - duration - thumbnail_url - videos - description - language - recorded - related_urls - speakers - tags - title tags: - title: XtremePython 2024 dir: xtremepython-2024 youtube_list: - https://www.youtube.com/playlist?list=PL9XJvIlpSqocUf9t_YcRHmNza5WkTACvJ related_urls: - label: Conference Website url: https://xtremepython.dev/2024/ language: eng dates: begin: 2024-11-19 end: 2024-11-19 default: 2024-11-19 minimal_download: false issue: xxx overwrite: # all: true # takes precedence over add_new_files and existing_files_fields add_new_files: true existing_files_fields: - duration - thumbnail_url - videos - description - language - recorded - related_urls - speakers - tags - title tags: - title: XtremePython 2023 dir: xtremepython-2023 youtube_list: - https://www.youtube.com/playlist?list=PL9XJvIlpSqoeSy5TcB0FGG_-6q3NeQNPl related_urls: - label: Conference Website url: https://xtremepython.dev/2023/ language: eng dates: begin: 2023-04-16 end: 2023-04-16 default: 2023-04-16 minimal_download: false issue: xxx overwrite: # all: true # takes precedence over add_new_files and existing_files_fields add_new_files: true existing_files_fields: - duration - thumbnail_url - videos - description - language - recorded - related_urls - speakers - tags - title tags: - title: XtremePython 2022 dir: xtremepython-2022 youtube_list: - https://www.youtube.com/playlist?list=PL9XJvIlpSqoc4ZggVIefeet8wjoMDHhms related_urls: - label: Conference Website url: https://xtremepython.dev/2022/ language: eng dates: begin: 2022-12-27 end: 2022-12-27 default: 2022-12-27 minimal_download: false issue: xxx overwrite: # all: true # takes precedence over add_new_files and existing_files_fields add_new_files: true existing_files_fields: - duration - thumbnail_url - videos - description - language - recorded - related_urls - speakers - tags - title tags: - title: XtremePython 2021 dir: xtremepython-2021 youtube_list: - https://www.youtube.com/playlist?list=PL9XJvIlpSqoeT58uNpmwLE8RA2vkr-klo related_urls: - label: Conference Website url: https://xtremepython.dev/2021/ language: eng dates: begin: 2021-11-25 end: 2021-11-25 default: 2021-11-25 minimal_download: false issue: xxx overwrite: # all: true # takes precedence over add_new_files and existing_files_fields add_new_files: true existing_files_fields: - duration - thumbnail_url - videos - description - language - recorded - related_urls - speakers - tags - title tags: - title: Posit:Conf 2024 dir: positconf-2024 youtube_list: - https://www.youtube.com/playlist?list=PL9HYL-VRX0oSFkdF4fJeY63eGDvgofcbn related_urls: - label: Conference Website url: https://web.archive.org/web/20240804081019/https://posit.co/conference/ language: eng dates: begin: 2024-08-12 end: 2024-08-14 default: 2024-08-12 minimal_download: false issue: xxx overwrite: # all: true # takes precedence over add_new_files and existing_files_fields add_new_files: true existing_files_fields: - duration - thumbnail_url - videos - description - language - recorded - related_urls - speakers - tags - title tags: - title: Posit:Conf 2023 dir: positconf-2023 youtube_list: - https://www.youtube.com/playlist?list=PL9HYL-VRX0oRFZslRGHwHuwea7SvAATHp related_urls: - label: Conference Website url: https://web.archive.org/web/20230902080800/https://posit.co/conference language: eng dates: begin: 2023-09-17 end: 2023-09-20 default: 2023-09-17 minimal_download: false issue: xxx overwrite: # all: true # takes precedence over add_new_files and existing_files_fields add_new_files: true existing_files_fields: - duration - thumbnail_url - videos - description - language - recorded - related_urls - speakers - tags - title tags: - title: PyTorch Day 2025 dir: pytorchday-2025 youtube_list: - https://www.youtube.com/playlist?list=PL_lsbAsL_o2DBxQBRA5SoqnTL_inXCOLU related_urls: - label: Conference Website url: https://events.linuxfoundation.org/pytorch-day-france/ language: eng dates: begin: 2025-05-07 end: 2025-05-07 default: 2025-05-07 minimal_download: false issue: xxx overwrite: # all: true # takes precedence over add_new_files and existing_files_fields add_new_files: true existing_files_fields: - duration - thumbnail_url - videos - description - language - recorded - related_urls - speakers - tags - title tags: - title: PyTorch Conference 2024 dir: pytorchconf-2024 youtube_list: - https://www.youtube.com/playlist?list=PL_lsbAsL_o2B_znuvm-pDtV_cRhpqZb8l related_urls: - label: Conference Website url: https://pytorch.org/event/pytorch-conference-2024/ language: eng dates: begin: 2024-09-18 end: 2024-09-19 default: 2024-09-18 minimal_download: false issue: xxx overwrite: # all: true # takes precedence over add_new_files and existing_files_fields add_new_files: true existing_files_fields: - duration - thumbnail_url - videos - description - language - recorded - related_urls - speakers - tags - title tags: - title: PyTorch Conference 2023 dir: pytorchconf-2023 youtube_list: - https://www.youtube.com/playlist?list=PL_lsbAsL_o2BivkGLiDfHY9VqWlaNoZ2O related_urls: - label: Conference Website url: https://pytorch.org/event/pytorch-conference-2023/ language: eng dates: begin: 2023-10-16 end: 2023-10-17 default: 2023-10-16 minimal_download: false issue: xxx overwrite: # all: true # takes precedence over add_new_files and existing_files_fields add_new_files: true existing_files_fields: - duration - thumbnail_url - videos - description - language - recorded - related_urls - speakers - tags - title tags: - title: PyTorch Conference 2019 dir: pytorchconf-2019 youtube_list: - https://www.youtube.com/playlist?list=PL_lsbAsL_o2BY-RrqVDKDcywKnuUTp-f3 language: eng dates: begin: 2019-10-16 end: 2019-10-17 default: 2019-10-16 minimal_download: false issue: xxx overwrite: # all: true # takes precedence over add_new_files and existing_files_fields add_new_files: true existing_files_fields: - duration - thumbnail_url - videos - description - language - recorded - related_urls - speakers - tags - title tags: # ISO_639-3 language codes https://en.wikipedia.org/wiki/ISO_639-3 # languages = { # 'ita': 'Italian', # 'zho': 'Chinese', # 'por': 'Portuguese', # 'ukr': 'Ukrainian', # 'deu': 'German', # 'eng': 'English', # 'rus': 'Russian', # 'fra': 'French', # 'spa': 'Spanish', # 'eus': 'Basque', # 'cat': 'Catalan', # 'glg': 'Galician', # 'kor': 'Korean', # 'lit': 'Lithuanian', # 'jpn': 'Japanese', # 'ces': 'Czech', # 'pol': 'Polish', # 'heb': 'Hebrew', # 'tha': 'Thai', # } ~~~ Scraped with [pyvideo_scrape](https://github.com/pyvideo/pyvideo_scrape) --- pytorchconf-2024/category.json | 3 ++ ...training-davis-wertheimer-linsong-chu.json | 24 ++++++++++++++++ ...a-loading-andrew-ho-ahmad-sharif-meta.json | 24 ++++++++++++++++ ...in-shades-emmanuel-acheampong-robomua.json | 24 ++++++++++++++++ ...e-pytorch-ecosystem-bharath-ramsundar.json | 24 ++++++++++++++++ ...endent-shapes-in-pt2-edward-yang-meta.json | 24 ++++++++++++++++ ...el-j-pienaar-t-chen-m-zolotukhin-p-wu.json | 24 ++++++++++++++++ ...ergen-nachin-mengtao-martin-yuan-meta.json | 24 ++++++++++++++++ ...training-faster-daniel-han-unsloth-ai.json | 24 ++++++++++++++++ ...nd-for-torchinductor-jason-ansel-meta.json | 24 ++++++++++++++++ ...dy-maanav-dalal-yulong-wang-microsoft.json | 24 ++++++++++++++++ ...nhancing-back-eikan-wang-min-jean-cho.json | 24 ++++++++++++++++ ...t-jerry-liu-co-founder-ceo-llamaindex.json | 24 ++++++++++++++++ .../videos/keynote-community-awards.json | 24 ++++++++++++++++ ...ormac-brick-principal-engineer-google.json | 24 ++++++++++++++++ ...f-llms-sebastian-raschka-lightning-ai.json | 24 ++++++++++++++++ ...of-language-modeling-hanna-hajishirzi.json | 24 ++++++++++++++++ ...varshney-s-hooker-a-madry-r-bommasani.json | 24 ++++++++++++++++ ...panel-discussion-scaling-benchmarking.json | 24 ++++++++++++++++ ...-p-wu-w-constable-k-khandelwal-m-yuan.json | 24 ++++++++++++++++ ...eous-computing-ion-stoica-uc-berkeley.json | 24 ++++++++++++++++ .../keynote-welcome-back-opening-remarks.json | 24 ++++++++++++++++ ...executive-director-pytorch-foundation.json | 24 ++++++++++++++++ ...aluation-tool-chip-huyen-voltron-data.json | 24 ++++++++++++++++ ...extension-points-alban-desmaison-meta.json | 24 ++++++++++++++++ ...rnel-libraries-on-rocm-jeff-daily-amd.json | 24 ++++++++++++++++ ...ick-smyth-dan-fernandez-srishti-hegde.json | 24 ++++++++++++++++ ...h-community-resources-tu-zong-zesheng.json | 24 ++++++++++++++++ ...rch-fx-simplify-zifei-xu-tristan-webb.json | 24 ++++++++++++++++ ...e-shailvi-wakhlu-shailvi-ventures-llc.json | 24 ++++++++++++++++ ...ild-lessons-learned-f-omar-sanseviero.json | 24 ++++++++++++++++ ...or-running-generative-a-pareena-verma.json | 28 +++++++++++++++++++ ...hon-c-cuda-operators-richard-zou-meta.json | 24 ++++++++++++++++ ...with-streamingdataset-saaketh-narayan.json | 24 ++++++++++++++++ ...ch-the-performa-yanbo-liang-horace-he.json | 24 ++++++++++++++++ ...for-automated-egyptian-h-susi-gentsch.json | 24 ++++++++++++++++ ...-what-could-go-wrong-nicolas-hug-meta.json | 24 ++++++++++++++++ ...rontier-exascale-system-vineeth-gutta.json | 24 ++++++++++++++++ ...d-pipelining-howard-huang-ke-wen-meta.json | 24 ++++++++++++++++ ...en-lai-kimish-patel-cemal-bilgin-meta.json | 24 ++++++++++++++++ ...pes-in-pytorch-vasiliy-kuznetsov-meta.json | 24 ++++++++++++++++ ...acity-using-pytorch-syed-ahmed-nvidia.json | 24 ++++++++++++++++ ...ch-low-light-denoising-alexis-baudron.json | 24 ++++++++++++++++ ...in-pytorch-jeffrey-wan-horace-he-meta.json | 24 ++++++++++++++++ ...-executorch-olivia-liu-vaun-puri-meta.json | 24 ++++++++++++++++ ...nux-cpus-sunita-nadampalli-amazon-aws.json | 24 ++++++++++++++++ ...ch-release-process-andrey-talman-meta.json | 24 ++++++++++++++++ ...la-auto-sharding-yeounoh-chung-google.json | 24 ++++++++++++++++ ...-minimal-accuracy-loss-jesse-cai-meta.json | 24 ++++++++++++++++ ...with-thunder-luca-antiga-lightning-ai.json | 24 ++++++++++++++++ ...ructure-sahan-paliskara-catherine-lee.json | 24 ++++++++++++++++ ...n-manlaibaatar-avik-chaudhuri-yidi-wu.json | 24 ++++++++++++++++ ...fsdp-l-chu-a-viros-i-martin-b-vaughan.json | 24 ++++++++++++++++ ...-spencer-whitman-vincent-gonguet-meta.json | 24 ++++++++++++++++ ...-and-compilers-jacques-pienaar-google.json | 24 ++++++++++++++++ ...x-and-mojo-mikhail-zolotukhin-modular.json | 24 ++++++++++++++++ ...y-a-conneau-moderated-by-k-khandelwal.json | 24 ++++++++++++++++ ...r-3-suvaditya-mukherjee-shireen-chand.json | 24 ++++++++++++++++ .../pytorch-conference-2024-highlights.json | 28 +++++++++++++++++++ ...-npu-acceleration-felix-baum-qualcomm.json | 24 ++++++++++++++++ ...aying-ooms-mark-saroufim-jane-xu-meta.json | 24 ++++++++++++++++ ...e-innovation-with-sea-anush-elangovan.json | 24 ++++++++++++++++ ...-pytorch-and-intel-kismat-singh-intel.json | 24 ++++++++++++++++ ...cting-the-ai-orchestra-taylor-dolezal.json | 24 ++++++++++++++++ ...language-models-mudhakar-srivatsa-ibm.json | 24 ++++++++++++++++ ...lerating-the-ai-lifecycle-luca-antiga.json | 24 ++++++++++++++++ ...mance-pyt-robert-suderman-ian-nordeng.json | 24 ++++++++++++++++ ...th-arms-global-comp-gian-marco-iodice.json | 24 ++++++++++++++++ ...del-alignment-gerald-shen-jimmy-zhang.json | 24 ++++++++++++++++ ...orm-mark-sherwood-shauheen-zahirazami.json | 24 ++++++++++++++++ ...m-ubiquity-jack-khuu-jesse-white-meta.json | 24 ++++++++++++++++ pytorchconf-2024/videos/startup-showcase.json | 24 ++++++++++++++++ ...e-of-pytorch-ji-li-damien-sereni-meta.json | 24 ++++++++++++++++ ...ce-llm-framework-wing-lian-axolotl-ai.json | 24 ++++++++++++++++ ...ve-datasets-and-models-aaron-gokaslan.json | 24 ++++++++++++++++ ...-ecosystem-arthur-zucker-hugging-face.json | 24 ++++++++++++++++ ...f-the-llama-ecosystem-joe-spisak-meta.json | 24 ++++++++++++++++ ...with-torch-compile-pragaash-ponnusamy.json | 24 ++++++++++++++++ ...-feng-chien-chin-huang-simon-fan-meta.json | 24 ++++++++++++++++ ...erformance-imp-jiong-gong-leslie-fang.json | 24 ++++++++++++++++ ...3d-parallel-wanchao-liang-linsong-chu.json | 24 ++++++++++++++++ ...-in-native-pytorch-evan-smothers-meta.json | 24 ++++++++++++++++ ...orch-mihir-patel-brian-chu-databricks.json | 24 ++++++++++++++++ ...performance-constraints-philip-tillet.json | 24 ++++++++++++++++ ...ml-compilation-tianqi-chen-cmu-octoai.json | 24 ++++++++++++++++ ...-inference-workload-mark-moyou-nvidia.json | 24 ++++++++++++++++ ...plainable-large-languag-rashmi-nagpal.json | 24 ++++++++++++++++ ...-woosuk-kwon-xiaoxuan-liu-uc-berkeley.json | 24 ++++++++++++++++ ...-mini-summit-kartikay-khandelwal-meta.json | 24 ++++++++++++++++ 89 files changed, 2123 insertions(+) create mode 100644 pytorchconf-2024/category.json create mode 100644 pytorchconf-2024/videos/a-distributed-stateful-dataloader-for-large-scale-pretraining-davis-wertheimer-linsong-chu.json create mode 100644 pytorchconf-2024/videos/blobs-to-clips-efficient-end-to-end-video-data-loading-andrew-ho-ahmad-sharif-meta.json create mode 100644 pytorchconf-2024/videos/building-pytorch-computer-vision-algorithms-for-100-skin-shades-emmanuel-acheampong-robomua.json create mode 100644 pytorchconf-2024/videos/building-scientific-computing-infrastructure-software-with-the-pytorch-ecosystem-bharath-ramsundar.json create mode 100644 pytorchconf-2024/videos/data-dependent-shapes-in-pt2-edward-yang-meta.json create mode 100644 pytorchconf-2024/videos/dl-compiler-panel-discussion-p-tillet-j-ansel-j-pienaar-t-chen-m-zolotukhin-p-wu.json create mode 100644 pytorchconf-2024/videos/executorch-beta-and-on-device-generative-ai-support-mergen-nachin-mengtao-martin-yuan-meta.json create mode 100644 pytorchconf-2024/videos/hacks-to-make-llm-training-faster-daniel-han-unsloth-ai.json create mode 100644 pytorchconf-2024/videos/halide-a-halide-backend-for-torchinductor-jason-ansel-meta.json create mode 100644 pytorchconf-2024/videos/implementing-a-custom-torch-compile-backend-a-case-study-maanav-dalal-yulong-wang-microsoft.json create mode 100644 pytorchconf-2024/videos/intel-gpu-in-upstream-pytorch-expanding-gpu-choices-and-enhancing-back-eikan-wang-min-jean-cho.json create mode 100644 pytorchconf-2024/videos/keynote-building-an-advanced-knowledge-assistant-jerry-liu-co-founder-ceo-llamaindex.json create mode 100644 pytorchconf-2024/videos/keynote-community-awards.json create mode 100644 pytorchconf-2024/videos/keynote-enabling-generative-ai-on-the-edge-cormac-brick-principal-engineer-google.json create mode 100644 pytorchconf-2024/videos/keynote-navigating-the-architectural-timeline-of-llms-sebastian-raschka-lightning-ai.json create mode 100644 pytorchconf-2024/videos/keynote-open-language-models-olmo-accelerating-the-science-of-language-modeling-hanna-hajishirzi.json create mode 100644 pytorchconf-2024/videos/keynote-panel-discussion-responsible-ai-k-rooney-k-varshney-s-hooker-a-madry-r-bommasani.json create mode 100644 pytorchconf-2024/videos/keynote-panel-discussion-scaling-benchmarking.json create mode 100644 pytorchconf-2024/videos/keynote-pytorch-technical-deep-dive-p-bialecki-p-wu-w-constable-k-khandelwal-m-yuan.json create mode 100644 pytorchconf-2024/videos/keynote-ray-a-distributed-framework-for-heterogeneous-computing-ion-stoica-uc-berkeley.json create mode 100644 pytorchconf-2024/videos/keynote-welcome-back-opening-remarks.json create mode 100644 pytorchconf-2024/videos/keynote-welcome-opening-remarks-matt-white-executive-director-pytorch-foundation.json create mode 100644 pytorchconf-2024/videos/keynote-why-you-should-think-twice-before-paying-for-an-evaluation-tool-chip-huyen-voltron-data.json create mode 100644 pytorchconf-2024/videos/lightning-talk-a-whirlwind-tour-of-pytorch-extension-points-alban-desmaison-meta.json create mode 100644 pytorchconf-2024/videos/lightning-talk-aotriton-ahead-of-time-triton-kernel-libraries-on-rocm-jeff-daily-amd.json create mode 100644 pytorchconf-2024/videos/lightning-talk-beyond-zero-eliminating-vulnerabili-patrick-smyth-dan-fernandez-srishti-hegde.json create mode 100644 pytorchconf-2024/videos/lightning-talk-building-and-supporting-the-chinese-pytorch-community-resources-tu-zong-zesheng.json create mode 100644 pytorchconf-2024/videos/lightning-talk-d-matrix-llm-compression-flow-based-on-torch-fx-simplify-zifei-xu-tristan-webb.json create mode 100644 pytorchconf-2024/videos/lightning-talk-debiasing-the-data-lifecycle-shailvi-wakhlu-shailvi-ventures-llc.json create mode 100644 pytorchconf-2024/videos/lightning-talk-distributing-a-million-open-models-in-the-wild-lessons-learned-f-omar-sanseviero.json create mode 100644 pytorchconf-2024/videos/lightning-talk-empowering-developers-tools-and-resources-for-running-generative-a-pareena-verma.json create mode 100644 pytorchconf-2024/videos/lightning-talk-extending-pytorch-with-custom-python-c-cuda-operators-richard-zou-meta.json create mode 100644 pytorchconf-2024/videos/lightning-talk-fast-scalable-distributed-training-with-streamingdataset-saaketh-narayan.json create mode 100644 pytorchconf-2024/videos/lightning-talk-flexattention-the-flexibility-of-pytorch-the-performa-yanbo-liang-horace-he.json create mode 100644 pytorchconf-2024/videos/lightning-talk-hieroglyph2text-a-pytorch-powered-pipeline-for-automated-egyptian-h-susi-gentsch.json create mode 100644 pytorchconf-2024/videos/lightning-talk-implementing-and-using-iterable-datasets-what-could-go-wrong-nicolas-hug-meta.json create mode 100644 pytorchconf-2024/videos/lightning-talk-in-transit-machine-learning-using-pytorch-on-frontier-exascale-system-vineeth-gutta.json create mode 100644 pytorchconf-2024/videos/lightning-talk-introduction-to-torch-distributed-pipelining-howard-huang-ke-wen-meta.json create mode 100644 pytorchconf-2024/videos/lightning-talk-llms-on-edge-with-ai-accelerators-chen-lai-kimish-patel-cemal-bilgin-meta.json create mode 100644 pytorchconf-2024/videos/lightning-talk-low-precision-dtypes-in-pytorch-vasiliy-kuznetsov-meta.json create mode 100644 pytorchconf-2024/videos/lightning-talk-making-the-most-of-heterogeneous-memory-capacity-using-pytorch-syed-ahmed-nvidia.json create mode 100644 pytorchconf-2024/videos/lightning-talk-mobile-computational-photography-with-pytorch-low-light-denoising-alexis-baudron.json create mode 100644 pytorchconf-2024/videos/lightning-talk-new-activation-checkpointing-apis-in-pytorch-jeffrey-wan-horace-he-meta.json create mode 100644 pytorchconf-2024/videos/lightning-talk-on-device-profiling-and-debugging-with-executorch-olivia-liu-vaun-puri-meta.json create mode 100644 pytorchconf-2024/videos/lightning-talk-optimized-pytorch-inference-on-aarch64-linux-cpus-sunita-nadampalli-amazon-aws.json create mode 100644 pytorchconf-2024/videos/lightning-talk-pytorch-release-process-andrey-talman-meta.json create mode 100644 pytorchconf-2024/videos/lightning-talk-pytorch-xla-auto-sharding-yeounoh-chung-google.json create mode 100644 pytorchconf-2024/videos/lightning-talk-sparsifying-vision-transformers-with-minimal-accuracy-loss-jesse-cai-meta.json create mode 100644 pytorchconf-2024/videos/lightning-talk-understanding-and-optimizing-pytorch-models-with-thunder-luca-antiga-lightning-ai.json create mode 100644 pytorchconf-2024/videos/lightning-talk-what-s-new-for-pytorch-developer-infrastructure-sahan-paliskara-catherine-lee.json create mode 100644 pytorchconf-2024/videos/lightning-talk-whats-new-in-ex-angela-yi-tugsbayasgalan-manlaibaatar-avik-chaudhuri-yidi-wu.json create mode 100644 pytorchconf-2024/videos/maximizing-training-throughput-using-torch-compile-and-fsdp-l-chu-a-viros-i-martin-b-vaughan.json create mode 100644 pytorchconf-2024/videos/meta-llama-3-and-the-future-of-responsible-ai-development-spencer-whitman-vincent-gonguet-meta.json create mode 100644 pytorchconf-2024/videos/mlir-enabling-composition-of-kernels-and-compilers-jacques-pienaar-google.json create mode 100644 pytorchconf-2024/videos/mojo-lifting-pt-to-new-heights-with-max-and-mojo-mikhail-zolotukhin-modular.json create mode 100644 pytorchconf-2024/videos/panel-discussion-t-dettmers-h-schoelkopf-a-chowdhery-a-conneau-moderated-by-k-khandelwal.json create mode 100644 pytorchconf-2024/videos/pushing-the-performance-envelope-an-optimization-study-for-3-suvaditya-mukherjee-shireen-chand.json create mode 100644 pytorchconf-2024/videos/pytorch-conference-2024-highlights.json create mode 100644 pytorchconf-2024/videos/running-state-of-art-gen-ai-models-on-device-with-npu-acceleration-felix-baum-qualcomm.json create mode 100644 pytorchconf-2024/videos/slaying-ooms-mark-saroufim-jane-xu-meta.json create mode 100644 pytorchconf-2024/videos/sponsored-keynote-accelerating-ai-how-amd-and-pytorch-drive-innovation-with-sea-anush-elangovan.json create mode 100644 pytorchconf-2024/videos/sponsored-keynote-enabling-ai-everywhere-with-pytorch-and-intel-kismat-singh-intel.json create mode 100644 pytorchconf-2024/videos/sponsored-keynote-from-containers-to-cognition-conducting-the-ai-orchestra-taylor-dolezal.json create mode 100644 pytorchconf-2024/videos/sponsored-keynote-optimizing-ai-inference-for-large-language-models-mudhakar-srivatsa-ibm.json create mode 100644 pytorchconf-2024/videos/sponsored-keynote-the-lightning-ai-oss-stack-for-accelerating-the-ai-lifecycle-luca-antiga.json create mode 100644 pytorchconf-2024/videos/sponsored-session-accelerating-ai-innovation-high-performance-pyt-robert-suderman-ian-nordeng.json create mode 100644 pytorchconf-2024/videos/sponsored-session-democratizing-ai-powering-the-future-with-arms-global-comp-gian-marco-iodice.json create mode 100644 pytorchconf-2024/videos/sponsored-session-nemo-aligner-a-scalable-toolkit-for-model-alignment-gerald-shen-jimmy-zhang.json create mode 100644 pytorchconf-2024/videos/sponsored-session-pytorch-support-by-google-enabling-perform-mark-sherwood-shauheen-zahirazami.json create mode 100644 pytorchconf-2024/videos/sponsored-session-torchchat-a-showcase-of-pytorch-llm-ubiquity-jack-khuu-jesse-white-meta.json create mode 100644 pytorchconf-2024/videos/startup-showcase.json create mode 100644 pytorchconf-2024/videos/state-of-pytorch-ji-li-damien-sereni-meta.json create mode 100644 pytorchconf-2024/videos/the-challenges-of-building-an-opinionated-open-source-llm-framework-wing-lian-axolotl-ai.json create mode 100644 pytorchconf-2024/videos/the-impact-and-challenges-of-open-source-generative-datasets-and-models-aaron-gokaslan.json create mode 100644 pytorchconf-2024/videos/the-rise-of-transformers-in-the-growing-pytorch-ecosystem-arthur-zucker-hugging-face.json create mode 100644 pytorchconf-2024/videos/the-state-of-the-llama-ecosystem-joe-spisak-meta.json create mode 100644 pytorchconf-2024/videos/together-goes-brrr-threading-research-production-with-torch-compile-pragaash-ponnusamy.json create mode 100644 pytorchconf-2024/videos/torch-compile-for-autograd-ddp-and-fsdp-will-feng-chien-chin-huang-simon-fan-meta.json create mode 100644 pytorchconf-2024/videos/torchinductor-cpu-backend-advancements-new-features-and-performance-imp-jiong-gong-leslie-fang.json create mode 100644 pytorchconf-2024/videos/torchtitan-large-scale-llm-training-using-native-pytorch-3d-parallel-wanchao-liang-linsong-chu.json create mode 100644 pytorchconf-2024/videos/torchtune-easy-and-accessible-finetuning-in-native-pytorch-evan-smothers-meta.json create mode 100644 pytorchconf-2024/videos/training-moes-at-scale-with-pytorch-mihir-patel-brian-chu-databricks.json create mode 100644 pytorchconf-2024/videos/triton-maximizing-kernel-development-productivity-under-performance-constraints-philip-tillet.json create mode 100644 pytorchconf-2024/videos/tvm-universally-deploy-large-language-models-via-ml-compilation-tianqi-chen-cmu-octoai.json create mode 100644 pytorchconf-2024/videos/understanding-the-llm-inference-workload-mark-moyou-nvidia.json create mode 100644 pytorchconf-2024/videos/unlocking-the-enigma-crafting-unbiased-transparent-and-explainable-large-languag-rashmi-nagpal.json create mode 100644 pytorchconf-2024/videos/vllm-easy-fast-and-cheap-llm-serving-for-everyone-woosuk-kwon-xiaoxuan-liu-uc-berkeley.json create mode 100644 pytorchconf-2024/videos/welcome-to-the-pytorch-ecosystem-for-llm-fine-tuning-mini-summit-kartikay-khandelwal-meta.json diff --git a/pytorchconf-2024/category.json b/pytorchconf-2024/category.json new file mode 100644 index 000000000..dfd3d3012 --- /dev/null +++ b/pytorchconf-2024/category.json @@ -0,0 +1,3 @@ +{ + "title": "PyTorch Conference 2024" +} diff --git a/pytorchconf-2024/videos/a-distributed-stateful-dataloader-for-large-scale-pretraining-davis-wertheimer-linsong-chu.json b/pytorchconf-2024/videos/a-distributed-stateful-dataloader-for-large-scale-pretraining-davis-wertheimer-linsong-chu.json new file mode 100644 index 000000000..5da50199d --- /dev/null +++ b/pytorchconf-2024/videos/a-distributed-stateful-dataloader-for-large-scale-pretraining-davis-wertheimer-linsong-chu.json @@ -0,0 +1,24 @@ +{ + "description": "A Distributed Stateful Dataloader for Large-Scale Pretraining - Davis Wertheimer, IBM & Linsong Chu, IBM Research\n\nLarge-scale model pretraining crucially relies on specialized and dedicated dataloaders that can, for example, partition and stream data asynchronously across multiple processes and physical nodes. In this talk we discuss one of the torch-native dataloaders we built and use at IBM Research for addressing these needs. Intended for use in large-scale model pretraining, particularly in research settings where rapid iteration between datasets may be required, our dataloader is distributed, stateful, checkpointable, composable and rescalable \u2013 while remaining a simple extension of the existing PyTorch dataloading framework. It automatically and invisibly handles data sharding, shuffling, subdataset weighting, checkpoint saving and loading, and custom user-defined preprocessing functions, with minimal overhead and high throughput. We discuss these properties and how we achieved them, such as reducing overhead by implementing a custom LCG random number generator, and demonstrate proof of concept on production-scale training of a 7B parameter Llama model over 4 trillion tokens.", + "duration": 1433, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/VtT4rdph4Qs/maxresdefault.webp", + "title": "A Distributed Stateful Dataloader for Large-Scale Pretraining - Davis Wertheimer & Linsong Chu", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=VtT4rdph4Qs" + } + ] +} diff --git a/pytorchconf-2024/videos/blobs-to-clips-efficient-end-to-end-video-data-loading-andrew-ho-ahmad-sharif-meta.json b/pytorchconf-2024/videos/blobs-to-clips-efficient-end-to-end-video-data-loading-andrew-ho-ahmad-sharif-meta.json new file mode 100644 index 000000000..127dadba2 --- /dev/null +++ b/pytorchconf-2024/videos/blobs-to-clips-efficient-end-to-end-video-data-loading-andrew-ho-ahmad-sharif-meta.json @@ -0,0 +1,24 @@ +{ + "description": "Blobs to Clips: Efficient End-to-End Video Data Loading - Andrew Ho & Ahmad Sharif, Meta\n\nThe PyTorch team has improved training speed by an order of magnitude for teams at Meta working on Small-to-Large-Scale MultiModal Video models. In this talk we\u2019ll share our learnings on reducing GPU starvation by overcoming data loading challenges such as dealing with large distributed datasets, worker imbalance, compute-bottlenecks due to parallel video decoding and sampling, checkpointing, and debuggability. As part of our commitment to open-source, we are releasing a new decoding library and updating existing PyTorch libraries on GitHub, and invite feedback and contributions from the community.", + "duration": 1544, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/-cBplHNM4RA/maxresdefault.webp", + "title": "Blobs to Clips: Efficient End-to-End Video Data Loading - Andrew Ho & Ahmad Sharif, Meta", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=-cBplHNM4RA" + } + ] +} diff --git a/pytorchconf-2024/videos/building-pytorch-computer-vision-algorithms-for-100-skin-shades-emmanuel-acheampong-robomua.json b/pytorchconf-2024/videos/building-pytorch-computer-vision-algorithms-for-100-skin-shades-emmanuel-acheampong-robomua.json new file mode 100644 index 000000000..5741debbf --- /dev/null +++ b/pytorchconf-2024/videos/building-pytorch-computer-vision-algorithms-for-100-skin-shades-emmanuel-acheampong-robomua.json @@ -0,0 +1,24 @@ +{ + "description": "Building PyTorch Computer Vision Algorithms for 100 Skin Shades - Emmanuel Acheampong, roboMUA\n\nAt roboMUA we're leading the charge in building predictive AI models for diverse skin shades with the use of Convolutional Neural Networks (CNNs), and harnessing the power of Generative Adversarial Networks (GANs) specifically for generating realistic images of black hairstyles. Our session showcases PyTorch's versatility in both predictive and generative tasks, offering a comprehensive approach to inclusive AI. For predictive AI models, we leverage PyTorch's flexible framework to develop CNNs. Through innovative techniques in feature engineering and model architecture design, we demonstrate how PyTorch enables accurate prediction across 100 skin shades. Simultaneously, we showcase the transformative potential of GANs in the realm of black hairstyles. By training GANs on a curated dataset of diverse hair textures and styles, we illustrate how PyTorch facilitates the generation of lifelike images that celebrate the beauty and diversity of black hair. Attendees will gain insights into the data preprocessing, model training, and evaluation processes and and learn how PyTorch empowers developers to build inclusive solutions.", + "duration": 882, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/eGZ3-8ZbFj0/maxresdefault.webp", + "title": "Building PyTorch Computer Vision Algorithms for 100 Skin Shades - Emmanuel Acheampong, roboMUA", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=eGZ3-8ZbFj0" + } + ] +} diff --git a/pytorchconf-2024/videos/building-scientific-computing-infrastructure-software-with-the-pytorch-ecosystem-bharath-ramsundar.json b/pytorchconf-2024/videos/building-scientific-computing-infrastructure-software-with-the-pytorch-ecosystem-bharath-ramsundar.json new file mode 100644 index 000000000..46dad290b --- /dev/null +++ b/pytorchconf-2024/videos/building-scientific-computing-infrastructure-software-with-the-pytorch-ecosystem-bharath-ramsundar.json @@ -0,0 +1,24 @@ +{ + "description": "Building Scientific Computing Infrastructure Software with the PyTorch Ecosystem - Bharath Ramsundar, Deep Forest Sciences\n\nThe DeepChem library is a scientific computing library that implements deep learning infrastructure for drug discovery, materials discovery, and biology. The DeepChem community is one of the largest scientific open source projects built in PyTorch, with over 5K stars on Github and thousands of citations. The DeepChem community has learned a number of useful lessons for building and maintaining high quality scientific code built on top of PyTorch. In this talk, I will share our learnings with the PyTorch community and also highlight opportunities for improving scientific support in the ecosystem.", + "duration": 1685, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/Ru-CDK_79NQ/maxresdefault.webp", + "title": "Building Scientific Computing Infrastructure Software with the PyTorch Ecosystem - Bharath Ramsundar", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=Ru-CDK_79NQ" + } + ] +} diff --git a/pytorchconf-2024/videos/data-dependent-shapes-in-pt2-edward-yang-meta.json b/pytorchconf-2024/videos/data-dependent-shapes-in-pt2-edward-yang-meta.json new file mode 100644 index 000000000..8248a9cec --- /dev/null +++ b/pytorchconf-2024/videos/data-dependent-shapes-in-pt2-edward-yang-meta.json @@ -0,0 +1,24 @@ +{ + "description": "Data-Dependent Shapes in PT2 - Edward Yang, Meta\n\nData-dependent shapes are ubiquitous whenever you want to take advantage of sparsity in your data representation, whether it is in recommendation systems, mixture of experts or other use cases. We have made a lot of improvements to torch.compile's support for capturing and compiling data dependent shapes, but they also require some user knowledge to work with effectively. This talk will give an overview of PT2's facilities for data dependent compute and how to use them effectively.", + "duration": 1524, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/VYuFJUfyWSU/maxresdefault.webp", + "title": "Data-Dependent Shapes in PT2 - Edward Yang, Meta", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=VYuFJUfyWSU" + } + ] +} diff --git a/pytorchconf-2024/videos/dl-compiler-panel-discussion-p-tillet-j-ansel-j-pienaar-t-chen-m-zolotukhin-p-wu.json b/pytorchconf-2024/videos/dl-compiler-panel-discussion-p-tillet-j-ansel-j-pienaar-t-chen-m-zolotukhin-p-wu.json new file mode 100644 index 000000000..4de82e036 --- /dev/null +++ b/pytorchconf-2024/videos/dl-compiler-panel-discussion-p-tillet-j-ansel-j-pienaar-t-chen-m-zolotukhin-p-wu.json @@ -0,0 +1,24 @@ +{ + "description": "DL Compiler Panel Discussion - Philip Tillet, OpenAI; Jason Ansel, Meta; Jacques Pienaar, Google; Tianqi Chen, CMU & OctoAI; Mikhail Zolotukhin, Modular; Peng Wu, Meta\n\nSince the release of PyTorch 2 in 2023, torch.compile() has spurred significant new thinking around DL compiler designs at the framework level. In this session, we invite leaders in this space to share their insights based on real experiences of building DL compilers \u2013 Triton, TorchInductor, Halide, TVM, OpenXLA, and Mojo \u2013 and growing their ecosystems. We also invite a \u2018compiler user representative,\u2019 together.ai, to share their recent journey of redesigning the LLM inference stack around torch.compile(). Each leader will give a 10-minute lightning talk and an engaging panel discussion.", + "duration": 2131, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/YWDzHGx8PrY/maxresdefault.webp", + "title": "DL Compiler Panel Discussion - P. Tillet, J. Ansel, J. Pienaar, T. Chen, M. Zolotukhin, P. Wu", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=YWDzHGx8PrY" + } + ] +} diff --git a/pytorchconf-2024/videos/executorch-beta-and-on-device-generative-ai-support-mergen-nachin-mengtao-martin-yuan-meta.json b/pytorchconf-2024/videos/executorch-beta-and-on-device-generative-ai-support-mergen-nachin-mengtao-martin-yuan-meta.json new file mode 100644 index 000000000..83d57093d --- /dev/null +++ b/pytorchconf-2024/videos/executorch-beta-and-on-device-generative-ai-support-mergen-nachin-mengtao-martin-yuan-meta.json @@ -0,0 +1,24 @@ +{ + "description": "ExecuTorch Beta and on-Device Generative AI Support - Mergen Nachin & Mengtao (Martin) Yuan, Meta\n\nDuring this session, we will discuss real-life case studies focusing on the productionization of PyTorch models onto edge devices and welcome the community to begin adopting ExecuTorch. Since announcing the ExecuTorch MVP at the previous PTC, we have made significant progress in terms of stability, model coverage, accelerator performance, and developer experience, reaching a milestone that marks the transition to beta status. In addition to the above improvements, we continue to support generative AI models. Since the alpha launch that initially enabled support for LLama2/3 models, we have now expanded our capabilities to include multimodal use cases and developed mobile demo apps showcasing these new features.", + "duration": 1213, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/45yNTi7c1Q0/maxresdefault.webp", + "title": "ExecuTorch Beta and on-Device Generative AI Support - Mergen Nachin & Mengtao (Martin) Yuan, Meta", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=45yNTi7c1Q0" + } + ] +} diff --git a/pytorchconf-2024/videos/hacks-to-make-llm-training-faster-daniel-han-unsloth-ai.json b/pytorchconf-2024/videos/hacks-to-make-llm-training-faster-daniel-han-unsloth-ai.json new file mode 100644 index 000000000..19679e3e9 --- /dev/null +++ b/pytorchconf-2024/videos/hacks-to-make-llm-training-faster-daniel-han-unsloth-ai.json @@ -0,0 +1,24 @@ +{ + "description": "Hacks to Make LLM Training Faster - Daniel Han, Unsloth AI\n\nAs open-source LLMs have become more capable, a substantial ecosystem has developed around the fine-tuning of these models. A thriving community of researchers, developers, practitioners and hobbyists has emerged which focuses on topics ranging from memory efficiency, parameter-efficient fine-tuning and quantization to performance at scale and reproducible evaluations. The goal of this mini-summit is to bring this community together to discuss ideas, share knowledge and build connections.\n\nThe agenda features a keynote from Joe Spisak on the state of the Llama ecosystem followed by invited talks from the founders of Axolotl, Unsloth and torchtune. We conclude the summit with a riveting discussion on what\u2019s next for LLMs, fine-tuning and the PyTorch ecosystem with a fabulous panel of experts - Tim Dettmers (author of bitsandbytes and QLoRA), Hailey Schoelkopf (maintainer of LM Eval Harness at EleutherAI), Aakanksha Chowdhery (Lead author on PaLM and Gemini) and Alexis Conneau (Research Lead at OpenAI)", + "duration": 477, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/PdtKkc5jB4g/maxresdefault.webp", + "title": "Hacks to Make LLM Training Faster - Daniel Han, Unsloth AI", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=PdtKkc5jB4g" + } + ] +} diff --git a/pytorchconf-2024/videos/halide-a-halide-backend-for-torchinductor-jason-ansel-meta.json b/pytorchconf-2024/videos/halide-a-halide-backend-for-torchinductor-jason-ansel-meta.json new file mode 100644 index 000000000..e376e2a4a --- /dev/null +++ b/pytorchconf-2024/videos/halide-a-halide-backend-for-torchinductor-jason-ansel-meta.json @@ -0,0 +1,24 @@ +{ + "description": "[HALIDE] A Halide Backend for TorchInductor - Jason Ansel, Meta\n\nThis talk will focus on a new Halide backend for TorchInductor, which is in addition to the existing Triton and C++ backends. The Halide backend is meant to serve as a reference backend to make it easier to extend TorchInductor to support new backend compilers and hardware devices. Halide has been the inspiration (either in ideas or through forking) of numerous other compiler projects, so it is a good starting point for adding new backends that follow a Halide-like model.", + "duration": 489, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/hCvlWZEXRrg/maxresdefault.webp", + "title": "[HALIDE] A Halide Backend for TorchInductor - Jason Ansel, Meta", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=hCvlWZEXRrg" + } + ] +} diff --git a/pytorchconf-2024/videos/implementing-a-custom-torch-compile-backend-a-case-study-maanav-dalal-yulong-wang-microsoft.json b/pytorchconf-2024/videos/implementing-a-custom-torch-compile-backend-a-case-study-maanav-dalal-yulong-wang-microsoft.json new file mode 100644 index 000000000..2b0856958 --- /dev/null +++ b/pytorchconf-2024/videos/implementing-a-custom-torch-compile-backend-a-case-study-maanav-dalal-yulong-wang-microsoft.json @@ -0,0 +1,24 @@ +{ + "description": "Implementing a Custom Torch.Compile Backend - A Case Study - Maanav Dalal & Yulong Wang, Microsoft\n\nThis presentation will dive into the development of the ONNXRuntime (ORT) backend for torch.compile. We'll cover the implementation process, starting with a PyTorch 2.0 generated FX graph, highlighting the unique challenges encountered when serving ORT-specific scenarios and how we solved them. Attendees will gain insights into optimizing performance, overcoming integration hurdles, and achieving efficient execution. Whether you're a developer looking to extend PyTorch's capabilities for your own use cases, keen to learn about ONNX Runtime, or interested in backend performance optimization, and the many steps we've taken to get to where we are now, this session promises valuable takeaways and practical knowledge.", + "duration": 666, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/MnNO_13fLtU/maxresdefault.webp", + "title": "Implementing a Custom Torch.Compile Backend - A Case Study - Maanav Dalal & Yulong Wang, Microsoft", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=MnNO_13fLtU" + } + ] +} diff --git a/pytorchconf-2024/videos/intel-gpu-in-upstream-pytorch-expanding-gpu-choices-and-enhancing-back-eikan-wang-min-jean-cho.json b/pytorchconf-2024/videos/intel-gpu-in-upstream-pytorch-expanding-gpu-choices-and-enhancing-back-eikan-wang-min-jean-cho.json new file mode 100644 index 000000000..14b2d1272 --- /dev/null +++ b/pytorchconf-2024/videos/intel-gpu-in-upstream-pytorch-expanding-gpu-choices-and-enhancing-back-eikan-wang-min-jean-cho.json @@ -0,0 +1,24 @@ +{ + "description": "Intel GPU in Upstream PyTorch: Expanding GPU Choices and Enhancing Backend Flexibility - Eikan Wang & Min Jean Cho, Intel\n\nThe integration of Intel GPU support into PyTorch marks a pivotal enhancement for PyTorch device and runtime. We generalized the PyTorch device and runtime to accommodate streaming devices. The generalization not only facilitates the deployment of PyTorch on ubiquitous hardware but also makes the integration of different HW backends easier. In addition, PyTorch with Intel GPU supports various Intel GPUs from the data center to the client. It enriches and democratizes PyTorch HW ecosystem. Particularly in AIPC scenarios where Intel's integrated and discrete GPUs are prevalent, Pytorch with Intel GPU can deliver promising performance and improved OOB experience in the AIPC domain that can extend PyTorch's applicability significantly.", + "duration": 1398, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/HRdda_kVEh4/maxresdefault.webp", + "title": "Intel GPU in Upstream PyTorch: Expanding GPU Choices and Enhancing Back... Eikan Wang & Min Jean Cho", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=HRdda_kVEh4" + } + ] +} diff --git a/pytorchconf-2024/videos/keynote-building-an-advanced-knowledge-assistant-jerry-liu-co-founder-ceo-llamaindex.json b/pytorchconf-2024/videos/keynote-building-an-advanced-knowledge-assistant-jerry-liu-co-founder-ceo-llamaindex.json new file mode 100644 index 000000000..560e18ba1 --- /dev/null +++ b/pytorchconf-2024/videos/keynote-building-an-advanced-knowledge-assistant-jerry-liu-co-founder-ceo-llamaindex.json @@ -0,0 +1,24 @@ +{ + "description": "Keynote: Building an Advanced Knowledge Assistant - Jerry Liu, Co-Founder & CEO, LlamaIndex\n\nA huge promise for LLMs is being able to answer questions and solve tasks of arbitrary complexity over an arbitrary number of data sources. The world has started to shift from simple RAG stacks, which are mostly good for answering pointed questions, to agents that can more autonomously reason over a diverse set of inputs, and interleave retrieval and tool use to produce sophisticated outputs.\n\nBuilding a reliable multi-agent system is challenging. There's a core question of developer ergonomics and production deployment - what makes sense outside a notebook setting. In this talk we outline some core building blocks for building advanced research assistants, including advanced RAG modules, event-driven workflow orchestration, and more.", + "duration": 896, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/EjGdum0UdaI/maxresdefault.webp", + "title": "Keynote: Building an Advanced Knowledge Assistant - Jerry Liu, Co-Founder & CEO, LlamaIndex", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=EjGdum0UdaI" + } + ] +} diff --git a/pytorchconf-2024/videos/keynote-community-awards.json b/pytorchconf-2024/videos/keynote-community-awards.json new file mode 100644 index 000000000..754f88c69 --- /dev/null +++ b/pytorchconf-2024/videos/keynote-community-awards.json @@ -0,0 +1,24 @@ +{ + "description": "Keynote: Community Awards", + "duration": 359, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/1dSVYKPk3N0/maxresdefault.webp", + "title": "Keynote: Community Awards", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=1dSVYKPk3N0" + } + ] +} diff --git a/pytorchconf-2024/videos/keynote-enabling-generative-ai-on-the-edge-cormac-brick-principal-engineer-google.json b/pytorchconf-2024/videos/keynote-enabling-generative-ai-on-the-edge-cormac-brick-principal-engineer-google.json new file mode 100644 index 000000000..993a9f9ca --- /dev/null +++ b/pytorchconf-2024/videos/keynote-enabling-generative-ai-on-the-edge-cormac-brick-principal-engineer-google.json @@ -0,0 +1,24 @@ +{ + "description": "Keynote: Enabling Generative AI on the Edge - Cormac Brick, Principal Engineer, Google\n\nGenerative AI is no longer just in the cloud - recently it's also getting deployed on edge devices. A disruptive goal of this work is AI-powered applications that respond instantly, work offline, and protect user privacy by processing data locally. In this talk, we'll explore the cutting edge of edge-based generative AI, showcasing open models that are pushing the boundaries of what's possible today on the edge. We'll dive deep into the PyTorch ecosystem, looking at projects that are making it easier than ever to author, optimize, and deploy these models across a wide range of devices.", + "duration": 848, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/n2DsEyLbuJg/maxresdefault.webp", + "title": "Keynote: Enabling Generative AI on the Edge - Cormac Brick, Principal Engineer, Google", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=n2DsEyLbuJg" + } + ] +} diff --git a/pytorchconf-2024/videos/keynote-navigating-the-architectural-timeline-of-llms-sebastian-raschka-lightning-ai.json b/pytorchconf-2024/videos/keynote-navigating-the-architectural-timeline-of-llms-sebastian-raschka-lightning-ai.json new file mode 100644 index 000000000..cc41cf780 --- /dev/null +++ b/pytorchconf-2024/videos/keynote-navigating-the-architectural-timeline-of-llms-sebastian-raschka-lightning-ai.json @@ -0,0 +1,24 @@ +{ + "description": "Keynote: Navigating the Architectural Timeline of LLMs - Sebastian Raschka, Staff Research Engineer, Lightning AI\n\nThe evolution of large language models (LLMs) from the original Generative Pre-trained Transformer (GPT) series to the recent advancements seen in models like Llama 3 has been accompanied by several architectural and methodological innovations. This talk aims to catch attendees up on the latest AI and LLM development trends, highlighting the key changes and motivations that led to the development of recent state-of-the-art LLMs, such as Llama 3.1.\n\nSpecifically, this presentation explores key developments in attention mechanisms, such as sliding window attention, group query, multi-query attention, and FlashAttention, and explains their key motivations and advantages. In addition to exploring the structural changes, this presentation also reviews the recent \"tricks of the trade\" that have improved the training processes and performance of the latest LLMs. This includes the recent two-step pretraining approach in Llama 3.1 and applying knowledge distillation techniques using real datasets like Gemma 2 and synthetic data, as seen in Llama 3.1.\n\nMoreover, we will also examine the integration of system-level optimizations, such as the Mixture of the Expert method and the hybrid model Samba, which combines Mamba techniques with attention mechanisms and illustrates a broader trend toward more specialized and efficient architectures.\n\nThis talk will provide attendees with an understanding of the most notable transformations that have defined the architectural timeline of LLMs.", + "duration": 925, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/frkAt-gZVjc/maxresdefault.webp", + "title": "Keynote: Navigating the Architectural Timeline of LLMs - Sebastian Raschka, Lightning AI", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=frkAt-gZVjc" + } + ] +} diff --git a/pytorchconf-2024/videos/keynote-open-language-models-olmo-accelerating-the-science-of-language-modeling-hanna-hajishirzi.json b/pytorchconf-2024/videos/keynote-open-language-models-olmo-accelerating-the-science-of-language-modeling-hanna-hajishirzi.json new file mode 100644 index 000000000..a80179560 --- /dev/null +++ b/pytorchconf-2024/videos/keynote-open-language-models-olmo-accelerating-the-science-of-language-modeling-hanna-hajishirzi.json @@ -0,0 +1,24 @@ +{ + "description": "Keynote: Open Language Models (OLMo): Accelerating the Science of Language Modeling - Hanna Hajishirzi, Senior Director NLP Research, Allen Institute for AI\n\nOver the past few years, and especially since the deployment of ChatGPT in November 2022, neural language models with billions of parameters and trained on trillions of words are powering the fastest-growing computing applications in history and generating discussion and debate across society. However, AI scientists cannot study or improve those state-of-the-art models because the models' parameters, training data, code, and even documentation are not openly available. In this talk, I present our OLMo project toward building strong language models and making them fully open to researchers along with open-source code for data management, training, inference, and interaction. In particular, I describe DOLMa, a 3T token open dataset curated for training language models, Tulu, our instruction-tuned language model, and OLMo v1, a fully-open 7B parameter language model trained from scratch.", + "duration": 1036, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/GYWCYU-Xmnk/maxresdefault.webp", + "title": "Keynote: Open Language Models (OLMo): Accelerating the Science of Language Modeling Hanna Hajishirzi", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=GYWCYU-Xmnk" + } + ] +} diff --git a/pytorchconf-2024/videos/keynote-panel-discussion-responsible-ai-k-rooney-k-varshney-s-hooker-a-madry-r-bommasani.json b/pytorchconf-2024/videos/keynote-panel-discussion-responsible-ai-k-rooney-k-varshney-s-hooker-a-madry-r-bommasani.json new file mode 100644 index 000000000..be238d1e2 --- /dev/null +++ b/pytorchconf-2024/videos/keynote-panel-discussion-responsible-ai-k-rooney-k-varshney-s-hooker-a-madry-r-bommasani.json @@ -0,0 +1,24 @@ +{ + "description": "Keynote Panel Discussion: Responsible AI - Kate Rooney, CNBC; Kush Varshney, IBM T. J. Watson Research Center; Sara Hooker, C4AI; Aleksander Madry, OpenAI; Rishi Bommasani, Stanford University", + "duration": 1832, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi/FLUd0k2ctJE/maxresdefault.jpg", + "title": "Keynote Panel Discussion: Responsible AI - K. Rooney, K. Varshney, S. Hooker, A. Madry, R. Bommasani", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=FLUd0k2ctJE" + } + ] +} diff --git a/pytorchconf-2024/videos/keynote-panel-discussion-scaling-benchmarking.json b/pytorchconf-2024/videos/keynote-panel-discussion-scaling-benchmarking.json new file mode 100644 index 000000000..322f86834 --- /dev/null +++ b/pytorchconf-2024/videos/keynote-panel-discussion-scaling-benchmarking.json @@ -0,0 +1,24 @@ +{ + "description": "Keynote Panel Discussion: Scaling & Benchmarking - Anastasios Nikolas Angelopoulos, UC Berkeley/LMSYS; Lisa Dunlap, UC Berkeley; James Bradbury, Anthropic; Tri Dao, together.ai; Aparna Ramani & Soumith Chintala, Meta", + "duration": 2023, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/01a4tBLEzVM/maxresdefault.webp", + "title": "Keynote Panel Discussion: Scaling & Benchmarking", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=01a4tBLEzVM" + } + ] +} diff --git a/pytorchconf-2024/videos/keynote-pytorch-technical-deep-dive-p-bialecki-p-wu-w-constable-k-khandelwal-m-yuan.json b/pytorchconf-2024/videos/keynote-pytorch-technical-deep-dive-p-bialecki-p-wu-w-constable-k-khandelwal-m-yuan.json new file mode 100644 index 000000000..377371724 --- /dev/null +++ b/pytorchconf-2024/videos/keynote-pytorch-technical-deep-dive-p-bialecki-p-wu-w-constable-k-khandelwal-m-yuan.json @@ -0,0 +1,24 @@ +{ + "description": "Keynote: PyTorch Technical Deep Dive - Piotr Bialecki, NVIDIA; Peng Wu, Will Constable, Kartikay Khandelwal & Mengtao (Martin) Yuan, Meta\n\nThis Deep Dive provides an update on PyTorch development since last conference and dives into the key new features coming in PyTorch 2.5 and beyond. We will explore how advancements across a number of PyTorch features combine to better support the full model development lifecycle across training, fine-tuning, and deployment.", + "duration": 3055, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/d29dGW8FafI/maxresdefault.webp", + "title": "Keynote: PyTorch Technical Deep Dive - P. Bialecki, P. Wu, W. Constable, K. Khandelwal & M. Yuan", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=d29dGW8FafI" + } + ] +} diff --git a/pytorchconf-2024/videos/keynote-ray-a-distributed-framework-for-heterogeneous-computing-ion-stoica-uc-berkeley.json b/pytorchconf-2024/videos/keynote-ray-a-distributed-framework-for-heterogeneous-computing-ion-stoica-uc-berkeley.json new file mode 100644 index 000000000..b3c7ec7cc --- /dev/null +++ b/pytorchconf-2024/videos/keynote-ray-a-distributed-framework-for-heterogeneous-computing-ion-stoica-uc-berkeley.json @@ -0,0 +1,24 @@ +{ + "description": "Keynote: Ray: A Distributed Framework for Heterogeneous Computing - Ion Stoica, Professor, UC Berkeley\n\nRay has recently become the framework of choice for scaling machine learning workloads\u2014from data preprocessing, to training, fine-tuning, and serving. This talk will highlight Ray\u2019s key features responsible for its flexibility and generality, as well as its recent support for GPUs.", + "duration": 1006, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/E98Pv36f0YI/maxresdefault.webp", + "title": "Keynote: Ray: A Distributed Framework for Heterogeneous Computing - Ion Stoica, UC Berkeley", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=E98Pv36f0YI" + } + ] +} diff --git a/pytorchconf-2024/videos/keynote-welcome-back-opening-remarks.json b/pytorchconf-2024/videos/keynote-welcome-back-opening-remarks.json new file mode 100644 index 000000000..938690c79 --- /dev/null +++ b/pytorchconf-2024/videos/keynote-welcome-back-opening-remarks.json @@ -0,0 +1,24 @@ +{ + "description": "Keynote: Welcome Back & Opening Remarks", + "duration": 160, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/fabW4B5IlVg/maxresdefault.webp", + "title": "Keynote: Welcome Back & Opening Remarks", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=fabW4B5IlVg" + } + ] +} diff --git a/pytorchconf-2024/videos/keynote-welcome-opening-remarks-matt-white-executive-director-pytorch-foundation.json b/pytorchconf-2024/videos/keynote-welcome-opening-remarks-matt-white-executive-director-pytorch-foundation.json new file mode 100644 index 000000000..c61f44eb7 --- /dev/null +++ b/pytorchconf-2024/videos/keynote-welcome-opening-remarks-matt-white-executive-director-pytorch-foundation.json @@ -0,0 +1,24 @@ +{ + "description": "Keynote: Welcome & Opening Remarks - Matt White, Executive Director, PyTorch Foundation\n\nOver the past few years, and especially since the deployment of ChatGPT in November 2022, neural language models with billions of parameters and trained on trillions of words are powering the fastest-growing computing applications in history and generating discussion and debate across society. However, AI scientists cannot study or improve those state-of-the-art models because the models' parameters, training data, code, and even documentation are not openly available. In this talk, I present our OLMo project toward building strong language models and making them fully open to researchers along with open-source code for data management, training, inference, and interaction. In particular, I describe DOLMa, a 3T token open dataset curated for training language models, Tulu, our instruction-tuned language model, and OLMo v1, a fully-open 7B parameter language model trained from scratch.", + "duration": 626, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/9VFHuGaq7SU/maxresdefault.webp", + "title": "Keynote: Welcome & Opening Remarks - Matt White, Executive Director, PyTorch Foundation", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=9VFHuGaq7SU" + } + ] +} diff --git a/pytorchconf-2024/videos/keynote-why-you-should-think-twice-before-paying-for-an-evaluation-tool-chip-huyen-voltron-data.json b/pytorchconf-2024/videos/keynote-why-you-should-think-twice-before-paying-for-an-evaluation-tool-chip-huyen-voltron-data.json new file mode 100644 index 000000000..ae52f3870 --- /dev/null +++ b/pytorchconf-2024/videos/keynote-why-you-should-think-twice-before-paying-for-an-evaluation-tool-chip-huyen-voltron-data.json @@ -0,0 +1,24 @@ +{ + "description": "Keynote: Why You Should Think Twice Before Paying for an Evaluation Tool - Chip Huyen, VP of AI & OSS, Voltron Data\n\nOpen-ended evaluation is hard, and the number of evaluation tools has exploded in response to this challenge. However, if tools could solve evaluation, evaluation would have been solved by now. While the right tools can make your life easier, this talk discusses why you should think twice before outsourcing your evaluation to an external tool.", + "duration": 949, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/IbKKPeuieFw/maxresdefault.webp", + "title": "Keynote: Why You Should Think Twice Before Paying for an Evaluation Tool - Chip Huyen, Voltron Data", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=IbKKPeuieFw" + } + ] +} diff --git a/pytorchconf-2024/videos/lightning-talk-a-whirlwind-tour-of-pytorch-extension-points-alban-desmaison-meta.json b/pytorchconf-2024/videos/lightning-talk-a-whirlwind-tour-of-pytorch-extension-points-alban-desmaison-meta.json new file mode 100644 index 000000000..3bb1cc4d8 --- /dev/null +++ b/pytorchconf-2024/videos/lightning-talk-a-whirlwind-tour-of-pytorch-extension-points-alban-desmaison-meta.json @@ -0,0 +1,24 @@ +{ + "description": "Lightning Talk: A Whirlwind Tour of PyTorch Extension Points - Alban Desmaison, Meta\n\nJourney across the PyTorch stack and see all the extension points that exist from nn.Module to the c++ Dispatcher through autograd and subclasses. This sessions will cover example use cases and when each one should be used while pointing to reference for in-depth details.", + "duration": 848, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/1A54cUPXhZ4/maxresdefault.webp", + "title": "Lightning Talk: A Whirlwind Tour of PyTorch Extension Points - Alban Desmaison, Meta", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=1A54cUPXhZ4" + } + ] +} diff --git a/pytorchconf-2024/videos/lightning-talk-aotriton-ahead-of-time-triton-kernel-libraries-on-rocm-jeff-daily-amd.json b/pytorchconf-2024/videos/lightning-talk-aotriton-ahead-of-time-triton-kernel-libraries-on-rocm-jeff-daily-amd.json new file mode 100644 index 000000000..1ddea8040 --- /dev/null +++ b/pytorchconf-2024/videos/lightning-talk-aotriton-ahead-of-time-triton-kernel-libraries-on-rocm-jeff-daily-amd.json @@ -0,0 +1,24 @@ +{ + "description": "Lightning Talk: AOTriton: Ahead of Time Triton Kernel Libraries on ROCm - Jeff Daily, AMD\n\nScaled dot product attention provides significant acceleration of the transformer layer through fusion of the multihead attention layer. There are several different algorithms to achieve this but tiled attention through scaled dot product attention via Flash Attention is a very popular approach. In PyTorch on the ROCm platform this is currently achieved through ahead of time compiled (AOT) Triton kernels in a linkable archive. AMD\u2019s work to enable and package these kernels is done through AOTriton, which aims to use Triton\u2019s compiler and GPU kernels for faster development. AOTriton maintains an optimized set of tiling sizes and other parameters to provide optimized, pre-compiled Triton kernels. The differences between JIT and AOT are few but are very important. Despite this, prototyping kernels in Triton is much faster than template-based C++ libraries. In this presentation we will go into detail on the interaction layer between PyTorch and AOTriton, the structure of AOTriton and how to add new triton kernels to AOTriton.", + "duration": 692, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/bhplJt1XAMI/maxresdefault.webp", + "title": "Lightning Talk: AOTriton: Ahead of Time Triton Kernel Libraries on ROCm - Jeff Daily, AMD", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=bhplJt1XAMI" + } + ] +} diff --git a/pytorchconf-2024/videos/lightning-talk-beyond-zero-eliminating-vulnerabili-patrick-smyth-dan-fernandez-srishti-hegde.json b/pytorchconf-2024/videos/lightning-talk-beyond-zero-eliminating-vulnerabili-patrick-smyth-dan-fernandez-srishti-hegde.json new file mode 100644 index 000000000..07c2069cd --- /dev/null +++ b/pytorchconf-2024/videos/lightning-talk-beyond-zero-eliminating-vulnerabili-patrick-smyth-dan-fernandez-srishti-hegde.json @@ -0,0 +1,24 @@ +{ + "description": "Lightning Talk: Beyond Zero: Eliminating Vulnerabilities in PyTorch Container Images - Patrick Smyth, Dan Fernandez & Srishti Hegde, Chainguard\n\nContainer images are increasingly the future of production applications at scale, providing reproducibility, robustness, and transparency. As PyTorch images get deployed to production, however, security becomes a major concern. PyTorch has a large attack surface, and building secure PyTorch images can be a challenge. Currently, the official PyTorch runtime container image has 30 CVEs (known vulnerabilities) rated critical and 256 CVE rated high. Improving this situation could secure many deployments that incorporate PyTorch for cloud-based inference or training. In this fast-paced session, we'll take a deep dive on the official PyTorch image from a vulnerability mitigation perspective, looking hard at included packages, executables, and active CVE. We'll identify low-hanging fruit for increasing security, including stripping bloat and building fresh. We'll also talk about the next level of security practiced in Chainguard's PyTorch image builds, such as including SBOMs and going distroless. Finally, we'll consider emerging tools and approaches for analyzing AI artifacts such as models and how these systems can benefit PyTorch in production.", + "duration": 736, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/1klynk1dxYA/maxresdefault.webp", + "title": "Lightning Talk: Beyond Zero: Eliminating Vulnerabili... Patrick Smyth, Dan Fernandez & Srishti Hegde", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=1klynk1dxYA" + } + ] +} diff --git a/pytorchconf-2024/videos/lightning-talk-building-and-supporting-the-chinese-pytorch-community-resources-tu-zong-zesheng.json b/pytorchconf-2024/videos/lightning-talk-building-and-supporting-the-chinese-pytorch-community-resources-tu-zong-zesheng.json new file mode 100644 index 000000000..cdef56690 --- /dev/null +++ b/pytorchconf-2024/videos/lightning-talk-building-and-supporting-the-chinese-pytorch-community-resources-tu-zong-zesheng.json @@ -0,0 +1,24 @@ +{ + "description": "Lightning Talk: Building and Supporting the Chinese PyTorch Community: Resources, Tutorials, and Engagement - Zong Zesheng, Huawei\n\nDescription: This proposal aims to provide a comprehensive introduction to the Chinese PyTorch community, we hope to inspire more users to join and contribute, fostering a vibrant and inclusive environment for PyTorch enthusiasts in China. Chinese PyTorch Homepage Introduction to the official Chinese version of the PyTorch website, highlighting its features. Navigation tips and key sections, such as documentation, tutorials, and community events. Improve the connection of users from China with PyTorch Community. Localized Tutorials and Documentation The 2.x version not have Translated version, it hard to catch up with latest features of PyTorch if the beginner not good at English. We translated official documents and tutorials, covering everything from basic PyTorch concepts to advanced applications. Interactive tutorials No interactive tutorials(Like Google Colab) for Chinese students or beginners before, they have to setup environment before start with PyTorch, which might be hard for beginners. And now, an online notebook & tutorials are available to practice or tuning steps for beginners.", + "duration": 453, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/AXMN5oUyWKE/maxresdefault.webp", + "title": "Lightning Talk: Building and Supporting the Chinese PyTorch Community: Resources, Tu... Zong Zesheng", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=AXMN5oUyWKE" + } + ] +} diff --git a/pytorchconf-2024/videos/lightning-talk-d-matrix-llm-compression-flow-based-on-torch-fx-simplify-zifei-xu-tristan-webb.json b/pytorchconf-2024/videos/lightning-talk-d-matrix-llm-compression-flow-based-on-torch-fx-simplify-zifei-xu-tristan-webb.json new file mode 100644 index 000000000..f9e148902 --- /dev/null +++ b/pytorchconf-2024/videos/lightning-talk-d-matrix-llm-compression-flow-based-on-torch-fx-simplify-zifei-xu-tristan-webb.json @@ -0,0 +1,24 @@ +{ + "description": "Lightning Talk: d-Matrix LLM Compression Flow Based on Torch.Fx: Simplifying PTQ/QAT - Zifei Xu & Tristan Webb, d-Matrix Corporation\n\nWe introduce dmx-compressor, d-Matrix's open-source LLM compression toolkit that is modular, robust, efficient, and user-friendly. It utilizes symbolic tracing and fx.Transformer for network compression while keeping the model a first-class citizen in PyTorch for the user, despite prevalent graph dynamism in LLMs. It achieves this by maintaining both the original nn.Module and a just-in-time (JIT) traced and transformed fx.GraphModule representation behind the scenes, in conjunction with an abstraction that cleanly decouples network compression from the original model graph definition. This design allows the FXIR to dynamically adapt to diverse forward call signatures and flow-control arguments throughout quantization-aware training and post-training quantization written in plain PyTorch, yielding a compressed FXIR fully compatible with application-level APIs like the Hugging Face pipeline. We also provide a graph visualizer based on fx.Interpreter for ease of debugging. We believe this project shall empower the community to build efficient LLMs for deployment on custom hardware accelerators and contribute to the PyTorch ecosystem.", + "duration": 822, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi/cw6dSlmMwj8/maxresdefault.jpg", + "title": "Lightning Talk: d-Matrix LLM Compression Flow Based on Torch.Fx: Simplify... Zifei Xu & Tristan Webb", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=cw6dSlmMwj8" + } + ] +} diff --git a/pytorchconf-2024/videos/lightning-talk-debiasing-the-data-lifecycle-shailvi-wakhlu-shailvi-ventures-llc.json b/pytorchconf-2024/videos/lightning-talk-debiasing-the-data-lifecycle-shailvi-wakhlu-shailvi-ventures-llc.json new file mode 100644 index 000000000..c68c58459 --- /dev/null +++ b/pytorchconf-2024/videos/lightning-talk-debiasing-the-data-lifecycle-shailvi-wakhlu-shailvi-ventures-llc.json @@ -0,0 +1,24 @@ +{ + "description": "Lightning Talk: Debiasing the Data Lifecycle - Shailvi Wakhlu, Shailvi Ventures LLC\n\nBiased data, results in biased decision-making. Making sure that at every step of the data lifecycle, we make conscious attempts to debias the data is an important responsibility for all data scientists. In this talk, I highlight the typical data lifecycle, and how to prevent biases at every step. ---- The key takeaways from my talk include: 1) Understanding the data lifecycle 2) What are the typical ways biases creep in 3) How we can proactively prevent and fix biases in data", + "duration": 661, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi/V9NVEceY9Wg/maxresdefault.jpg", + "title": "Lightning Talk: Debiasing the Data Lifecycle - Shailvi Wakhlu, Shailvi Ventures LLC", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=V9NVEceY9Wg" + } + ] +} diff --git a/pytorchconf-2024/videos/lightning-talk-distributing-a-million-open-models-in-the-wild-lessons-learned-f-omar-sanseviero.json b/pytorchconf-2024/videos/lightning-talk-distributing-a-million-open-models-in-the-wild-lessons-learned-f-omar-sanseviero.json new file mode 100644 index 000000000..63df10978 --- /dev/null +++ b/pytorchconf-2024/videos/lightning-talk-distributing-a-million-open-models-in-the-wild-lessons-learned-f-omar-sanseviero.json @@ -0,0 +1,24 @@ +{ + "description": "Lightning Talk: Distributing a Million Open Models in the Wild: Lessons Learned from the Hugging Face Hub - Omar Sanseviero, Hugging Face\n\nThe Hugging Face Hub has over 300,000 PyTorch models. Distributing such number of models poses challenges. In this talk, Omar will share how the community has tackled these challenges, including techniques to ensure torch model security and tooling for researchers to share their models. He'll also take attendees on a journey through the evolution of torch models distributed by the community, highlighting new trends and directions. Attending this talk will give attendees practical insights into the latest developments in model distribution and ecosystem trends.", + "duration": 579, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/t3AHZzz8Ong/maxresdefault.webp", + "title": "Lightning Talk: Distributing a Million Open Models in the Wild: Lessons Learned f... Omar Sanseviero", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=t3AHZzz8Ong" + } + ] +} diff --git a/pytorchconf-2024/videos/lightning-talk-empowering-developers-tools-and-resources-for-running-generative-a-pareena-verma.json b/pytorchconf-2024/videos/lightning-talk-empowering-developers-tools-and-resources-for-running-generative-a-pareena-verma.json new file mode 100644 index 000000000..d35489e6c --- /dev/null +++ b/pytorchconf-2024/videos/lightning-talk-empowering-developers-tools-and-resources-for-running-generative-a-pareena-verma.json @@ -0,0 +1,28 @@ +{ + "description": "Lightning Talk: Empowering Developers: Tools and Resources for Running Generative AI on Arm CPUs - Pareena Verma, Arm\n\nAs the demand for accessible and scalable AI solutions grows, leveraging CPUs for generative AI offers significant advantages in cost, energy efficiency and widespread availability. This sessions aims to equip developers with the ecosystem of tools, resources and technical content needed to effectively run generative AI use cases on Arm CPUs. We have launched a range of easily digestible tutorials for developers, part of our Learning Paths on https://learn.arm.com/, which demonstrate how you can easily and efficiently run small and large language models on Arm-based devices. Learn about end-to-end workflows to accelerate PyTorch based sentiment analysis models from Hugging Face on Arm servers with optimizations in Arm Compute Library kernels for fp32 and bfloat16. Use the new KleidiAI library to accelerate LLMs with AI frameworks and build an Android chat app on your Arm mobile device with ExecuTorch, and XNNPACK. Find out about our roadmap for learning content demonstrating the feasibility and successful deployment of generative AI on Arm-based devices. Help us shape the support that we offer developers.", + "duration": 738, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + }, + { + "label": "https://learn.arm.com/,", + "url": "https://learn.arm.com/," + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/wkzBHF9CL5s/maxresdefault.webp", + "title": "Lightning Talk: Empowering Developers: Tools and Resources for Running Generative A... Pareena Verma", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=wkzBHF9CL5s" + } + ] +} diff --git a/pytorchconf-2024/videos/lightning-talk-extending-pytorch-with-custom-python-c-cuda-operators-richard-zou-meta.json b/pytorchconf-2024/videos/lightning-talk-extending-pytorch-with-custom-python-c-cuda-operators-richard-zou-meta.json new file mode 100644 index 000000000..4e112aaf1 --- /dev/null +++ b/pytorchconf-2024/videos/lightning-talk-extending-pytorch-with-custom-python-c-cuda-operators-richard-zou-meta.json @@ -0,0 +1,24 @@ +{ + "description": "Lightning Talk: Extending PyTorch with Custom Python/C++/CUDA Operators - Richard Zou, Meta\n\nIn this talk, we'll go over the new recommended APIs to extend PyTorch with custom Python/C++/CUDA operators. Users have been able to extend PyTorch with custom operators for years but we have updated our guidance for creating custom operators that compose with torch.compile, autograd, and other PyTorch subsystems.", + "duration": 511, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/LI3h8aVchwo/maxresdefault.webp", + "title": "Lightning Talk: Extending PyTorch with Custom Python/C++/CUDA Operators - Richard Zou, Meta", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=LI3h8aVchwo" + } + ] +} diff --git a/pytorchconf-2024/videos/lightning-talk-fast-scalable-distributed-training-with-streamingdataset-saaketh-narayan.json b/pytorchconf-2024/videos/lightning-talk-fast-scalable-distributed-training-with-streamingdataset-saaketh-narayan.json new file mode 100644 index 000000000..bd066bec5 --- /dev/null +++ b/pytorchconf-2024/videos/lightning-talk-fast-scalable-distributed-training-with-streamingdataset-saaketh-narayan.json @@ -0,0 +1,24 @@ +{ + "description": "Lightning Talk: Fast, Scalable Distributed Training with StreamingDataset - Saaketh Narayan, Databricks\n\nStreamingDataset makes training on large datasets from cloud storage as fast, cheap, and scalable as possible. It\u2019s specially designed for multi-node, distributed training for large models \u2014 maximizing correctness guarantees, performance, and ease of use. Key features include elastically deterministic training, instant mid-epoch resumption, effective shuffling, high training throughput, and flexible data mixing, among other features. When training with StreamingDataset, the data shards are written to cloud storage in MDS, our file format that allows for low-latency random access to samples. By being as efficient as possible with shard downloads and shuffling, StreamingDataset minimizes egress costs while ensuring that dataloading never bottlenecks model training. StreamingDataset powers training for LLMs with over 100 billion parameters like DBRX, to advanced diffusion models, to two-tower recommendation models, and more, scaling to training jobs on thousands of GPUs with ease. Join us to learn how StreamingDataset can elevate your distributed model training experience.", + "duration": 423, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/C1TpUZjqTfA/maxresdefault.webp", + "title": "Lightning Talk: Fast, Scalable Distributed Training with StreamingDataset - Saaketh Narayan", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=C1TpUZjqTfA" + } + ] +} diff --git a/pytorchconf-2024/videos/lightning-talk-flexattention-the-flexibility-of-pytorch-the-performa-yanbo-liang-horace-he.json b/pytorchconf-2024/videos/lightning-talk-flexattention-the-flexibility-of-pytorch-the-performa-yanbo-liang-horace-he.json new file mode 100644 index 000000000..ce8bb06ee --- /dev/null +++ b/pytorchconf-2024/videos/lightning-talk-flexattention-the-flexibility-of-pytorch-the-performa-yanbo-liang-horace-he.json @@ -0,0 +1,24 @@ +{ + "description": "Lightning Talk: FlexAttention - The Flexibility of PyTorch + The Performance of FlashAttention - Yanbo Liang & Horace He, Meta\n\nIntroducing a novel abstraction leveraging the PyTorch compiler stack to enable custom, user-defined attention mechanisms. This new API supports dynamic modifications to attention scores within SDPA, providing both runtime and memory efficiency through kernel fusion with the FlashAttention algorithm.", + "duration": 1060, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/ju-KlcuWlbk/maxresdefault.webp", + "title": "Lightning Talk: FlexAttention - The Flexibility of PyTorch + The Performa... Yanbo Liang & Horace He", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=ju-KlcuWlbk" + } + ] +} diff --git a/pytorchconf-2024/videos/lightning-talk-hieroglyph2text-a-pytorch-powered-pipeline-for-automated-egyptian-h-susi-gentsch.json b/pytorchconf-2024/videos/lightning-talk-hieroglyph2text-a-pytorch-powered-pipeline-for-automated-egyptian-h-susi-gentsch.json new file mode 100644 index 000000000..3a5bd593c --- /dev/null +++ b/pytorchconf-2024/videos/lightning-talk-hieroglyph2text-a-pytorch-powered-pipeline-for-automated-egyptian-h-susi-gentsch.json @@ -0,0 +1,24 @@ +{ + "description": "Lightning Talk: HieroGlyph2Text: A PyTorch-Powered Pipeline for Automated Egyptian Hieroglyph Translation from Image - Susi Gentsch, University of Bonn\n\nHieroGlyph2Text is an innovative PyTorch-powered pipeline that automates the detection, classification, and attempts translation of Egyptian hieroglyphs from large image inputs. It addresses the challenge of decoding and translating ancient hieroglyphic inscriptions, traditionally a time-consuming and specialized task. This pipeline leverages PyTorch to create custom models: 1. Object Detection: YOLOv8 accurately detects individual hieroglyphs within images. 2. Image Classification: A custom ResNet model built using PyTorch achieves state-of-the-art accuracy in assigning Gardiner Codes to hieroglyphs. 3. Translation: The classified Gardiner Codes outputs from the ResNet model are integrated with Llama3, a large language model (LLM), using Retrieval-Augmented Generation (RAG) and a custom dataset based upon Gardiner Codes and their respective description and ideogram. Key highlights include accurate hieroglyph detection and state-of-the-art classification performance through an optimized ResNet model. This pipeline lays the groundwork for collaboration with subject matter experts to refine the translation process and democratize access to ancient Egyptian hieroglyphic knowledge.", + "duration": 564, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/eDjHKOb9Tn4/maxresdefault.webp", + "title": "Lightning Talk: HieroGlyph2Text: A PyTorch-Powered Pipeline for Automated Egyptian H... Susi Gentsch", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=eDjHKOb9Tn4" + } + ] +} diff --git a/pytorchconf-2024/videos/lightning-talk-implementing-and-using-iterable-datasets-what-could-go-wrong-nicolas-hug-meta.json b/pytorchconf-2024/videos/lightning-talk-implementing-and-using-iterable-datasets-what-could-go-wrong-nicolas-hug-meta.json new file mode 100644 index 000000000..8f2fd171e --- /dev/null +++ b/pytorchconf-2024/videos/lightning-talk-implementing-and-using-iterable-datasets-what-could-go-wrong-nicolas-hug-meta.json @@ -0,0 +1,24 @@ +{ + "description": "Lightning Talk: Implementing and Using Iterable Datasets: What Could Go Wrong? - Nicolas Hug, Meta\n\nPyTorch supports two kinds of datasets: Iterable datasets and indexable \"map-style\" datasets. Iterable datasets can be more flexible and potentially faster than their indexable cousins. They are also much harder to use correctly, and can easily lead to silently wrong results. This talk is a quick and fun intro to some of the traps that Iterable datasets lay out for you, with some tips to help you avoid them.", + "duration": 674, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/Kx0BoPa9juQ/maxresdefault.webp", + "title": "Lightning Talk: Implementing and Using Iterable Datasets: What Could Go Wrong? - Nicolas Hug, Meta", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=Kx0BoPa9juQ" + } + ] +} diff --git a/pytorchconf-2024/videos/lightning-talk-in-transit-machine-learning-using-pytorch-on-frontier-exascale-system-vineeth-gutta.json b/pytorchconf-2024/videos/lightning-talk-in-transit-machine-learning-using-pytorch-on-frontier-exascale-system-vineeth-gutta.json new file mode 100644 index 000000000..1382e7cbe --- /dev/null +++ b/pytorchconf-2024/videos/lightning-talk-in-transit-machine-learning-using-pytorch-on-frontier-exascale-system-vineeth-gutta.json @@ -0,0 +1,24 @@ +{ + "description": "Lightning Talk: In-Transit Machine Learning Using PyTorch on Frontier Exascale System - Vineeth Gutta, University of Delaware\n\nTraditional ML workflows use offline training where the data is stored on disk and is subsequently loaded into accelerator (CPU,GPU, etc) memory during training or inference. We recently devised a novel and scalable in-transit ML workflow for a plasma-physics application (chosen as 1 out of 8 compelling codes in the country) for the world\u2019s fastest supercomputer, Frontier) with an aim to build a high-energy laser particle accelerator. Data generated in distributed HPC systems like Frontier create volumes of data that is infeasible to store on HPC file systems. A mismatch between modern memory hierarchies occurs due to high volume and rate of data generation. Our novel ML workflow utilizes continuous learning where the data is consumed in batches as the simulation produces the data and then discards after each batch is trained. This in-transit workflow integrates particle-in-cell simulations with distributed ML training on PyTorch using DDP allows for an application coupling enabling the model to learn correlations between emitted radiation and particle dynamics within simulation in an unsupervised method. This workflow is demonstrated at scale on Frontier using 400 AMD MI250X GPUs", + "duration": 725, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/s4Gh2UX0EyA/maxresdefault.webp", + "title": "Lightning Talk: In-Transit Machine Learning Using PyTorch on Frontier Exascale System- Vineeth Gutta", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=s4Gh2UX0EyA" + } + ] +} diff --git a/pytorchconf-2024/videos/lightning-talk-introduction-to-torch-distributed-pipelining-howard-huang-ke-wen-meta.json b/pytorchconf-2024/videos/lightning-talk-introduction-to-torch-distributed-pipelining-howard-huang-ke-wen-meta.json new file mode 100644 index 000000000..b78bacae8 --- /dev/null +++ b/pytorchconf-2024/videos/lightning-talk-introduction-to-torch-distributed-pipelining-howard-huang-ke-wen-meta.json @@ -0,0 +1,24 @@ +{ + "description": "Lightning Talk: Introduction to Torch.Distributed.Pipelining - Howard Huang & Ke Wen, Meta\n\nPipeline parallelism is a technique employed in distributed deep learning that enhances model execution by dividing the model into distinct segments, or \"stages.\" As large language models and other memory-intensive models become more common, pipeline parallelism has grown increasingly important for several key areas: - Executing large-scale training jobs. - Enhancing performance in bandwidth-limited clusters. - Supporting large model inference. In this talk, we will introduce the `torch.distributed.pipelining` package which provides users a seamless way of applying pipeline parallelism. We will demonstrate the following features: - Splitting of model code based on simple specification. - Support for pipeline schedules, including GPipe, 1F1B, Interleaved 1F1B and Looped BFS, and providing the infrastructure for writing customized schedules. - Composability with other PyTorch parallel techniques such as data parallel (DDP, FSDP) or tensor parallel. - Out of the box integration with Hugging Face models for efficient inference.", + "duration": 765, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/YieoyQll4_4/maxresdefault.webp", + "title": "Lightning Talk: Introduction to Torch.Distributed.Pipelining - Howard Huang & Ke Wen, Meta", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=YieoyQll4_4" + } + ] +} diff --git a/pytorchconf-2024/videos/lightning-talk-llms-on-edge-with-ai-accelerators-chen-lai-kimish-patel-cemal-bilgin-meta.json b/pytorchconf-2024/videos/lightning-talk-llms-on-edge-with-ai-accelerators-chen-lai-kimish-patel-cemal-bilgin-meta.json new file mode 100644 index 000000000..10062ca0a --- /dev/null +++ b/pytorchconf-2024/videos/lightning-talk-llms-on-edge-with-ai-accelerators-chen-lai-kimish-patel-cemal-bilgin-meta.json @@ -0,0 +1,24 @@ +{ + "description": "Lightning Talk: LLMs on Edge with AI Accelerators - Chen Lai, Kimish Patel & Cemal Bilgin, Meta\n\nLLMs are known to be compute heavy and consume lots of resources (almost all resources on phones), including memory and power. A natural thought is to leverage the AI hardware accelerators, for example, Apple Neural Engine (ANE) on Apple devices and HTP on Qualcomm SoCs, to make it run fast and efficiently. Only by optimizing the model latency, memory consumption and power usage to a certain level will users be interested in installing the models on their devices. In this session, we\u2019d like to introduce how we leverage these AI accelerators within the PyTorch ecosystem to achieve the state-of-art performance for llama3 on device, via ExecuTorch and the partnership with Apple and Qualcomm. Hardware companies usually have their own AI accelerators. Likely they have different characteristics, one may support a list of different operators than others, and one may only support static shapes (like HTP). However, transformers-based optimization can be generic. We\u2019ll discuss in more detail how we apply the generic optimization as well as the backend specific optimization. The techniques we applied here are not just for LLMs, but can be applied to other transformer-based models.", + "duration": 729, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/8p8Pntnr3GU/maxresdefault.webp", + "title": "Lightning Talk: LLMs on Edge with AI Accelerators - Chen Lai, Kimish Patel & Cemal Bilgin, Meta", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=8p8Pntnr3GU" + } + ] +} diff --git a/pytorchconf-2024/videos/lightning-talk-low-precision-dtypes-in-pytorch-vasiliy-kuznetsov-meta.json b/pytorchconf-2024/videos/lightning-talk-low-precision-dtypes-in-pytorch-vasiliy-kuznetsov-meta.json new file mode 100644 index 000000000..38c415cc1 --- /dev/null +++ b/pytorchconf-2024/videos/lightning-talk-low-precision-dtypes-in-pytorch-vasiliy-kuznetsov-meta.json @@ -0,0 +1,24 @@ +{ + "description": "Lightning Talk: Low Precision Dtypes in PyTorch - Vasiliy Kuznetsov, Meta\n\nThis talk deep dives into the new native PyTorch float8 training library, and previews PyTorch's strategy for supporting upcoming low precision dtypes such as float6, float4 and MX for efficient training and inference.", + "duration": 530, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/xcKwEZ77Cps/maxresdefault.webp", + "title": "Lightning Talk: Low Precision Dtypes in PyTorch - Vasiliy Kuznetsov, Meta", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=xcKwEZ77Cps" + } + ] +} diff --git a/pytorchconf-2024/videos/lightning-talk-making-the-most-of-heterogeneous-memory-capacity-using-pytorch-syed-ahmed-nvidia.json b/pytorchconf-2024/videos/lightning-talk-making-the-most-of-heterogeneous-memory-capacity-using-pytorch-syed-ahmed-nvidia.json new file mode 100644 index 000000000..ebf0a1428 --- /dev/null +++ b/pytorchconf-2024/videos/lightning-talk-making-the-most-of-heterogeneous-memory-capacity-using-pytorch-syed-ahmed-nvidia.json @@ -0,0 +1,24 @@ +{ + "description": "Lightning Talk: Making the Most of Heterogeneous Memory Capacity Using PyTorch - Syed Ahmed, NVIDIA Corporation\n\nMemory intensive deep learning workloads require efficient use of all kinds of memories that are available in a system. In this session, we will discuss how we can utilize such heterogeneous memory through memory pools in PyTorch. We will show how to mix-and-match different CUDA system allocators in the same PyTorch program using memory pools. Consequently, this API unlocks new use cases such as Extended GPU Memory (EGM) based all-gathers, Unified Virtual Memory (UVM), and NVLink Sharp (NVLS) reductions. New NVIDIA architectures accelerate such use cases with high-bandwidth and low-latency interconnects in the hardware, driven by extended functionality of CUDA system allocators in the software. Learn how to use these techniques on memory-intensive deep learning models like LLMs, and discover new CUDA features powered by PyTorch.", + "duration": 686, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/srQOK1UusZ4/maxresdefault.webp", + "title": "Lightning Talk: Making the Most of Heterogeneous Memory Capacity Using PyTorch - Syed Ahmed, NVIDIA", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=srQOK1UusZ4" + } + ] +} diff --git a/pytorchconf-2024/videos/lightning-talk-mobile-computational-photography-with-pytorch-low-light-denoising-alexis-baudron.json b/pytorchconf-2024/videos/lightning-talk-mobile-computational-photography-with-pytorch-low-light-denoising-alexis-baudron.json new file mode 100644 index 000000000..928b6c573 --- /dev/null +++ b/pytorchconf-2024/videos/lightning-talk-mobile-computational-photography-with-pytorch-low-light-denoising-alexis-baudron.json @@ -0,0 +1,24 @@ +{ + "description": "Lightning Talk: Mobile Computational Photography with PyTorch: Low-Light Denoising - Alexis Baudron, Sony\n\nOver the last decade, smartphone cameras have improved significantly, becoming the primary device people use for capturing everyday moments and high-quality photographs. This progress is largely due to advances in computational photography and novel image sensors. Computational photography enables great images from compact mobile cameras, enhancing photos through various techniques such as multi-shot merging. Despite these advancements, challenges such as noise, artifacts, and distortions persist, especially in low-light conditions where limited light increases noise levels. In this lightning talk, we will explore how PyTorch can be used to design and optimize deep learning networks for real-time low-light denoising. We will dive into noise modeling, data generation, physics-aware models, and advanced network architectures for effective denoising in challenging low-light scenarios. Attendees will gain practical insights into the latest advancements in mobile computational photography using PyTorch.", + "duration": 507, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/hjOtsOqPw3Y/maxresdefault.webp", + "title": "Lightning Talk: Mobile Computational Photography with PyTorch: Low-Light Denoising - Alexis Baudron", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=hjOtsOqPw3Y" + } + ] +} diff --git a/pytorchconf-2024/videos/lightning-talk-new-activation-checkpointing-apis-in-pytorch-jeffrey-wan-horace-he-meta.json b/pytorchconf-2024/videos/lightning-talk-new-activation-checkpointing-apis-in-pytorch-jeffrey-wan-horace-he-meta.json new file mode 100644 index 000000000..d86d9aaa3 --- /dev/null +++ b/pytorchconf-2024/videos/lightning-talk-new-activation-checkpointing-apis-in-pytorch-jeffrey-wan-horace-he-meta.json @@ -0,0 +1,24 @@ +{ + "description": "Lightning Talk: New Activation Checkpointing APIs in PyTorch - Jeffrey Wan & Horace He, Meta\n\nActivation checkpointing is a commonly used technique to reduce memory usage during model training by reducing the number of activations saved for backward. Instead of keeping tensors needed for backward alive until they are used in gradient computation during backward, those tensors are recomputed during the backward pass. This talk will introduce new activation checkpoint APIs that can help achieve a better trade off between memory savings and compute overhead that recomputing introduces.", + "duration": 750, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/v3gsrJtGLiA/maxresdefault.webp", + "title": "Lightning Talk: New Activation Checkpointing APIs in PyTorch - Jeffrey Wan & Horace He, Meta", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=v3gsrJtGLiA" + } + ] +} diff --git a/pytorchconf-2024/videos/lightning-talk-on-device-profiling-and-debugging-with-executorch-olivia-liu-vaun-puri-meta.json b/pytorchconf-2024/videos/lightning-talk-on-device-profiling-and-debugging-with-executorch-olivia-liu-vaun-puri-meta.json new file mode 100644 index 000000000..f272770fb --- /dev/null +++ b/pytorchconf-2024/videos/lightning-talk-on-device-profiling-and-debugging-with-executorch-olivia-liu-vaun-puri-meta.json @@ -0,0 +1,24 @@ +{ + "description": "Lightning Talk: On-Device Profiling and Debugging with ExecuTorch - Olivia Liu & Vaun Puri, Meta\n\nHigh developer velocity is crucial to shipping new ML-enabled experiences from a server-trained model to a customers\u2019 device. ExecuTorch is an on-device runtime that seamlessly integrates with the PyTorch stack with a focus on developer productivity. We present the ExecuTorch Dev Tools and highlight key features that tighten the iteration loop when optimizing models for deployment and execution on edge devices. We demonstrate how ExecuTorch\u2019s built-in profiler and bundled tools tackle key pain-points, such as: 1. Examining the memory footprint of an ExecuTorch program ahead-of-time; 2. Collecting runtime performance metrics and intermediate outputs for accuracy analysis; 3. Correlating runtime data with the underlying graph of an exported model.", + "duration": 852, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/_e6E6fnEIDQ/maxresdefault.webp", + "title": "Lightning Talk: On-Device Profiling and Debugging with ExecuTorch - Olivia Liu & Vaun Puri, Meta", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=_e6E6fnEIDQ" + } + ] +} diff --git a/pytorchconf-2024/videos/lightning-talk-optimized-pytorch-inference-on-aarch64-linux-cpus-sunita-nadampalli-amazon-aws.json b/pytorchconf-2024/videos/lightning-talk-optimized-pytorch-inference-on-aarch64-linux-cpus-sunita-nadampalli-amazon-aws.json new file mode 100644 index 000000000..f00b2fb5e --- /dev/null +++ b/pytorchconf-2024/videos/lightning-talk-optimized-pytorch-inference-on-aarch64-linux-cpus-sunita-nadampalli-amazon-aws.json @@ -0,0 +1,24 @@ +{ + "description": "Lightning Talk: Optimized PyTorch Inference on aarch64 Linux CPUs - Sunita Nadampalli, Amazon (AWS)\n\nIn the last 2 years we've optimized performance of PyTorch on Arm processors. The optimizations have included changes to ATen, C10, MKLDNN operators, GEMM backend, and Torch inductor. In many cases instead of writing our own kernel we integrated the Arm compute library, used fastmath kernels with format types like bf16, implemented operator caching, selected optimal backend based on the input context etc. Through these optimizations we improved performance by over 2x. In this presentation first we will talk about how we went across this process, what those optimizations are, performance numbers for AWS Graviton3 processors for around 75 models, and CI/CD workflow details. Next, we will walk through a sample PyTorch application showing basic usage, how to tune runtime and the resulting speed up. At the end of the presentation attendees will learn about PyTorch performance optimizations on Arm processors, how to use them, and the areas where they can collaborate to further improve PyTorch for aarch64 CPUs.", + "duration": 790, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/_9xohmSa2G8/maxresdefault.webp", + "title": "Lightning Talk: Optimized PyTorch Inference on aarch64 Linux CPUs - Sunita Nadampalli, Amazon (AWS)", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=_9xohmSa2G8" + } + ] +} diff --git a/pytorchconf-2024/videos/lightning-talk-pytorch-release-process-andrey-talman-meta.json b/pytorchconf-2024/videos/lightning-talk-pytorch-release-process-andrey-talman-meta.json new file mode 100644 index 000000000..019dd636b --- /dev/null +++ b/pytorchconf-2024/videos/lightning-talk-pytorch-release-process-andrey-talman-meta.json @@ -0,0 +1,24 @@ +{ + "description": "Lightning Talk: PyTorch Release Process - Andrey Talman, Meta\n\nI would like to present and quickly discuss PyTorch Release process, how it happens. What are milestones. What is our cherry-picking criteria, how we validate the release.", + "duration": 573, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/4REnTJO3_ps/maxresdefault.webp", + "title": "Lightning Talk: PyTorch Release Process - Andrey Talman, Meta", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=4REnTJO3_ps" + } + ] +} diff --git a/pytorchconf-2024/videos/lightning-talk-pytorch-xla-auto-sharding-yeounoh-chung-google.json b/pytorchconf-2024/videos/lightning-talk-pytorch-xla-auto-sharding-yeounoh-chung-google.json new file mode 100644 index 000000000..f2a6e2f92 --- /dev/null +++ b/pytorchconf-2024/videos/lightning-talk-pytorch-xla-auto-sharding-yeounoh-chung-google.json @@ -0,0 +1,24 @@ +{ + "description": "Lightning Talk: PyTorch/XLA Auto-Sharding - Yeounoh Chung, Google\n\nPyTorch/XLA recently launched the new PyTorch/XLA SPMD feature as a first-step to automate ML workloads parallelization using GSPMD. It turns out that the performance largely depends on the quality of sharding hints provided by the user \u2013 and it requires a correct and deep understanding of model architectures and much expertise to come up with optimal sharding hints. To address this problem, we propose to integrate PyTorch/XLA SPMD with XLA's auto sharding service that allows the XLA compiler to shard and optimize the whole model without any user input.", + "duration": 461, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/gNfyS9asA7A/maxresdefault.webp", + "title": "Lightning Talk: PyTorch/XLA Auto-Sharding - Yeounoh Chung, Google", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=gNfyS9asA7A" + } + ] +} diff --git a/pytorchconf-2024/videos/lightning-talk-sparsifying-vision-transformers-with-minimal-accuracy-loss-jesse-cai-meta.json b/pytorchconf-2024/videos/lightning-talk-sparsifying-vision-transformers-with-minimal-accuracy-loss-jesse-cai-meta.json new file mode 100644 index 000000000..a2c3a9c09 --- /dev/null +++ b/pytorchconf-2024/videos/lightning-talk-sparsifying-vision-transformers-with-minimal-accuracy-loss-jesse-cai-meta.json @@ -0,0 +1,24 @@ +{ + "description": "Lightning Talk: Sparsifying Vision Transformers with Minimal Accuracy Loss - Jesse Cai, Meta\n\nSparsity, like quantization, is an approximate model optimization technique, where we trade some model accuracy for increased performance.\n\nIn this talk we'll explore how to minimize the accuracy degradation of sparsifying Vision Transformer (ViT) based models to GPU accelerable sparsity patterns like block sparsity and semi-structured sparsity.\n\nWe'll cover the best techniques to ensure a less-than 5% loss in accuracy when:\n- training a sparse model from scratch\n- pruning and retraining an existing dense model\n- zero-shot/one-shot pruning a dense model\n\nWe've collected these techniques into a single repository, torchao, so that model optimization enthusiasts like you can sparsify your models with just a few lines of code.", + "duration": 841, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/mqoIfs5hTxA/maxresdefault.webp", + "title": "Lightning Talk: Sparsifying Vision Transformers with Minimal Accuracy Loss - Jesse Cai, Meta", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=mqoIfs5hTxA" + } + ] +} diff --git a/pytorchconf-2024/videos/lightning-talk-understanding-and-optimizing-pytorch-models-with-thunder-luca-antiga-lightning-ai.json b/pytorchconf-2024/videos/lightning-talk-understanding-and-optimizing-pytorch-models-with-thunder-luca-antiga-lightning-ai.json new file mode 100644 index 000000000..e8667a142 --- /dev/null +++ b/pytorchconf-2024/videos/lightning-talk-understanding-and-optimizing-pytorch-models-with-thunder-luca-antiga-lightning-ai.json @@ -0,0 +1,24 @@ +{ + "description": "Lightning Talk: Understanding and Optimizing PyTorch Models with Thunder - Luca Antiga, Lightning AI\n\nA hallmark feature of PyTorch is the natural expression of computation. This enables practitioners to implement AI models with ease. However, it prompts the question how to optimize the workload for a given hardware setup because those optimizations clutter our code and are tricky to combine. Lightning Thunder provides a Python-to-Python compiler to scale and optimize PyTorch programs that focuses on usability, understandability, and extensibility. A key tool in delivering on these goals is the composability of transformations: without changing the user code, we can stack quantization, distributing the computation across multiple GPUs, dispatching to optimized kernels, offloading, and other pluggable optimizations. Lightning Thunder flourishes in the PyTorch ecosystem: with PyTorch eager and with executors like torch.compile and nvFuser. It also dispatches to libraries like cuDNN, TransformerEngine, Apex, OpenAI Triton. The ability to apply multiple optimizations just-in-time leads to significant compounded speed-ups over unoptimized code out of the box. Luca will discuss the design of Thunder and demonstrate applications on training and inference for large language and multimodal models.", + "duration": 837, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/5dQ423cFfLc/maxresdefault.webp", + "title": "Lightning Talk: Understanding and Optimizing PyTorch Models with Thunder - Luca Antiga, Lightning AI", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=5dQ423cFfLc" + } + ] +} diff --git a/pytorchconf-2024/videos/lightning-talk-what-s-new-for-pytorch-developer-infrastructure-sahan-paliskara-catherine-lee.json b/pytorchconf-2024/videos/lightning-talk-what-s-new-for-pytorch-developer-infrastructure-sahan-paliskara-catherine-lee.json new file mode 100644 index 000000000..6e142c2d9 --- /dev/null +++ b/pytorchconf-2024/videos/lightning-talk-what-s-new-for-pytorch-developer-infrastructure-sahan-paliskara-catherine-lee.json @@ -0,0 +1,24 @@ +{ + "description": "Lightning Talk: What's New for PyTorch Developer Infrastructure - Sahan Paliskara & Catherine Lee, Meta\n\nHaving a chat about all of the work being done to continue supporting PyTorch's Developer Infrastructure needs including updates around Target Determination, Releases, and OSS Tooling.", + "duration": 671, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/xlhBDySu7jM/maxresdefault.webp", + "title": "Lightning Talk: What's New for PyTorch Developer Infrastructure - Sahan Paliskara & Catherine Lee", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=xlhBDySu7jM" + } + ] +} diff --git a/pytorchconf-2024/videos/lightning-talk-whats-new-in-ex-angela-yi-tugsbayasgalan-manlaibaatar-avik-chaudhuri-yidi-wu.json b/pytorchconf-2024/videos/lightning-talk-whats-new-in-ex-angela-yi-tugsbayasgalan-manlaibaatar-avik-chaudhuri-yidi-wu.json new file mode 100644 index 000000000..35788fe2a --- /dev/null +++ b/pytorchconf-2024/videos/lightning-talk-whats-new-in-ex-angela-yi-tugsbayasgalan-manlaibaatar-avik-chaudhuri-yidi-wu.json @@ -0,0 +1,24 @@ +{ + "description": "Lightning Talk: What\u2019s New in Export? - Angela Yi, Tugsbayasgalan Manlaibaatar, Avik Chaudhuri & Yidi Wu, Meta\n\nThis talk discusses updates we've made to torch.export this past year: (a) Non-strict mode, an alternative tracing mode which in practice covers more programs than TorchDynamo without compromising important soundness guarantees (b) Better dynamic shapes specifications through generating suggested fixes and runtime assertions (c) Control flow operators such as cond, map, and associative scan (d) A shift in the export generated IR, which will enable both training and inference (e) An unflattener, which will reconstruct the eager module structure from the flattened exported graph", + "duration": 814, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/ap1TOf7p0iA/maxresdefault.webp", + "title": "Lightning Talk: What\u2019s New in Ex... Angela Yi, Tugsbayasgalan Manlaibaatar, Avik Chaudhuri & Yidi Wu", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=ap1TOf7p0iA" + } + ] +} diff --git a/pytorchconf-2024/videos/maximizing-training-throughput-using-torch-compile-and-fsdp-l-chu-a-viros-i-martin-b-vaughan.json b/pytorchconf-2024/videos/maximizing-training-throughput-using-torch-compile-and-fsdp-l-chu-a-viros-i-martin-b-vaughan.json new file mode 100644 index 000000000..0eb19b628 --- /dev/null +++ b/pytorchconf-2024/videos/maximizing-training-throughput-using-torch-compile-and-fsdp-l-chu-a-viros-i-martin-b-vaughan.json @@ -0,0 +1,24 @@ +{ + "description": "Maximizing Training Throughput Using Torch.Compile and FSDP - Linsong Chu & Antoni Viros i Martin, IBM Research; Brian Vaughan, IBM\n\ntorch.compile is a graph compilation technique that improves GPU utilization. A key challenge in getting torch.compile to perform well is to minimize (or eliminate) graph breaks, however, this isn't trivial as even the Llama implementation provided by Meta has many graph breaks resulting in reduced training throughput. In this talk we discuss 1. how we addressed these challenges in order to train a model using torch.compile 2. how we combined torch.compile with FSDP and selective activation checkpointing to achieve the maximum throughput for training 3. model quality comparison between models trained with compile and no-compile, and lastly 4. the best setup we have for different model sizes in the Llama family that achieves the maximum throughput and MFU number (e.g. 68% MFU for the 7B model on A100 GPUs!)", + "duration": 220, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/_CuLeABf_fM/maxresdefault.webp", + "title": "Maximizing Training Throughput Using Torch.Compile and FSDP - L. Chu, A. Viros i Martin, B. Vaughan", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=_CuLeABf_fM" + } + ] +} diff --git a/pytorchconf-2024/videos/meta-llama-3-and-the-future-of-responsible-ai-development-spencer-whitman-vincent-gonguet-meta.json b/pytorchconf-2024/videos/meta-llama-3-and-the-future-of-responsible-ai-development-spencer-whitman-vincent-gonguet-meta.json new file mode 100644 index 000000000..7bdd120de --- /dev/null +++ b/pytorchconf-2024/videos/meta-llama-3-and-the-future-of-responsible-ai-development-spencer-whitman-vincent-gonguet-meta.json @@ -0,0 +1,24 @@ +{ + "description": "Meta Llama 3 and the Future of Responsible AI Development - Spencer Whitman & Vincent Gonguet, Meta\n\nAs AI models become increasingly powerful and pervasive, trust and safety have become top priorities. Join us for a timely talk on Llama 3, our latest foundation model, and the cutting-edge trust and safety models and tools we've developed to ensure responsible AI development. In this talk, we'll dive into: \u2022The advancements of Llama 3 and its applications \u2022Our innovative trust and safety approaches, including toxicity detection and mitigation \u2022The open-source tools and resources we're sharing to empower the community Discover how Meta is pushing the boundaries of trust and safety and learn how you can integrate these solutions into your own projects. Let's build a safer, more responsible AI future together!", + "duration": 1251, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/XOIuFIl2-Ao/maxresdefault.webp", + "title": "Meta Llama 3 and the Future of Responsible AI Development - Spencer Whitman & Vincent Gonguet, Meta", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=XOIuFIl2-Ao" + } + ] +} diff --git a/pytorchconf-2024/videos/mlir-enabling-composition-of-kernels-and-compilers-jacques-pienaar-google.json b/pytorchconf-2024/videos/mlir-enabling-composition-of-kernels-and-compilers-jacques-pienaar-google.json new file mode 100644 index 000000000..98597e37b --- /dev/null +++ b/pytorchconf-2024/videos/mlir-enabling-composition-of-kernels-and-compilers-jacques-pienaar-google.json @@ -0,0 +1,24 @@ +{ + "description": "[MLIR] Enabling Composition of Kernels and Compilers - Jacques Pienaar, Google\n\nHand written kernels and compilers have been part of the toolbox to provide efficient and broad coverage. These approaches have often been positioned as being at odds with one another - and indeed the software solutions either side have sometimes made it such. MLIR, since inception, aimed to enable general, beneficial composition instead. Rather than treating kernels as a black box escape hatch, treat it as a peer in solving the serving needs. This is not magic and requires consideration of how best to combine. In this talk I'll present the approach and effect of this both in IREE and OpenXLA.", + "duration": 672, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/Dx1fAE9fk8s/maxresdefault.webp", + "title": "[MLIR] Enabling Composition of Kernels and Compilers - Jacques Pienaar, Google", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=Dx1fAE9fk8s" + } + ] +} diff --git a/pytorchconf-2024/videos/mojo-lifting-pt-to-new-heights-with-max-and-mojo-mikhail-zolotukhin-modular.json b/pytorchconf-2024/videos/mojo-lifting-pt-to-new-heights-with-max-and-mojo-mikhail-zolotukhin-modular.json new file mode 100644 index 000000000..e1d1de8eb --- /dev/null +++ b/pytorchconf-2024/videos/mojo-lifting-pt-to-new-heights-with-max-and-mojo-mikhail-zolotukhin-modular.json @@ -0,0 +1,24 @@ +{ + "description": "[MOJO] Lifting PT to New Heights with MAX and Mojo - Mikhail Zolotukhin, Modular\n\nIn this talk we'll peek into Modular's inference engine: how it builds on and works with PyTorch and what is unique about it. We will look into how Mojo language can be used to define performant kernels and what optimizations the inference engine can perform. We will also talk briefly about our experience of developing a third party backend for torch.compile.", + "duration": 572, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/JmHKhc6EGpg/maxresdefault.webp", + "title": "[MOJO] Lifting PT to New Heights with MAX and Mojo - Mikhail Zolotukhin, Modular", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=JmHKhc6EGpg" + } + ] +} diff --git a/pytorchconf-2024/videos/panel-discussion-t-dettmers-h-schoelkopf-a-chowdhery-a-conneau-moderated-by-k-khandelwal.json b/pytorchconf-2024/videos/panel-discussion-t-dettmers-h-schoelkopf-a-chowdhery-a-conneau-moderated-by-k-khandelwal.json new file mode 100644 index 000000000..99aee6bbe --- /dev/null +++ b/pytorchconf-2024/videos/panel-discussion-t-dettmers-h-schoelkopf-a-chowdhery-a-conneau-moderated-by-k-khandelwal.json @@ -0,0 +1,24 @@ +{ + "description": "Panel Discussion - Tim Dettmers, AI2/Carnegie Melon; Hailey Schoelkopf, EleutherAI; Aakanksha Chowdhery, Meta; Alexis Conneau, OpenAI; Moderated by Kartikay Khandelwal, Meta\n\nSpeakers: Tim Dettmers, Hailey Schoelkopf, Aakanksha Chowdhery, Alexis Conneau\nAs open-source LLMs have become more capable, a substantial ecosystem has developed around the fine-tuning of these models. A thriving community of researchers, developers, practitioners and hobbyists has emerged which focuses on topics ranging from memory efficiency, parameter-efficient fine-tuning and quantization to performance at scale and reproducible evaluations. The goal of this mini-summit is to bring this community together to discuss ideas, share knowledge and build connections.\n\nThe agenda features a keynote from Joe Spisak on the state of the Llama ecosystem followed by invited talks from the founders of Axolotl, Unsloth and torchtune. We conclude the summit with a riveting discussion on what\u2019s next for LLMs, fine-tuning and the PyTorch ecosystem with a fabulous panel of experts - Tim Dettmers (author of bitsandbytes and QLoRA), Hailey Schoelkopf (maintainer of LM Eval Harness at EleutherAI), Aakanksha Chowdhery (Lead author on PaLM and Gemini) and Alexis Conneau (Research Lead at OpenAI)", + "duration": 2160, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/1VwOB7OYm4U/maxresdefault.webp", + "title": "Panel Discussion - T. Dettmers, H. Schoelkopf, A. Chowdhery, A. Conneau, Moderated by K. Khandelwal", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=1VwOB7OYm4U" + } + ] +} diff --git a/pytorchconf-2024/videos/pushing-the-performance-envelope-an-optimization-study-for-3-suvaditya-mukherjee-shireen-chand.json b/pytorchconf-2024/videos/pushing-the-performance-envelope-an-optimization-study-for-3-suvaditya-mukherjee-shireen-chand.json new file mode 100644 index 000000000..45397adb0 --- /dev/null +++ b/pytorchconf-2024/videos/pushing-the-performance-envelope-an-optimization-study-for-3-suvaditya-mukherjee-shireen-chand.json @@ -0,0 +1,24 @@ +{ + "description": "Pushing the Performance Envelope: An Optimization Study for 3D Generative Modelling with PyTorch - Suvaditya Mukherjee & Shireen Chand, University of Southern California\n\nThis work explores performance optimization strategies for training 3D generative models using PyTorch. We focus on training Variational Autoencoders (VAEs) on the ShapeNet dataset, a popular benchmark for this task. Our objective is to achieve high-fidelity reconstructions while minimizing the computational footprint and training time. We focus on: 1) Large-scale 3D dataset loading strategies using PyTorch & Google Cloud Storage Buckets 2) Implementation details and insights for 3D VAEs using PyTorch 2.x 3) Training using Automatic Mixed-precision regimes 4) Optimized training using torch.compile and different quantization techniques (as supported) - Dynamic Quantization - Static Quantization - Static Quantization-aware Training 5) Comparative Benchmark over several experiments performed with a focus on execution time and memory footprint Through this comprehensive study, we present a comparative analysis of the performance gains achieved by our optimized models. Our findings present empirical insights into the trade-offs between model accuracy, computational complexity, and hardware resource utilization.", + "duration": 1153, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/Xm9Bex3LE_8/maxresdefault.webp", + "title": "Pushing the Performance Envelope: An Optimization Study for 3... Suvaditya Mukherjee & Shireen Chand", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=Xm9Bex3LE_8" + } + ] +} diff --git a/pytorchconf-2024/videos/pytorch-conference-2024-highlights.json b/pytorchconf-2024/videos/pytorch-conference-2024-highlights.json new file mode 100644 index 000000000..4e5abd751 --- /dev/null +++ b/pytorchconf-2024/videos/pytorch-conference-2024-highlights.json @@ -0,0 +1,28 @@ +{ + "description": "Relive the excitement of the PyTorch Conference 2024, where AI innovators and enthusiasts gathered to explore cutting-edge advancements, foster community connections, and shape the future of machine learning.\n\nView all of the conference sessions on the PyTorch YouTube channel: https://www.youtube.com/playlist?list=PL_lsbAsL_o2B_znuvm-pDtV_cRhpqZb8l", + "duration": 100, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + }, + { + "label": "https://www.youtube.com/playlist?list=PL_lsbAsL_o2B_znuvm-pDtV_cRhpqZb8l", + "url": "https://www.youtube.com/playlist?list=PL_lsbAsL_o2B_znuvm-pDtV_cRhpqZb8l" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/tUG3UWSP2ZU/maxresdefault.webp", + "title": "PyTorch Conference 2024 Highlights", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=tUG3UWSP2ZU" + } + ] +} diff --git a/pytorchconf-2024/videos/running-state-of-art-gen-ai-models-on-device-with-npu-acceleration-felix-baum-qualcomm.json b/pytorchconf-2024/videos/running-state-of-art-gen-ai-models-on-device-with-npu-acceleration-felix-baum-qualcomm.json new file mode 100644 index 000000000..21006aca1 --- /dev/null +++ b/pytorchconf-2024/videos/running-state-of-art-gen-ai-models-on-device-with-npu-acceleration-felix-baum-qualcomm.json @@ -0,0 +1,24 @@ +{ + "description": "Running State-of-Art Gen AI Models on-Device with NPU Acceleration - Felix Baum, Qualcomm\n\nSince the boom of generative AI, the industry is now moving towards on-device AI inferencing, as it is not only a trend but a necessity now in order to save costs, achieve the best inference performance, ultra-low latency at the lowest power possible. In this session we go over the new features added on the Qualcomm AI Stack and how it works with the public release of ExecuTorch 1.0. We will discuss how to run traditional workloads as well as GenAI use cases including the latest version of Llama on the Mobile device while using Qualcomm Hexagon NPU.", + "duration": 1460, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/wd57g2IM3C4/maxresdefault.webp", + "title": "Running State-of-Art Gen AI Models on-Device with NPU Acceleration - Felix Baum, Qualcomm", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=wd57g2IM3C4" + } + ] +} diff --git a/pytorchconf-2024/videos/slaying-ooms-mark-saroufim-jane-xu-meta.json b/pytorchconf-2024/videos/slaying-ooms-mark-saroufim-jane-xu-meta.json new file mode 100644 index 000000000..3561a7bc3 --- /dev/null +++ b/pytorchconf-2024/videos/slaying-ooms-mark-saroufim-jane-xu-meta.json @@ -0,0 +1,24 @@ +{ + "description": "Slaying OOMs - Mark Saroufim & Jane Xu, Meta\n\nHave you ever hit an OOM (and wished you had more VRAM)? Who hasn't! Hop on the bus with us and feel the road become smoother as we talk about stacking together techniques like FSDP2 + QLoRa + CPU Offloading + Fused ADAM (thanks Intel) + more in PyTorch native. We will give an overview of these techniques as well as the hard edges we solved in their composition. Curious for more? Or...still OOMing? We also plan on discussing our more researchy work on offloading, pagedness, and low precision optimizers.", + "duration": 1541, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/xzBcBJ8_rzM/maxresdefault.webp", + "title": "Slaying OOMs - Mark Saroufim & Jane Xu, Meta", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=xzBcBJ8_rzM" + } + ] +} diff --git a/pytorchconf-2024/videos/sponsored-keynote-accelerating-ai-how-amd-and-pytorch-drive-innovation-with-sea-anush-elangovan.json b/pytorchconf-2024/videos/sponsored-keynote-accelerating-ai-how-amd-and-pytorch-drive-innovation-with-sea-anush-elangovan.json new file mode 100644 index 000000000..14c7990a4 --- /dev/null +++ b/pytorchconf-2024/videos/sponsored-keynote-accelerating-ai-how-amd-and-pytorch-drive-innovation-with-sea-anush-elangovan.json @@ -0,0 +1,24 @@ +{ + "description": "Sponsored Keynote: Accelerating AI: How AMD and PyTorch Drive Innovation with Seamless Day-0 Support and High Performance - Anush Elangovan, CVP Software Development, AMD\n\nIn this keynote presentation, we explore the robust collaboration between AMD and PyTorch that is propelling advancements in artificial intelligence and machine learning. Discover how AMD's commitment to Day-0 PyTorch support ensures that PyTorch users benefit from cutting-edge performance enhancements and out-of-the-box compatibility. We delve into the technical synergies that make AMD hardware an ideal choice for PyTorch frameworks, showcasing real-world examples of accelerated workflows and breakthrough AI applications. Join us to learn how this dynamic partnership is enabling researchers, developers, and data scientists to push the boundaries of innovation and achieve unprecedented results in their AI projects.", + "duration": 343, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/paR73OPcdNw/maxresdefault.webp", + "title": "Sponsored Keynote: Accelerating AI: How AMD and PyTorch Drive Innovation with Sea... Anush Elangovan", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=paR73OPcdNw" + } + ] +} diff --git a/pytorchconf-2024/videos/sponsored-keynote-enabling-ai-everywhere-with-pytorch-and-intel-kismat-singh-intel.json b/pytorchconf-2024/videos/sponsored-keynote-enabling-ai-everywhere-with-pytorch-and-intel-kismat-singh-intel.json new file mode 100644 index 000000000..739653fab --- /dev/null +++ b/pytorchconf-2024/videos/sponsored-keynote-enabling-ai-everywhere-with-pytorch-and-intel-kismat-singh-intel.json @@ -0,0 +1,24 @@ +{ + "description": "Sponsored Keynote: Enabling AI Everywhere with PyTorch and Intel - Kismat Singh,VP of Engineering for AI Frameworks, Intel\n\nUnlocking the availability of and access to generative AI technologies has great societal value. In this keynote, Kismat Singh will present how open software built on industry-standard frameworks such as PyTorch, and ubiquitous hardware from Intel that forms a large part of the current installed base across edge, PC and cloud are keys to democratizing AI and allowing new solutions to be implemented across industries ranging from healthcare, telecommunication, industrial and more. Kismat will share his thoughts on how software acceleration, flexibility and security are important factors in deploying AI applications in production and what he sees as challenges with those projects. He will also discuss Open Platform for Enterprise AI (OPEA), a new Linux Foundation AI and Data project that gives developers access to open source, standardized, modular, and heterogenous retrieval-augmented generation (RAG) pipelines that they can use for their enterprise-grade Generative AI deployments. Lastly, he will share some exciting Intel contributed features recently upstreamed into PyTorch. He will end the keynote by stating what he believes to be the future of AI and the part each of us will play in it!", + "duration": 358, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/tGEtpXaoXuk/maxresdefault.webp", + "title": "Sponsored Keynote: Enabling AI Everywhere with PyTorch and Intel - Kismat Singh, Intel", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=tGEtpXaoXuk" + } + ] +} diff --git a/pytorchconf-2024/videos/sponsored-keynote-from-containers-to-cognition-conducting-the-ai-orchestra-taylor-dolezal.json b/pytorchconf-2024/videos/sponsored-keynote-from-containers-to-cognition-conducting-the-ai-orchestra-taylor-dolezal.json new file mode 100644 index 000000000..26ea06f66 --- /dev/null +++ b/pytorchconf-2024/videos/sponsored-keynote-from-containers-to-cognition-conducting-the-ai-orchestra-taylor-dolezal.json @@ -0,0 +1,24 @@ +{ + "description": "Sponsored Keynote: From Containers to Cognition: Conducting the AI Orchestra - Taylor Dolezal, Head of Ecosystem, Cloud Native Computing Foundation\n\nLet's explore the powerful harmony created when the CNCF and PyTorch communities join forces. This keynote highlights how the collaboration between cloud native experts and AI innovators is orchestrating a new era of technological symphonies. We'll touch on critical initiatives and shared victories that demonstrate the strength of this partnership. To illustrate the creative potential of this alliance, we'll briefly showcase a demo of how containerized workloads can produce unexpected melodies. Join us for this exploration of community-driven innovation, where containers and cognition come together to compose the future of technology.", + "duration": 306, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/MWmOeXI17Kg/maxresdefault.webp", + "title": "Sponsored Keynote: From Containers to Cognition: Conducting the AI Orchestra - Taylor Dolezal", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=MWmOeXI17Kg" + } + ] +} diff --git a/pytorchconf-2024/videos/sponsored-keynote-optimizing-ai-inference-for-large-language-models-mudhakar-srivatsa-ibm.json b/pytorchconf-2024/videos/sponsored-keynote-optimizing-ai-inference-for-large-language-models-mudhakar-srivatsa-ibm.json new file mode 100644 index 000000000..9ec26daed --- /dev/null +++ b/pytorchconf-2024/videos/sponsored-keynote-optimizing-ai-inference-for-large-language-models-mudhakar-srivatsa-ibm.json @@ -0,0 +1,24 @@ +{ + "description": "Sponsored Keynote: Optimizing AI Inference for Large Language Models - Mudhakar Srivatsa, Distinguished Engineer, IBM\n\nThis talk will cover two new ways IBM has optimized generative AI inferencing with PyTorch: speculative decoding and Triton kernel development. Speculative decoding leverages predictive modeling to reduce latency by anticipating potential outputs, streamlining the inference process without sacrificing accuracy. IBM Research's team developed new speculative architectures and open sourced speculators for LLama3 models. It will also discuss various Triton kernels to accelerate inference, one of which was contributed to vLLM for accelerating MoE models. Finally, it will share a glimpse of IBM's AI hardware work, including how the IBM Artificial Intelligence Unit (AIU) could integrate into the PyTorch stack.", + "duration": 349, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/ZpNyKs4EpVI/maxresdefault.webp", + "title": "Sponsored Keynote: Optimizing AI Inference for Large Language Models - Mudhakar Srivatsa, IBM", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=ZpNyKs4EpVI" + } + ] +} diff --git a/pytorchconf-2024/videos/sponsored-keynote-the-lightning-ai-oss-stack-for-accelerating-the-ai-lifecycle-luca-antiga.json b/pytorchconf-2024/videos/sponsored-keynote-the-lightning-ai-oss-stack-for-accelerating-the-ai-lifecycle-luca-antiga.json new file mode 100644 index 000000000..86df77233 --- /dev/null +++ b/pytorchconf-2024/videos/sponsored-keynote-the-lightning-ai-oss-stack-for-accelerating-the-ai-lifecycle-luca-antiga.json @@ -0,0 +1,24 @@ +{ + "description": "Sponsored Keynote: The Lightning AI OSS Stack for Accelerating the AI Lifecycle - Luca Antiga, CTO, Lightning AI\n\nWe introduce the Lightning AI open source stack, a high-performance stack for training, fine-tuning, and deploying AI systems that augments the PyTorch ecosystem.\n\nToday PyTorch Lightning powers training workloads across the industry, from small-scale research to large-scale training endeavors. The package has reached 130M total downloads in June 2024, 2x since early 2023. PyTorch Lightning 2.4 features support for 2D parallelism via DTensors, first introduced in PyTorch 2.3.\n\nThe open source stack is completed by Fabric (lightweight building blocks for scaling training workloads), LitGPT (library for pre-training, fine-tuning, serving LLMs), LitData (parallel data processing and streaming data loading), LitServe (lightweight, high-performance serving framework), TorchMetrics (de-facto standard in deep learning metrics), and the recently released Thunder compiler. Together, these packages provide a low-friction, high-performance stack to democratize and accelerate the AI lifecycle.\n\nThe stack is optimized to run on Lightning Studios, a PyTorch native, fully integrated AI development environment on the cloud.", + "duration": 368, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/jAtrviDGpwE/maxresdefault.webp", + "title": "Sponsored Keynote: The Lightning AI OSS Stack for Accelerating the AI Lifecycle - Luca Antiga", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=jAtrviDGpwE" + } + ] +} diff --git a/pytorchconf-2024/videos/sponsored-session-accelerating-ai-innovation-high-performance-pyt-robert-suderman-ian-nordeng.json b/pytorchconf-2024/videos/sponsored-session-accelerating-ai-innovation-high-performance-pyt-robert-suderman-ian-nordeng.json new file mode 100644 index 000000000..803de28c1 --- /dev/null +++ b/pytorchconf-2024/videos/sponsored-session-accelerating-ai-innovation-high-performance-pyt-robert-suderman-ian-nordeng.json @@ -0,0 +1,24 @@ +{ + "description": "Sponsored Session: Accelerating AI Innovation: High Performance PyTorch at AMD - Robert Suderman & Ian Nordeng, AMD\n\nExplore the powerful collaboration between AMD and PyTorch, driving advancements in AI and machine learning. Learn how AMD\u2019s Day-0 PyTorch support delivers cutting-edge performance and seamless compatibility.\n\nThis session will highlight the technical synergies that make AMD hardware ideal choice for PyTorch frameworks, with real-world examples of accelerated workflows and breakthrough AI applications. Attendees will gain insights into how this dynamic partnership is enabling researchers, developers, and data scientists to push the boundaries of innovation and achieve unprecedented results in AI projects.", + "duration": 1710, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/h0LG3JbA8kc/maxresdefault.webp", + "title": "Sponsored Session: Accelerating AI Innovation: High Performance PyT... Robert Suderman & Ian Nordeng", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=h0LG3JbA8kc" + } + ] +} diff --git a/pytorchconf-2024/videos/sponsored-session-democratizing-ai-powering-the-future-with-arms-global-comp-gian-marco-iodice.json b/pytorchconf-2024/videos/sponsored-session-democratizing-ai-powering-the-future-with-arms-global-comp-gian-marco-iodice.json new file mode 100644 index 000000000..390c7aba5 --- /dev/null +++ b/pytorchconf-2024/videos/sponsored-session-democratizing-ai-powering-the-future-with-arms-global-comp-gian-marco-iodice.json @@ -0,0 +1,24 @@ +{ + "description": "Sponsored Session: Democratizing AI: Powering the Future with Arm\u2019s Global Compute Ecosystem - Gian Marco Iodice, Arm\n\nArm is excited to be at the center of the world's largest compute ecosystem at the dawn of the AI era. A key tenant of our mission is to democratize AI capabilities, empowering millions of developers to put advanced AI features into the hands of billions of users.\n\nIn this presentation, we'll explore how Arm is enabling the world\u2019s leading open-source AI frameworks to leverage power-efficient Arm-based computing platforms and Arm architecture features, as a tool for enabling fast and secure AI workloads. The session focuses on how our strategic partnership with the Pytorch and Executorch community is enabling a seamless and transparent developer experience, to run workloads everywhere from cloud to edge. This session will highlight some of our optimized libraries, upstreamed contributions and a wealth of AI-related developer material to build the future of AI on Arm.", + "duration": 1823, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/0D_SAurBA0s/maxresdefault.webp", + "title": "Sponsored Session: Democratizing AI: Powering the Future with Arm\u2019s Global Comp... Gian Marco Iodice", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=0D_SAurBA0s" + } + ] +} diff --git a/pytorchconf-2024/videos/sponsored-session-nemo-aligner-a-scalable-toolkit-for-model-alignment-gerald-shen-jimmy-zhang.json b/pytorchconf-2024/videos/sponsored-session-nemo-aligner-a-scalable-toolkit-for-model-alignment-gerald-shen-jimmy-zhang.json new file mode 100644 index 000000000..0a5c1d26e --- /dev/null +++ b/pytorchconf-2024/videos/sponsored-session-nemo-aligner-a-scalable-toolkit-for-model-alignment-gerald-shen-jimmy-zhang.json @@ -0,0 +1,24 @@ +{ + "description": "Sponsored Session: NeMo-Aligner: A Scalable Toolkit for Model Alignment - Gerald Shen & Jimmy Zhang, NVIDIA\n\nAligning AI models with human values and preferences is essential for making them safe and helpful. However, building an efficient and scalable toolkit for alignment can be challenging, especially when applied to state of the art foundation models with billions or trillions of parameters. NeMo-Aligner is an open-source, optimized and scalable toolkit that implements alignment algorithms such as Reinforcement Learning from Human Feedback (RLHF), Direct Preference Optimization (DPO), SteerLM and Self-Play Fine Tuning (SPIN). This talk will introduce NeMo-Aligner and show the steps we took to design and optimize the toolkit around various alignment algorithms. In particular, we discuss the RLHF implementation where we observe close to 7x speedup and excellent scaling performance by adding TRT-LLM integration, carefully orchestrating communication and utilizing fast training kernels. We\u2019re able to align state-of-the-art open source models with NeMo-Aligner and hope our framework can enable the community to performantly customize, fine-tune and align foundational models at any scale.", + "duration": 1275, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/rROnWTNF-_A/maxresdefault.webp", + "title": "Sponsored Session: NeMo-Aligner: A Scalable Toolkit for Model Alignment - Gerald Shen & Jimmy Zhang", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=rROnWTNF-_A" + } + ] +} diff --git a/pytorchconf-2024/videos/sponsored-session-pytorch-support-by-google-enabling-perform-mark-sherwood-shauheen-zahirazami.json b/pytorchconf-2024/videos/sponsored-session-pytorch-support-by-google-enabling-perform-mark-sherwood-shauheen-zahirazami.json new file mode 100644 index 000000000..828a755ce --- /dev/null +++ b/pytorchconf-2024/videos/sponsored-session-pytorch-support-by-google-enabling-perform-mark-sherwood-shauheen-zahirazami.json @@ -0,0 +1,24 @@ +{ + "description": "Sponsored Session: PyTorch Support by Google Enabling Performance from Cloud to Edge - Mark Sherwood & Shauheen Zahirazami, Google\n\nIn this session we will cover various ways teams at google are working to help the Pytorch community achieve performance and scale from cloud to edge. We will cover how Google Cloud customers can use PyTorch and OpenXLA to get competitive performance for their ML workloads. We\u2019ll also cover how Google AI Edge Torch works with Pytorch to help developers integrate LLMs, vision models and more to easily create new edge applications that can run on a wide set of devices.", + "duration": 1157, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/CzviSkYjEkQ/maxresdefault.webp", + "title": "Sponsored Session: PyTorch Support by Google Enabling Perform... Mark Sherwood & Shauheen Zahirazami", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=CzviSkYjEkQ" + } + ] +} diff --git a/pytorchconf-2024/videos/sponsored-session-torchchat-a-showcase-of-pytorch-llm-ubiquity-jack-khuu-jesse-white-meta.json b/pytorchconf-2024/videos/sponsored-session-torchchat-a-showcase-of-pytorch-llm-ubiquity-jack-khuu-jesse-white-meta.json new file mode 100644 index 000000000..795069270 --- /dev/null +++ b/pytorchconf-2024/videos/sponsored-session-torchchat-a-showcase-of-pytorch-llm-ubiquity-jack-khuu-jesse-white-meta.json @@ -0,0 +1,24 @@ +{ + "description": "Sponsored Session: Torchchat: A Showcase of PyTorch LLM Ubiquity - Jack Khuu & Jesse White, Meta\n\nThis talk explores the journey of enabling LLMs in the PyTorch ecosystem, as well as how the teams behind AOT Inductor, ExecuTorch, and torchao collaborated to create torchchat, a showcase of PyTorch\u2019s ability to run LLM inference everywhere.\n\nTorchchat demonstrates the ubiquity, simplicity, and quality of PyTorch\u2019s LLM support through performant, reproducible implementations for not only Python environments, but on desktop, server, and on-device as-well.\n\nAll of our work is open source and available on GitHub.", + "duration": 1459, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/hSJ9oD6tbeU/maxresdefault.webp", + "title": "Sponsored Session: Torchchat: A Showcase of PyTorch LLM Ubiquity - Jack Khuu & Jesse White, Meta", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=hSJ9oD6tbeU" + } + ] +} diff --git a/pytorchconf-2024/videos/startup-showcase.json b/pytorchconf-2024/videos/startup-showcase.json new file mode 100644 index 000000000..420b6e979 --- /dev/null +++ b/pytorchconf-2024/videos/startup-showcase.json @@ -0,0 +1,24 @@ +{ + "description": "Startup Showcase\n\nModerators: Chappy Asel\nJudges: Kevin Crosby, Astasia Myers, Rajko Radovanovic, Vijay Reddy, Vig Sachidananda, Simon Tiu\nThe PyTorch Conference Startup Showcase is giving emerging companies the chance to pitch to a panel of VCs looking to support AI/ML startups with high growth potential, and meet some of the best AI focused Engineers in the Industry. This is an exciting and unique opportunity for early-stage founders to showcase their ideas and breakthroughs, connect with leading VCs, and increase visibility in the generative AI and machine learning industry.\n\nThe winning startup will be announced at the Flare Party taking place after the Startup Showcase.\n\nFinalists:\n- Remix Inc.\n- Cartesia\n- OpenBabylon\n- Remyx AI\n- A2 Labs, Inc.\n- QuicSnap\n- Iso AI\n- CTGT\n- Creao.ai", + "duration": 3555, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/xAePG2YVz7c/maxresdefault.webp", + "title": "Startup Showcase", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=xAePG2YVz7c" + } + ] +} diff --git a/pytorchconf-2024/videos/state-of-pytorch-ji-li-damien-sereni-meta.json b/pytorchconf-2024/videos/state-of-pytorch-ji-li-damien-sereni-meta.json new file mode 100644 index 000000000..617620a5e --- /dev/null +++ b/pytorchconf-2024/videos/state-of-pytorch-ji-li-damien-sereni-meta.json @@ -0,0 +1,24 @@ +{ + "description": "State of PyTorch - Ji Li & Damien Sereni, Meta\n\nThis talk gives a run through of who builds PyTorch, new and upcoming improvements to the framework and how to get involved. All thanks to our awesome community of contributors, partners and ecosystem tools.", + "duration": 1237, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/ibgW_ali0Gc/maxresdefault.webp", + "title": "State of PyTorch - Ji Li & Damien Sereni, Meta", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=ibgW_ali0Gc" + } + ] +} diff --git a/pytorchconf-2024/videos/the-challenges-of-building-an-opinionated-open-source-llm-framework-wing-lian-axolotl-ai.json b/pytorchconf-2024/videos/the-challenges-of-building-an-opinionated-open-source-llm-framework-wing-lian-axolotl-ai.json new file mode 100644 index 000000000..1e84c623c --- /dev/null +++ b/pytorchconf-2024/videos/the-challenges-of-building-an-opinionated-open-source-llm-framework-wing-lian-axolotl-ai.json @@ -0,0 +1,24 @@ +{ + "description": "The Challenges of Building an Opinionated Open Source LLM Framework - Wing Lian, Axolotl AI\n\nAs open-source LLMs have become more capable, a substantial ecosystem has developed around the fine-tuning of these models. A thriving community of researchers, developers, practitioners and hobbyists has emerged which focuses on topics ranging from memory efficiency, parameter-efficient fine-tuning and quantization to performance at scale and reproducible evaluations. The goal of this mini-summit is to bring this community together to discuss ideas, share knowledge and build connections.\n\nThe agenda features a keynote from Joe Spisak on the state of the Llama ecosystem followed by invited talks from the founders of Axolotl, Unsloth and torchtune. We conclude the summit with a riveting discussion on what\u2019s next for LLMs, fine-tuning and the PyTorch ecosystem with a fabulous panel of experts - Tim Dettmers (author of bitsandbytes and QLoRA), Hailey Schoelkopf (maintainer of LM Eval Harness at EleutherAI), Aakanksha Chowdhery (Lead author on PaLM and Gemini) and Alexis Conneau (Research Lead at OpenAI)", + "duration": 557, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/bp9xioJv0gw/maxresdefault.webp", + "title": "The Challenges of Building an Opinionated Open Source LLM Framework - Wing Lian, Axolotl AI", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=bp9xioJv0gw" + } + ] +} diff --git a/pytorchconf-2024/videos/the-impact-and-challenges-of-open-source-generative-datasets-and-models-aaron-gokaslan.json b/pytorchconf-2024/videos/the-impact-and-challenges-of-open-source-generative-datasets-and-models-aaron-gokaslan.json new file mode 100644 index 000000000..e39d4d424 --- /dev/null +++ b/pytorchconf-2024/videos/the-impact-and-challenges-of-open-source-generative-datasets-and-models-aaron-gokaslan.json @@ -0,0 +1,24 @@ +{ + "description": "The Impact and Challenges of Open Source Generative Datasets and Models - Aaron Gokaslan, Cornell University\n\nOpen source generative models like OpenGPT2, BLOOM, and others have been pivotal in advancing AI technology. These models leverage extensive text data to achieve advanced linguistic capabilities. However, the trend towards proprietary tools and closed large language models is growing, posing unique challenges in open-source AI development. This discussion will explore the intricacies of training such models, the hurdles in dataset management, and the regulation of open-source contributions. We'll explore how to effectively iterate on collected data, prepare for extensive training sessions, and coordinate research across large open-source organizations. We will discuss the challenges of generative models in three different modalities: text, image, and genomics. The talk will draw from the speaker\u2019s personal experience on working on OpenWebText, OpenGPT2, BLOOM, CommonCanvas, Caduceus, and other generative models. We will also cover the changing AI environment and how the future of open souce is threatened by onerous regulation, ever increasing compute costs, and the commoditization of previously open data.", + "duration": 1815, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/15ECoIeBf3c/maxresdefault.webp", + "title": "The Impact and Challenges of Open Source Generative Datasets and Models - Aaron Gokaslan", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=15ECoIeBf3c" + } + ] +} diff --git a/pytorchconf-2024/videos/the-rise-of-transformers-in-the-growing-pytorch-ecosystem-arthur-zucker-hugging-face.json b/pytorchconf-2024/videos/the-rise-of-transformers-in-the-growing-pytorch-ecosystem-arthur-zucker-hugging-face.json new file mode 100644 index 000000000..4fb108cd7 --- /dev/null +++ b/pytorchconf-2024/videos/the-rise-of-transformers-in-the-growing-pytorch-ecosystem-arthur-zucker-hugging-face.json @@ -0,0 +1,24 @@ +{ + "description": "The Rise of `Transformers` in the Growing PyTorch Ecosystem - Arthur Zucker, Hugging Face\n\nExplore how the `tranformers` library grows and adapts to the fast paced and ever-changing AI field to bring the best to the AI community", + "duration": 1117, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/P-asaQVmA3o/maxresdefault.webp", + "title": "The Rise of `Transformers` in the Growing PyTorch Ecosystem - Arthur Zucker, Hugging Face", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=P-asaQVmA3o" + } + ] +} diff --git a/pytorchconf-2024/videos/the-state-of-the-llama-ecosystem-joe-spisak-meta.json b/pytorchconf-2024/videos/the-state-of-the-llama-ecosystem-joe-spisak-meta.json new file mode 100644 index 000000000..f15037eb4 --- /dev/null +++ b/pytorchconf-2024/videos/the-state-of-the-llama-ecosystem-joe-spisak-meta.json @@ -0,0 +1,24 @@ +{ + "description": "The State of the Llama Ecosystem - Joe Spisak, Meta\n\nAs open-source LLMs have become more capable, a substantial ecosystem has developed around the fine-tuning of these models. A thriving community of researchers, developers, practitioners and hobbyists has emerged which focuses on topics ranging from memory efficiency, parameter-efficient fine-tuning and quantization to performance at scale and reproducible evaluations. The goal of this mini-summit is to bring this community together to discuss ideas, share knowledge and build connections.\n\nThe agenda features a keynote from Joe Spisak on the state of the Llama ecosystem followed by invited talks from the founders of Axolotl, Unsloth and torchtune. We conclude the summit with a riveting discussion on what\u2019s next for LLMs, fine-tuning and the PyTorch ecosystem with a fabulous panel of experts - Tim Dettmers (author of bitsandbytes and QLoRA), Hailey Schoelkopf (maintainer of LM Eval Harness at EleutherAI), Aakanksha Chowdhery (Lead author on PaLM and Gemini) and Alexis Conneau (Research Lead at OpenAI)", + "duration": 1044, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/9_gmwWCdAlU/maxresdefault.webp", + "title": "The State of the Llama Ecosystem - Joe Spisak, Meta", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=9_gmwWCdAlU" + } + ] +} diff --git a/pytorchconf-2024/videos/together-goes-brrr-threading-research-production-with-torch-compile-pragaash-ponnusamy.json b/pytorchconf-2024/videos/together-goes-brrr-threading-research-production-with-torch-compile-pragaash-ponnusamy.json new file mode 100644 index 000000000..37d5c78d1 --- /dev/null +++ b/pytorchconf-2024/videos/together-goes-brrr-threading-research-production-with-torch-compile-pragaash-ponnusamy.json @@ -0,0 +1,24 @@ +{ + "description": "Together Goes Brrr: Threading Research & Production with Torch Compile - Pragaash Ponnusamy, together.ai\n\nThe deployment of large language models for inference at scale is inherently complex, often requiring intricate optimizations across compute-bound and memory-bound regimes. This talk explores how PyTorch's torch.compile has revolutionized the optimization landscape for LLM serving at Together AI. Through its sophisticated Dynamo tracer and Inductor backend, torch.compile has transformed the approach to critical performance bottlenecks in both prefill and decode phases of inference. We examine how automatic vertical fusion, epilogue optimization, and adaptive kernel generation across batch sizes for GEMV and GEMM workloads, addressing key efficiency concerns, from CUDA graph captures and optimized all-reduce strategies to custom kernel registrations. The presentation highlights Together AI's journey in leveraging torch.compile to streamline the transition from research to production, significantly simplifying the deployment process for even custom architectures. By automating many performance-critical optimizations, torch.compile has not only enhanced inference efficiency but also democratized high-performance LLM deployment. We'll conclude by sharing key lessons learned and best practices gleaned from Together AI's experience in deploying torch.compile to production, serving billions of user queries and navigating the complexities of large-scale LLM inference.", + "duration": 566, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/zn0Pm2Pv3O0/maxresdefault.webp", + "title": "Together Goes Brrr: Threading Research & Production with Torch Compile - Pragaash Ponnusamy", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=zn0Pm2Pv3O0" + } + ] +} diff --git a/pytorchconf-2024/videos/torch-compile-for-autograd-ddp-and-fsdp-will-feng-chien-chin-huang-simon-fan-meta.json b/pytorchconf-2024/videos/torch-compile-for-autograd-ddp-and-fsdp-will-feng-chien-chin-huang-simon-fan-meta.json new file mode 100644 index 000000000..66cb0a5f8 --- /dev/null +++ b/pytorchconf-2024/videos/torch-compile-for-autograd-ddp-and-fsdp-will-feng-chien-chin-huang-simon-fan-meta.json @@ -0,0 +1,24 @@ +{ + "description": "Torch.Compile for Autograd, DDP and FSDP - Will Feng , Chien-Chin Huang & Simon Fan, Meta\n\nIn this talk, we will present the latest advancements in torch.compile for distributed training via DDP and FSDP. We will first introduce Compiled Autograd, a torch.compile mode to fully capture the backpropagation step, including the communication collective operators used in distributed. We will then cover the improvements this new approach brought to Compiled DDP/FSDP, notably by removing DDP/FSDP graph breaks which brings the potential of improving compute/communication overlap.", + "duration": 1343, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/1XuibaVRewc/maxresdefault.webp", + "title": "Torch.Compile for Autograd, DDP and FSDP - Will Feng , Chien-Chin Huang & Simon Fan, Meta", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=1XuibaVRewc" + } + ] +} diff --git a/pytorchconf-2024/videos/torchinductor-cpu-backend-advancements-new-features-and-performance-imp-jiong-gong-leslie-fang.json b/pytorchconf-2024/videos/torchinductor-cpu-backend-advancements-new-features-and-performance-imp-jiong-gong-leslie-fang.json new file mode 100644 index 000000000..b344f7367 --- /dev/null +++ b/pytorchconf-2024/videos/torchinductor-cpu-backend-advancements-new-features-and-performance-imp-jiong-gong-leslie-fang.json @@ -0,0 +1,24 @@ +{ + "description": "TorchInductor CPU Backend Advancements: New Features and Performance Improvements - Jiong Gong & Leslie Fang, Intel\n\nThis presentation provides an update on the latest advancements in the TorchInductor CPU backend since the last conference to bring best-in-class CPU performance for broad DL workloads. We will discuss new features and performance enhancements, including: \u2022 Max-autotune support with codegen for GEMMs, boosting performance for GEMM-related operations \u2022 Enhanced vectorized codegen support, now covering all data types beyond floating points with flexible vector factors, and optimized loop scheduling \u2022 Comprehensive quantization support, including weight-only-quantization (WoQ), and optimizations for dynamic quantization and quantization-aware training \u2022 Improved Attention support, featuring attention masks and optimizating SoftMax via flash attention v2 etc. \u2022 AOTInductor support, enabling high-performance inference with frozen weights \u2022 Native Windows support, with improved vectorization capabilities These advancements, combined with ongoing optimizations, have resulted in significant performance improvements since PyTorch 2.1, demonstrated through extensive benchmarks and large language models (LLMs).", + "duration": 1312, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/VwmOgzXtxYw/maxresdefault.webp", + "title": "TorchInductor CPU Backend Advancements: New Features and Performance Imp... Jiong Gong & Leslie Fang", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=VwmOgzXtxYw" + } + ] +} diff --git a/pytorchconf-2024/videos/torchtitan-large-scale-llm-training-using-native-pytorch-3d-parallel-wanchao-liang-linsong-chu.json b/pytorchconf-2024/videos/torchtitan-large-scale-llm-training-using-native-pytorch-3d-parallel-wanchao-liang-linsong-chu.json new file mode 100644 index 000000000..3d49f2774 --- /dev/null +++ b/pytorchconf-2024/videos/torchtitan-large-scale-llm-training-using-native-pytorch-3d-parallel-wanchao-liang-linsong-chu.json @@ -0,0 +1,24 @@ +{ + "description": "Torchtitan: Large-Scale LLM Training Using Native PyTorch 3D Parallelism - Wanchao Liang, Meta & Linsong Chu, IBM Research\n\ntorchtitan is a proof-of-concept for Large-scale LLM training using native PyTorch. It is a repo that showcases PyTorch's latest distributed training features in a clean, minimal codebase. We show-cased end to end large scale training features enablement: 1. 3D/4D Parallelism 2. Efficient distributed checkpoint save/load/resharding 3. Many efficient training techniques including Float8, torch.compile, activation checkpoint, etc.", + "duration": 1460, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/WsNEBxPDljU/maxresdefault.webp", + "title": "Torchtitan: Large-Scale LLM Training Using Native PyTorch 3D Parallel... Wanchao Liang & Linsong Chu", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=WsNEBxPDljU" + } + ] +} diff --git a/pytorchconf-2024/videos/torchtune-easy-and-accessible-finetuning-in-native-pytorch-evan-smothers-meta.json b/pytorchconf-2024/videos/torchtune-easy-and-accessible-finetuning-in-native-pytorch-evan-smothers-meta.json new file mode 100644 index 000000000..d90d2bd5b --- /dev/null +++ b/pytorchconf-2024/videos/torchtune-easy-and-accessible-finetuning-in-native-pytorch-evan-smothers-meta.json @@ -0,0 +1,24 @@ +{ + "description": "torchtune: Easy and Accessible Finetuning in Native PyTorch - Evan Smothers, Meta\n\nAs open-source LLMs have become more capable, a substantial ecosystem has developed around the fine-tuning of these models. A thriving community of researchers, developers, practitioners and hobbyists has emerged which focuses on topics ranging from memory efficiency, parameter-efficient fine-tuning and quantization to performance at scale and reproducible evaluations. The goal of this mini-summit is to bring this community together to discuss ideas, share knowledge and build connections.\n\nThe agenda features a keynote from Joe Spisak on the state of the Llama ecosystem followed by invited talks from the founders of Axolotl, Unsloth and torchtune. We conclude the summit with a riveting discussion on what\u2019s next for LLMs, fine-tuning and the PyTorch ecosystem with a fabulous panel of experts - Tim Dettmers (author of bitsandbytes and QLoRA), Hailey Schoelkopf (maintainer of LM Eval Harness at EleutherAI), Aakanksha Chowdhery (Lead author on PaLM and Gemini) and Alexis Conneau (Research Lead at OpenAI)", + "duration": 958, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/43X9E25-Qg0/maxresdefault.webp", + "title": "torchtune: Easy and Accessible Finetuning in Native PyTorch - Evan Smothers, Meta", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=43X9E25-Qg0" + } + ] +} diff --git a/pytorchconf-2024/videos/training-moes-at-scale-with-pytorch-mihir-patel-brian-chu-databricks.json b/pytorchconf-2024/videos/training-moes-at-scale-with-pytorch-mihir-patel-brian-chu-databricks.json new file mode 100644 index 000000000..fa02c680a --- /dev/null +++ b/pytorchconf-2024/videos/training-moes-at-scale-with-pytorch-mihir-patel-brian-chu-databricks.json @@ -0,0 +1,24 @@ +{ + "description": "Training MoEs at Scale with PyTorch - Mihir Patel & Brian Chu, Databricks\n\nMixture-of-Experts MoE (models) are becoming an increasingly popular architecture choice for large language models (LLMs). In this talk, we describe how to train MoE models with PyTorch. After discussing various performance tradeoffs, we use PyTorch distributed tools like DTensor to build custom parallelism approaches, including expert parallelism via MegaBlocks. We then show how to get near linear scaling to thousands of GPUs, combining PyTorch FSDP and HSDP with our parallelism strategies. We discuss many of the challenges of training at scale, including communication bottlenecks, hardware failures, and networking challenges. We further improve training at scale setups using tools like PyTorch Distributed Checkpointing for rapid saving and loading. We then highlight further optimizations to minimize challenges only present at scale, such as object store failures for large checkpoints.", + "duration": 1132, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/f2OxAWRCmPo/maxresdefault.webp", + "title": "Training MoEs at Scale with PyTorch - Mihir Patel & Brian Chu, Databricks", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=f2OxAWRCmPo" + } + ] +} diff --git a/pytorchconf-2024/videos/triton-maximizing-kernel-development-productivity-under-performance-constraints-philip-tillet.json b/pytorchconf-2024/videos/triton-maximizing-kernel-development-productivity-under-performance-constraints-philip-tillet.json new file mode 100644 index 000000000..8c8316742 --- /dev/null +++ b/pytorchconf-2024/videos/triton-maximizing-kernel-development-productivity-under-performance-constraints-philip-tillet.json @@ -0,0 +1,24 @@ +{ + "description": "[TRITON] Maximizing Kernel Development Productivity Under Performance Constraints - Philip Tillet, OpenAI\n\nMachine Learning research workflows are often bottlenecked by the development of compute kernels for new algorithms and GPU architectures. This process can be daunting, and often requires a careful trade-off between productivity and performance. In this talk, we will discuss how Triton -- a mid-level programming language for kernel development -- approaches this multi-objective optimization problem, and the design decisions that were made to that effect.", + "duration": 601, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/h449o7UijHk/maxresdefault.webp", + "title": "[TRITON] Maximizing Kernel Development Productivity Under Performance Constraints - Philip Tillet", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=h449o7UijHk" + } + ] +} diff --git a/pytorchconf-2024/videos/tvm-universally-deploy-large-language-models-via-ml-compilation-tianqi-chen-cmu-octoai.json b/pytorchconf-2024/videos/tvm-universally-deploy-large-language-models-via-ml-compilation-tianqi-chen-cmu-octoai.json new file mode 100644 index 000000000..4d110be30 --- /dev/null +++ b/pytorchconf-2024/videos/tvm-universally-deploy-large-language-models-via-ml-compilation-tianqi-chen-cmu-octoai.json @@ -0,0 +1,24 @@ +{ + "description": "[TVM] Universally Deploy Large-language Models via ML Compilation - Tianqi Chen, CMU & OctoAI\n\nDeploying deep learning models on various devices has become an important topic. Machine learning compilation is an emerging field that leverages compiler and automatic search techniques to accelerate AI models. ML compilation brings a unique set of challenges: emerging machine learning models; increasing hardware specialization brings a diverse set of acceleration primitives; growing tension between flexibility and performance. In this talk. I then discuss our experience in bringing foundational models to a variety of devices and hardware environments through machine learning compilation.", + "duration": 650, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/n66LLI_ebx0/maxresdefault.webp", + "title": "[TVM] Universally Deploy Large-language Models via ML Compilation - Tianqi Chen, CMU & OctoAI", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=n66LLI_ebx0" + } + ] +} diff --git a/pytorchconf-2024/videos/understanding-the-llm-inference-workload-mark-moyou-nvidia.json b/pytorchconf-2024/videos/understanding-the-llm-inference-workload-mark-moyou-nvidia.json new file mode 100644 index 000000000..9aba442c0 --- /dev/null +++ b/pytorchconf-2024/videos/understanding-the-llm-inference-workload-mark-moyou-nvidia.json @@ -0,0 +1,24 @@ +{ + "description": "Understanding the LLM Inference Workload - Mark Moyou, NVIDIA\n\nUnderstanding how to effectively size a production grade LLM deployment requires understanding of the model(s), the compute hardware, quantization and parallelization methods, KV Cache budgets, input and output token length predictions, model adapter management and much more. - Why LLM inference is different to standard deep learning inference - Current and future NVIDIA GPU overview - which GPU(s) for which models and why - Understanding the importance of building inference engines - Deep recap on the attention mechanism along with different types of popular attention mechanisms used in production - Deep dive on KV Cache and managing KV Cache budgets - Parallelism (reducing latency) - mainly tensor parallelism, but data, sequence, pipeline, and expert parallelism will be highlighted - Quantization methods on weights, activations, and KV Cache to reduce engine sizes for more effective GPU utilization - Increasing throughput with inflight batching and other techniques - Detailed performance analysis of LLM deployments looking at Time to first token, inter-token latencies, llm deployment characterizations, and more that can help reduce deployment costs", + "duration": 2054, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/z2M8gKGYws4/maxresdefault.webp", + "title": "Understanding the LLM Inference Workload - Mark Moyou, NVIDIA", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=z2M8gKGYws4" + } + ] +} diff --git a/pytorchconf-2024/videos/unlocking-the-enigma-crafting-unbiased-transparent-and-explainable-large-languag-rashmi-nagpal.json b/pytorchconf-2024/videos/unlocking-the-enigma-crafting-unbiased-transparent-and-explainable-large-languag-rashmi-nagpal.json new file mode 100644 index 000000000..975f79ddf --- /dev/null +++ b/pytorchconf-2024/videos/unlocking-the-enigma-crafting-unbiased-transparent-and-explainable-large-languag-rashmi-nagpal.json @@ -0,0 +1,24 @@ +{ + "description": "Unlocking the Enigma: Crafting Unbiased, Transparent, and Explainable Large Language Models - Rashmi Nagpal, Patchstack\n\nIn an era where artificial intelligence reigns supreme, the statistics are both perplexing and thought-provoking \u2013 only a mere 13% of large language models manage to transcend the realms of research and enter the practical world of production. Who bears the responsibility when these models err, spewing out biased or discriminatory outputs? It's time to demystify the complex landscape of machine learning ethics and carve a path towards a brighter, more accountable future! In this talk, firstly, we will navigate the profound impacts of large language models across diverse domains, from the lifesaving advances in medicine to safeguarding our nations through enhanced security protocols. Secondly, as we marvel at data-driven decisions laid by these models, we will confront the darker shadows cast by \u2013 the looming spectre of bias in the data. Finally, we will delve deep into the art of building interpretable models and navigating the maze of ethical considerations. Through a live demonstration in PyTorch, we will witness how to craft unbiased, transparent, and explainable models.", + "duration": 2998, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/vGHb4jO2_Ks/maxresdefault.webp", + "title": "Unlocking the Enigma: Crafting Unbiased, Transparent, and Explainable Large Languag... Rashmi Nagpal", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=vGHb4jO2_Ks" + } + ] +} diff --git a/pytorchconf-2024/videos/vllm-easy-fast-and-cheap-llm-serving-for-everyone-woosuk-kwon-xiaoxuan-liu-uc-berkeley.json b/pytorchconf-2024/videos/vllm-easy-fast-and-cheap-llm-serving-for-everyone-woosuk-kwon-xiaoxuan-liu-uc-berkeley.json new file mode 100644 index 000000000..307847e02 --- /dev/null +++ b/pytorchconf-2024/videos/vllm-easy-fast-and-cheap-llm-serving-for-everyone-woosuk-kwon-xiaoxuan-liu-uc-berkeley.json @@ -0,0 +1,24 @@ +{ + "description": "vLLM: Easy, Fast, and Cheap LLM Serving for Everyone - Woosuk Kwon & Xiaoxuan Liu, UC Berkeley\n\nWe will present vLLM, an open-source high-performance LLM inference engine built on top of PyTorch. Starting as a research project at UC Berkeley, vLLM has been one of the fastest and most popular LLM inference solutions in industry, reaching 20K+ stars and 350+ contributors. In this talk, we will cover how vLLM adopts various LLM inference optimizations and how it supports various AI accelerators such as AMD GPUs, Google TPUs, and AWS Inferentia. Also, we will discuss how vLLM benefits from PyTorch 2 and its ecosystem.", + "duration": 1412, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/9ih0EmcXRHE/maxresdefault.webp", + "title": "vLLM: Easy, Fast, and Cheap LLM Serving for Everyone - Woosuk Kwon & Xiaoxuan Liu, UC Berkeley", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=9ih0EmcXRHE" + } + ] +} diff --git a/pytorchconf-2024/videos/welcome-to-the-pytorch-ecosystem-for-llm-fine-tuning-mini-summit-kartikay-khandelwal-meta.json b/pytorchconf-2024/videos/welcome-to-the-pytorch-ecosystem-for-llm-fine-tuning-mini-summit-kartikay-khandelwal-meta.json new file mode 100644 index 000000000..378111517 --- /dev/null +++ b/pytorchconf-2024/videos/welcome-to-the-pytorch-ecosystem-for-llm-fine-tuning-mini-summit-kartikay-khandelwal-meta.json @@ -0,0 +1,24 @@ +{ + "description": "Welcome to the PyTorch Ecosystem for LLM Fine-tuning Mini Summit - Kartikay Khandelwal, Meta\n\nAs open-source LLMs have become more capable, a substantial ecosystem has developed around the fine-tuning of these models. A thriving community of researchers, developers, practitioners and hobbyists has emerged which focuses on topics ranging from memory efficiency, parameter-efficient fine-tuning and quantization to performance at scale and reproducible evaluations. The goal of this mini-summit is to bring this community together to discuss ideas, share knowledge and build connections.\n\nThe agenda features a keynote from Joe Spisak on the state of the Llama ecosystem followed by invited talks from the founders of Axolotl, Unsloth and torchtune. We conclude the summit with a riveting discussion on what\u2019s next for LLMs, fine-tuning and the PyTorch ecosystem with a fabulous panel of experts - Tim Dettmers (author of bitsandbytes and QLoRA), Hailey Schoelkopf (maintainer of LM Eval Harness at EleutherAI), Aakanksha Chowdhery (Lead author on PaLM and Gemini) and Alexis Conneau (Research Lead at OpenAI)", + "duration": 81, + "language": "eng", + "recorded": "2024-09-18", + "related_urls": [ + { + "label": "Conference Website", + "url": "https://pytorch.org/event/pytorch-conference-2024/" + } + ], + "speakers": [ + "TODO" + ], + "tags": [], + "thumbnail_url": "https://i.ytimg.com/vi_webp/Pe_VT5ReB3U/maxresdefault.webp", + "title": "Welcome to the PyTorch Ecosystem for LLM Fine-tuning Mini Summit - Kartikay Khandelwal, Meta", + "videos": [ + { + "type": "youtube", + "url": "https://www.youtube.com/watch?v=Pe_VT5ReB3U" + } + ] +} From 77f0948bc48c527d97fe47dbef3659ac9bdf4c81 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Ezequiel=20Leonardo=20Casta=C3=B1o?= <14986783+ELC@users.noreply.github.com> Date: Fri, 21 Nov 2025 19:30:41 -0300 Subject: [PATCH 2/8] Sample: process pytorchconf-2024 first file - extract speakers and clean titles --- ...large-scale-pretraining-davis-wertheimer-linsong-chu.json | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-) diff --git a/pytorchconf-2024/videos/a-distributed-stateful-dataloader-for-large-scale-pretraining-davis-wertheimer-linsong-chu.json b/pytorchconf-2024/videos/a-distributed-stateful-dataloader-for-large-scale-pretraining-davis-wertheimer-linsong-chu.json index 5da50199d..2caaa1b76 100644 --- a/pytorchconf-2024/videos/a-distributed-stateful-dataloader-for-large-scale-pretraining-davis-wertheimer-linsong-chu.json +++ b/pytorchconf-2024/videos/a-distributed-stateful-dataloader-for-large-scale-pretraining-davis-wertheimer-linsong-chu.json @@ -10,11 +10,12 @@ } ], "speakers": [ - "TODO" + "Davis Wertheimer", + "Linsong Chu" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/VtT4rdph4Qs/maxresdefault.webp", - "title": "A Distributed Stateful Dataloader for Large-Scale Pretraining - Davis Wertheimer & Linsong Chu", + "title": "A Distributed Stateful Dataloader for Large-Scale Pretraining", "videos": [ { "type": "youtube", From 7dd74fa07f8f627b65ea0e2ae05cd644ce1abb24 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Ezequiel=20Leonardo=20Casta=C3=B1o?= <14986783+ELC@users.noreply.github.com> Date: Fri, 21 Nov 2025 19:34:30 -0300 Subject: [PATCH 3/8] Process pytorchconf-2024 files (20/88): extract speakers and clean titles --- ...d-video-data-loading-andrew-ho-ahmad-sharif-meta.json | 5 +++-- ...-for-100-skin-shades-emmanuel-acheampong-robomua.json | 4 ++-- ...are-with-the-pytorch-ecosystem-bharath-ramsundar.json | 4 ++-- .../data-dependent-shapes-in-pt2-edward-yang-meta.json | 4 ++-- ...illet-j-ansel-j-pienaar-t-chen-m-zolotukhin-p-wu.json | 9 +++++++-- ...i-support-mergen-nachin-mengtao-martin-yuan-meta.json | 5 +++-- ...o-make-llm-training-faster-daniel-han-unsloth-ai.json | 4 ++-- ...alide-backend-for-torchinductor-jason-ansel-meta.json | 4 ++-- ...-a-case-study-maanav-dalal-yulong-wang-microsoft.json | 5 +++-- ...oices-and-enhancing-back-eikan-wang-min-jean-cho.json | 5 +++-- ...ge-assistant-jerry-liu-co-founder-ceo-llamaindex.json | 4 ++-- pytorchconf-2024/videos/keynote-community-awards.json | 4 +--- ...-the-edge-cormac-brick-principal-engineer-google.json | 4 ++-- ...-timeline-of-llms-sebastian-raschka-lightning-ai.json | 4 ++-- 14 files changed, 36 insertions(+), 29 deletions(-) diff --git a/pytorchconf-2024/videos/blobs-to-clips-efficient-end-to-end-video-data-loading-andrew-ho-ahmad-sharif-meta.json b/pytorchconf-2024/videos/blobs-to-clips-efficient-end-to-end-video-data-loading-andrew-ho-ahmad-sharif-meta.json index 127dadba2..9af20c723 100644 --- a/pytorchconf-2024/videos/blobs-to-clips-efficient-end-to-end-video-data-loading-andrew-ho-ahmad-sharif-meta.json +++ b/pytorchconf-2024/videos/blobs-to-clips-efficient-end-to-end-video-data-loading-andrew-ho-ahmad-sharif-meta.json @@ -10,11 +10,12 @@ } ], "speakers": [ - "TODO" + "Andrew Ho", + "Ahmad Sharif" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/-cBplHNM4RA/maxresdefault.webp", - "title": "Blobs to Clips: Efficient End-to-End Video Data Loading - Andrew Ho & Ahmad Sharif, Meta", + "title": "Blobs to Clips: Efficient End-to-End Video Data Loading", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/building-pytorch-computer-vision-algorithms-for-100-skin-shades-emmanuel-acheampong-robomua.json b/pytorchconf-2024/videos/building-pytorch-computer-vision-algorithms-for-100-skin-shades-emmanuel-acheampong-robomua.json index 5741debbf..0db0bfc21 100644 --- a/pytorchconf-2024/videos/building-pytorch-computer-vision-algorithms-for-100-skin-shades-emmanuel-acheampong-robomua.json +++ b/pytorchconf-2024/videos/building-pytorch-computer-vision-algorithms-for-100-skin-shades-emmanuel-acheampong-robomua.json @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Emmanuel Acheampong" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/eGZ3-8ZbFj0/maxresdefault.webp", - "title": "Building PyTorch Computer Vision Algorithms for 100 Skin Shades - Emmanuel Acheampong, roboMUA", + "title": "Building PyTorch Computer Vision Algorithms for 100 Skin Shades", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/building-scientific-computing-infrastructure-software-with-the-pytorch-ecosystem-bharath-ramsundar.json b/pytorchconf-2024/videos/building-scientific-computing-infrastructure-software-with-the-pytorch-ecosystem-bharath-ramsundar.json index 46dad290b..13b5d5982 100644 --- a/pytorchconf-2024/videos/building-scientific-computing-infrastructure-software-with-the-pytorch-ecosystem-bharath-ramsundar.json +++ b/pytorchconf-2024/videos/building-scientific-computing-infrastructure-software-with-the-pytorch-ecosystem-bharath-ramsundar.json @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Bharath Ramsundar" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/Ru-CDK_79NQ/maxresdefault.webp", - "title": "Building Scientific Computing Infrastructure Software with the PyTorch Ecosystem - Bharath Ramsundar", + "title": "Building Scientific Computing Infrastructure Software with the PyTorch Ecosystem", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/data-dependent-shapes-in-pt2-edward-yang-meta.json b/pytorchconf-2024/videos/data-dependent-shapes-in-pt2-edward-yang-meta.json index 8248a9cec..0c3e776d7 100644 --- a/pytorchconf-2024/videos/data-dependent-shapes-in-pt2-edward-yang-meta.json +++ b/pytorchconf-2024/videos/data-dependent-shapes-in-pt2-edward-yang-meta.json @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Edward Yang" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/VYuFJUfyWSU/maxresdefault.webp", - "title": "Data-Dependent Shapes in PT2 - Edward Yang, Meta", + "title": "Data-Dependent Shapes in PT2", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/dl-compiler-panel-discussion-p-tillet-j-ansel-j-pienaar-t-chen-m-zolotukhin-p-wu.json b/pytorchconf-2024/videos/dl-compiler-panel-discussion-p-tillet-j-ansel-j-pienaar-t-chen-m-zolotukhin-p-wu.json index 4de82e036..35e32524b 100644 --- a/pytorchconf-2024/videos/dl-compiler-panel-discussion-p-tillet-j-ansel-j-pienaar-t-chen-m-zolotukhin-p-wu.json +++ b/pytorchconf-2024/videos/dl-compiler-panel-discussion-p-tillet-j-ansel-j-pienaar-t-chen-m-zolotukhin-p-wu.json @@ -10,11 +10,16 @@ } ], "speakers": [ - "TODO" + "Philip Tillet", + "Jason Ansel", + "Jacques Pienaar", + "Tianqi Chen", + "Mikhail Zolotukhin", + "Peng Wu" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/YWDzHGx8PrY/maxresdefault.webp", - "title": "DL Compiler Panel Discussion - P. Tillet, J. Ansel, J. Pienaar, T. Chen, M. Zolotukhin, P. Wu", + "title": "DL Compiler Panel Discussion", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/executorch-beta-and-on-device-generative-ai-support-mergen-nachin-mengtao-martin-yuan-meta.json b/pytorchconf-2024/videos/executorch-beta-and-on-device-generative-ai-support-mergen-nachin-mengtao-martin-yuan-meta.json index 83d57093d..01acf3a8d 100644 --- a/pytorchconf-2024/videos/executorch-beta-and-on-device-generative-ai-support-mergen-nachin-mengtao-martin-yuan-meta.json +++ b/pytorchconf-2024/videos/executorch-beta-and-on-device-generative-ai-support-mergen-nachin-mengtao-martin-yuan-meta.json @@ -10,11 +10,12 @@ } ], "speakers": [ - "TODO" + "Mergen Nachin", + "Mengtao Yuan" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/45yNTi7c1Q0/maxresdefault.webp", - "title": "ExecuTorch Beta and on-Device Generative AI Support - Mergen Nachin & Mengtao (Martin) Yuan, Meta", + "title": "ExecuTorch Beta and on-Device Generative AI Support", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/hacks-to-make-llm-training-faster-daniel-han-unsloth-ai.json b/pytorchconf-2024/videos/hacks-to-make-llm-training-faster-daniel-han-unsloth-ai.json index 19679e3e9..58cd2c8bd 100644 --- a/pytorchconf-2024/videos/hacks-to-make-llm-training-faster-daniel-han-unsloth-ai.json +++ b/pytorchconf-2024/videos/hacks-to-make-llm-training-faster-daniel-han-unsloth-ai.json @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Daniel Han" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/PdtKkc5jB4g/maxresdefault.webp", - "title": "Hacks to Make LLM Training Faster - Daniel Han, Unsloth AI", + "title": "Hacks to Make LLM Training Faster", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/halide-a-halide-backend-for-torchinductor-jason-ansel-meta.json b/pytorchconf-2024/videos/halide-a-halide-backend-for-torchinductor-jason-ansel-meta.json index e376e2a4a..edbf1fe34 100644 --- a/pytorchconf-2024/videos/halide-a-halide-backend-for-torchinductor-jason-ansel-meta.json +++ b/pytorchconf-2024/videos/halide-a-halide-backend-for-torchinductor-jason-ansel-meta.json @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Jason Ansel" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/hCvlWZEXRrg/maxresdefault.webp", - "title": "[HALIDE] A Halide Backend for TorchInductor - Jason Ansel, Meta", + "title": "[HALIDE] A Halide Backend for TorchInductor", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/implementing-a-custom-torch-compile-backend-a-case-study-maanav-dalal-yulong-wang-microsoft.json b/pytorchconf-2024/videos/implementing-a-custom-torch-compile-backend-a-case-study-maanav-dalal-yulong-wang-microsoft.json index 2b0856958..ca9f37341 100644 --- a/pytorchconf-2024/videos/implementing-a-custom-torch-compile-backend-a-case-study-maanav-dalal-yulong-wang-microsoft.json +++ b/pytorchconf-2024/videos/implementing-a-custom-torch-compile-backend-a-case-study-maanav-dalal-yulong-wang-microsoft.json @@ -10,11 +10,12 @@ } ], "speakers": [ - "TODO" + "Maanav Dalal", + "Yulong Wang" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/MnNO_13fLtU/maxresdefault.webp", - "title": "Implementing a Custom Torch.Compile Backend - A Case Study - Maanav Dalal & Yulong Wang, Microsoft", + "title": "Implementing a Custom Torch.Compile Backend - A Case Study", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/intel-gpu-in-upstream-pytorch-expanding-gpu-choices-and-enhancing-back-eikan-wang-min-jean-cho.json b/pytorchconf-2024/videos/intel-gpu-in-upstream-pytorch-expanding-gpu-choices-and-enhancing-back-eikan-wang-min-jean-cho.json index 14b2d1272..9bd2023e5 100644 --- a/pytorchconf-2024/videos/intel-gpu-in-upstream-pytorch-expanding-gpu-choices-and-enhancing-back-eikan-wang-min-jean-cho.json +++ b/pytorchconf-2024/videos/intel-gpu-in-upstream-pytorch-expanding-gpu-choices-and-enhancing-back-eikan-wang-min-jean-cho.json @@ -10,11 +10,12 @@ } ], "speakers": [ - "TODO" + "Eikan Wang", + "Min Jean Cho" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/HRdda_kVEh4/maxresdefault.webp", - "title": "Intel GPU in Upstream PyTorch: Expanding GPU Choices and Enhancing Back... Eikan Wang & Min Jean Cho", + "title": "Intel GPU in Upstream PyTorch: Expanding GPU Choices and Enhancing Backend Flexibility", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/keynote-building-an-advanced-knowledge-assistant-jerry-liu-co-founder-ceo-llamaindex.json b/pytorchconf-2024/videos/keynote-building-an-advanced-knowledge-assistant-jerry-liu-co-founder-ceo-llamaindex.json index 560e18ba1..e4a3a61cd 100644 --- a/pytorchconf-2024/videos/keynote-building-an-advanced-knowledge-assistant-jerry-liu-co-founder-ceo-llamaindex.json +++ b/pytorchconf-2024/videos/keynote-building-an-advanced-knowledge-assistant-jerry-liu-co-founder-ceo-llamaindex.json @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Jerry Liu" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/EjGdum0UdaI/maxresdefault.webp", - "title": "Keynote: Building an Advanced Knowledge Assistant - Jerry Liu, Co-Founder & CEO, LlamaIndex", + "title": "Keynote: Building an Advanced Knowledge Assistant", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/keynote-community-awards.json b/pytorchconf-2024/videos/keynote-community-awards.json index 754f88c69..0c0f7b6ce 100644 --- a/pytorchconf-2024/videos/keynote-community-awards.json +++ b/pytorchconf-2024/videos/keynote-community-awards.json @@ -9,9 +9,7 @@ "url": "https://pytorch.org/event/pytorch-conference-2024/" } ], - "speakers": [ - "TODO" - ], + "speakers": [], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/1dSVYKPk3N0/maxresdefault.webp", "title": "Keynote: Community Awards", diff --git a/pytorchconf-2024/videos/keynote-enabling-generative-ai-on-the-edge-cormac-brick-principal-engineer-google.json b/pytorchconf-2024/videos/keynote-enabling-generative-ai-on-the-edge-cormac-brick-principal-engineer-google.json index 993a9f9ca..b3e7764fe 100644 --- a/pytorchconf-2024/videos/keynote-enabling-generative-ai-on-the-edge-cormac-brick-principal-engineer-google.json +++ b/pytorchconf-2024/videos/keynote-enabling-generative-ai-on-the-edge-cormac-brick-principal-engineer-google.json @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Cormac Brick" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/n2DsEyLbuJg/maxresdefault.webp", - "title": "Keynote: Enabling Generative AI on the Edge - Cormac Brick, Principal Engineer, Google", + "title": "Keynote: Enabling Generative AI on the Edge", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/keynote-navigating-the-architectural-timeline-of-llms-sebastian-raschka-lightning-ai.json b/pytorchconf-2024/videos/keynote-navigating-the-architectural-timeline-of-llms-sebastian-raschka-lightning-ai.json index cc41cf780..5b476a885 100644 --- a/pytorchconf-2024/videos/keynote-navigating-the-architectural-timeline-of-llms-sebastian-raschka-lightning-ai.json +++ b/pytorchconf-2024/videos/keynote-navigating-the-architectural-timeline-of-llms-sebastian-raschka-lightning-ai.json @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Sebastian Raschka" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/frkAt-gZVjc/maxresdefault.webp", - "title": "Keynote: Navigating the Architectural Timeline of LLMs - Sebastian Raschka, Lightning AI", + "title": "Keynote: Navigating the Architectural Timeline of LLMs", "videos": [ { "type": "youtube", From 6d18d37addbdad4beb7cd890238d82c9ec6e7ba9 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Ezequiel=20Leonardo=20Casta=C3=B1o?= <14986783+ELC@users.noreply.github.com> Date: Fri, 21 Nov 2025 19:35:12 -0300 Subject: [PATCH 4/8] Complete all 12 keynote files in pytorchconf-2024 --- ...he-science-of-language-modeling-hanna-hajishirzi.json | 4 ++-- ...k-rooney-k-varshney-s-hooker-a-madry-r-bommasani.json | 8 ++++++-- .../keynote-panel-discussion-scaling-benchmarking.json | 9 +++++++-- ...-p-bialecki-p-wu-w-constable-k-khandelwal-m-yuan.json | 8 ++++++-- ...r-heterogeneous-computing-ion-stoica-uc-berkeley.json | 4 ++-- .../videos/keynote-welcome-back-opening-remarks.json | 4 +--- ...matt-white-executive-director-pytorch-foundation.json | 4 ++-- ...g-for-an-evaluation-tool-chip-huyen-voltron-data.json | 4 ++-- 8 files changed, 28 insertions(+), 17 deletions(-) diff --git a/pytorchconf-2024/videos/keynote-open-language-models-olmo-accelerating-the-science-of-language-modeling-hanna-hajishirzi.json b/pytorchconf-2024/videos/keynote-open-language-models-olmo-accelerating-the-science-of-language-modeling-hanna-hajishirzi.json index a80179560..548cf4d1c 100644 --- a/pytorchconf-2024/videos/keynote-open-language-models-olmo-accelerating-the-science-of-language-modeling-hanna-hajishirzi.json +++ b/pytorchconf-2024/videos/keynote-open-language-models-olmo-accelerating-the-science-of-language-modeling-hanna-hajishirzi.json @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Hanna Hajishirzi" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/GYWCYU-Xmnk/maxresdefault.webp", - "title": "Keynote: Open Language Models (OLMo): Accelerating the Science of Language Modeling Hanna Hajishirzi", + "title": "Keynote: Open Language Models (OLMo): Accelerating the Science of Language Modeling", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/keynote-panel-discussion-responsible-ai-k-rooney-k-varshney-s-hooker-a-madry-r-bommasani.json b/pytorchconf-2024/videos/keynote-panel-discussion-responsible-ai-k-rooney-k-varshney-s-hooker-a-madry-r-bommasani.json index be238d1e2..fd53115a7 100644 --- a/pytorchconf-2024/videos/keynote-panel-discussion-responsible-ai-k-rooney-k-varshney-s-hooker-a-madry-r-bommasani.json +++ b/pytorchconf-2024/videos/keynote-panel-discussion-responsible-ai-k-rooney-k-varshney-s-hooker-a-madry-r-bommasani.json @@ -10,11 +10,15 @@ } ], "speakers": [ - "TODO" + "Kate Rooney", + "Kush Varshney", + "Sara Hooker", + "Aleksander Madry", + "Rishi Bommasani" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi/FLUd0k2ctJE/maxresdefault.jpg", - "title": "Keynote Panel Discussion: Responsible AI - K. Rooney, K. Varshney, S. Hooker, A. Madry, R. Bommasani", + "title": "Keynote Panel Discussion: Responsible AI", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/keynote-panel-discussion-scaling-benchmarking.json b/pytorchconf-2024/videos/keynote-panel-discussion-scaling-benchmarking.json index 322f86834..5c3f06529 100644 --- a/pytorchconf-2024/videos/keynote-panel-discussion-scaling-benchmarking.json +++ b/pytorchconf-2024/videos/keynote-panel-discussion-scaling-benchmarking.json @@ -1,5 +1,5 @@ { - "description": "Keynote Panel Discussion: Scaling & Benchmarking - Anastasios Nikolas Angelopoulos, UC Berkeley/LMSYS; Lisa Dunlap, UC Berkeley; James Bradbury, Anthropic; Tri Dao, together.ai; Aparna Ramani & Soumith Chintala, Meta", + "description": "Keynote Panel Discussion: Scaling & Benchmarking", "duration": 2023, "language": "eng", "recorded": "2024-09-18", @@ -10,7 +10,12 @@ } ], "speakers": [ - "TODO" + "Anastasios Angelopoulos", + "Lisa Dunlap", + "James Bradbury", + "Tri Dao", + "Aparna Ramani", + "Soumith Chintala" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/01a4tBLEzVM/maxresdefault.webp", diff --git a/pytorchconf-2024/videos/keynote-pytorch-technical-deep-dive-p-bialecki-p-wu-w-constable-k-khandelwal-m-yuan.json b/pytorchconf-2024/videos/keynote-pytorch-technical-deep-dive-p-bialecki-p-wu-w-constable-k-khandelwal-m-yuan.json index 377371724..5dff0a593 100644 --- a/pytorchconf-2024/videos/keynote-pytorch-technical-deep-dive-p-bialecki-p-wu-w-constable-k-khandelwal-m-yuan.json +++ b/pytorchconf-2024/videos/keynote-pytorch-technical-deep-dive-p-bialecki-p-wu-w-constable-k-khandelwal-m-yuan.json @@ -10,11 +10,15 @@ } ], "speakers": [ - "TODO" + "Piotr Bialecki", + "Peng Wu", + "Will Constable", + "Kartikay Khandelwal", + "Mengtao Yuan" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/d29dGW8FafI/maxresdefault.webp", - "title": "Keynote: PyTorch Technical Deep Dive - P. Bialecki, P. Wu, W. Constable, K. Khandelwal & M. Yuan", + "title": "Keynote: PyTorch Technical Deep Dive", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/keynote-ray-a-distributed-framework-for-heterogeneous-computing-ion-stoica-uc-berkeley.json b/pytorchconf-2024/videos/keynote-ray-a-distributed-framework-for-heterogeneous-computing-ion-stoica-uc-berkeley.json index b3c7ec7cc..a318524d5 100644 --- a/pytorchconf-2024/videos/keynote-ray-a-distributed-framework-for-heterogeneous-computing-ion-stoica-uc-berkeley.json +++ b/pytorchconf-2024/videos/keynote-ray-a-distributed-framework-for-heterogeneous-computing-ion-stoica-uc-berkeley.json @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Ion Stoica" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/E98Pv36f0YI/maxresdefault.webp", - "title": "Keynote: Ray: A Distributed Framework for Heterogeneous Computing - Ion Stoica, UC Berkeley", + "title": "Keynote: Ray: A Distributed Framework for Heterogeneous Computing", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/keynote-welcome-back-opening-remarks.json b/pytorchconf-2024/videos/keynote-welcome-back-opening-remarks.json index 938690c79..c0cccf9e7 100644 --- a/pytorchconf-2024/videos/keynote-welcome-back-opening-remarks.json +++ b/pytorchconf-2024/videos/keynote-welcome-back-opening-remarks.json @@ -9,9 +9,7 @@ "url": "https://pytorch.org/event/pytorch-conference-2024/" } ], - "speakers": [ - "TODO" - ], + "speakers": [], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/fabW4B5IlVg/maxresdefault.webp", "title": "Keynote: Welcome Back & Opening Remarks", diff --git a/pytorchconf-2024/videos/keynote-welcome-opening-remarks-matt-white-executive-director-pytorch-foundation.json b/pytorchconf-2024/videos/keynote-welcome-opening-remarks-matt-white-executive-director-pytorch-foundation.json index c61f44eb7..70c76672a 100644 --- a/pytorchconf-2024/videos/keynote-welcome-opening-remarks-matt-white-executive-director-pytorch-foundation.json +++ b/pytorchconf-2024/videos/keynote-welcome-opening-remarks-matt-white-executive-director-pytorch-foundation.json @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Matt White" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/9VFHuGaq7SU/maxresdefault.webp", - "title": "Keynote: Welcome & Opening Remarks - Matt White, Executive Director, PyTorch Foundation", + "title": "Keynote: Welcome & Opening Remarks", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/keynote-why-you-should-think-twice-before-paying-for-an-evaluation-tool-chip-huyen-voltron-data.json b/pytorchconf-2024/videos/keynote-why-you-should-think-twice-before-paying-for-an-evaluation-tool-chip-huyen-voltron-data.json index ae52f3870..2cd66a8ca 100644 --- a/pytorchconf-2024/videos/keynote-why-you-should-think-twice-before-paying-for-an-evaluation-tool-chip-huyen-voltron-data.json +++ b/pytorchconf-2024/videos/keynote-why-you-should-think-twice-before-paying-for-an-evaluation-tool-chip-huyen-voltron-data.json @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Chip Huyen" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/IbKKPeuieFw/maxresdefault.webp", - "title": "Keynote: Why You Should Think Twice Before Paying for an Evaluation Tool - Chip Huyen, Voltron Data", + "title": "Keynote: Why You Should Think Twice Before Paying for an Evaluation Tool", "videos": [ { "type": "youtube", From 8be251b4928e10b471c79577b64bae2ddc01a275 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Ezequiel=20Leonardo=20Casta=C3=B1o?= <14986783+ELC@users.noreply.github.com> Date: Fri, 21 Nov 2025 19:45:33 -0300 Subject: [PATCH 5/8] Process pytorchconf-2024: Extract speakers, clean titles/descriptions for 50+ files - Processed panel discussion, startup showcase, and regular session files - Processed sponsored keynotes and sponsored sessions - Processed first batch of lightning talks (5 files) - Extracted speakers from titles and descriptions - Removed author names and organization info from titles - Removed title prefixes from descriptions - Cleaned up Lightning Talk: prefixes from titles --- ...extension-points-alban-desmaison-meta.json | 6 ++-- ...rnel-libraries-on-rocm-jeff-daily-amd.json | 6 ++-- ...ick-smyth-dan-fernandez-srishti-hegde.json | 8 ++++-- ...h-community-resources-tu-zong-zesheng.json | 6 ++-- ...e-shailvi-wakhlu-shailvi-ventures-llc.json | 6 ++-- ...y-a-conneau-moderated-by-k-khandelwal.json | 9 ++++-- ...r-3-suvaditya-mukherjee-shireen-chand.json | 7 +++-- .../pytorch-conference-2024-highlights.json | 28 ------------------- ...-npu-acceleration-felix-baum-qualcomm.json | 6 ++-- ...aying-ooms-mark-saroufim-jane-xu-meta.json | 7 +++-- ...e-innovation-with-sea-anush-elangovan.json | 6 ++-- ...-pytorch-and-intel-kismat-singh-intel.json | 6 ++-- ...cting-the-ai-orchestra-taylor-dolezal.json | 6 ++-- ...language-models-mudhakar-srivatsa-ibm.json | 6 ++-- ...lerating-the-ai-lifecycle-luca-antiga.json | 6 ++-- ...mance-pyt-robert-suderman-ian-nordeng.json | 7 +++-- ...th-arms-global-comp-gian-marco-iodice.json | 6 ++-- ...del-alignment-gerald-shen-jimmy-zhang.json | 7 +++-- ...orm-mark-sherwood-shauheen-zahirazami.json | 7 +++-- ...m-ubiquity-jack-khuu-jesse-white-meta.json | 7 +++-- pytorchconf-2024/videos/startup-showcase.json | 24 ---------------- ...e-of-pytorch-ji-li-damien-sereni-meta.json | 7 +++-- ...ce-llm-framework-wing-lian-axolotl-ai.json | 6 ++-- ...ve-datasets-and-models-aaron-gokaslan.json | 6 ++-- ...-ecosystem-arthur-zucker-hugging-face.json | 6 ++-- ...f-the-llama-ecosystem-joe-spisak-meta.json | 6 ++-- ...with-torch-compile-pragaash-ponnusamy.json | 6 ++-- ...-feng-chien-chin-huang-simon-fan-meta.json | 8 ++++-- ...erformance-imp-jiong-gong-leslie-fang.json | 7 +++-- ...3d-parallel-wanchao-liang-linsong-chu.json | 7 +++-- ...-in-native-pytorch-evan-smothers-meta.json | 6 ++-- ...orch-mihir-patel-brian-chu-databricks.json | 7 +++-- ...performance-constraints-philip-tillet.json | 6 ++-- ...ml-compilation-tianqi-chen-cmu-octoai.json | 6 ++-- ...-inference-workload-mark-moyou-nvidia.json | 6 ++-- ...plainable-large-languag-rashmi-nagpal.json | 6 ++-- ...-woosuk-kwon-xiaoxuan-liu-uc-berkeley.json | 7 +++-- 37 files changed, 123 insertions(+), 157 deletions(-) delete mode 100644 pytorchconf-2024/videos/pytorch-conference-2024-highlights.json delete mode 100644 pytorchconf-2024/videos/startup-showcase.json diff --git a/pytorchconf-2024/videos/lightning-talk-a-whirlwind-tour-of-pytorch-extension-points-alban-desmaison-meta.json b/pytorchconf-2024/videos/lightning-talk-a-whirlwind-tour-of-pytorch-extension-points-alban-desmaison-meta.json index 3bb1cc4d8..cf07826b9 100644 --- a/pytorchconf-2024/videos/lightning-talk-a-whirlwind-tour-of-pytorch-extension-points-alban-desmaison-meta.json +++ b/pytorchconf-2024/videos/lightning-talk-a-whirlwind-tour-of-pytorch-extension-points-alban-desmaison-meta.json @@ -1,5 +1,5 @@ { - "description": "Lightning Talk: A Whirlwind Tour of PyTorch Extension Points - Alban Desmaison, Meta\n\nJourney across the PyTorch stack and see all the extension points that exist from nn.Module to the c++ Dispatcher through autograd and subclasses. This sessions will cover example use cases and when each one should be used while pointing to reference for in-depth details.", + "description": "Journey across the PyTorch stack and see all the extension points that exist from nn.Module to the c++ Dispatcher through autograd and subclasses. This sessions will cover example use cases and when each one should be used while pointing to reference for in-depth details.", "duration": 848, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Alban Desmaison" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/1A54cUPXhZ4/maxresdefault.webp", - "title": "Lightning Talk: A Whirlwind Tour of PyTorch Extension Points - Alban Desmaison, Meta", + "title": "A Whirlwind Tour of PyTorch Extension Points", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/lightning-talk-aotriton-ahead-of-time-triton-kernel-libraries-on-rocm-jeff-daily-amd.json b/pytorchconf-2024/videos/lightning-talk-aotriton-ahead-of-time-triton-kernel-libraries-on-rocm-jeff-daily-amd.json index 1ddea8040..3add9dee6 100644 --- a/pytorchconf-2024/videos/lightning-talk-aotriton-ahead-of-time-triton-kernel-libraries-on-rocm-jeff-daily-amd.json +++ b/pytorchconf-2024/videos/lightning-talk-aotriton-ahead-of-time-triton-kernel-libraries-on-rocm-jeff-daily-amd.json @@ -1,5 +1,5 @@ { - "description": "Lightning Talk: AOTriton: Ahead of Time Triton Kernel Libraries on ROCm - Jeff Daily, AMD\n\nScaled dot product attention provides significant acceleration of the transformer layer through fusion of the multihead attention layer. There are several different algorithms to achieve this but tiled attention through scaled dot product attention via Flash Attention is a very popular approach. In PyTorch on the ROCm platform this is currently achieved through ahead of time compiled (AOT) Triton kernels in a linkable archive. AMD\u2019s work to enable and package these kernels is done through AOTriton, which aims to use Triton\u2019s compiler and GPU kernels for faster development. AOTriton maintains an optimized set of tiling sizes and other parameters to provide optimized, pre-compiled Triton kernels. The differences between JIT and AOT are few but are very important. Despite this, prototyping kernels in Triton is much faster than template-based C++ libraries. In this presentation we will go into detail on the interaction layer between PyTorch and AOTriton, the structure of AOTriton and how to add new triton kernels to AOTriton.", + "description": "Scaled dot product attention provides significant acceleration of the transformer layer through fusion of the multihead attention layer. There are several different algorithms to achieve this but tiled attention through scaled dot product attention via Flash Attention is a very popular approach. In PyTorch on the ROCm platform this is currently achieved through ahead of time compiled (AOT) Triton kernels in a linkable archive. AMD\u2019s work to enable and package these kernels is done through AOTriton, which aims to use Triton\u2019s compiler and GPU kernels for faster development. AOTriton maintains an optimized set of tiling sizes and other parameters to provide optimized, pre-compiled Triton kernels. The differences between JIT and AOT are few but are very important. Despite this, prototyping kernels in Triton is much faster than template-based C++ libraries. In this presentation we will go into detail on the interaction layer between PyTorch and AOTriton, the structure of AOTriton and how to add new triton kernels to AOTriton.", "duration": 692, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Jeff Daily" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/bhplJt1XAMI/maxresdefault.webp", - "title": "Lightning Talk: AOTriton: Ahead of Time Triton Kernel Libraries on ROCm - Jeff Daily, AMD", + "title": "AOTriton: Ahead of Time Triton Kernel Libraries on ROCm", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/lightning-talk-beyond-zero-eliminating-vulnerabili-patrick-smyth-dan-fernandez-srishti-hegde.json b/pytorchconf-2024/videos/lightning-talk-beyond-zero-eliminating-vulnerabili-patrick-smyth-dan-fernandez-srishti-hegde.json index 07c2069cd..a118a616e 100644 --- a/pytorchconf-2024/videos/lightning-talk-beyond-zero-eliminating-vulnerabili-patrick-smyth-dan-fernandez-srishti-hegde.json +++ b/pytorchconf-2024/videos/lightning-talk-beyond-zero-eliminating-vulnerabili-patrick-smyth-dan-fernandez-srishti-hegde.json @@ -1,5 +1,5 @@ { - "description": "Lightning Talk: Beyond Zero: Eliminating Vulnerabilities in PyTorch Container Images - Patrick Smyth, Dan Fernandez & Srishti Hegde, Chainguard\n\nContainer images are increasingly the future of production applications at scale, providing reproducibility, robustness, and transparency. As PyTorch images get deployed to production, however, security becomes a major concern. PyTorch has a large attack surface, and building secure PyTorch images can be a challenge. Currently, the official PyTorch runtime container image has 30 CVEs (known vulnerabilities) rated critical and 256 CVE rated high. Improving this situation could secure many deployments that incorporate PyTorch for cloud-based inference or training. In this fast-paced session, we'll take a deep dive on the official PyTorch image from a vulnerability mitigation perspective, looking hard at included packages, executables, and active CVE. We'll identify low-hanging fruit for increasing security, including stripping bloat and building fresh. We'll also talk about the next level of security practiced in Chainguard's PyTorch image builds, such as including SBOMs and going distroless. Finally, we'll consider emerging tools and approaches for analyzing AI artifacts such as models and how these systems can benefit PyTorch in production.", + "description": "Container images are increasingly the future of production applications at scale, providing reproducibility, robustness, and transparency. As PyTorch images get deployed to production, however, security becomes a major concern. PyTorch has a large attack surface, and building secure PyTorch images can be a challenge. Currently, the official PyTorch runtime container image has 30 CVEs (known vulnerabilities) rated critical and 256 CVE rated high. Improving this situation could secure many deployments that incorporate PyTorch for cloud-based inference or training. In this fast-paced session, we'll take a deep dive on the official PyTorch image from a vulnerability mitigation perspective, looking hard at included packages, executables, and active CVE. We'll identify low-hanging fruit for increasing security, including stripping bloat and building fresh. We'll also talk about the next level of security practiced in Chainguard's PyTorch image builds, such as including SBOMs and going distroless. Finally, we'll consider emerging tools and approaches for analyzing AI artifacts such as models and how these systems can benefit PyTorch in production.", "duration": 736, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,13 @@ } ], "speakers": [ - "TODO" + "Patrick Smyth", + "Dan Fernandez", + "Srishti Hegde" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/1klynk1dxYA/maxresdefault.webp", - "title": "Lightning Talk: Beyond Zero: Eliminating Vulnerabili... Patrick Smyth, Dan Fernandez & Srishti Hegde", + "title": "Beyond Zero: Eliminating Vulnerabilities in PyTorch Container Images", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/lightning-talk-building-and-supporting-the-chinese-pytorch-community-resources-tu-zong-zesheng.json b/pytorchconf-2024/videos/lightning-talk-building-and-supporting-the-chinese-pytorch-community-resources-tu-zong-zesheng.json index cdef56690..a410cd6c1 100644 --- a/pytorchconf-2024/videos/lightning-talk-building-and-supporting-the-chinese-pytorch-community-resources-tu-zong-zesheng.json +++ b/pytorchconf-2024/videos/lightning-talk-building-and-supporting-the-chinese-pytorch-community-resources-tu-zong-zesheng.json @@ -1,5 +1,5 @@ { - "description": "Lightning Talk: Building and Supporting the Chinese PyTorch Community: Resources, Tutorials, and Engagement - Zong Zesheng, Huawei\n\nDescription: This proposal aims to provide a comprehensive introduction to the Chinese PyTorch community, we hope to inspire more users to join and contribute, fostering a vibrant and inclusive environment for PyTorch enthusiasts in China. Chinese PyTorch Homepage Introduction to the official Chinese version of the PyTorch website, highlighting its features. Navigation tips and key sections, such as documentation, tutorials, and community events. Improve the connection of users from China with PyTorch Community. Localized Tutorials and Documentation The 2.x version not have Translated version, it hard to catch up with latest features of PyTorch if the beginner not good at English. We translated official documents and tutorials, covering everything from basic PyTorch concepts to advanced applications. Interactive tutorials No interactive tutorials(Like Google Colab) for Chinese students or beginners before, they have to setup environment before start with PyTorch, which might be hard for beginners. And now, an online notebook & tutorials are available to practice or tuning steps for beginners.", + "description": "Description: This proposal aims to provide a comprehensive introduction to the Chinese PyTorch community, we hope to inspire more users to join and contribute, fostering a vibrant and inclusive environment for PyTorch enthusiasts in China. Chinese PyTorch Homepage Introduction to the official Chinese version of the PyTorch website, highlighting its features. Navigation tips and key sections, such as documentation, tutorials, and community events. Improve the connection of users from China with PyTorch Community. Localized Tutorials and Documentation The 2.x version not have Translated version, it hard to catch up with latest features of PyTorch if the beginner not good at English. We translated official documents and tutorials, covering everything from basic PyTorch concepts to advanced applications. Interactive tutorials No interactive tutorials(Like Google Colab) for Chinese students or beginners before, they have to setup environment before start with PyTorch, which might be hard for beginners. And now, an online notebook & tutorials are available to practice or tuning steps for beginners.", "duration": 453, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Zong Zesheng" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/AXMN5oUyWKE/maxresdefault.webp", - "title": "Lightning Talk: Building and Supporting the Chinese PyTorch Community: Resources, Tu... Zong Zesheng", + "title": "Building and Supporting the Chinese PyTorch Community: Resources, Tutorials, and Engagement", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/lightning-talk-debiasing-the-data-lifecycle-shailvi-wakhlu-shailvi-ventures-llc.json b/pytorchconf-2024/videos/lightning-talk-debiasing-the-data-lifecycle-shailvi-wakhlu-shailvi-ventures-llc.json index c68c58459..8bafbb9b1 100644 --- a/pytorchconf-2024/videos/lightning-talk-debiasing-the-data-lifecycle-shailvi-wakhlu-shailvi-ventures-llc.json +++ b/pytorchconf-2024/videos/lightning-talk-debiasing-the-data-lifecycle-shailvi-wakhlu-shailvi-ventures-llc.json @@ -1,5 +1,5 @@ { - "description": "Lightning Talk: Debiasing the Data Lifecycle - Shailvi Wakhlu, Shailvi Ventures LLC\n\nBiased data, results in biased decision-making. Making sure that at every step of the data lifecycle, we make conscious attempts to debias the data is an important responsibility for all data scientists. In this talk, I highlight the typical data lifecycle, and how to prevent biases at every step. ---- The key takeaways from my talk include: 1) Understanding the data lifecycle 2) What are the typical ways biases creep in 3) How we can proactively prevent and fix biases in data", + "description": "Biased data, results in biased decision-making. Making sure that at every step of the data lifecycle, we make conscious attempts to debias the data is an important responsibility for all data scientists. In this talk, I highlight the typical data lifecycle, and how to prevent biases at every step. ---- The key takeaways from my talk include: 1) Understanding the data lifecycle 2) What are the typical ways biases creep in 3) How we can proactively prevent and fix biases in data", "duration": 661, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Shailvi Wakhlu" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi/V9NVEceY9Wg/maxresdefault.jpg", - "title": "Lightning Talk: Debiasing the Data Lifecycle - Shailvi Wakhlu, Shailvi Ventures LLC", + "title": "Debiasing the Data Lifecycle", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/panel-discussion-t-dettmers-h-schoelkopf-a-chowdhery-a-conneau-moderated-by-k-khandelwal.json b/pytorchconf-2024/videos/panel-discussion-t-dettmers-h-schoelkopf-a-chowdhery-a-conneau-moderated-by-k-khandelwal.json index 99aee6bbe..96759d049 100644 --- a/pytorchconf-2024/videos/panel-discussion-t-dettmers-h-schoelkopf-a-chowdhery-a-conneau-moderated-by-k-khandelwal.json +++ b/pytorchconf-2024/videos/panel-discussion-t-dettmers-h-schoelkopf-a-chowdhery-a-conneau-moderated-by-k-khandelwal.json @@ -1,5 +1,5 @@ { - "description": "Panel Discussion - Tim Dettmers, AI2/Carnegie Melon; Hailey Schoelkopf, EleutherAI; Aakanksha Chowdhery, Meta; Alexis Conneau, OpenAI; Moderated by Kartikay Khandelwal, Meta\n\nSpeakers: Tim Dettmers, Hailey Schoelkopf, Aakanksha Chowdhery, Alexis Conneau\nAs open-source LLMs have become more capable, a substantial ecosystem has developed around the fine-tuning of these models. A thriving community of researchers, developers, practitioners and hobbyists has emerged which focuses on topics ranging from memory efficiency, parameter-efficient fine-tuning and quantization to performance at scale and reproducible evaluations. The goal of this mini-summit is to bring this community together to discuss ideas, share knowledge and build connections.\n\nThe agenda features a keynote from Joe Spisak on the state of the Llama ecosystem followed by invited talks from the founders of Axolotl, Unsloth and torchtune. We conclude the summit with a riveting discussion on what\u2019s next for LLMs, fine-tuning and the PyTorch ecosystem with a fabulous panel of experts - Tim Dettmers (author of bitsandbytes and QLoRA), Hailey Schoelkopf (maintainer of LM Eval Harness at EleutherAI), Aakanksha Chowdhery (Lead author on PaLM and Gemini) and Alexis Conneau (Research Lead at OpenAI)", + "description": "Speakers: Tim Dettmers, Hailey Schoelkopf, Aakanksha Chowdhery, Alexis Conneau\nAs open-source LLMs have become more capable, a substantial ecosystem has developed around the fine-tuning of these models. A thriving community of researchers, developers, practitioners and hobbyists has emerged which focuses on topics ranging from memory efficiency, parameter-efficient fine-tuning and quantization to performance at scale and reproducible evaluations. The goal of this mini-summit is to bring this community together to discuss ideas, share knowledge and build connections.\n\nThe agenda features a keynote from Joe Spisak on the state of the Llama ecosystem followed by invited talks from the founders of Axolotl, Unsloth and torchtune. We conclude the summit with a riveting discussion on what\u2019s next for LLMs, fine-tuning and the PyTorch ecosystem with a fabulous panel of experts - Tim Dettmers (author of bitsandbytes and QLoRA), Hailey Schoelkopf (maintainer of LM Eval Harness at EleutherAI), Aakanksha Chowdhery (Lead author on PaLM and Gemini) and Alexis Conneau (Research Lead at OpenAI)", "duration": 2160, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,14 @@ } ], "speakers": [ - "TODO" + "Tim Dettmers", + "Hailey Schoelkopf", + "Aakanksha Chowdhery", + "Alexis Conneau" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/1VwOB7OYm4U/maxresdefault.webp", - "title": "Panel Discussion - T. Dettmers, H. Schoelkopf, A. Chowdhery, A. Conneau, Moderated by K. Khandelwal", + "title": "Panel Discussion", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/pushing-the-performance-envelope-an-optimization-study-for-3-suvaditya-mukherjee-shireen-chand.json b/pytorchconf-2024/videos/pushing-the-performance-envelope-an-optimization-study-for-3-suvaditya-mukherjee-shireen-chand.json index 45397adb0..903353e34 100644 --- a/pytorchconf-2024/videos/pushing-the-performance-envelope-an-optimization-study-for-3-suvaditya-mukherjee-shireen-chand.json +++ b/pytorchconf-2024/videos/pushing-the-performance-envelope-an-optimization-study-for-3-suvaditya-mukherjee-shireen-chand.json @@ -1,5 +1,5 @@ { - "description": "Pushing the Performance Envelope: An Optimization Study for 3D Generative Modelling with PyTorch - Suvaditya Mukherjee & Shireen Chand, University of Southern California\n\nThis work explores performance optimization strategies for training 3D generative models using PyTorch. We focus on training Variational Autoencoders (VAEs) on the ShapeNet dataset, a popular benchmark for this task. Our objective is to achieve high-fidelity reconstructions while minimizing the computational footprint and training time. We focus on: 1) Large-scale 3D dataset loading strategies using PyTorch & Google Cloud Storage Buckets 2) Implementation details and insights for 3D VAEs using PyTorch 2.x 3) Training using Automatic Mixed-precision regimes 4) Optimized training using torch.compile and different quantization techniques (as supported) - Dynamic Quantization - Static Quantization - Static Quantization-aware Training 5) Comparative Benchmark over several experiments performed with a focus on execution time and memory footprint Through this comprehensive study, we present a comparative analysis of the performance gains achieved by our optimized models. Our findings present empirical insights into the trade-offs between model accuracy, computational complexity, and hardware resource utilization.", + "description": "This work explores performance optimization strategies for training 3D generative models using PyTorch. We focus on training Variational Autoencoders (VAEs) on the ShapeNet dataset, a popular benchmark for this task. Our objective is to achieve high-fidelity reconstructions while minimizing the computational footprint and training time. We focus on: 1) Large-scale 3D dataset loading strategies using PyTorch & Google Cloud Storage Buckets 2) Implementation details and insights for 3D VAEs using PyTorch 2.x 3) Training using Automatic Mixed-precision regimes 4) Optimized training using torch.compile and different quantization techniques (as supported) - Dynamic Quantization - Static Quantization - Static Quantization-aware Training 5) Comparative Benchmark over several experiments performed with a focus on execution time and memory footprint Through this comprehensive study, we present a comparative analysis of the performance gains achieved by our optimized models. Our findings present empirical insights into the trade-offs between model accuracy, computational complexity, and hardware resource utilization.", "duration": 1153, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,12 @@ } ], "speakers": [ - "TODO" + "Suvaditya Mukherjee", + "Shireen Chand" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/Xm9Bex3LE_8/maxresdefault.webp", - "title": "Pushing the Performance Envelope: An Optimization Study for 3... Suvaditya Mukherjee & Shireen Chand", + "title": "Pushing the Performance Envelope: An Optimization Study for 3D Generative Modelling with PyTorch", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/pytorch-conference-2024-highlights.json b/pytorchconf-2024/videos/pytorch-conference-2024-highlights.json deleted file mode 100644 index 4e5abd751..000000000 --- a/pytorchconf-2024/videos/pytorch-conference-2024-highlights.json +++ /dev/null @@ -1,28 +0,0 @@ -{ - "description": "Relive the excitement of the PyTorch Conference 2024, where AI innovators and enthusiasts gathered to explore cutting-edge advancements, foster community connections, and shape the future of machine learning.\n\nView all of the conference sessions on the PyTorch YouTube channel: https://www.youtube.com/playlist?list=PL_lsbAsL_o2B_znuvm-pDtV_cRhpqZb8l", - "duration": 100, - "language": "eng", - "recorded": "2024-09-18", - "related_urls": [ - { - "label": "Conference Website", - "url": "https://pytorch.org/event/pytorch-conference-2024/" - }, - { - "label": "https://www.youtube.com/playlist?list=PL_lsbAsL_o2B_znuvm-pDtV_cRhpqZb8l", - "url": "https://www.youtube.com/playlist?list=PL_lsbAsL_o2B_znuvm-pDtV_cRhpqZb8l" - } - ], - "speakers": [ - "TODO" - ], - "tags": [], - "thumbnail_url": "https://i.ytimg.com/vi_webp/tUG3UWSP2ZU/maxresdefault.webp", - "title": "PyTorch Conference 2024 Highlights", - "videos": [ - { - "type": "youtube", - "url": "https://www.youtube.com/watch?v=tUG3UWSP2ZU" - } - ] -} diff --git a/pytorchconf-2024/videos/running-state-of-art-gen-ai-models-on-device-with-npu-acceleration-felix-baum-qualcomm.json b/pytorchconf-2024/videos/running-state-of-art-gen-ai-models-on-device-with-npu-acceleration-felix-baum-qualcomm.json index 21006aca1..43201a0a4 100644 --- a/pytorchconf-2024/videos/running-state-of-art-gen-ai-models-on-device-with-npu-acceleration-felix-baum-qualcomm.json +++ b/pytorchconf-2024/videos/running-state-of-art-gen-ai-models-on-device-with-npu-acceleration-felix-baum-qualcomm.json @@ -1,5 +1,5 @@ { - "description": "Running State-of-Art Gen AI Models on-Device with NPU Acceleration - Felix Baum, Qualcomm\n\nSince the boom of generative AI, the industry is now moving towards on-device AI inferencing, as it is not only a trend but a necessity now in order to save costs, achieve the best inference performance, ultra-low latency at the lowest power possible. In this session we go over the new features added on the Qualcomm AI Stack and how it works with the public release of ExecuTorch 1.0. We will discuss how to run traditional workloads as well as GenAI use cases including the latest version of Llama on the Mobile device while using Qualcomm Hexagon NPU.", + "description": "Since the boom of generative AI, the industry is now moving towards on-device AI inferencing, as it is not only a trend but a necessity now in order to save costs, achieve the best inference performance, ultra-low latency at the lowest power possible. In this session we go over the new features added on the Qualcomm AI Stack and how it works with the public release of ExecuTorch 1.0. We will discuss how to run traditional workloads as well as GenAI use cases including the latest version of Llama on the Mobile device while using Qualcomm Hexagon NPU.", "duration": 1460, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Felix Baum" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/wd57g2IM3C4/maxresdefault.webp", - "title": "Running State-of-Art Gen AI Models on-Device with NPU Acceleration - Felix Baum, Qualcomm", + "title": "Running State-of-Art Gen AI Models on-Device with NPU Acceleration", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/slaying-ooms-mark-saroufim-jane-xu-meta.json b/pytorchconf-2024/videos/slaying-ooms-mark-saroufim-jane-xu-meta.json index 3561a7bc3..2de2ee0df 100644 --- a/pytorchconf-2024/videos/slaying-ooms-mark-saroufim-jane-xu-meta.json +++ b/pytorchconf-2024/videos/slaying-ooms-mark-saroufim-jane-xu-meta.json @@ -1,5 +1,5 @@ { - "description": "Slaying OOMs - Mark Saroufim & Jane Xu, Meta\n\nHave you ever hit an OOM (and wished you had more VRAM)? Who hasn't! Hop on the bus with us and feel the road become smoother as we talk about stacking together techniques like FSDP2 + QLoRa + CPU Offloading + Fused ADAM (thanks Intel) + more in PyTorch native. We will give an overview of these techniques as well as the hard edges we solved in their composition. Curious for more? Or...still OOMing? We also plan on discussing our more researchy work on offloading, pagedness, and low precision optimizers.", + "description": "Have you ever hit an OOM (and wished you had more VRAM)? Who hasn't! Hop on the bus with us and feel the road become smoother as we talk about stacking together techniques like FSDP2 + QLoRa + CPU Offloading + Fused ADAM (thanks Intel) + more in PyTorch native. We will give an overview of these techniques as well as the hard edges we solved in their composition. Curious for more? Or...still OOMing? We also plan on discussing our more researchy work on offloading, pagedness, and low precision optimizers.", "duration": 1541, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,12 @@ } ], "speakers": [ - "TODO" + "Mark Saroufim", + "Jane Xu" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/xzBcBJ8_rzM/maxresdefault.webp", - "title": "Slaying OOMs - Mark Saroufim & Jane Xu, Meta", + "title": "Slaying OOMs", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/sponsored-keynote-accelerating-ai-how-amd-and-pytorch-drive-innovation-with-sea-anush-elangovan.json b/pytorchconf-2024/videos/sponsored-keynote-accelerating-ai-how-amd-and-pytorch-drive-innovation-with-sea-anush-elangovan.json index 14c7990a4..bc11cff13 100644 --- a/pytorchconf-2024/videos/sponsored-keynote-accelerating-ai-how-amd-and-pytorch-drive-innovation-with-sea-anush-elangovan.json +++ b/pytorchconf-2024/videos/sponsored-keynote-accelerating-ai-how-amd-and-pytorch-drive-innovation-with-sea-anush-elangovan.json @@ -1,5 +1,5 @@ { - "description": "Sponsored Keynote: Accelerating AI: How AMD and PyTorch Drive Innovation with Seamless Day-0 Support and High Performance - Anush Elangovan, CVP Software Development, AMD\n\nIn this keynote presentation, we explore the robust collaboration between AMD and PyTorch that is propelling advancements in artificial intelligence and machine learning. Discover how AMD's commitment to Day-0 PyTorch support ensures that PyTorch users benefit from cutting-edge performance enhancements and out-of-the-box compatibility. We delve into the technical synergies that make AMD hardware an ideal choice for PyTorch frameworks, showcasing real-world examples of accelerated workflows and breakthrough AI applications. Join us to learn how this dynamic partnership is enabling researchers, developers, and data scientists to push the boundaries of innovation and achieve unprecedented results in their AI projects.", + "description": "In this keynote presentation, we explore the robust collaboration between AMD and PyTorch that is propelling advancements in artificial intelligence and machine learning. Discover how AMD's commitment to Day-0 PyTorch support ensures that PyTorch users benefit from cutting-edge performance enhancements and out-of-the-box compatibility. We delve into the technical synergies that make AMD hardware an ideal choice for PyTorch frameworks, showcasing real-world examples of accelerated workflows and breakthrough AI applications. Join us to learn how this dynamic partnership is enabling researchers, developers, and data scientists to push the boundaries of innovation and achieve unprecedented results in their AI projects.", "duration": 343, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Anush Elangovan" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/paR73OPcdNw/maxresdefault.webp", - "title": "Sponsored Keynote: Accelerating AI: How AMD and PyTorch Drive Innovation with Sea... Anush Elangovan", + "title": "Accelerating AI: How AMD and PyTorch Drive Innovation with Seamless Day-0 Support and High Performance", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/sponsored-keynote-enabling-ai-everywhere-with-pytorch-and-intel-kismat-singh-intel.json b/pytorchconf-2024/videos/sponsored-keynote-enabling-ai-everywhere-with-pytorch-and-intel-kismat-singh-intel.json index 739653fab..af0f634b8 100644 --- a/pytorchconf-2024/videos/sponsored-keynote-enabling-ai-everywhere-with-pytorch-and-intel-kismat-singh-intel.json +++ b/pytorchconf-2024/videos/sponsored-keynote-enabling-ai-everywhere-with-pytorch-and-intel-kismat-singh-intel.json @@ -1,5 +1,5 @@ { - "description": "Sponsored Keynote: Enabling AI Everywhere with PyTorch and Intel - Kismat Singh,VP of Engineering for AI Frameworks, Intel\n\nUnlocking the availability of and access to generative AI technologies has great societal value. In this keynote, Kismat Singh will present how open software built on industry-standard frameworks such as PyTorch, and ubiquitous hardware from Intel that forms a large part of the current installed base across edge, PC and cloud are keys to democratizing AI and allowing new solutions to be implemented across industries ranging from healthcare, telecommunication, industrial and more. Kismat will share his thoughts on how software acceleration, flexibility and security are important factors in deploying AI applications in production and what he sees as challenges with those projects. He will also discuss Open Platform for Enterprise AI (OPEA), a new Linux Foundation AI and Data project that gives developers access to open source, standardized, modular, and heterogenous retrieval-augmented generation (RAG) pipelines that they can use for their enterprise-grade Generative AI deployments. Lastly, he will share some exciting Intel contributed features recently upstreamed into PyTorch. He will end the keynote by stating what he believes to be the future of AI and the part each of us will play in it!", + "description": "Unlocking the availability of and access to generative AI technologies has great societal value. In this keynote, Kismat Singh will present how open software built on industry-standard frameworks such as PyTorch, and ubiquitous hardware from Intel that forms a large part of the current installed base across edge, PC and cloud are keys to democratizing AI and allowing new solutions to be implemented across industries ranging from healthcare, telecommunication, industrial and more. Kismat will share his thoughts on how software acceleration, flexibility and security are important factors in deploying AI applications in production and what he sees as challenges with those projects. He will also discuss Open Platform for Enterprise AI (OPEA), a new Linux Foundation AI and Data project that gives developers access to open source, standardized, modular, and heterogenous retrieval-augmented generation (RAG) pipelines that they can use for their enterprise-grade Generative AI deployments. Lastly, he will share some exciting Intel contributed features recently upstreamed into PyTorch. He will end the keynote by stating what he believes to be the future of AI and the part each of us will play in it!", "duration": 358, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Kismat Singh" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/tGEtpXaoXuk/maxresdefault.webp", - "title": "Sponsored Keynote: Enabling AI Everywhere with PyTorch and Intel - Kismat Singh, Intel", + "title": "Enabling AI Everywhere with PyTorch and Intel", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/sponsored-keynote-from-containers-to-cognition-conducting-the-ai-orchestra-taylor-dolezal.json b/pytorchconf-2024/videos/sponsored-keynote-from-containers-to-cognition-conducting-the-ai-orchestra-taylor-dolezal.json index 26ea06f66..27a04e867 100644 --- a/pytorchconf-2024/videos/sponsored-keynote-from-containers-to-cognition-conducting-the-ai-orchestra-taylor-dolezal.json +++ b/pytorchconf-2024/videos/sponsored-keynote-from-containers-to-cognition-conducting-the-ai-orchestra-taylor-dolezal.json @@ -1,5 +1,5 @@ { - "description": "Sponsored Keynote: From Containers to Cognition: Conducting the AI Orchestra - Taylor Dolezal, Head of Ecosystem, Cloud Native Computing Foundation\n\nLet's explore the powerful harmony created when the CNCF and PyTorch communities join forces. This keynote highlights how the collaboration between cloud native experts and AI innovators is orchestrating a new era of technological symphonies. We'll touch on critical initiatives and shared victories that demonstrate the strength of this partnership. To illustrate the creative potential of this alliance, we'll briefly showcase a demo of how containerized workloads can produce unexpected melodies. Join us for this exploration of community-driven innovation, where containers and cognition come together to compose the future of technology.", + "description": "Let's explore the powerful harmony created when the CNCF and PyTorch communities join forces. This keynote highlights how the collaboration between cloud native experts and AI innovators is orchestrating a new era of technological symphonies. We'll touch on critical initiatives and shared victories that demonstrate the strength of this partnership. To illustrate the creative potential of this alliance, we'll briefly showcase a demo of how containerized workloads can produce unexpected melodies. Join us for this exploration of community-driven innovation, where containers and cognition come together to compose the future of technology.", "duration": 306, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Taylor Dolezal" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/MWmOeXI17Kg/maxresdefault.webp", - "title": "Sponsored Keynote: From Containers to Cognition: Conducting the AI Orchestra - Taylor Dolezal", + "title": "From Containers to Cognition: Conducting the AI Orchestra", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/sponsored-keynote-optimizing-ai-inference-for-large-language-models-mudhakar-srivatsa-ibm.json b/pytorchconf-2024/videos/sponsored-keynote-optimizing-ai-inference-for-large-language-models-mudhakar-srivatsa-ibm.json index 9ec26daed..ed2eb11d4 100644 --- a/pytorchconf-2024/videos/sponsored-keynote-optimizing-ai-inference-for-large-language-models-mudhakar-srivatsa-ibm.json +++ b/pytorchconf-2024/videos/sponsored-keynote-optimizing-ai-inference-for-large-language-models-mudhakar-srivatsa-ibm.json @@ -1,5 +1,5 @@ { - "description": "Sponsored Keynote: Optimizing AI Inference for Large Language Models - Mudhakar Srivatsa, Distinguished Engineer, IBM\n\nThis talk will cover two new ways IBM has optimized generative AI inferencing with PyTorch: speculative decoding and Triton kernel development. Speculative decoding leverages predictive modeling to reduce latency by anticipating potential outputs, streamlining the inference process without sacrificing accuracy. IBM Research's team developed new speculative architectures and open sourced speculators for LLama3 models. It will also discuss various Triton kernels to accelerate inference, one of which was contributed to vLLM for accelerating MoE models. Finally, it will share a glimpse of IBM's AI hardware work, including how the IBM Artificial Intelligence Unit (AIU) could integrate into the PyTorch stack.", + "description": "This talk will cover two new ways IBM has optimized generative AI inferencing with PyTorch: speculative decoding and Triton kernel development. Speculative decoding leverages predictive modeling to reduce latency by anticipating potential outputs, streamlining the inference process without sacrificing accuracy. IBM Research's team developed new speculative architectures and open sourced speculators for LLama3 models. It will also discuss various Triton kernels to accelerate inference, one of which was contributed to vLLM for accelerating MoE models. Finally, it will share a glimpse of IBM's AI hardware work, including how the IBM Artificial Intelligence Unit (AIU) could integrate into the PyTorch stack.", "duration": 349, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Mudhakar Srivatsa" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/ZpNyKs4EpVI/maxresdefault.webp", - "title": "Sponsored Keynote: Optimizing AI Inference for Large Language Models - Mudhakar Srivatsa, IBM", + "title": "Optimizing AI Inference for Large Language Models", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/sponsored-keynote-the-lightning-ai-oss-stack-for-accelerating-the-ai-lifecycle-luca-antiga.json b/pytorchconf-2024/videos/sponsored-keynote-the-lightning-ai-oss-stack-for-accelerating-the-ai-lifecycle-luca-antiga.json index 86df77233..892cda6cd 100644 --- a/pytorchconf-2024/videos/sponsored-keynote-the-lightning-ai-oss-stack-for-accelerating-the-ai-lifecycle-luca-antiga.json +++ b/pytorchconf-2024/videos/sponsored-keynote-the-lightning-ai-oss-stack-for-accelerating-the-ai-lifecycle-luca-antiga.json @@ -1,5 +1,5 @@ { - "description": "Sponsored Keynote: The Lightning AI OSS Stack for Accelerating the AI Lifecycle - Luca Antiga, CTO, Lightning AI\n\nWe introduce the Lightning AI open source stack, a high-performance stack for training, fine-tuning, and deploying AI systems that augments the PyTorch ecosystem.\n\nToday PyTorch Lightning powers training workloads across the industry, from small-scale research to large-scale training endeavors. The package has reached 130M total downloads in June 2024, 2x since early 2023. PyTorch Lightning 2.4 features support for 2D parallelism via DTensors, first introduced in PyTorch 2.3.\n\nThe open source stack is completed by Fabric (lightweight building blocks for scaling training workloads), LitGPT (library for pre-training, fine-tuning, serving LLMs), LitData (parallel data processing and streaming data loading), LitServe (lightweight, high-performance serving framework), TorchMetrics (de-facto standard in deep learning metrics), and the recently released Thunder compiler. Together, these packages provide a low-friction, high-performance stack to democratize and accelerate the AI lifecycle.\n\nThe stack is optimized to run on Lightning Studios, a PyTorch native, fully integrated AI development environment on the cloud.", + "description": "We introduce the Lightning AI open source stack, a high-performance stack for training, fine-tuning, and deploying AI systems that augments the PyTorch ecosystem.\n\nToday PyTorch Lightning powers training workloads across the industry, from small-scale research to large-scale training endeavors. The package has reached 130M total downloads in June 2024, 2x since early 2023. PyTorch Lightning 2.4 features support for 2D parallelism via DTensors, first introduced in PyTorch 2.3.\n\nThe open source stack is completed by Fabric (lightweight building blocks for scaling training workloads), LitGPT (library for pre-training, fine-tuning, serving LLMs), LitData (parallel data processing and streaming data loading), LitServe (lightweight, high-performance serving framework), TorchMetrics (de-facto standard in deep learning metrics), and the recently released Thunder compiler. Together, these packages provide a low-friction, high-performance stack to democratize and accelerate the AI lifecycle.\n\nThe stack is optimized to run on Lightning Studios, a PyTorch native, fully integrated AI development environment on the cloud.", "duration": 368, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Luca Antiga" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/jAtrviDGpwE/maxresdefault.webp", - "title": "Sponsored Keynote: The Lightning AI OSS Stack for Accelerating the AI Lifecycle - Luca Antiga", + "title": "The Lightning AI OSS Stack for Accelerating the AI Lifecycle", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/sponsored-session-accelerating-ai-innovation-high-performance-pyt-robert-suderman-ian-nordeng.json b/pytorchconf-2024/videos/sponsored-session-accelerating-ai-innovation-high-performance-pyt-robert-suderman-ian-nordeng.json index 803de28c1..b966d886c 100644 --- a/pytorchconf-2024/videos/sponsored-session-accelerating-ai-innovation-high-performance-pyt-robert-suderman-ian-nordeng.json +++ b/pytorchconf-2024/videos/sponsored-session-accelerating-ai-innovation-high-performance-pyt-robert-suderman-ian-nordeng.json @@ -1,5 +1,5 @@ { - "description": "Sponsored Session: Accelerating AI Innovation: High Performance PyTorch at AMD - Robert Suderman & Ian Nordeng, AMD\n\nExplore the powerful collaboration between AMD and PyTorch, driving advancements in AI and machine learning. Learn how AMD\u2019s Day-0 PyTorch support delivers cutting-edge performance and seamless compatibility.\n\nThis session will highlight the technical synergies that make AMD hardware ideal choice for PyTorch frameworks, with real-world examples of accelerated workflows and breakthrough AI applications. Attendees will gain insights into how this dynamic partnership is enabling researchers, developers, and data scientists to push the boundaries of innovation and achieve unprecedented results in AI projects.", + "description": "Explore the powerful collaboration between AMD and PyTorch, driving advancements in AI and machine learning. Learn how AMD\u2019s Day-0 PyTorch support delivers cutting-edge performance and seamless compatibility.\n\nThis session will highlight the technical synergies that make AMD hardware ideal choice for PyTorch frameworks, with real-world examples of accelerated workflows and breakthrough AI applications. Attendees will gain insights into how this dynamic partnership is enabling researchers, developers, and data scientists to push the boundaries of innovation and achieve unprecedented results in AI projects.", "duration": 1710, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,12 @@ } ], "speakers": [ - "TODO" + "Robert Suderman", + "Ian Nordeng" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/h0LG3JbA8kc/maxresdefault.webp", - "title": "Sponsored Session: Accelerating AI Innovation: High Performance PyT... Robert Suderman & Ian Nordeng", + "title": "Accelerating AI Innovation: High Performance PyTorch at AMD", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/sponsored-session-democratizing-ai-powering-the-future-with-arms-global-comp-gian-marco-iodice.json b/pytorchconf-2024/videos/sponsored-session-democratizing-ai-powering-the-future-with-arms-global-comp-gian-marco-iodice.json index 390c7aba5..59979cec8 100644 --- a/pytorchconf-2024/videos/sponsored-session-democratizing-ai-powering-the-future-with-arms-global-comp-gian-marco-iodice.json +++ b/pytorchconf-2024/videos/sponsored-session-democratizing-ai-powering-the-future-with-arms-global-comp-gian-marco-iodice.json @@ -1,5 +1,5 @@ { - "description": "Sponsored Session: Democratizing AI: Powering the Future with Arm\u2019s Global Compute Ecosystem - Gian Marco Iodice, Arm\n\nArm is excited to be at the center of the world's largest compute ecosystem at the dawn of the AI era. A key tenant of our mission is to democratize AI capabilities, empowering millions of developers to put advanced AI features into the hands of billions of users.\n\nIn this presentation, we'll explore how Arm is enabling the world\u2019s leading open-source AI frameworks to leverage power-efficient Arm-based computing platforms and Arm architecture features, as a tool for enabling fast and secure AI workloads. The session focuses on how our strategic partnership with the Pytorch and Executorch community is enabling a seamless and transparent developer experience, to run workloads everywhere from cloud to edge. This session will highlight some of our optimized libraries, upstreamed contributions and a wealth of AI-related developer material to build the future of AI on Arm.", + "description": "Arm is excited to be at the center of the world's largest compute ecosystem at the dawn of the AI era. A key tenant of our mission is to democratize AI capabilities, empowering millions of developers to put advanced AI features into the hands of billions of users.\n\nIn this presentation, we'll explore how Arm is enabling the world\u2019s leading open-source AI frameworks to leverage power-efficient Arm-based computing platforms and Arm architecture features, as a tool for enabling fast and secure AI workloads. The session focuses on how our strategic partnership with the Pytorch and Executorch community is enabling a seamless and transparent developer experience, to run workloads everywhere from cloud to edge. This session will highlight some of our optimized libraries, upstreamed contributions and a wealth of AI-related developer material to build the future of AI on Arm.", "duration": 1823, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Gian Marco Iodice" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/0D_SAurBA0s/maxresdefault.webp", - "title": "Sponsored Session: Democratizing AI: Powering the Future with Arm\u2019s Global Comp... Gian Marco Iodice", + "title": "Democratizing AI: Powering the Future with Arm\u2019s Global Compute Ecosystem", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/sponsored-session-nemo-aligner-a-scalable-toolkit-for-model-alignment-gerald-shen-jimmy-zhang.json b/pytorchconf-2024/videos/sponsored-session-nemo-aligner-a-scalable-toolkit-for-model-alignment-gerald-shen-jimmy-zhang.json index 0a5c1d26e..f825d1af2 100644 --- a/pytorchconf-2024/videos/sponsored-session-nemo-aligner-a-scalable-toolkit-for-model-alignment-gerald-shen-jimmy-zhang.json +++ b/pytorchconf-2024/videos/sponsored-session-nemo-aligner-a-scalable-toolkit-for-model-alignment-gerald-shen-jimmy-zhang.json @@ -1,5 +1,5 @@ { - "description": "Sponsored Session: NeMo-Aligner: A Scalable Toolkit for Model Alignment - Gerald Shen & Jimmy Zhang, NVIDIA\n\nAligning AI models with human values and preferences is essential for making them safe and helpful. However, building an efficient and scalable toolkit for alignment can be challenging, especially when applied to state of the art foundation models with billions or trillions of parameters. NeMo-Aligner is an open-source, optimized and scalable toolkit that implements alignment algorithms such as Reinforcement Learning from Human Feedback (RLHF), Direct Preference Optimization (DPO), SteerLM and Self-Play Fine Tuning (SPIN). This talk will introduce NeMo-Aligner and show the steps we took to design and optimize the toolkit around various alignment algorithms. In particular, we discuss the RLHF implementation where we observe close to 7x speedup and excellent scaling performance by adding TRT-LLM integration, carefully orchestrating communication and utilizing fast training kernels. We\u2019re able to align state-of-the-art open source models with NeMo-Aligner and hope our framework can enable the community to performantly customize, fine-tune and align foundational models at any scale.", + "description": "Aligning AI models with human values and preferences is essential for making them safe and helpful. However, building an efficient and scalable toolkit for alignment can be challenging, especially when applied to state of the art foundation models with billions or trillions of parameters. NeMo-Aligner is an open-source, optimized and scalable toolkit that implements alignment algorithms such as Reinforcement Learning from Human Feedback (RLHF), Direct Preference Optimization (DPO), SteerLM and Self-Play Fine Tuning (SPIN). This talk will introduce NeMo-Aligner and show the steps we took to design and optimize the toolkit around various alignment algorithms. In particular, we discuss the RLHF implementation where we observe close to 7x speedup and excellent scaling performance by adding TRT-LLM integration, carefully orchestrating communication and utilizing fast training kernels. We\u2019re able to align state-of-the-art open source models with NeMo-Aligner and hope our framework can enable the community to performantly customize, fine-tune and align foundational models at any scale.", "duration": 1275, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,12 @@ } ], "speakers": [ - "TODO" + "Gerald Shen", + "Jimmy Zhang" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/rROnWTNF-_A/maxresdefault.webp", - "title": "Sponsored Session: NeMo-Aligner: A Scalable Toolkit for Model Alignment - Gerald Shen & Jimmy Zhang", + "title": "NeMo-Aligner: A Scalable Toolkit for Model Alignment", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/sponsored-session-pytorch-support-by-google-enabling-perform-mark-sherwood-shauheen-zahirazami.json b/pytorchconf-2024/videos/sponsored-session-pytorch-support-by-google-enabling-perform-mark-sherwood-shauheen-zahirazami.json index 828a755ce..cc116b319 100644 --- a/pytorchconf-2024/videos/sponsored-session-pytorch-support-by-google-enabling-perform-mark-sherwood-shauheen-zahirazami.json +++ b/pytorchconf-2024/videos/sponsored-session-pytorch-support-by-google-enabling-perform-mark-sherwood-shauheen-zahirazami.json @@ -1,5 +1,5 @@ { - "description": "Sponsored Session: PyTorch Support by Google Enabling Performance from Cloud to Edge - Mark Sherwood & Shauheen Zahirazami, Google\n\nIn this session we will cover various ways teams at google are working to help the Pytorch community achieve performance and scale from cloud to edge. We will cover how Google Cloud customers can use PyTorch and OpenXLA to get competitive performance for their ML workloads. We\u2019ll also cover how Google AI Edge Torch works with Pytorch to help developers integrate LLMs, vision models and more to easily create new edge applications that can run on a wide set of devices.", + "description": "In this session we will cover various ways teams at google are working to help the Pytorch community achieve performance and scale from cloud to edge. We will cover how Google Cloud customers can use PyTorch and OpenXLA to get competitive performance for their ML workloads. We\u2019ll also cover how Google AI Edge Torch works with Pytorch to help developers integrate LLMs, vision models and more to easily create new edge applications that can run on a wide set of devices.", "duration": 1157, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,12 @@ } ], "speakers": [ - "TODO" + "Mark Sherwood", + "Shauheen Zahirazami" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/CzviSkYjEkQ/maxresdefault.webp", - "title": "Sponsored Session: PyTorch Support by Google Enabling Perform... Mark Sherwood & Shauheen Zahirazami", + "title": "PyTorch Support by Google Enabling Performance from Cloud to Edge", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/sponsored-session-torchchat-a-showcase-of-pytorch-llm-ubiquity-jack-khuu-jesse-white-meta.json b/pytorchconf-2024/videos/sponsored-session-torchchat-a-showcase-of-pytorch-llm-ubiquity-jack-khuu-jesse-white-meta.json index 795069270..949225214 100644 --- a/pytorchconf-2024/videos/sponsored-session-torchchat-a-showcase-of-pytorch-llm-ubiquity-jack-khuu-jesse-white-meta.json +++ b/pytorchconf-2024/videos/sponsored-session-torchchat-a-showcase-of-pytorch-llm-ubiquity-jack-khuu-jesse-white-meta.json @@ -1,5 +1,5 @@ { - "description": "Sponsored Session: Torchchat: A Showcase of PyTorch LLM Ubiquity - Jack Khuu & Jesse White, Meta\n\nThis talk explores the journey of enabling LLMs in the PyTorch ecosystem, as well as how the teams behind AOT Inductor, ExecuTorch, and torchao collaborated to create torchchat, a showcase of PyTorch\u2019s ability to run LLM inference everywhere.\n\nTorchchat demonstrates the ubiquity, simplicity, and quality of PyTorch\u2019s LLM support through performant, reproducible implementations for not only Python environments, but on desktop, server, and on-device as-well.\n\nAll of our work is open source and available on GitHub.", + "description": "This talk explores the journey of enabling LLMs in the PyTorch ecosystem, as well as how the teams behind AOT Inductor, ExecuTorch, and torchao collaborated to create torchchat, a showcase of PyTorch\u2019s ability to run LLM inference everywhere.\n\nTorchchat demonstrates the ubiquity, simplicity, and quality of PyTorch\u2019s LLM support through performant, reproducible implementations for not only Python environments, but on desktop, server, and on-device as-well.\n\nAll of our work is open source and available on GitHub.", "duration": 1459, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,12 @@ } ], "speakers": [ - "TODO" + "Jack Khuu", + "Jesse White" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/hSJ9oD6tbeU/maxresdefault.webp", - "title": "Sponsored Session: Torchchat: A Showcase of PyTorch LLM Ubiquity - Jack Khuu & Jesse White, Meta", + "title": "Torchchat: A Showcase of PyTorch LLM Ubiquity", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/startup-showcase.json b/pytorchconf-2024/videos/startup-showcase.json deleted file mode 100644 index 420b6e979..000000000 --- a/pytorchconf-2024/videos/startup-showcase.json +++ /dev/null @@ -1,24 +0,0 @@ -{ - "description": "Startup Showcase\n\nModerators: Chappy Asel\nJudges: Kevin Crosby, Astasia Myers, Rajko Radovanovic, Vijay Reddy, Vig Sachidananda, Simon Tiu\nThe PyTorch Conference Startup Showcase is giving emerging companies the chance to pitch to a panel of VCs looking to support AI/ML startups with high growth potential, and meet some of the best AI focused Engineers in the Industry. This is an exciting and unique opportunity for early-stage founders to showcase their ideas and breakthroughs, connect with leading VCs, and increase visibility in the generative AI and machine learning industry.\n\nThe winning startup will be announced at the Flare Party taking place after the Startup Showcase.\n\nFinalists:\n- Remix Inc.\n- Cartesia\n- OpenBabylon\n- Remyx AI\n- A2 Labs, Inc.\n- QuicSnap\n- Iso AI\n- CTGT\n- Creao.ai", - "duration": 3555, - "language": "eng", - "recorded": "2024-09-18", - "related_urls": [ - { - "label": "Conference Website", - "url": "https://pytorch.org/event/pytorch-conference-2024/" - } - ], - "speakers": [ - "TODO" - ], - "tags": [], - "thumbnail_url": "https://i.ytimg.com/vi_webp/xAePG2YVz7c/maxresdefault.webp", - "title": "Startup Showcase", - "videos": [ - { - "type": "youtube", - "url": "https://www.youtube.com/watch?v=xAePG2YVz7c" - } - ] -} diff --git a/pytorchconf-2024/videos/state-of-pytorch-ji-li-damien-sereni-meta.json b/pytorchconf-2024/videos/state-of-pytorch-ji-li-damien-sereni-meta.json index 617620a5e..bff4ba287 100644 --- a/pytorchconf-2024/videos/state-of-pytorch-ji-li-damien-sereni-meta.json +++ b/pytorchconf-2024/videos/state-of-pytorch-ji-li-damien-sereni-meta.json @@ -1,5 +1,5 @@ { - "description": "State of PyTorch - Ji Li & Damien Sereni, Meta\n\nThis talk gives a run through of who builds PyTorch, new and upcoming improvements to the framework and how to get involved. All thanks to our awesome community of contributors, partners and ecosystem tools.", + "description": "This talk gives a run through of who builds PyTorch, new and upcoming improvements to the framework and how to get involved. All thanks to our awesome community of contributors, partners and ecosystem tools.", "duration": 1237, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,12 @@ } ], "speakers": [ - "TODO" + "Ji Li", + "Damien Sereni" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/ibgW_ali0Gc/maxresdefault.webp", - "title": "State of PyTorch - Ji Li & Damien Sereni, Meta", + "title": "State of PyTorch", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/the-challenges-of-building-an-opinionated-open-source-llm-framework-wing-lian-axolotl-ai.json b/pytorchconf-2024/videos/the-challenges-of-building-an-opinionated-open-source-llm-framework-wing-lian-axolotl-ai.json index 1e84c623c..eadf36b55 100644 --- a/pytorchconf-2024/videos/the-challenges-of-building-an-opinionated-open-source-llm-framework-wing-lian-axolotl-ai.json +++ b/pytorchconf-2024/videos/the-challenges-of-building-an-opinionated-open-source-llm-framework-wing-lian-axolotl-ai.json @@ -1,5 +1,5 @@ { - "description": "The Challenges of Building an Opinionated Open Source LLM Framework - Wing Lian, Axolotl AI\n\nAs open-source LLMs have become more capable, a substantial ecosystem has developed around the fine-tuning of these models. A thriving community of researchers, developers, practitioners and hobbyists has emerged which focuses on topics ranging from memory efficiency, parameter-efficient fine-tuning and quantization to performance at scale and reproducible evaluations. The goal of this mini-summit is to bring this community together to discuss ideas, share knowledge and build connections.\n\nThe agenda features a keynote from Joe Spisak on the state of the Llama ecosystem followed by invited talks from the founders of Axolotl, Unsloth and torchtune. We conclude the summit with a riveting discussion on what\u2019s next for LLMs, fine-tuning and the PyTorch ecosystem with a fabulous panel of experts - Tim Dettmers (author of bitsandbytes and QLoRA), Hailey Schoelkopf (maintainer of LM Eval Harness at EleutherAI), Aakanksha Chowdhery (Lead author on PaLM and Gemini) and Alexis Conneau (Research Lead at OpenAI)", + "description": "As open-source LLMs have become more capable, a substantial ecosystem has developed around the fine-tuning of these models. A thriving community of researchers, developers, practitioners and hobbyists has emerged which focuses on topics ranging from memory efficiency, parameter-efficient fine-tuning and quantization to performance at scale and reproducible evaluations. The goal of this mini-summit is to bring this community together to discuss ideas, share knowledge and build connections.\n\nThe agenda features a keynote from Joe Spisak on the state of the Llama ecosystem followed by invited talks from the founders of Axolotl, Unsloth and torchtune. We conclude the summit with a riveting discussion on what\u2019s next for LLMs, fine-tuning and the PyTorch ecosystem with a fabulous panel of experts - Tim Dettmers (author of bitsandbytes and QLoRA), Hailey Schoelkopf (maintainer of LM Eval Harness at EleutherAI), Aakanksha Chowdhery (Lead author on PaLM and Gemini) and Alexis Conneau (Research Lead at OpenAI)", "duration": 557, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Wing Lian" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/bp9xioJv0gw/maxresdefault.webp", - "title": "The Challenges of Building an Opinionated Open Source LLM Framework - Wing Lian, Axolotl AI", + "title": "The Challenges of Building an Opinionated Open Source LLM Framework", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/the-impact-and-challenges-of-open-source-generative-datasets-and-models-aaron-gokaslan.json b/pytorchconf-2024/videos/the-impact-and-challenges-of-open-source-generative-datasets-and-models-aaron-gokaslan.json index e39d4d424..324d40249 100644 --- a/pytorchconf-2024/videos/the-impact-and-challenges-of-open-source-generative-datasets-and-models-aaron-gokaslan.json +++ b/pytorchconf-2024/videos/the-impact-and-challenges-of-open-source-generative-datasets-and-models-aaron-gokaslan.json @@ -1,5 +1,5 @@ { - "description": "The Impact and Challenges of Open Source Generative Datasets and Models - Aaron Gokaslan, Cornell University\n\nOpen source generative models like OpenGPT2, BLOOM, and others have been pivotal in advancing AI technology. These models leverage extensive text data to achieve advanced linguistic capabilities. However, the trend towards proprietary tools and closed large language models is growing, posing unique challenges in open-source AI development. This discussion will explore the intricacies of training such models, the hurdles in dataset management, and the regulation of open-source contributions. We'll explore how to effectively iterate on collected data, prepare for extensive training sessions, and coordinate research across large open-source organizations. We will discuss the challenges of generative models in three different modalities: text, image, and genomics. The talk will draw from the speaker\u2019s personal experience on working on OpenWebText, OpenGPT2, BLOOM, CommonCanvas, Caduceus, and other generative models. We will also cover the changing AI environment and how the future of open souce is threatened by onerous regulation, ever increasing compute costs, and the commoditization of previously open data.", + "description": "Open source generative models like OpenGPT2, BLOOM, and others have been pivotal in advancing AI technology. These models leverage extensive text data to achieve advanced linguistic capabilities. However, the trend towards proprietary tools and closed large language models is growing, posing unique challenges in open-source AI development. This discussion will explore the intricacies of training such models, the hurdles in dataset management, and the regulation of open-source contributions. We'll explore how to effectively iterate on collected data, prepare for extensive training sessions, and coordinate research across large open-source organizations. We will discuss the challenges of generative models in three different modalities: text, image, and genomics. The talk will draw from the speaker\u2019s personal experience on working on OpenWebText, OpenGPT2, BLOOM, CommonCanvas, Caduceus, and other generative models. We will also cover the changing AI environment and how the future of open souce is threatened by onerous regulation, ever increasing compute costs, and the commoditization of previously open data.", "duration": 1815, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Aaron Gokaslan" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/15ECoIeBf3c/maxresdefault.webp", - "title": "The Impact and Challenges of Open Source Generative Datasets and Models - Aaron Gokaslan", + "title": "The Impact and Challenges of Open Source Generative Datasets and Models", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/the-rise-of-transformers-in-the-growing-pytorch-ecosystem-arthur-zucker-hugging-face.json b/pytorchconf-2024/videos/the-rise-of-transformers-in-the-growing-pytorch-ecosystem-arthur-zucker-hugging-face.json index 4fb108cd7..69538970a 100644 --- a/pytorchconf-2024/videos/the-rise-of-transformers-in-the-growing-pytorch-ecosystem-arthur-zucker-hugging-face.json +++ b/pytorchconf-2024/videos/the-rise-of-transformers-in-the-growing-pytorch-ecosystem-arthur-zucker-hugging-face.json @@ -1,5 +1,5 @@ { - "description": "The Rise of `Transformers` in the Growing PyTorch Ecosystem - Arthur Zucker, Hugging Face\n\nExplore how the `tranformers` library grows and adapts to the fast paced and ever-changing AI field to bring the best to the AI community", + "description": "Explore how the `tranformers` library grows and adapts to the fast paced and ever-changing AI field to bring the best to the AI community", "duration": 1117, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Arthur Zucker" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/P-asaQVmA3o/maxresdefault.webp", - "title": "The Rise of `Transformers` in the Growing PyTorch Ecosystem - Arthur Zucker, Hugging Face", + "title": "The Rise of `Transformers` in the Growing PyTorch Ecosystem", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/the-state-of-the-llama-ecosystem-joe-spisak-meta.json b/pytorchconf-2024/videos/the-state-of-the-llama-ecosystem-joe-spisak-meta.json index f15037eb4..41cd96901 100644 --- a/pytorchconf-2024/videos/the-state-of-the-llama-ecosystem-joe-spisak-meta.json +++ b/pytorchconf-2024/videos/the-state-of-the-llama-ecosystem-joe-spisak-meta.json @@ -1,5 +1,5 @@ { - "description": "The State of the Llama Ecosystem - Joe Spisak, Meta\n\nAs open-source LLMs have become more capable, a substantial ecosystem has developed around the fine-tuning of these models. A thriving community of researchers, developers, practitioners and hobbyists has emerged which focuses on topics ranging from memory efficiency, parameter-efficient fine-tuning and quantization to performance at scale and reproducible evaluations. The goal of this mini-summit is to bring this community together to discuss ideas, share knowledge and build connections.\n\nThe agenda features a keynote from Joe Spisak on the state of the Llama ecosystem followed by invited talks from the founders of Axolotl, Unsloth and torchtune. We conclude the summit with a riveting discussion on what\u2019s next for LLMs, fine-tuning and the PyTorch ecosystem with a fabulous panel of experts - Tim Dettmers (author of bitsandbytes and QLoRA), Hailey Schoelkopf (maintainer of LM Eval Harness at EleutherAI), Aakanksha Chowdhery (Lead author on PaLM and Gemini) and Alexis Conneau (Research Lead at OpenAI)", + "description": "As open-source LLMs have become more capable, a substantial ecosystem has developed around the fine-tuning of these models. A thriving community of researchers, developers, practitioners and hobbyists has emerged which focuses on topics ranging from memory efficiency, parameter-efficient fine-tuning and quantization to performance at scale and reproducible evaluations. The goal of this mini-summit is to bring this community together to discuss ideas, share knowledge and build connections.\n\nThe agenda features a keynote from Joe Spisak on the state of the Llama ecosystem followed by invited talks from the founders of Axolotl, Unsloth and torchtune. We conclude the summit with a riveting discussion on what\u2019s next for LLMs, fine-tuning and the PyTorch ecosystem with a fabulous panel of experts - Tim Dettmers (author of bitsandbytes and QLoRA), Hailey Schoelkopf (maintainer of LM Eval Harness at EleutherAI), Aakanksha Chowdhery (Lead author on PaLM and Gemini) and Alexis Conneau (Research Lead at OpenAI)", "duration": 1044, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Joe Spisak" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/9_gmwWCdAlU/maxresdefault.webp", - "title": "The State of the Llama Ecosystem - Joe Spisak, Meta", + "title": "The State of the Llama Ecosystem", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/together-goes-brrr-threading-research-production-with-torch-compile-pragaash-ponnusamy.json b/pytorchconf-2024/videos/together-goes-brrr-threading-research-production-with-torch-compile-pragaash-ponnusamy.json index 37d5c78d1..b2e4d8841 100644 --- a/pytorchconf-2024/videos/together-goes-brrr-threading-research-production-with-torch-compile-pragaash-ponnusamy.json +++ b/pytorchconf-2024/videos/together-goes-brrr-threading-research-production-with-torch-compile-pragaash-ponnusamy.json @@ -1,5 +1,5 @@ { - "description": "Together Goes Brrr: Threading Research & Production with Torch Compile - Pragaash Ponnusamy, together.ai\n\nThe deployment of large language models for inference at scale is inherently complex, often requiring intricate optimizations across compute-bound and memory-bound regimes. This talk explores how PyTorch's torch.compile has revolutionized the optimization landscape for LLM serving at Together AI. Through its sophisticated Dynamo tracer and Inductor backend, torch.compile has transformed the approach to critical performance bottlenecks in both prefill and decode phases of inference. We examine how automatic vertical fusion, epilogue optimization, and adaptive kernel generation across batch sizes for GEMV and GEMM workloads, addressing key efficiency concerns, from CUDA graph captures and optimized all-reduce strategies to custom kernel registrations. The presentation highlights Together AI's journey in leveraging torch.compile to streamline the transition from research to production, significantly simplifying the deployment process for even custom architectures. By automating many performance-critical optimizations, torch.compile has not only enhanced inference efficiency but also democratized high-performance LLM deployment. We'll conclude by sharing key lessons learned and best practices gleaned from Together AI's experience in deploying torch.compile to production, serving billions of user queries and navigating the complexities of large-scale LLM inference.", + "description": "The deployment of large language models for inference at scale is inherently complex, often requiring intricate optimizations across compute-bound and memory-bound regimes. This talk explores how PyTorch's torch.compile has revolutionized the optimization landscape for LLM serving at Together AI. Through its sophisticated Dynamo tracer and Inductor backend, torch.compile has transformed the approach to critical performance bottlenecks in both prefill and decode phases of inference. We examine how automatic vertical fusion, epilogue optimization, and adaptive kernel generation across batch sizes for GEMV and GEMM workloads, addressing key efficiency concerns, from CUDA graph captures and optimized all-reduce strategies to custom kernel registrations. The presentation highlights Together AI's journey in leveraging torch.compile to streamline the transition from research to production, significantly simplifying the deployment process for even custom architectures. By automating many performance-critical optimizations, torch.compile has not only enhanced inference efficiency but also democratized high-performance LLM deployment. We'll conclude by sharing key lessons learned and best practices gleaned from Together AI's experience in deploying torch.compile to production, serving billions of user queries and navigating the complexities of large-scale LLM inference.", "duration": 566, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Pragaash Ponnusamy" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/zn0Pm2Pv3O0/maxresdefault.webp", - "title": "Together Goes Brrr: Threading Research & Production with Torch Compile - Pragaash Ponnusamy", + "title": "Together Goes Brrr: Threading Research & Production with Torch Compile", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/torch-compile-for-autograd-ddp-and-fsdp-will-feng-chien-chin-huang-simon-fan-meta.json b/pytorchconf-2024/videos/torch-compile-for-autograd-ddp-and-fsdp-will-feng-chien-chin-huang-simon-fan-meta.json index 66cb0a5f8..225b221f8 100644 --- a/pytorchconf-2024/videos/torch-compile-for-autograd-ddp-and-fsdp-will-feng-chien-chin-huang-simon-fan-meta.json +++ b/pytorchconf-2024/videos/torch-compile-for-autograd-ddp-and-fsdp-will-feng-chien-chin-huang-simon-fan-meta.json @@ -1,5 +1,5 @@ { - "description": "Torch.Compile for Autograd, DDP and FSDP - Will Feng , Chien-Chin Huang & Simon Fan, Meta\n\nIn this talk, we will present the latest advancements in torch.compile for distributed training via DDP and FSDP. We will first introduce Compiled Autograd, a torch.compile mode to fully capture the backpropagation step, including the communication collective operators used in distributed. We will then cover the improvements this new approach brought to Compiled DDP/FSDP, notably by removing DDP/FSDP graph breaks which brings the potential of improving compute/communication overlap.", + "description": "In this talk, we will present the latest advancements in torch.compile for distributed training via DDP and FSDP. We will first introduce Compiled Autograd, a torch.compile mode to fully capture the backpropagation step, including the communication collective operators used in distributed. We will then cover the improvements this new approach brought to Compiled DDP/FSDP, notably by removing DDP/FSDP graph breaks which brings the potential of improving compute/communication overlap.", "duration": 1343, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,13 @@ } ], "speakers": [ - "TODO" + "Will Feng", + "Chien-Chin Huang", + "Simon Fan" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/1XuibaVRewc/maxresdefault.webp", - "title": "Torch.Compile for Autograd, DDP and FSDP - Will Feng , Chien-Chin Huang & Simon Fan, Meta", + "title": "Torch.Compile for Autograd, DDP and FSDP", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/torchinductor-cpu-backend-advancements-new-features-and-performance-imp-jiong-gong-leslie-fang.json b/pytorchconf-2024/videos/torchinductor-cpu-backend-advancements-new-features-and-performance-imp-jiong-gong-leslie-fang.json index b344f7367..d20caa4fc 100644 --- a/pytorchconf-2024/videos/torchinductor-cpu-backend-advancements-new-features-and-performance-imp-jiong-gong-leslie-fang.json +++ b/pytorchconf-2024/videos/torchinductor-cpu-backend-advancements-new-features-and-performance-imp-jiong-gong-leslie-fang.json @@ -1,5 +1,5 @@ { - "description": "TorchInductor CPU Backend Advancements: New Features and Performance Improvements - Jiong Gong & Leslie Fang, Intel\n\nThis presentation provides an update on the latest advancements in the TorchInductor CPU backend since the last conference to bring best-in-class CPU performance for broad DL workloads. We will discuss new features and performance enhancements, including: \u2022 Max-autotune support with codegen for GEMMs, boosting performance for GEMM-related operations \u2022 Enhanced vectorized codegen support, now covering all data types beyond floating points with flexible vector factors, and optimized loop scheduling \u2022 Comprehensive quantization support, including weight-only-quantization (WoQ), and optimizations for dynamic quantization and quantization-aware training \u2022 Improved Attention support, featuring attention masks and optimizating SoftMax via flash attention v2 etc. \u2022 AOTInductor support, enabling high-performance inference with frozen weights \u2022 Native Windows support, with improved vectorization capabilities These advancements, combined with ongoing optimizations, have resulted in significant performance improvements since PyTorch 2.1, demonstrated through extensive benchmarks and large language models (LLMs).", + "description": "This presentation provides an update on the latest advancements in the TorchInductor CPU backend since the last conference to bring best-in-class CPU performance for broad DL workloads. We will discuss new features and performance enhancements, including: \u2022 Max-autotune support with codegen for GEMMs, boosting performance for GEMM-related operations \u2022 Enhanced vectorized codegen support, now covering all data types beyond floating points with flexible vector factors, and optimized loop scheduling \u2022 Comprehensive quantization support, including weight-only-quantization (WoQ), and optimizations for dynamic quantization and quantization-aware training \u2022 Improved Attention support, featuring attention masks and optimizating SoftMax via flash attention v2 etc. \u2022 AOTInductor support, enabling high-performance inference with frozen weights \u2022 Native Windows support, with improved vectorization capabilities These advancements, combined with ongoing optimizations, have resulted in significant performance improvements since PyTorch 2.1, demonstrated through extensive benchmarks and large language models (LLMs).", "duration": 1312, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,12 @@ } ], "speakers": [ - "TODO" + "Jiong Gong", + "Leslie Fang" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/VwmOgzXtxYw/maxresdefault.webp", - "title": "TorchInductor CPU Backend Advancements: New Features and Performance Imp... Jiong Gong & Leslie Fang", + "title": "TorchInductor CPU Backend Advancements: New Features and Performance Improvements", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/torchtitan-large-scale-llm-training-using-native-pytorch-3d-parallel-wanchao-liang-linsong-chu.json b/pytorchconf-2024/videos/torchtitan-large-scale-llm-training-using-native-pytorch-3d-parallel-wanchao-liang-linsong-chu.json index 3d49f2774..672f407e5 100644 --- a/pytorchconf-2024/videos/torchtitan-large-scale-llm-training-using-native-pytorch-3d-parallel-wanchao-liang-linsong-chu.json +++ b/pytorchconf-2024/videos/torchtitan-large-scale-llm-training-using-native-pytorch-3d-parallel-wanchao-liang-linsong-chu.json @@ -1,5 +1,5 @@ { - "description": "Torchtitan: Large-Scale LLM Training Using Native PyTorch 3D Parallelism - Wanchao Liang, Meta & Linsong Chu, IBM Research\n\ntorchtitan is a proof-of-concept for Large-scale LLM training using native PyTorch. It is a repo that showcases PyTorch's latest distributed training features in a clean, minimal codebase. We show-cased end to end large scale training features enablement: 1. 3D/4D Parallelism 2. Efficient distributed checkpoint save/load/resharding 3. Many efficient training techniques including Float8, torch.compile, activation checkpoint, etc.", + "description": "torchtitan is a proof-of-concept for Large-scale LLM training using native PyTorch. It is a repo that showcases PyTorch's latest distributed training features in a clean, minimal codebase. We show-cased end to end large scale training features enablement: 1. 3D/4D Parallelism 2. Efficient distributed checkpoint save/load/resharding 3. Many efficient training techniques including Float8, torch.compile, activation checkpoint, etc.", "duration": 1460, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,12 @@ } ], "speakers": [ - "TODO" + "Wanchao Liang", + "Linsong Chu" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/WsNEBxPDljU/maxresdefault.webp", - "title": "Torchtitan: Large-Scale LLM Training Using Native PyTorch 3D Parallel... Wanchao Liang & Linsong Chu", + "title": "Torchtitan: Large-Scale LLM Training Using Native PyTorch 3D Parallelism", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/torchtune-easy-and-accessible-finetuning-in-native-pytorch-evan-smothers-meta.json b/pytorchconf-2024/videos/torchtune-easy-and-accessible-finetuning-in-native-pytorch-evan-smothers-meta.json index d90d2bd5b..01044c26c 100644 --- a/pytorchconf-2024/videos/torchtune-easy-and-accessible-finetuning-in-native-pytorch-evan-smothers-meta.json +++ b/pytorchconf-2024/videos/torchtune-easy-and-accessible-finetuning-in-native-pytorch-evan-smothers-meta.json @@ -1,5 +1,5 @@ { - "description": "torchtune: Easy and Accessible Finetuning in Native PyTorch - Evan Smothers, Meta\n\nAs open-source LLMs have become more capable, a substantial ecosystem has developed around the fine-tuning of these models. A thriving community of researchers, developers, practitioners and hobbyists has emerged which focuses on topics ranging from memory efficiency, parameter-efficient fine-tuning and quantization to performance at scale and reproducible evaluations. The goal of this mini-summit is to bring this community together to discuss ideas, share knowledge and build connections.\n\nThe agenda features a keynote from Joe Spisak on the state of the Llama ecosystem followed by invited talks from the founders of Axolotl, Unsloth and torchtune. We conclude the summit with a riveting discussion on what\u2019s next for LLMs, fine-tuning and the PyTorch ecosystem with a fabulous panel of experts - Tim Dettmers (author of bitsandbytes and QLoRA), Hailey Schoelkopf (maintainer of LM Eval Harness at EleutherAI), Aakanksha Chowdhery (Lead author on PaLM and Gemini) and Alexis Conneau (Research Lead at OpenAI)", + "description": "As open-source LLMs have become more capable, a substantial ecosystem has developed around the fine-tuning of these models. A thriving community of researchers, developers, practitioners and hobbyists has emerged which focuses on topics ranging from memory efficiency, parameter-efficient fine-tuning and quantization to performance at scale and reproducible evaluations. The goal of this mini-summit is to bring this community together to discuss ideas, share knowledge and build connections.\n\nThe agenda features a keynote from Joe Spisak on the state of the Llama ecosystem followed by invited talks from the founders of Axolotl, Unsloth and torchtune. We conclude the summit with a riveting discussion on what\u2019s next for LLMs, fine-tuning and the PyTorch ecosystem with a fabulous panel of experts - Tim Dettmers (author of bitsandbytes and QLoRA), Hailey Schoelkopf (maintainer of LM Eval Harness at EleutherAI), Aakanksha Chowdhery (Lead author on PaLM and Gemini) and Alexis Conneau (Research Lead at OpenAI)", "duration": 958, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Evan Smothers" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/43X9E25-Qg0/maxresdefault.webp", - "title": "torchtune: Easy and Accessible Finetuning in Native PyTorch - Evan Smothers, Meta", + "title": "torchtune: Easy and Accessible Finetuning in Native PyTorch", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/training-moes-at-scale-with-pytorch-mihir-patel-brian-chu-databricks.json b/pytorchconf-2024/videos/training-moes-at-scale-with-pytorch-mihir-patel-brian-chu-databricks.json index fa02c680a..6450167a6 100644 --- a/pytorchconf-2024/videos/training-moes-at-scale-with-pytorch-mihir-patel-brian-chu-databricks.json +++ b/pytorchconf-2024/videos/training-moes-at-scale-with-pytorch-mihir-patel-brian-chu-databricks.json @@ -1,5 +1,5 @@ { - "description": "Training MoEs at Scale with PyTorch - Mihir Patel & Brian Chu, Databricks\n\nMixture-of-Experts MoE (models) are becoming an increasingly popular architecture choice for large language models (LLMs). In this talk, we describe how to train MoE models with PyTorch. After discussing various performance tradeoffs, we use PyTorch distributed tools like DTensor to build custom parallelism approaches, including expert parallelism via MegaBlocks. We then show how to get near linear scaling to thousands of GPUs, combining PyTorch FSDP and HSDP with our parallelism strategies. We discuss many of the challenges of training at scale, including communication bottlenecks, hardware failures, and networking challenges. We further improve training at scale setups using tools like PyTorch Distributed Checkpointing for rapid saving and loading. We then highlight further optimizations to minimize challenges only present at scale, such as object store failures for large checkpoints.", + "description": "Mixture-of-Experts MoE (models) are becoming an increasingly popular architecture choice for large language models (LLMs). In this talk, we describe how to train MoE models with PyTorch. After discussing various performance tradeoffs, we use PyTorch distributed tools like DTensor to build custom parallelism approaches, including expert parallelism via MegaBlocks. We then show how to get near linear scaling to thousands of GPUs, combining PyTorch FSDP and HSDP with our parallelism strategies. We discuss many of the challenges of training at scale, including communication bottlenecks, hardware failures, and networking challenges. We further improve training at scale setups using tools like PyTorch Distributed Checkpointing for rapid saving and loading. We then highlight further optimizations to minimize challenges only present at scale, such as object store failures for large checkpoints.", "duration": 1132, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,12 @@ } ], "speakers": [ - "TODO" + "Mihir Patel", + "Brian Chu" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/f2OxAWRCmPo/maxresdefault.webp", - "title": "Training MoEs at Scale with PyTorch - Mihir Patel & Brian Chu, Databricks", + "title": "Training MoEs at Scale with PyTorch", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/triton-maximizing-kernel-development-productivity-under-performance-constraints-philip-tillet.json b/pytorchconf-2024/videos/triton-maximizing-kernel-development-productivity-under-performance-constraints-philip-tillet.json index 8c8316742..b7fc6a97d 100644 --- a/pytorchconf-2024/videos/triton-maximizing-kernel-development-productivity-under-performance-constraints-philip-tillet.json +++ b/pytorchconf-2024/videos/triton-maximizing-kernel-development-productivity-under-performance-constraints-philip-tillet.json @@ -1,5 +1,5 @@ { - "description": "[TRITON] Maximizing Kernel Development Productivity Under Performance Constraints - Philip Tillet, OpenAI\n\nMachine Learning research workflows are often bottlenecked by the development of compute kernels for new algorithms and GPU architectures. This process can be daunting, and often requires a careful trade-off between productivity and performance. In this talk, we will discuss how Triton -- a mid-level programming language for kernel development -- approaches this multi-objective optimization problem, and the design decisions that were made to that effect.", + "description": "Machine Learning research workflows are often bottlenecked by the development of compute kernels for new algorithms and GPU architectures. This process can be daunting, and often requires a careful trade-off between productivity and performance. In this talk, we will discuss how Triton -- a mid-level programming language for kernel development -- approaches this multi-objective optimization problem, and the design decisions that were made to that effect.", "duration": 601, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Philip Tillet" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/h449o7UijHk/maxresdefault.webp", - "title": "[TRITON] Maximizing Kernel Development Productivity Under Performance Constraints - Philip Tillet", + "title": "Maximizing Kernel Development Productivity Under Performance Constraints", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/tvm-universally-deploy-large-language-models-via-ml-compilation-tianqi-chen-cmu-octoai.json b/pytorchconf-2024/videos/tvm-universally-deploy-large-language-models-via-ml-compilation-tianqi-chen-cmu-octoai.json index 4d110be30..ac26dbe79 100644 --- a/pytorchconf-2024/videos/tvm-universally-deploy-large-language-models-via-ml-compilation-tianqi-chen-cmu-octoai.json +++ b/pytorchconf-2024/videos/tvm-universally-deploy-large-language-models-via-ml-compilation-tianqi-chen-cmu-octoai.json @@ -1,5 +1,5 @@ { - "description": "[TVM] Universally Deploy Large-language Models via ML Compilation - Tianqi Chen, CMU & OctoAI\n\nDeploying deep learning models on various devices has become an important topic. Machine learning compilation is an emerging field that leverages compiler and automatic search techniques to accelerate AI models. ML compilation brings a unique set of challenges: emerging machine learning models; increasing hardware specialization brings a diverse set of acceleration primitives; growing tension between flexibility and performance. In this talk. I then discuss our experience in bringing foundational models to a variety of devices and hardware environments through machine learning compilation.", + "description": "Deploying deep learning models on various devices has become an important topic. Machine learning compilation is an emerging field that leverages compiler and automatic search techniques to accelerate AI models. ML compilation brings a unique set of challenges: emerging machine learning models; increasing hardware specialization brings a diverse set of acceleration primitives; growing tension between flexibility and performance. In this talk. I then discuss our experience in bringing foundational models to a variety of devices and hardware environments through machine learning compilation.", "duration": 650, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Tianqi Chen" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/n66LLI_ebx0/maxresdefault.webp", - "title": "[TVM] Universally Deploy Large-language Models via ML Compilation - Tianqi Chen, CMU & OctoAI", + "title": "Universally Deploy Large-language Models via ML Compilation", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/understanding-the-llm-inference-workload-mark-moyou-nvidia.json b/pytorchconf-2024/videos/understanding-the-llm-inference-workload-mark-moyou-nvidia.json index 9aba442c0..1635f196e 100644 --- a/pytorchconf-2024/videos/understanding-the-llm-inference-workload-mark-moyou-nvidia.json +++ b/pytorchconf-2024/videos/understanding-the-llm-inference-workload-mark-moyou-nvidia.json @@ -1,5 +1,5 @@ { - "description": "Understanding the LLM Inference Workload - Mark Moyou, NVIDIA\n\nUnderstanding how to effectively size a production grade LLM deployment requires understanding of the model(s), the compute hardware, quantization and parallelization methods, KV Cache budgets, input and output token length predictions, model adapter management and much more. - Why LLM inference is different to standard deep learning inference - Current and future NVIDIA GPU overview - which GPU(s) for which models and why - Understanding the importance of building inference engines - Deep recap on the attention mechanism along with different types of popular attention mechanisms used in production - Deep dive on KV Cache and managing KV Cache budgets - Parallelism (reducing latency) - mainly tensor parallelism, but data, sequence, pipeline, and expert parallelism will be highlighted - Quantization methods on weights, activations, and KV Cache to reduce engine sizes for more effective GPU utilization - Increasing throughput with inflight batching and other techniques - Detailed performance analysis of LLM deployments looking at Time to first token, inter-token latencies, llm deployment characterizations, and more that can help reduce deployment costs", + "description": "Understanding how to effectively size a production grade LLM deployment requires understanding of the model(s), the compute hardware, quantization and parallelization methods, KV Cache budgets, input and output token length predictions, model adapter management and much more. - Why LLM inference is different to standard deep learning inference - Current and future NVIDIA GPU overview - which GPU(s) for which models and why - Understanding the importance of building inference engines - Deep recap on the attention mechanism along with different types of popular attention mechanisms used in production - Deep dive on KV Cache and managing KV Cache budgets - Parallelism (reducing latency) - mainly tensor parallelism, but data, sequence, pipeline, and expert parallelism will be highlighted - Quantization methods on weights, activations, and KV Cache to reduce engine sizes for more effective GPU utilization - Increasing throughput with inflight batching and other techniques - Detailed performance analysis of LLM deployments looking at Time to first token, inter-token latencies, llm deployment characterizations, and more that can help reduce deployment costs", "duration": 2054, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Mark Moyou" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/z2M8gKGYws4/maxresdefault.webp", - "title": "Understanding the LLM Inference Workload - Mark Moyou, NVIDIA", + "title": "Understanding the LLM Inference Workload", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/unlocking-the-enigma-crafting-unbiased-transparent-and-explainable-large-languag-rashmi-nagpal.json b/pytorchconf-2024/videos/unlocking-the-enigma-crafting-unbiased-transparent-and-explainable-large-languag-rashmi-nagpal.json index 975f79ddf..e67d1dfce 100644 --- a/pytorchconf-2024/videos/unlocking-the-enigma-crafting-unbiased-transparent-and-explainable-large-languag-rashmi-nagpal.json +++ b/pytorchconf-2024/videos/unlocking-the-enigma-crafting-unbiased-transparent-and-explainable-large-languag-rashmi-nagpal.json @@ -1,5 +1,5 @@ { - "description": "Unlocking the Enigma: Crafting Unbiased, Transparent, and Explainable Large Language Models - Rashmi Nagpal, Patchstack\n\nIn an era where artificial intelligence reigns supreme, the statistics are both perplexing and thought-provoking \u2013 only a mere 13% of large language models manage to transcend the realms of research and enter the practical world of production. Who bears the responsibility when these models err, spewing out biased or discriminatory outputs? It's time to demystify the complex landscape of machine learning ethics and carve a path towards a brighter, more accountable future! In this talk, firstly, we will navigate the profound impacts of large language models across diverse domains, from the lifesaving advances in medicine to safeguarding our nations through enhanced security protocols. Secondly, as we marvel at data-driven decisions laid by these models, we will confront the darker shadows cast by \u2013 the looming spectre of bias in the data. Finally, we will delve deep into the art of building interpretable models and navigating the maze of ethical considerations. Through a live demonstration in PyTorch, we will witness how to craft unbiased, transparent, and explainable models.", + "description": "In an era where artificial intelligence reigns supreme, the statistics are both perplexing and thought-provoking \u2013 only a mere 13% of large language models manage to transcend the realms of research and enter the practical world of production. Who bears the responsibility when these models err, spewing out biased or discriminatory outputs? It's time to demystify the complex landscape of machine learning ethics and carve a path towards a brighter, more accountable future! In this talk, firstly, we will navigate the profound impacts of large language models across diverse domains, from the lifesaving advances in medicine to safeguarding our nations through enhanced security protocols. Secondly, as we marvel at data-driven decisions laid by these models, we will confront the darker shadows cast by \u2013 the looming spectre of bias in the data. Finally, we will delve deep into the art of building interpretable models and navigating the maze of ethical considerations. Through a live demonstration in PyTorch, we will witness how to craft unbiased, transparent, and explainable models.", "duration": 2998, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Rashmi Nagpal" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/vGHb4jO2_Ks/maxresdefault.webp", - "title": "Unlocking the Enigma: Crafting Unbiased, Transparent, and Explainable Large Languag... Rashmi Nagpal", + "title": "Unlocking the Enigma: Crafting Unbiased, Transparent, and Explainable Large Language Models", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/vllm-easy-fast-and-cheap-llm-serving-for-everyone-woosuk-kwon-xiaoxuan-liu-uc-berkeley.json b/pytorchconf-2024/videos/vllm-easy-fast-and-cheap-llm-serving-for-everyone-woosuk-kwon-xiaoxuan-liu-uc-berkeley.json index 307847e02..60e678f5e 100644 --- a/pytorchconf-2024/videos/vllm-easy-fast-and-cheap-llm-serving-for-everyone-woosuk-kwon-xiaoxuan-liu-uc-berkeley.json +++ b/pytorchconf-2024/videos/vllm-easy-fast-and-cheap-llm-serving-for-everyone-woosuk-kwon-xiaoxuan-liu-uc-berkeley.json @@ -1,5 +1,5 @@ { - "description": "vLLM: Easy, Fast, and Cheap LLM Serving for Everyone - Woosuk Kwon & Xiaoxuan Liu, UC Berkeley\n\nWe will present vLLM, an open-source high-performance LLM inference engine built on top of PyTorch. Starting as a research project at UC Berkeley, vLLM has been one of the fastest and most popular LLM inference solutions in industry, reaching 20K+ stars and 350+ contributors. In this talk, we will cover how vLLM adopts various LLM inference optimizations and how it supports various AI accelerators such as AMD GPUs, Google TPUs, and AWS Inferentia. Also, we will discuss how vLLM benefits from PyTorch 2 and its ecosystem.", + "description": "We will present vLLM, an open-source high-performance LLM inference engine built on top of PyTorch. Starting as a research project at UC Berkeley, vLLM has been one of the fastest and most popular LLM inference solutions in industry, reaching 20K+ stars and 350+ contributors. In this talk, we will cover how vLLM adopts various LLM inference optimizations and how it supports various AI accelerators such as AMD GPUs, Google TPUs, and AWS Inferentia. Also, we will discuss how vLLM benefits from PyTorch 2 and its ecosystem.", "duration": 1412, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,12 @@ } ], "speakers": [ - "TODO" + "Woosuk Kwon", + "Xiaoxuan Liu" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/9ih0EmcXRHE/maxresdefault.webp", - "title": "vLLM: Easy, Fast, and Cheap LLM Serving for Everyone - Woosuk Kwon & Xiaoxuan Liu, UC Berkeley", + "title": "vLLM: Easy, Fast, and Cheap LLM Serving for Everyone", "videos": [ { "type": "youtube", From f0f907f229d277c8830baf20147e17cf0eda254c Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Ezequiel=20Leonardo=20Casta=C3=B1o?= <14986783+ELC@users.noreply.github.com> Date: Fri, 21 Nov 2025 19:51:41 -0300 Subject: [PATCH 6/8] Complete processing of pytorchconf-2024: All 90 JSON files - Extracted speakers from titles and descriptions - Removed author names and organizations from titles - Removed title prefixes from descriptions - Removed Lightning Talk: and Sponsored Keynote: prefixes from titles - Processed all regular sessions, keynotes, sponsored sessions, lightning talks - Cleaned up panel discussions, startup showcase, and other special sessions --- ...based-on-torch-fx-simplify-zifei-xu-tristan-webb.json | 7 ++++--- ...ls-in-the-wild-lessons-learned-f-omar-sanseviero.json | 6 +++--- ...resources-for-running-generative-a-pareena-verma.json | 6 +++--- ...-custom-python-c-cuda-operators-richard-zou-meta.json | 6 +++--- ...d-training-with-streamingdataset-saaketh-narayan.json | 6 +++--- ...ty-of-pytorch-the-performa-yanbo-liang-horace-he.json | 7 ++++--- ...d-pipeline-for-automated-egyptian-h-susi-gentsch.json | 6 +++--- ...le-datasets-what-could-go-wrong-nicolas-hug-meta.json | 6 +++--- ...ytorch-on-frontier-exascale-system-vineeth-gutta.json | 6 +++--- ...-distributed-pipelining-howard-huang-ke-wen-meta.json | 7 ++++--- ...lerators-chen-lai-kimish-patel-cemal-bilgin-meta.json | 8 +++++--- ...ecision-dtypes-in-pytorch-vasiliy-kuznetsov-meta.json | 6 +++--- ...-memory-capacity-using-pytorch-syed-ahmed-nvidia.json | 6 +++--- ...-with-pytorch-low-light-denoising-alexis-baudron.json | 6 +++--- ...nting-apis-in-pytorch-jeffrey-wan-horace-he-meta.json | 7 ++++--- ...ugging-with-executorch-olivia-liu-vaun-puri-meta.json | 7 ++++--- ...-aarch64-linux-cpus-sunita-nadampalli-amazon-aws.json | 6 +++--- ...-talk-pytorch-release-process-andrey-talman-meta.json | 6 +++--- ...k-pytorch-xla-auto-sharding-yeounoh-chung-google.json | 6 +++--- ...ormers-with-minimal-accuracy-loss-jesse-cai-meta.json | 6 +++--- ...rch-models-with-thunder-luca-antiga-lightning-ai.json | 6 +++--- ...per-infrastructure-sahan-paliskara-catherine-lee.json | 7 ++++--- ...gsbayasgalan-manlaibaatar-avik-chaudhuri-yidi-wu.json | 9 ++++++--- 23 files changed, 80 insertions(+), 69 deletions(-) diff --git a/pytorchconf-2024/videos/lightning-talk-d-matrix-llm-compression-flow-based-on-torch-fx-simplify-zifei-xu-tristan-webb.json b/pytorchconf-2024/videos/lightning-talk-d-matrix-llm-compression-flow-based-on-torch-fx-simplify-zifei-xu-tristan-webb.json index f9e148902..d5a492fee 100644 --- a/pytorchconf-2024/videos/lightning-talk-d-matrix-llm-compression-flow-based-on-torch-fx-simplify-zifei-xu-tristan-webb.json +++ b/pytorchconf-2024/videos/lightning-talk-d-matrix-llm-compression-flow-based-on-torch-fx-simplify-zifei-xu-tristan-webb.json @@ -1,5 +1,5 @@ { - "description": "Lightning Talk: d-Matrix LLM Compression Flow Based on Torch.Fx: Simplifying PTQ/QAT - Zifei Xu & Tristan Webb, d-Matrix Corporation\n\nWe introduce dmx-compressor, d-Matrix's open-source LLM compression toolkit that is modular, robust, efficient, and user-friendly. It utilizes symbolic tracing and fx.Transformer for network compression while keeping the model a first-class citizen in PyTorch for the user, despite prevalent graph dynamism in LLMs. It achieves this by maintaining both the original nn.Module and a just-in-time (JIT) traced and transformed fx.GraphModule representation behind the scenes, in conjunction with an abstraction that cleanly decouples network compression from the original model graph definition. This design allows the FXIR to dynamically adapt to diverse forward call signatures and flow-control arguments throughout quantization-aware training and post-training quantization written in plain PyTorch, yielding a compressed FXIR fully compatible with application-level APIs like the Hugging Face pipeline. We also provide a graph visualizer based on fx.Interpreter for ease of debugging. We believe this project shall empower the community to build efficient LLMs for deployment on custom hardware accelerators and contribute to the PyTorch ecosystem.", + "description": "We introduce dmx-compressor, d-Matrix's open-source LLM compression toolkit that is modular, robust, efficient, and user-friendly. It utilizes symbolic tracing and fx.Transformer for network compression while keeping the model a first-class citizen in PyTorch for the user, despite prevalent graph dynamism in LLMs. It achieves this by maintaining both the original nn.Module and a just-in-time (JIT) traced and transformed fx.GraphModule representation behind the scenes, in conjunction with an abstraction that cleanly decouples network compression from the original model graph definition. This design allows the FXIR to dynamically adapt to diverse forward call signatures and flow-control arguments throughout quantization-aware training and post-training quantization written in plain PyTorch, yielding a compressed FXIR fully compatible with application-level APIs like the Hugging Face pipeline. We also provide a graph visualizer based on fx.Interpreter for ease of debugging. We believe this project shall empower the community to build efficient LLMs for deployment on custom hardware accelerators and contribute to the PyTorch ecosystem.", "duration": 822, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,12 @@ } ], "speakers": [ - "TODO" + "Zifei Xu", + "Tristan Webb" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi/cw6dSlmMwj8/maxresdefault.jpg", - "title": "Lightning Talk: d-Matrix LLM Compression Flow Based on Torch.Fx: Simplify... Zifei Xu & Tristan Webb", + "title": "d-Matrix LLM Compression Flow Based on Torch.Fx: Simplifying PTQ/QAT", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/lightning-talk-distributing-a-million-open-models-in-the-wild-lessons-learned-f-omar-sanseviero.json b/pytorchconf-2024/videos/lightning-talk-distributing-a-million-open-models-in-the-wild-lessons-learned-f-omar-sanseviero.json index 63df10978..37914c830 100644 --- a/pytorchconf-2024/videos/lightning-talk-distributing-a-million-open-models-in-the-wild-lessons-learned-f-omar-sanseviero.json +++ b/pytorchconf-2024/videos/lightning-talk-distributing-a-million-open-models-in-the-wild-lessons-learned-f-omar-sanseviero.json @@ -1,5 +1,5 @@ { - "description": "Lightning Talk: Distributing a Million Open Models in the Wild: Lessons Learned from the Hugging Face Hub - Omar Sanseviero, Hugging Face\n\nThe Hugging Face Hub has over 300,000 PyTorch models. Distributing such number of models poses challenges. In this talk, Omar will share how the community has tackled these challenges, including techniques to ensure torch model security and tooling for researchers to share their models. He'll also take attendees on a journey through the evolution of torch models distributed by the community, highlighting new trends and directions. Attending this talk will give attendees practical insights into the latest developments in model distribution and ecosystem trends.", + "description": "The Hugging Face Hub has over 300,000 PyTorch models. Distributing such number of models poses challenges. In this talk, Omar will share how the community has tackled these challenges, including techniques to ensure torch model security and tooling for researchers to share their models. He'll also take attendees on a journey through the evolution of torch models distributed by the community, highlighting new trends and directions. Attending this talk will give attendees practical insights into the latest developments in model distribution and ecosystem trends.", "duration": 579, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Omar Sanseviero" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/t3AHZzz8Ong/maxresdefault.webp", - "title": "Lightning Talk: Distributing a Million Open Models in the Wild: Lessons Learned f... Omar Sanseviero", + "title": "Distributing a Million Open Models in the Wild: Lessons Learned from the Hugging Face Hub", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/lightning-talk-empowering-developers-tools-and-resources-for-running-generative-a-pareena-verma.json b/pytorchconf-2024/videos/lightning-talk-empowering-developers-tools-and-resources-for-running-generative-a-pareena-verma.json index d35489e6c..eaa9b9659 100644 --- a/pytorchconf-2024/videos/lightning-talk-empowering-developers-tools-and-resources-for-running-generative-a-pareena-verma.json +++ b/pytorchconf-2024/videos/lightning-talk-empowering-developers-tools-and-resources-for-running-generative-a-pareena-verma.json @@ -1,5 +1,5 @@ { - "description": "Lightning Talk: Empowering Developers: Tools and Resources for Running Generative AI on Arm CPUs - Pareena Verma, Arm\n\nAs the demand for accessible and scalable AI solutions grows, leveraging CPUs for generative AI offers significant advantages in cost, energy efficiency and widespread availability. This sessions aims to equip developers with the ecosystem of tools, resources and technical content needed to effectively run generative AI use cases on Arm CPUs. We have launched a range of easily digestible tutorials for developers, part of our Learning Paths on https://learn.arm.com/, which demonstrate how you can easily and efficiently run small and large language models on Arm-based devices. Learn about end-to-end workflows to accelerate PyTorch based sentiment analysis models from Hugging Face on Arm servers with optimizations in Arm Compute Library kernels for fp32 and bfloat16. Use the new KleidiAI library to accelerate LLMs with AI frameworks and build an Android chat app on your Arm mobile device with ExecuTorch, and XNNPACK. Find out about our roadmap for learning content demonstrating the feasibility and successful deployment of generative AI on Arm-based devices. Help us shape the support that we offer developers.", + "description": "As the demand for accessible and scalable AI solutions grows, leveraging CPUs for generative AI offers significant advantages in cost, energy efficiency and widespread availability. This sessions aims to equip developers with the ecosystem of tools, resources and technical content needed to effectively run generative AI use cases on Arm CPUs. We have launched a range of easily digestible tutorials for developers, part of our Learning Paths on https://learn.arm.com/, which demonstrate how you can easily and efficiently run small and large language models on Arm-based devices. Learn about end-to-end workflows to accelerate PyTorch based sentiment analysis models from Hugging Face on Arm servers with optimizations in Arm Compute Library kernels for fp32 and bfloat16. Use the new KleidiAI library to accelerate LLMs with AI frameworks and build an Android chat app on your Arm mobile device with ExecuTorch, and XNNPACK. Find out about our roadmap for learning content demonstrating the feasibility and successful deployment of generative AI on Arm-based devices. Help us shape the support that we offer developers.", "duration": 738, "language": "eng", "recorded": "2024-09-18", @@ -14,11 +14,11 @@ } ], "speakers": [ - "TODO" + "Pareena Verma" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/wkzBHF9CL5s/maxresdefault.webp", - "title": "Lightning Talk: Empowering Developers: Tools and Resources for Running Generative A... Pareena Verma", + "title": "Empowering Developers: Tools and Resources for Running Generative AI on Arm CPUs", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/lightning-talk-extending-pytorch-with-custom-python-c-cuda-operators-richard-zou-meta.json b/pytorchconf-2024/videos/lightning-talk-extending-pytorch-with-custom-python-c-cuda-operators-richard-zou-meta.json index 4e112aaf1..28050a7ac 100644 --- a/pytorchconf-2024/videos/lightning-talk-extending-pytorch-with-custom-python-c-cuda-operators-richard-zou-meta.json +++ b/pytorchconf-2024/videos/lightning-talk-extending-pytorch-with-custom-python-c-cuda-operators-richard-zou-meta.json @@ -1,5 +1,5 @@ { - "description": "Lightning Talk: Extending PyTorch with Custom Python/C++/CUDA Operators - Richard Zou, Meta\n\nIn this talk, we'll go over the new recommended APIs to extend PyTorch with custom Python/C++/CUDA operators. Users have been able to extend PyTorch with custom operators for years but we have updated our guidance for creating custom operators that compose with torch.compile, autograd, and other PyTorch subsystems.", + "description": "In this talk, we'll go over the new recommended APIs to extend PyTorch with custom Python/C++/CUDA operators. Users have been able to extend PyTorch with custom operators for years but we have updated our guidance for creating custom operators that compose with torch.compile, autograd, and other PyTorch subsystems.", "duration": 511, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Richard Zou" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/LI3h8aVchwo/maxresdefault.webp", - "title": "Lightning Talk: Extending PyTorch with Custom Python/C++/CUDA Operators - Richard Zou, Meta", + "title": "Extending PyTorch with Custom Python/C++/CUDA Operators", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/lightning-talk-fast-scalable-distributed-training-with-streamingdataset-saaketh-narayan.json b/pytorchconf-2024/videos/lightning-talk-fast-scalable-distributed-training-with-streamingdataset-saaketh-narayan.json index bd066bec5..9a63f0c39 100644 --- a/pytorchconf-2024/videos/lightning-talk-fast-scalable-distributed-training-with-streamingdataset-saaketh-narayan.json +++ b/pytorchconf-2024/videos/lightning-talk-fast-scalable-distributed-training-with-streamingdataset-saaketh-narayan.json @@ -1,5 +1,5 @@ { - "description": "Lightning Talk: Fast, Scalable Distributed Training with StreamingDataset - Saaketh Narayan, Databricks\n\nStreamingDataset makes training on large datasets from cloud storage as fast, cheap, and scalable as possible. It\u2019s specially designed for multi-node, distributed training for large models \u2014 maximizing correctness guarantees, performance, and ease of use. Key features include elastically deterministic training, instant mid-epoch resumption, effective shuffling, high training throughput, and flexible data mixing, among other features. When training with StreamingDataset, the data shards are written to cloud storage in MDS, our file format that allows for low-latency random access to samples. By being as efficient as possible with shard downloads and shuffling, StreamingDataset minimizes egress costs while ensuring that dataloading never bottlenecks model training. StreamingDataset powers training for LLMs with over 100 billion parameters like DBRX, to advanced diffusion models, to two-tower recommendation models, and more, scaling to training jobs on thousands of GPUs with ease. Join us to learn how StreamingDataset can elevate your distributed model training experience.", + "description": "StreamingDataset makes training on large datasets from cloud storage as fast, cheap, and scalable as possible. It\u2019s specially designed for multi-node, distributed training for large models \u2014 maximizing correctness guarantees, performance, and ease of use. Key features include elastically deterministic training, instant mid-epoch resumption, effective shuffling, high training throughput, and flexible data mixing, among other features. When training with StreamingDataset, the data shards are written to cloud storage in MDS, our file format that allows for low-latency random access to samples. By being as efficient as possible with shard downloads and shuffling, StreamingDataset minimizes egress costs while ensuring that dataloading never bottlenecks model training. StreamingDataset powers training for LLMs with over 100 billion parameters like DBRX, to advanced diffusion models, to two-tower recommendation models, and more, scaling to training jobs on thousands of GPUs with ease. Join us to learn how StreamingDataset can elevate your distributed model training experience.", "duration": 423, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Saaketh Narayan" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/C1TpUZjqTfA/maxresdefault.webp", - "title": "Lightning Talk: Fast, Scalable Distributed Training with StreamingDataset - Saaketh Narayan", + "title": "Fast, Scalable Distributed Training with StreamingDataset", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/lightning-talk-flexattention-the-flexibility-of-pytorch-the-performa-yanbo-liang-horace-he.json b/pytorchconf-2024/videos/lightning-talk-flexattention-the-flexibility-of-pytorch-the-performa-yanbo-liang-horace-he.json index ce8bb06ee..b53280314 100644 --- a/pytorchconf-2024/videos/lightning-talk-flexattention-the-flexibility-of-pytorch-the-performa-yanbo-liang-horace-he.json +++ b/pytorchconf-2024/videos/lightning-talk-flexattention-the-flexibility-of-pytorch-the-performa-yanbo-liang-horace-he.json @@ -1,5 +1,5 @@ { - "description": "Lightning Talk: FlexAttention - The Flexibility of PyTorch + The Performance of FlashAttention - Yanbo Liang & Horace He, Meta\n\nIntroducing a novel abstraction leveraging the PyTorch compiler stack to enable custom, user-defined attention mechanisms. This new API supports dynamic modifications to attention scores within SDPA, providing both runtime and memory efficiency through kernel fusion with the FlashAttention algorithm.", + "description": "Introducing a novel abstraction leveraging the PyTorch compiler stack to enable custom, user-defined attention mechanisms. This new API supports dynamic modifications to attention scores within SDPA, providing both runtime and memory efficiency through kernel fusion with the FlashAttention algorithm.", "duration": 1060, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,12 @@ } ], "speakers": [ - "TODO" + "Yanbo Liang", + "Horace He" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/ju-KlcuWlbk/maxresdefault.webp", - "title": "Lightning Talk: FlexAttention - The Flexibility of PyTorch + The Performa... Yanbo Liang & Horace He", + "title": "FlexAttention - The Flexibility of PyTorch + The Performance of FlashAttention", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/lightning-talk-hieroglyph2text-a-pytorch-powered-pipeline-for-automated-egyptian-h-susi-gentsch.json b/pytorchconf-2024/videos/lightning-talk-hieroglyph2text-a-pytorch-powered-pipeline-for-automated-egyptian-h-susi-gentsch.json index 3a5bd593c..20d79f855 100644 --- a/pytorchconf-2024/videos/lightning-talk-hieroglyph2text-a-pytorch-powered-pipeline-for-automated-egyptian-h-susi-gentsch.json +++ b/pytorchconf-2024/videos/lightning-talk-hieroglyph2text-a-pytorch-powered-pipeline-for-automated-egyptian-h-susi-gentsch.json @@ -1,5 +1,5 @@ { - "description": "Lightning Talk: HieroGlyph2Text: A PyTorch-Powered Pipeline for Automated Egyptian Hieroglyph Translation from Image - Susi Gentsch, University of Bonn\n\nHieroGlyph2Text is an innovative PyTorch-powered pipeline that automates the detection, classification, and attempts translation of Egyptian hieroglyphs from large image inputs. It addresses the challenge of decoding and translating ancient hieroglyphic inscriptions, traditionally a time-consuming and specialized task. This pipeline leverages PyTorch to create custom models: 1. Object Detection: YOLOv8 accurately detects individual hieroglyphs within images. 2. Image Classification: A custom ResNet model built using PyTorch achieves state-of-the-art accuracy in assigning Gardiner Codes to hieroglyphs. 3. Translation: The classified Gardiner Codes outputs from the ResNet model are integrated with Llama3, a large language model (LLM), using Retrieval-Augmented Generation (RAG) and a custom dataset based upon Gardiner Codes and their respective description and ideogram. Key highlights include accurate hieroglyph detection and state-of-the-art classification performance through an optimized ResNet model. This pipeline lays the groundwork for collaboration with subject matter experts to refine the translation process and democratize access to ancient Egyptian hieroglyphic knowledge.", + "description": "HieroGlyph2Text is an innovative PyTorch-powered pipeline that automates the detection, classification, and attempts translation of Egyptian hieroglyphs from large image inputs. It addresses the challenge of decoding and translating ancient hieroglyphic inscriptions, traditionally a time-consuming and specialized task. This pipeline leverages PyTorch to create custom models: 1. Object Detection: YOLOv8 accurately detects individual hieroglyphs within images. 2. Image Classification: A custom ResNet model built using PyTorch achieves state-of-the-art accuracy in assigning Gardiner Codes to hieroglyphs. 3. Translation: The classified Gardiner Codes outputs from the ResNet model are integrated with Llama3, a large language model (LLM), using Retrieval-Augmented Generation (RAG) and a custom dataset based upon Gardiner Codes and their respective description and ideogram. Key highlights include accurate hieroglyph detection and state-of-the-art classification performance through an optimized ResNet model. This pipeline lays the groundwork for collaboration with subject matter experts to refine the translation process and democratize access to ancient Egyptian hieroglyphic knowledge.", "duration": 564, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Susi Gentsch" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/eDjHKOb9Tn4/maxresdefault.webp", - "title": "Lightning Talk: HieroGlyph2Text: A PyTorch-Powered Pipeline for Automated Egyptian H... Susi Gentsch", + "title": "HieroGlyph2Text: A PyTorch-Powered Pipeline for Automated Egyptian Hieroglyph Translation from Image", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/lightning-talk-implementing-and-using-iterable-datasets-what-could-go-wrong-nicolas-hug-meta.json b/pytorchconf-2024/videos/lightning-talk-implementing-and-using-iterable-datasets-what-could-go-wrong-nicolas-hug-meta.json index 8f2fd171e..739f8b100 100644 --- a/pytorchconf-2024/videos/lightning-talk-implementing-and-using-iterable-datasets-what-could-go-wrong-nicolas-hug-meta.json +++ b/pytorchconf-2024/videos/lightning-talk-implementing-and-using-iterable-datasets-what-could-go-wrong-nicolas-hug-meta.json @@ -1,5 +1,5 @@ { - "description": "Lightning Talk: Implementing and Using Iterable Datasets: What Could Go Wrong? - Nicolas Hug, Meta\n\nPyTorch supports two kinds of datasets: Iterable datasets and indexable \"map-style\" datasets. Iterable datasets can be more flexible and potentially faster than their indexable cousins. They are also much harder to use correctly, and can easily lead to silently wrong results. This talk is a quick and fun intro to some of the traps that Iterable datasets lay out for you, with some tips to help you avoid them.", + "description": "PyTorch supports two kinds of datasets: Iterable datasets and indexable \"map-style\" datasets. Iterable datasets can be more flexible and potentially faster than their indexable cousins. They are also much harder to use correctly, and can easily lead to silently wrong results. This talk is a quick and fun intro to some of the traps that Iterable datasets lay out for you, with some tips to help you avoid them.", "duration": 674, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Nicolas Hug" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/Kx0BoPa9juQ/maxresdefault.webp", - "title": "Lightning Talk: Implementing and Using Iterable Datasets: What Could Go Wrong? - Nicolas Hug, Meta", + "title": "Implementing and Using Iterable Datasets: What Could Go Wrong?", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/lightning-talk-in-transit-machine-learning-using-pytorch-on-frontier-exascale-system-vineeth-gutta.json b/pytorchconf-2024/videos/lightning-talk-in-transit-machine-learning-using-pytorch-on-frontier-exascale-system-vineeth-gutta.json index 1382e7cbe..a53d5c400 100644 --- a/pytorchconf-2024/videos/lightning-talk-in-transit-machine-learning-using-pytorch-on-frontier-exascale-system-vineeth-gutta.json +++ b/pytorchconf-2024/videos/lightning-talk-in-transit-machine-learning-using-pytorch-on-frontier-exascale-system-vineeth-gutta.json @@ -1,5 +1,5 @@ { - "description": "Lightning Talk: In-Transit Machine Learning Using PyTorch on Frontier Exascale System - Vineeth Gutta, University of Delaware\n\nTraditional ML workflows use offline training where the data is stored on disk and is subsequently loaded into accelerator (CPU,GPU, etc) memory during training or inference. We recently devised a novel and scalable in-transit ML workflow for a plasma-physics application (chosen as 1 out of 8 compelling codes in the country) for the world\u2019s fastest supercomputer, Frontier) with an aim to build a high-energy laser particle accelerator. Data generated in distributed HPC systems like Frontier create volumes of data that is infeasible to store on HPC file systems. A mismatch between modern memory hierarchies occurs due to high volume and rate of data generation. Our novel ML workflow utilizes continuous learning where the data is consumed in batches as the simulation produces the data and then discards after each batch is trained. This in-transit workflow integrates particle-in-cell simulations with distributed ML training on PyTorch using DDP allows for an application coupling enabling the model to learn correlations between emitted radiation and particle dynamics within simulation in an unsupervised method. This workflow is demonstrated at scale on Frontier using 400 AMD MI250X GPUs", + "description": "Traditional ML workflows use offline training where the data is stored on disk and is subsequently loaded into accelerator (CPU,GPU, etc) memory during training or inference. We recently devised a novel and scalable in-transit ML workflow for a plasma-physics application (chosen as 1 out of 8 compelling codes in the country) for the world\u2019s fastest supercomputer, Frontier) with an aim to build a high-energy laser particle accelerator. Data generated in distributed HPC systems like Frontier create volumes of data that is infeasible to store on HPC file systems. A mismatch between modern memory hierarchies occurs due to high volume and rate of data generation. Our novel ML workflow utilizes continuous learning where the data is consumed in batches as the simulation produces the data and then discards after each batch is trained. This in-transit workflow integrates particle-in-cell simulations with distributed ML training on PyTorch using DDP allows for an application coupling enabling the model to learn correlations between emitted radiation and particle dynamics within simulation in an unsupervised method. This workflow is demonstrated at scale on Frontier using 400 AMD MI250X GPUs", "duration": 725, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Vineeth Gutta" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/s4Gh2UX0EyA/maxresdefault.webp", - "title": "Lightning Talk: In-Transit Machine Learning Using PyTorch on Frontier Exascale System- Vineeth Gutta", + "title": "In-Transit Machine Learning Using PyTorch on Frontier Exascale System", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/lightning-talk-introduction-to-torch-distributed-pipelining-howard-huang-ke-wen-meta.json b/pytorchconf-2024/videos/lightning-talk-introduction-to-torch-distributed-pipelining-howard-huang-ke-wen-meta.json index b78bacae8..cec4a2dca 100644 --- a/pytorchconf-2024/videos/lightning-talk-introduction-to-torch-distributed-pipelining-howard-huang-ke-wen-meta.json +++ b/pytorchconf-2024/videos/lightning-talk-introduction-to-torch-distributed-pipelining-howard-huang-ke-wen-meta.json @@ -1,5 +1,5 @@ { - "description": "Lightning Talk: Introduction to Torch.Distributed.Pipelining - Howard Huang & Ke Wen, Meta\n\nPipeline parallelism is a technique employed in distributed deep learning that enhances model execution by dividing the model into distinct segments, or \"stages.\" As large language models and other memory-intensive models become more common, pipeline parallelism has grown increasingly important for several key areas: - Executing large-scale training jobs. - Enhancing performance in bandwidth-limited clusters. - Supporting large model inference. In this talk, we will introduce the `torch.distributed.pipelining` package which provides users a seamless way of applying pipeline parallelism. We will demonstrate the following features: - Splitting of model code based on simple specification. - Support for pipeline schedules, including GPipe, 1F1B, Interleaved 1F1B and Looped BFS, and providing the infrastructure for writing customized schedules. - Composability with other PyTorch parallel techniques such as data parallel (DDP, FSDP) or tensor parallel. - Out of the box integration with Hugging Face models for efficient inference.", + "description": "Pipeline parallelism is a technique employed in distributed deep learning that enhances model execution by dividing the model into distinct segments, or \"stages.\" As large language models and other memory-intensive models become more common, pipeline parallelism has grown increasingly important for several key areas: - Executing large-scale training jobs. - Enhancing performance in bandwidth-limited clusters. - Supporting large model inference. In this talk, we will introduce the `torch.distributed.pipelining` package which provides users a seamless way of applying pipeline parallelism. We will demonstrate the following features: - Splitting of model code based on simple specification. - Support for pipeline schedules, including GPipe, 1F1B, Interleaved 1F1B and Looped BFS, and providing the infrastructure for writing customized schedules. - Composability with other PyTorch parallel techniques such as data parallel (DDP, FSDP) or tensor parallel. - Out of the box integration with Hugging Face models for efficient inference.", "duration": 765, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,12 @@ } ], "speakers": [ - "TODO" + "Howard Huang", + "Ke Wen" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/YieoyQll4_4/maxresdefault.webp", - "title": "Lightning Talk: Introduction to Torch.Distributed.Pipelining - Howard Huang & Ke Wen, Meta", + "title": "Introduction to Torch.Distributed.Pipelining", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/lightning-talk-llms-on-edge-with-ai-accelerators-chen-lai-kimish-patel-cemal-bilgin-meta.json b/pytorchconf-2024/videos/lightning-talk-llms-on-edge-with-ai-accelerators-chen-lai-kimish-patel-cemal-bilgin-meta.json index 10062ca0a..c46f9c1b0 100644 --- a/pytorchconf-2024/videos/lightning-talk-llms-on-edge-with-ai-accelerators-chen-lai-kimish-patel-cemal-bilgin-meta.json +++ b/pytorchconf-2024/videos/lightning-talk-llms-on-edge-with-ai-accelerators-chen-lai-kimish-patel-cemal-bilgin-meta.json @@ -1,5 +1,5 @@ { - "description": "Lightning Talk: LLMs on Edge with AI Accelerators - Chen Lai, Kimish Patel & Cemal Bilgin, Meta\n\nLLMs are known to be compute heavy and consume lots of resources (almost all resources on phones), including memory and power. A natural thought is to leverage the AI hardware accelerators, for example, Apple Neural Engine (ANE) on Apple devices and HTP on Qualcomm SoCs, to make it run fast and efficiently. Only by optimizing the model latency, memory consumption and power usage to a certain level will users be interested in installing the models on their devices. In this session, we\u2019d like to introduce how we leverage these AI accelerators within the PyTorch ecosystem to achieve the state-of-art performance for llama3 on device, via ExecuTorch and the partnership with Apple and Qualcomm. Hardware companies usually have their own AI accelerators. Likely they have different characteristics, one may support a list of different operators than others, and one may only support static shapes (like HTP). However, transformers-based optimization can be generic. We\u2019ll discuss in more detail how we apply the generic optimization as well as the backend specific optimization. The techniques we applied here are not just for LLMs, but can be applied to other transformer-based models.", + "description": "LLMs are known to be compute heavy and consume lots of resources (almost all resources on phones), including memory and power. A natural thought is to leverage the AI hardware accelerators, for example, Apple Neural Engine (ANE) on Apple devices and HTP on Qualcomm SoCs, to make it run fast and efficiently. Only by optimizing the model latency, memory consumption and power usage to a certain level will users be interested in installing the models on their devices. In this session, we\u2019d like to introduce how we leverage these AI accelerators within the PyTorch ecosystem to achieve the state-of-art performance for llama3 on device, via ExecuTorch and the partnership with Apple and Qualcomm. Hardware companies usually have their own AI accelerators. Likely they have different characteristics, one may support a list of different operators than others, and one may only support static shapes (like HTP). However, transformers-based optimization can be generic. We\u2019ll discuss in more detail how we apply the generic optimization as well as the backend specific optimization. The techniques we applied here are not just for LLMs, but can be applied to other transformer-based models.", "duration": 729, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,13 @@ } ], "speakers": [ - "TODO" + "Chen Lai", + "Kimish Patel", + "Cemal Bilgin" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/8p8Pntnr3GU/maxresdefault.webp", - "title": "Lightning Talk: LLMs on Edge with AI Accelerators - Chen Lai, Kimish Patel & Cemal Bilgin, Meta", + "title": "LLMs on Edge with AI Accelerators", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/lightning-talk-low-precision-dtypes-in-pytorch-vasiliy-kuznetsov-meta.json b/pytorchconf-2024/videos/lightning-talk-low-precision-dtypes-in-pytorch-vasiliy-kuznetsov-meta.json index 38c415cc1..579d0fb20 100644 --- a/pytorchconf-2024/videos/lightning-talk-low-precision-dtypes-in-pytorch-vasiliy-kuznetsov-meta.json +++ b/pytorchconf-2024/videos/lightning-talk-low-precision-dtypes-in-pytorch-vasiliy-kuznetsov-meta.json @@ -1,5 +1,5 @@ { - "description": "Lightning Talk: Low Precision Dtypes in PyTorch - Vasiliy Kuznetsov, Meta\n\nThis talk deep dives into the new native PyTorch float8 training library, and previews PyTorch's strategy for supporting upcoming low precision dtypes such as float6, float4 and MX for efficient training and inference.", + "description": "This talk deep dives into the new native PyTorch float8 training library, and previews PyTorch's strategy for supporting upcoming low precision dtypes such as float6, float4 and MX for efficient training and inference.", "duration": 530, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Vasiliy Kuznetsov" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/xcKwEZ77Cps/maxresdefault.webp", - "title": "Lightning Talk: Low Precision Dtypes in PyTorch - Vasiliy Kuznetsov, Meta", + "title": "Low Precision Dtypes in PyTorch", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/lightning-talk-making-the-most-of-heterogeneous-memory-capacity-using-pytorch-syed-ahmed-nvidia.json b/pytorchconf-2024/videos/lightning-talk-making-the-most-of-heterogeneous-memory-capacity-using-pytorch-syed-ahmed-nvidia.json index ebf0a1428..b9ff20708 100644 --- a/pytorchconf-2024/videos/lightning-talk-making-the-most-of-heterogeneous-memory-capacity-using-pytorch-syed-ahmed-nvidia.json +++ b/pytorchconf-2024/videos/lightning-talk-making-the-most-of-heterogeneous-memory-capacity-using-pytorch-syed-ahmed-nvidia.json @@ -1,5 +1,5 @@ { - "description": "Lightning Talk: Making the Most of Heterogeneous Memory Capacity Using PyTorch - Syed Ahmed, NVIDIA Corporation\n\nMemory intensive deep learning workloads require efficient use of all kinds of memories that are available in a system. In this session, we will discuss how we can utilize such heterogeneous memory through memory pools in PyTorch. We will show how to mix-and-match different CUDA system allocators in the same PyTorch program using memory pools. Consequently, this API unlocks new use cases such as Extended GPU Memory (EGM) based all-gathers, Unified Virtual Memory (UVM), and NVLink Sharp (NVLS) reductions. New NVIDIA architectures accelerate such use cases with high-bandwidth and low-latency interconnects in the hardware, driven by extended functionality of CUDA system allocators in the software. Learn how to use these techniques on memory-intensive deep learning models like LLMs, and discover new CUDA features powered by PyTorch.", + "description": "Memory intensive deep learning workloads require efficient use of all kinds of memories that are available in a system. In this session, we will discuss how we can utilize such heterogeneous memory through memory pools in PyTorch. We will show how to mix-and-match different CUDA system allocators in the same PyTorch program using memory pools. Consequently, this API unlocks new use cases such as Extended GPU Memory (EGM) based all-gathers, Unified Virtual Memory (UVM), and NVLink Sharp (NVLS) reductions. New NVIDIA architectures accelerate such use cases with high-bandwidth and low-latency interconnects in the hardware, driven by extended functionality of CUDA system allocators in the software. Learn how to use these techniques on memory-intensive deep learning models like LLMs, and discover new CUDA features powered by PyTorch.", "duration": 686, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Syed Ahmed" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/srQOK1UusZ4/maxresdefault.webp", - "title": "Lightning Talk: Making the Most of Heterogeneous Memory Capacity Using PyTorch - Syed Ahmed, NVIDIA", + "title": "Making the Most of Heterogeneous Memory Capacity Using PyTorch", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/lightning-talk-mobile-computational-photography-with-pytorch-low-light-denoising-alexis-baudron.json b/pytorchconf-2024/videos/lightning-talk-mobile-computational-photography-with-pytorch-low-light-denoising-alexis-baudron.json index 928b6c573..6902dd77f 100644 --- a/pytorchconf-2024/videos/lightning-talk-mobile-computational-photography-with-pytorch-low-light-denoising-alexis-baudron.json +++ b/pytorchconf-2024/videos/lightning-talk-mobile-computational-photography-with-pytorch-low-light-denoising-alexis-baudron.json @@ -1,5 +1,5 @@ { - "description": "Lightning Talk: Mobile Computational Photography with PyTorch: Low-Light Denoising - Alexis Baudron, Sony\n\nOver the last decade, smartphone cameras have improved significantly, becoming the primary device people use for capturing everyday moments and high-quality photographs. This progress is largely due to advances in computational photography and novel image sensors. Computational photography enables great images from compact mobile cameras, enhancing photos through various techniques such as multi-shot merging. Despite these advancements, challenges such as noise, artifacts, and distortions persist, especially in low-light conditions where limited light increases noise levels. In this lightning talk, we will explore how PyTorch can be used to design and optimize deep learning networks for real-time low-light denoising. We will dive into noise modeling, data generation, physics-aware models, and advanced network architectures for effective denoising in challenging low-light scenarios. Attendees will gain practical insights into the latest advancements in mobile computational photography using PyTorch.", + "description": "Over the last decade, smartphone cameras have improved significantly, becoming the primary device people use for capturing everyday moments and high-quality photographs. This progress is largely due to advances in computational photography and novel image sensors. Computational photography enables great images from compact mobile cameras, enhancing photos through various techniques such as multi-shot merging. Despite these advancements, challenges such as noise, artifacts, and distortions persist, especially in low-light conditions where limited light increases noise levels. In this lightning talk, we will explore how PyTorch can be used to design and optimize deep learning networks for real-time low-light denoising. We will dive into noise modeling, data generation, physics-aware models, and advanced network architectures for effective denoising in challenging low-light scenarios. Attendees will gain practical insights into the latest advancements in mobile computational photography using PyTorch.", "duration": 507, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Alexis Baudron" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/hjOtsOqPw3Y/maxresdefault.webp", - "title": "Lightning Talk: Mobile Computational Photography with PyTorch: Low-Light Denoising - Alexis Baudron", + "title": "Mobile Computational Photography with PyTorch: Low-Light Denoising", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/lightning-talk-new-activation-checkpointing-apis-in-pytorch-jeffrey-wan-horace-he-meta.json b/pytorchconf-2024/videos/lightning-talk-new-activation-checkpointing-apis-in-pytorch-jeffrey-wan-horace-he-meta.json index d86d9aaa3..dd57c5387 100644 --- a/pytorchconf-2024/videos/lightning-talk-new-activation-checkpointing-apis-in-pytorch-jeffrey-wan-horace-he-meta.json +++ b/pytorchconf-2024/videos/lightning-talk-new-activation-checkpointing-apis-in-pytorch-jeffrey-wan-horace-he-meta.json @@ -1,5 +1,5 @@ { - "description": "Lightning Talk: New Activation Checkpointing APIs in PyTorch - Jeffrey Wan & Horace He, Meta\n\nActivation checkpointing is a commonly used technique to reduce memory usage during model training by reducing the number of activations saved for backward. Instead of keeping tensors needed for backward alive until they are used in gradient computation during backward, those tensors are recomputed during the backward pass. This talk will introduce new activation checkpoint APIs that can help achieve a better trade off between memory savings and compute overhead that recomputing introduces.", + "description": "Activation checkpointing is a commonly used technique to reduce memory usage during model training by reducing the number of activations saved for backward. Instead of keeping tensors needed for backward alive until they are used in gradient computation during backward, those tensors are recomputed during the backward pass. This talk will introduce new activation checkpoint APIs that can help achieve a better trade off between memory savings and compute overhead that recomputing introduces.", "duration": 750, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,12 @@ } ], "speakers": [ - "TODO" + "Jeffrey Wan", + "Horace He" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/v3gsrJtGLiA/maxresdefault.webp", - "title": "Lightning Talk: New Activation Checkpointing APIs in PyTorch - Jeffrey Wan & Horace He, Meta", + "title": "New Activation Checkpointing APIs in PyTorch", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/lightning-talk-on-device-profiling-and-debugging-with-executorch-olivia-liu-vaun-puri-meta.json b/pytorchconf-2024/videos/lightning-talk-on-device-profiling-and-debugging-with-executorch-olivia-liu-vaun-puri-meta.json index f272770fb..36858fe8f 100644 --- a/pytorchconf-2024/videos/lightning-talk-on-device-profiling-and-debugging-with-executorch-olivia-liu-vaun-puri-meta.json +++ b/pytorchconf-2024/videos/lightning-talk-on-device-profiling-and-debugging-with-executorch-olivia-liu-vaun-puri-meta.json @@ -1,5 +1,5 @@ { - "description": "Lightning Talk: On-Device Profiling and Debugging with ExecuTorch - Olivia Liu & Vaun Puri, Meta\n\nHigh developer velocity is crucial to shipping new ML-enabled experiences from a server-trained model to a customers\u2019 device. ExecuTorch is an on-device runtime that seamlessly integrates with the PyTorch stack with a focus on developer productivity. We present the ExecuTorch Dev Tools and highlight key features that tighten the iteration loop when optimizing models for deployment and execution on edge devices. We demonstrate how ExecuTorch\u2019s built-in profiler and bundled tools tackle key pain-points, such as: 1. Examining the memory footprint of an ExecuTorch program ahead-of-time; 2. Collecting runtime performance metrics and intermediate outputs for accuracy analysis; 3. Correlating runtime data with the underlying graph of an exported model.", + "description": "High developer velocity is crucial to shipping new ML-enabled experiences from a server-trained model to a customers\u2019 device. ExecuTorch is an on-device runtime that seamlessly integrates with the PyTorch stack with a focus on developer productivity. We present the ExecuTorch Dev Tools and highlight key features that tighten the iteration loop when optimizing models for deployment and execution on edge devices. We demonstrate how ExecuTorch\u2019s built-in profiler and bundled tools tackle key pain-points, such as: 1. Examining the memory footprint of an ExecuTorch program ahead-of-time; 2. Collecting runtime performance metrics and intermediate outputs for accuracy analysis; 3. Correlating runtime data with the underlying graph of an exported model.", "duration": 852, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,12 @@ } ], "speakers": [ - "TODO" + "Olivia Liu", + "Vaun Puri" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/_e6E6fnEIDQ/maxresdefault.webp", - "title": "Lightning Talk: On-Device Profiling and Debugging with ExecuTorch - Olivia Liu & Vaun Puri, Meta", + "title": "On-Device Profiling and Debugging with ExecuTorch", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/lightning-talk-optimized-pytorch-inference-on-aarch64-linux-cpus-sunita-nadampalli-amazon-aws.json b/pytorchconf-2024/videos/lightning-talk-optimized-pytorch-inference-on-aarch64-linux-cpus-sunita-nadampalli-amazon-aws.json index f00b2fb5e..8e57caebf 100644 --- a/pytorchconf-2024/videos/lightning-talk-optimized-pytorch-inference-on-aarch64-linux-cpus-sunita-nadampalli-amazon-aws.json +++ b/pytorchconf-2024/videos/lightning-talk-optimized-pytorch-inference-on-aarch64-linux-cpus-sunita-nadampalli-amazon-aws.json @@ -1,5 +1,5 @@ { - "description": "Lightning Talk: Optimized PyTorch Inference on aarch64 Linux CPUs - Sunita Nadampalli, Amazon (AWS)\n\nIn the last 2 years we've optimized performance of PyTorch on Arm processors. The optimizations have included changes to ATen, C10, MKLDNN operators, GEMM backend, and Torch inductor. In many cases instead of writing our own kernel we integrated the Arm compute library, used fastmath kernels with format types like bf16, implemented operator caching, selected optimal backend based on the input context etc. Through these optimizations we improved performance by over 2x. In this presentation first we will talk about how we went across this process, what those optimizations are, performance numbers for AWS Graviton3 processors for around 75 models, and CI/CD workflow details. Next, we will walk through a sample PyTorch application showing basic usage, how to tune runtime and the resulting speed up. At the end of the presentation attendees will learn about PyTorch performance optimizations on Arm processors, how to use them, and the areas where they can collaborate to further improve PyTorch for aarch64 CPUs.", + "description": "In the last 2 years we've optimized performance of PyTorch on Arm processors. The optimizations have included changes to ATen, C10, MKLDNN operators, GEMM backend, and Torch inductor. In many cases instead of writing our own kernel we integrated the Arm compute library, used fastmath kernels with format types like bf16, implemented operator caching, selected optimal backend based on the input context etc. Through these optimizations we improved performance by over 2x. In this presentation first we will talk about how we went across this process, what those optimizations are, performance numbers for AWS Graviton3 processors for around 75 models, and CI/CD workflow details. Next, we will walk through a sample PyTorch application showing basic usage, how to tune runtime and the resulting speed up. At the end of the presentation attendees will learn about PyTorch performance optimizations on Arm processors, how to use them, and the areas where they can collaborate to further improve PyTorch for aarch64 CPUs.", "duration": 790, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Sunita Nadampalli" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/_9xohmSa2G8/maxresdefault.webp", - "title": "Lightning Talk: Optimized PyTorch Inference on aarch64 Linux CPUs - Sunita Nadampalli, Amazon (AWS)", + "title": "Optimized PyTorch Inference on aarch64 Linux CPUs", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/lightning-talk-pytorch-release-process-andrey-talman-meta.json b/pytorchconf-2024/videos/lightning-talk-pytorch-release-process-andrey-talman-meta.json index 019dd636b..6e0de648a 100644 --- a/pytorchconf-2024/videos/lightning-talk-pytorch-release-process-andrey-talman-meta.json +++ b/pytorchconf-2024/videos/lightning-talk-pytorch-release-process-andrey-talman-meta.json @@ -1,5 +1,5 @@ { - "description": "Lightning Talk: PyTorch Release Process - Andrey Talman, Meta\n\nI would like to present and quickly discuss PyTorch Release process, how it happens. What are milestones. What is our cherry-picking criteria, how we validate the release.", + "description": "I would like to present and quickly discuss PyTorch Release process, how it happens. What are milestones. What is our cherry-picking criteria, how we validate the release.", "duration": 573, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Andrey Talman" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/4REnTJO3_ps/maxresdefault.webp", - "title": "Lightning Talk: PyTorch Release Process - Andrey Talman, Meta", + "title": "PyTorch Release Process", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/lightning-talk-pytorch-xla-auto-sharding-yeounoh-chung-google.json b/pytorchconf-2024/videos/lightning-talk-pytorch-xla-auto-sharding-yeounoh-chung-google.json index f2a6e2f92..1f16ece47 100644 --- a/pytorchconf-2024/videos/lightning-talk-pytorch-xla-auto-sharding-yeounoh-chung-google.json +++ b/pytorchconf-2024/videos/lightning-talk-pytorch-xla-auto-sharding-yeounoh-chung-google.json @@ -1,5 +1,5 @@ { - "description": "Lightning Talk: PyTorch/XLA Auto-Sharding - Yeounoh Chung, Google\n\nPyTorch/XLA recently launched the new PyTorch/XLA SPMD feature as a first-step to automate ML workloads parallelization using GSPMD. It turns out that the performance largely depends on the quality of sharding hints provided by the user \u2013 and it requires a correct and deep understanding of model architectures and much expertise to come up with optimal sharding hints. To address this problem, we propose to integrate PyTorch/XLA SPMD with XLA's auto sharding service that allows the XLA compiler to shard and optimize the whole model without any user input.", + "description": "PyTorch/XLA recently launched the new PyTorch/XLA SPMD feature as a first-step to automate ML workloads parallelization using GSPMD. It turns out that the performance largely depends on the quality of sharding hints provided by the user \u2013 and it requires a correct and deep understanding of model architectures and much expertise to come up with optimal sharding hints. To address this problem, we propose to integrate PyTorch/XLA SPMD with XLA's auto sharding service that allows the XLA compiler to shard and optimize the whole model without any user input.", "duration": 461, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Yeounoh Chung" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/gNfyS9asA7A/maxresdefault.webp", - "title": "Lightning Talk: PyTorch/XLA Auto-Sharding - Yeounoh Chung, Google", + "title": "PyTorch/XLA Auto-Sharding", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/lightning-talk-sparsifying-vision-transformers-with-minimal-accuracy-loss-jesse-cai-meta.json b/pytorchconf-2024/videos/lightning-talk-sparsifying-vision-transformers-with-minimal-accuracy-loss-jesse-cai-meta.json index a2c3a9c09..cae9d698c 100644 --- a/pytorchconf-2024/videos/lightning-talk-sparsifying-vision-transformers-with-minimal-accuracy-loss-jesse-cai-meta.json +++ b/pytorchconf-2024/videos/lightning-talk-sparsifying-vision-transformers-with-minimal-accuracy-loss-jesse-cai-meta.json @@ -1,5 +1,5 @@ { - "description": "Lightning Talk: Sparsifying Vision Transformers with Minimal Accuracy Loss - Jesse Cai, Meta\n\nSparsity, like quantization, is an approximate model optimization technique, where we trade some model accuracy for increased performance.\n\nIn this talk we'll explore how to minimize the accuracy degradation of sparsifying Vision Transformer (ViT) based models to GPU accelerable sparsity patterns like block sparsity and semi-structured sparsity.\n\nWe'll cover the best techniques to ensure a less-than 5% loss in accuracy when:\n- training a sparse model from scratch\n- pruning and retraining an existing dense model\n- zero-shot/one-shot pruning a dense model\n\nWe've collected these techniques into a single repository, torchao, so that model optimization enthusiasts like you can sparsify your models with just a few lines of code.", + "description": "Sparsity, like quantization, is an approximate model optimization technique, where we trade some model accuracy for increased performance.\n\nIn this talk we'll explore how to minimize the accuracy degradation of sparsifying Vision Transformer (ViT) based models to GPU accelerable sparsity patterns like block sparsity and semi-structured sparsity.\n\nWe'll cover the best techniques to ensure a less-than 5% loss in accuracy when:\n- training a sparse model from scratch\n- pruning and retraining an existing dense model\n- zero-shot/one-shot pruning a dense model\n\nWe've collected these techniques into a single repository, torchao, so that model optimization enthusiasts like you can sparsify your models with just a few lines of code.", "duration": 841, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Jesse Cai" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/mqoIfs5hTxA/maxresdefault.webp", - "title": "Lightning Talk: Sparsifying Vision Transformers with Minimal Accuracy Loss - Jesse Cai, Meta", + "title": "Sparsifying Vision Transformers with Minimal Accuracy Loss", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/lightning-talk-understanding-and-optimizing-pytorch-models-with-thunder-luca-antiga-lightning-ai.json b/pytorchconf-2024/videos/lightning-talk-understanding-and-optimizing-pytorch-models-with-thunder-luca-antiga-lightning-ai.json index e8667a142..b73a0424e 100644 --- a/pytorchconf-2024/videos/lightning-talk-understanding-and-optimizing-pytorch-models-with-thunder-luca-antiga-lightning-ai.json +++ b/pytorchconf-2024/videos/lightning-talk-understanding-and-optimizing-pytorch-models-with-thunder-luca-antiga-lightning-ai.json @@ -1,5 +1,5 @@ { - "description": "Lightning Talk: Understanding and Optimizing PyTorch Models with Thunder - Luca Antiga, Lightning AI\n\nA hallmark feature of PyTorch is the natural expression of computation. This enables practitioners to implement AI models with ease. However, it prompts the question how to optimize the workload for a given hardware setup because those optimizations clutter our code and are tricky to combine. Lightning Thunder provides a Python-to-Python compiler to scale and optimize PyTorch programs that focuses on usability, understandability, and extensibility. A key tool in delivering on these goals is the composability of transformations: without changing the user code, we can stack quantization, distributing the computation across multiple GPUs, dispatching to optimized kernels, offloading, and other pluggable optimizations. Lightning Thunder flourishes in the PyTorch ecosystem: with PyTorch eager and with executors like torch.compile and nvFuser. It also dispatches to libraries like cuDNN, TransformerEngine, Apex, OpenAI Triton. The ability to apply multiple optimizations just-in-time leads to significant compounded speed-ups over unoptimized code out of the box. Luca will discuss the design of Thunder and demonstrate applications on training and inference for large language and multimodal models.", + "description": "A hallmark feature of PyTorch is the natural expression of computation. This enables practitioners to implement AI models with ease. However, it prompts the question how to optimize the workload for a given hardware setup because those optimizations clutter our code and are tricky to combine. Lightning Thunder provides a Python-to-Python compiler to scale and optimize PyTorch programs that focuses on usability, understandability, and extensibility. A key tool in delivering on these goals is the composability of transformations: without changing the user code, we can stack quantization, distributing the computation across multiple GPUs, dispatching to optimized kernels, offloading, and other pluggable optimizations. Lightning Thunder flourishes in the PyTorch ecosystem: with PyTorch eager and with executors like torch.compile and nvFuser. It also dispatches to libraries like cuDNN, TransformerEngine, Apex, OpenAI Triton. The ability to apply multiple optimizations just-in-time leads to significant compounded speed-ups over unoptimized code out of the box. Luca will discuss the design of Thunder and demonstrate applications on training and inference for large language and multimodal models.", "duration": 837, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Luca Antiga" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/5dQ423cFfLc/maxresdefault.webp", - "title": "Lightning Talk: Understanding and Optimizing PyTorch Models with Thunder - Luca Antiga, Lightning AI", + "title": "Understanding and Optimizing PyTorch Models with Thunder", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/lightning-talk-what-s-new-for-pytorch-developer-infrastructure-sahan-paliskara-catherine-lee.json b/pytorchconf-2024/videos/lightning-talk-what-s-new-for-pytorch-developer-infrastructure-sahan-paliskara-catherine-lee.json index 6e142c2d9..50dba65fa 100644 --- a/pytorchconf-2024/videos/lightning-talk-what-s-new-for-pytorch-developer-infrastructure-sahan-paliskara-catherine-lee.json +++ b/pytorchconf-2024/videos/lightning-talk-what-s-new-for-pytorch-developer-infrastructure-sahan-paliskara-catherine-lee.json @@ -1,5 +1,5 @@ { - "description": "Lightning Talk: What's New for PyTorch Developer Infrastructure - Sahan Paliskara & Catherine Lee, Meta\n\nHaving a chat about all of the work being done to continue supporting PyTorch's Developer Infrastructure needs including updates around Target Determination, Releases, and OSS Tooling.", + "description": "Having a chat about all of the work being done to continue supporting PyTorch's Developer Infrastructure needs including updates around Target Determination, Releases, and OSS Tooling.", "duration": 671, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,12 @@ } ], "speakers": [ - "TODO" + "Sahan Paliskara", + "Catherine Lee" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/xlhBDySu7jM/maxresdefault.webp", - "title": "Lightning Talk: What's New for PyTorch Developer Infrastructure - Sahan Paliskara & Catherine Lee", + "title": "What's New for PyTorch Developer Infrastructure", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/lightning-talk-whats-new-in-ex-angela-yi-tugsbayasgalan-manlaibaatar-avik-chaudhuri-yidi-wu.json b/pytorchconf-2024/videos/lightning-talk-whats-new-in-ex-angela-yi-tugsbayasgalan-manlaibaatar-avik-chaudhuri-yidi-wu.json index 35788fe2a..30b954d24 100644 --- a/pytorchconf-2024/videos/lightning-talk-whats-new-in-ex-angela-yi-tugsbayasgalan-manlaibaatar-avik-chaudhuri-yidi-wu.json +++ b/pytorchconf-2024/videos/lightning-talk-whats-new-in-ex-angela-yi-tugsbayasgalan-manlaibaatar-avik-chaudhuri-yidi-wu.json @@ -1,5 +1,5 @@ { - "description": "Lightning Talk: What\u2019s New in Export? - Angela Yi, Tugsbayasgalan Manlaibaatar, Avik Chaudhuri & Yidi Wu, Meta\n\nThis talk discusses updates we've made to torch.export this past year: (a) Non-strict mode, an alternative tracing mode which in practice covers more programs than TorchDynamo without compromising important soundness guarantees (b) Better dynamic shapes specifications through generating suggested fixes and runtime assertions (c) Control flow operators such as cond, map, and associative scan (d) A shift in the export generated IR, which will enable both training and inference (e) An unflattener, which will reconstruct the eager module structure from the flattened exported graph", + "description": "This talk discusses updates we've made to torch.export this past year: (a) Non-strict mode, an alternative tracing mode which in practice covers more programs than TorchDynamo without compromising important soundness guarantees (b) Better dynamic shapes specifications through generating suggested fixes and runtime assertions (c) Control flow operators such as cond, map, and associative scan (d) A shift in the export generated IR, which will enable both training and inference (e) An unflattener, which will reconstruct the eager module structure from the flattened exported graph", "duration": 814, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,14 @@ } ], "speakers": [ - "TODO" + "Angela Yi", + "Tugsbayasgalan Manlaibaatar", + "Avik Chaudhuri", + "Yidi Wu" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/ap1TOf7p0iA/maxresdefault.webp", - "title": "Lightning Talk: What\u2019s New in Ex... Angela Yi, Tugsbayasgalan Manlaibaatar, Avik Chaudhuri & Yidi Wu", + "title": "What\u2019s New in Export?", "videos": [ { "type": "youtube", From 45ab6cd293ccdabb6abb25cbe29a3cd7cc18691b Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Ezequiel=20Leonardo=20Casta=C3=B1o?= <14986783+ELC@users.noreply.github.com> Date: Fri, 21 Nov 2025 19:55:26 -0300 Subject: [PATCH 7/8] Add 'Lightning Talk' tag to all 28 lightning talk files in pytorchconf-2024 --- ...tour-of-pytorch-extension-points-alban-desmaison-meta.json | 4 +++- ...f-time-triton-kernel-libraries-on-rocm-jeff-daily-amd.json | 4 +++- ...vulnerabili-patrick-smyth-dan-fernandez-srishti-hegde.json | 4 +++- ...e-chinese-pytorch-community-resources-tu-zong-zesheng.json | 4 +++- ...flow-based-on-torch-fx-simplify-zifei-xu-tristan-webb.json | 4 +++- ...he-data-lifecycle-shailvi-wakhlu-shailvi-ventures-llc.json | 4 +++- ...-models-in-the-wild-lessons-learned-f-omar-sanseviero.json | 4 +++- ...-and-resources-for-running-generative-a-pareena-verma.json | 4 +++- ...-with-custom-python-c-cuda-operators-richard-zou-meta.json | 4 +++- ...ibuted-training-with-streamingdataset-saaketh-narayan.json | 4 +++- ...ibility-of-pytorch-the-performa-yanbo-liang-horace-he.json | 4 +++- ...owered-pipeline-for-automated-egyptian-h-susi-gentsch.json | 4 +++- ...terable-datasets-what-could-go-wrong-nicolas-hug-meta.json | 4 +++- ...ing-pytorch-on-frontier-exascale-system-vineeth-gutta.json | 4 +++- ...torch-distributed-pipelining-howard-huang-ke-wen-meta.json | 4 +++- ...-accelerators-chen-lai-kimish-patel-cemal-bilgin-meta.json | 4 +++- ...ow-precision-dtypes-in-pytorch-vasiliy-kuznetsov-meta.json | 4 +++- ...neous-memory-capacity-using-pytorch-syed-ahmed-nvidia.json | 4 +++- ...raphy-with-pytorch-low-light-denoising-alexis-baudron.json | 4 +++- ...ckpointing-apis-in-pytorch-jeffrey-wan-horace-he-meta.json | 4 +++- ...d-debugging-with-executorch-olivia-liu-vaun-puri-meta.json | 4 +++- ...ce-on-aarch64-linux-cpus-sunita-nadampalli-amazon-aws.json | 4 +++- ...tning-talk-pytorch-release-process-andrey-talman-meta.json | 4 +++- ...g-talk-pytorch-xla-auto-sharding-yeounoh-chung-google.json | 4 +++- ...ransformers-with-minimal-accuracy-loss-jesse-cai-meta.json | 4 +++- ...-pytorch-models-with-thunder-luca-antiga-lightning-ai.json | 4 +++- ...eveloper-infrastructure-sahan-paliskara-catherine-lee.json | 4 +++- ...yi-tugsbayasgalan-manlaibaatar-avik-chaudhuri-yidi-wu.json | 4 +++- 28 files changed, 84 insertions(+), 28 deletions(-) diff --git a/pytorchconf-2024/videos/lightning-talk-a-whirlwind-tour-of-pytorch-extension-points-alban-desmaison-meta.json b/pytorchconf-2024/videos/lightning-talk-a-whirlwind-tour-of-pytorch-extension-points-alban-desmaison-meta.json index cf07826b9..b6e919bd0 100644 --- a/pytorchconf-2024/videos/lightning-talk-a-whirlwind-tour-of-pytorch-extension-points-alban-desmaison-meta.json +++ b/pytorchconf-2024/videos/lightning-talk-a-whirlwind-tour-of-pytorch-extension-points-alban-desmaison-meta.json @@ -12,7 +12,9 @@ "speakers": [ "Alban Desmaison" ], - "tags": [], + "tags": [ + "Lightning Talk" + ], "thumbnail_url": "https://i.ytimg.com/vi_webp/1A54cUPXhZ4/maxresdefault.webp", "title": "A Whirlwind Tour of PyTorch Extension Points", "videos": [ diff --git a/pytorchconf-2024/videos/lightning-talk-aotriton-ahead-of-time-triton-kernel-libraries-on-rocm-jeff-daily-amd.json b/pytorchconf-2024/videos/lightning-talk-aotriton-ahead-of-time-triton-kernel-libraries-on-rocm-jeff-daily-amd.json index 3add9dee6..3290f88b7 100644 --- a/pytorchconf-2024/videos/lightning-talk-aotriton-ahead-of-time-triton-kernel-libraries-on-rocm-jeff-daily-amd.json +++ b/pytorchconf-2024/videos/lightning-talk-aotriton-ahead-of-time-triton-kernel-libraries-on-rocm-jeff-daily-amd.json @@ -12,7 +12,9 @@ "speakers": [ "Jeff Daily" ], - "tags": [], + "tags": [ + "Lightning Talk" + ], "thumbnail_url": "https://i.ytimg.com/vi_webp/bhplJt1XAMI/maxresdefault.webp", "title": "AOTriton: Ahead of Time Triton Kernel Libraries on ROCm", "videos": [ diff --git a/pytorchconf-2024/videos/lightning-talk-beyond-zero-eliminating-vulnerabili-patrick-smyth-dan-fernandez-srishti-hegde.json b/pytorchconf-2024/videos/lightning-talk-beyond-zero-eliminating-vulnerabili-patrick-smyth-dan-fernandez-srishti-hegde.json index a118a616e..d8c7b3fe8 100644 --- a/pytorchconf-2024/videos/lightning-talk-beyond-zero-eliminating-vulnerabili-patrick-smyth-dan-fernandez-srishti-hegde.json +++ b/pytorchconf-2024/videos/lightning-talk-beyond-zero-eliminating-vulnerabili-patrick-smyth-dan-fernandez-srishti-hegde.json @@ -14,7 +14,9 @@ "Dan Fernandez", "Srishti Hegde" ], - "tags": [], + "tags": [ + "Lightning Talk" + ], "thumbnail_url": "https://i.ytimg.com/vi_webp/1klynk1dxYA/maxresdefault.webp", "title": "Beyond Zero: Eliminating Vulnerabilities in PyTorch Container Images", "videos": [ diff --git a/pytorchconf-2024/videos/lightning-talk-building-and-supporting-the-chinese-pytorch-community-resources-tu-zong-zesheng.json b/pytorchconf-2024/videos/lightning-talk-building-and-supporting-the-chinese-pytorch-community-resources-tu-zong-zesheng.json index a410cd6c1..2b50f461a 100644 --- a/pytorchconf-2024/videos/lightning-talk-building-and-supporting-the-chinese-pytorch-community-resources-tu-zong-zesheng.json +++ b/pytorchconf-2024/videos/lightning-talk-building-and-supporting-the-chinese-pytorch-community-resources-tu-zong-zesheng.json @@ -12,7 +12,9 @@ "speakers": [ "Zong Zesheng" ], - "tags": [], + "tags": [ + "Lightning Talk" + ], "thumbnail_url": "https://i.ytimg.com/vi_webp/AXMN5oUyWKE/maxresdefault.webp", "title": "Building and Supporting the Chinese PyTorch Community: Resources, Tutorials, and Engagement", "videos": [ diff --git a/pytorchconf-2024/videos/lightning-talk-d-matrix-llm-compression-flow-based-on-torch-fx-simplify-zifei-xu-tristan-webb.json b/pytorchconf-2024/videos/lightning-talk-d-matrix-llm-compression-flow-based-on-torch-fx-simplify-zifei-xu-tristan-webb.json index d5a492fee..eb573b7f9 100644 --- a/pytorchconf-2024/videos/lightning-talk-d-matrix-llm-compression-flow-based-on-torch-fx-simplify-zifei-xu-tristan-webb.json +++ b/pytorchconf-2024/videos/lightning-talk-d-matrix-llm-compression-flow-based-on-torch-fx-simplify-zifei-xu-tristan-webb.json @@ -13,7 +13,9 @@ "Zifei Xu", "Tristan Webb" ], - "tags": [], + "tags": [ + "Lightning Talk" + ], "thumbnail_url": "https://i.ytimg.com/vi/cw6dSlmMwj8/maxresdefault.jpg", "title": "d-Matrix LLM Compression Flow Based on Torch.Fx: Simplifying PTQ/QAT", "videos": [ diff --git a/pytorchconf-2024/videos/lightning-talk-debiasing-the-data-lifecycle-shailvi-wakhlu-shailvi-ventures-llc.json b/pytorchconf-2024/videos/lightning-talk-debiasing-the-data-lifecycle-shailvi-wakhlu-shailvi-ventures-llc.json index 8bafbb9b1..e10f0f41a 100644 --- a/pytorchconf-2024/videos/lightning-talk-debiasing-the-data-lifecycle-shailvi-wakhlu-shailvi-ventures-llc.json +++ b/pytorchconf-2024/videos/lightning-talk-debiasing-the-data-lifecycle-shailvi-wakhlu-shailvi-ventures-llc.json @@ -12,7 +12,9 @@ "speakers": [ "Shailvi Wakhlu" ], - "tags": [], + "tags": [ + "Lightning Talk" + ], "thumbnail_url": "https://i.ytimg.com/vi/V9NVEceY9Wg/maxresdefault.jpg", "title": "Debiasing the Data Lifecycle", "videos": [ diff --git a/pytorchconf-2024/videos/lightning-talk-distributing-a-million-open-models-in-the-wild-lessons-learned-f-omar-sanseviero.json b/pytorchconf-2024/videos/lightning-talk-distributing-a-million-open-models-in-the-wild-lessons-learned-f-omar-sanseviero.json index 37914c830..08d117ca1 100644 --- a/pytorchconf-2024/videos/lightning-talk-distributing-a-million-open-models-in-the-wild-lessons-learned-f-omar-sanseviero.json +++ b/pytorchconf-2024/videos/lightning-talk-distributing-a-million-open-models-in-the-wild-lessons-learned-f-omar-sanseviero.json @@ -12,7 +12,9 @@ "speakers": [ "Omar Sanseviero" ], - "tags": [], + "tags": [ + "Lightning Talk" + ], "thumbnail_url": "https://i.ytimg.com/vi_webp/t3AHZzz8Ong/maxresdefault.webp", "title": "Distributing a Million Open Models in the Wild: Lessons Learned from the Hugging Face Hub", "videos": [ diff --git a/pytorchconf-2024/videos/lightning-talk-empowering-developers-tools-and-resources-for-running-generative-a-pareena-verma.json b/pytorchconf-2024/videos/lightning-talk-empowering-developers-tools-and-resources-for-running-generative-a-pareena-verma.json index eaa9b9659..2eff22565 100644 --- a/pytorchconf-2024/videos/lightning-talk-empowering-developers-tools-and-resources-for-running-generative-a-pareena-verma.json +++ b/pytorchconf-2024/videos/lightning-talk-empowering-developers-tools-and-resources-for-running-generative-a-pareena-verma.json @@ -16,7 +16,9 @@ "speakers": [ "Pareena Verma" ], - "tags": [], + "tags": [ + "Lightning Talk" + ], "thumbnail_url": "https://i.ytimg.com/vi_webp/wkzBHF9CL5s/maxresdefault.webp", "title": "Empowering Developers: Tools and Resources for Running Generative AI on Arm CPUs", "videos": [ diff --git a/pytorchconf-2024/videos/lightning-talk-extending-pytorch-with-custom-python-c-cuda-operators-richard-zou-meta.json b/pytorchconf-2024/videos/lightning-talk-extending-pytorch-with-custom-python-c-cuda-operators-richard-zou-meta.json index 28050a7ac..2104544ca 100644 --- a/pytorchconf-2024/videos/lightning-talk-extending-pytorch-with-custom-python-c-cuda-operators-richard-zou-meta.json +++ b/pytorchconf-2024/videos/lightning-talk-extending-pytorch-with-custom-python-c-cuda-operators-richard-zou-meta.json @@ -12,7 +12,9 @@ "speakers": [ "Richard Zou" ], - "tags": [], + "tags": [ + "Lightning Talk" + ], "thumbnail_url": "https://i.ytimg.com/vi_webp/LI3h8aVchwo/maxresdefault.webp", "title": "Extending PyTorch with Custom Python/C++/CUDA Operators", "videos": [ diff --git a/pytorchconf-2024/videos/lightning-talk-fast-scalable-distributed-training-with-streamingdataset-saaketh-narayan.json b/pytorchconf-2024/videos/lightning-talk-fast-scalable-distributed-training-with-streamingdataset-saaketh-narayan.json index 9a63f0c39..77a2f3d65 100644 --- a/pytorchconf-2024/videos/lightning-talk-fast-scalable-distributed-training-with-streamingdataset-saaketh-narayan.json +++ b/pytorchconf-2024/videos/lightning-talk-fast-scalable-distributed-training-with-streamingdataset-saaketh-narayan.json @@ -12,7 +12,9 @@ "speakers": [ "Saaketh Narayan" ], - "tags": [], + "tags": [ + "Lightning Talk" + ], "thumbnail_url": "https://i.ytimg.com/vi_webp/C1TpUZjqTfA/maxresdefault.webp", "title": "Fast, Scalable Distributed Training with StreamingDataset", "videos": [ diff --git a/pytorchconf-2024/videos/lightning-talk-flexattention-the-flexibility-of-pytorch-the-performa-yanbo-liang-horace-he.json b/pytorchconf-2024/videos/lightning-talk-flexattention-the-flexibility-of-pytorch-the-performa-yanbo-liang-horace-he.json index b53280314..134926283 100644 --- a/pytorchconf-2024/videos/lightning-talk-flexattention-the-flexibility-of-pytorch-the-performa-yanbo-liang-horace-he.json +++ b/pytorchconf-2024/videos/lightning-talk-flexattention-the-flexibility-of-pytorch-the-performa-yanbo-liang-horace-he.json @@ -13,7 +13,9 @@ "Yanbo Liang", "Horace He" ], - "tags": [], + "tags": [ + "Lightning Talk" + ], "thumbnail_url": "https://i.ytimg.com/vi_webp/ju-KlcuWlbk/maxresdefault.webp", "title": "FlexAttention - The Flexibility of PyTorch + The Performance of FlashAttention", "videos": [ diff --git a/pytorchconf-2024/videos/lightning-talk-hieroglyph2text-a-pytorch-powered-pipeline-for-automated-egyptian-h-susi-gentsch.json b/pytorchconf-2024/videos/lightning-talk-hieroglyph2text-a-pytorch-powered-pipeline-for-automated-egyptian-h-susi-gentsch.json index 20d79f855..d58aa176f 100644 --- a/pytorchconf-2024/videos/lightning-talk-hieroglyph2text-a-pytorch-powered-pipeline-for-automated-egyptian-h-susi-gentsch.json +++ b/pytorchconf-2024/videos/lightning-talk-hieroglyph2text-a-pytorch-powered-pipeline-for-automated-egyptian-h-susi-gentsch.json @@ -12,7 +12,9 @@ "speakers": [ "Susi Gentsch" ], - "tags": [], + "tags": [ + "Lightning Talk" + ], "thumbnail_url": "https://i.ytimg.com/vi_webp/eDjHKOb9Tn4/maxresdefault.webp", "title": "HieroGlyph2Text: A PyTorch-Powered Pipeline for Automated Egyptian Hieroglyph Translation from Image", "videos": [ diff --git a/pytorchconf-2024/videos/lightning-talk-implementing-and-using-iterable-datasets-what-could-go-wrong-nicolas-hug-meta.json b/pytorchconf-2024/videos/lightning-talk-implementing-and-using-iterable-datasets-what-could-go-wrong-nicolas-hug-meta.json index 739f8b100..345b526e7 100644 --- a/pytorchconf-2024/videos/lightning-talk-implementing-and-using-iterable-datasets-what-could-go-wrong-nicolas-hug-meta.json +++ b/pytorchconf-2024/videos/lightning-talk-implementing-and-using-iterable-datasets-what-could-go-wrong-nicolas-hug-meta.json @@ -12,7 +12,9 @@ "speakers": [ "Nicolas Hug" ], - "tags": [], + "tags": [ + "Lightning Talk" + ], "thumbnail_url": "https://i.ytimg.com/vi_webp/Kx0BoPa9juQ/maxresdefault.webp", "title": "Implementing and Using Iterable Datasets: What Could Go Wrong?", "videos": [ diff --git a/pytorchconf-2024/videos/lightning-talk-in-transit-machine-learning-using-pytorch-on-frontier-exascale-system-vineeth-gutta.json b/pytorchconf-2024/videos/lightning-talk-in-transit-machine-learning-using-pytorch-on-frontier-exascale-system-vineeth-gutta.json index a53d5c400..bdd6864cd 100644 --- a/pytorchconf-2024/videos/lightning-talk-in-transit-machine-learning-using-pytorch-on-frontier-exascale-system-vineeth-gutta.json +++ b/pytorchconf-2024/videos/lightning-talk-in-transit-machine-learning-using-pytorch-on-frontier-exascale-system-vineeth-gutta.json @@ -12,7 +12,9 @@ "speakers": [ "Vineeth Gutta" ], - "tags": [], + "tags": [ + "Lightning Talk" + ], "thumbnail_url": "https://i.ytimg.com/vi_webp/s4Gh2UX0EyA/maxresdefault.webp", "title": "In-Transit Machine Learning Using PyTorch on Frontier Exascale System", "videos": [ diff --git a/pytorchconf-2024/videos/lightning-talk-introduction-to-torch-distributed-pipelining-howard-huang-ke-wen-meta.json b/pytorchconf-2024/videos/lightning-talk-introduction-to-torch-distributed-pipelining-howard-huang-ke-wen-meta.json index cec4a2dca..bb3955b38 100644 --- a/pytorchconf-2024/videos/lightning-talk-introduction-to-torch-distributed-pipelining-howard-huang-ke-wen-meta.json +++ b/pytorchconf-2024/videos/lightning-talk-introduction-to-torch-distributed-pipelining-howard-huang-ke-wen-meta.json @@ -13,7 +13,9 @@ "Howard Huang", "Ke Wen" ], - "tags": [], + "tags": [ + "Lightning Talk" + ], "thumbnail_url": "https://i.ytimg.com/vi_webp/YieoyQll4_4/maxresdefault.webp", "title": "Introduction to Torch.Distributed.Pipelining", "videos": [ diff --git a/pytorchconf-2024/videos/lightning-talk-llms-on-edge-with-ai-accelerators-chen-lai-kimish-patel-cemal-bilgin-meta.json b/pytorchconf-2024/videos/lightning-talk-llms-on-edge-with-ai-accelerators-chen-lai-kimish-patel-cemal-bilgin-meta.json index c46f9c1b0..5ff60315d 100644 --- a/pytorchconf-2024/videos/lightning-talk-llms-on-edge-with-ai-accelerators-chen-lai-kimish-patel-cemal-bilgin-meta.json +++ b/pytorchconf-2024/videos/lightning-talk-llms-on-edge-with-ai-accelerators-chen-lai-kimish-patel-cemal-bilgin-meta.json @@ -14,7 +14,9 @@ "Kimish Patel", "Cemal Bilgin" ], - "tags": [], + "tags": [ + "Lightning Talk" + ], "thumbnail_url": "https://i.ytimg.com/vi_webp/8p8Pntnr3GU/maxresdefault.webp", "title": "LLMs on Edge with AI Accelerators", "videos": [ diff --git a/pytorchconf-2024/videos/lightning-talk-low-precision-dtypes-in-pytorch-vasiliy-kuznetsov-meta.json b/pytorchconf-2024/videos/lightning-talk-low-precision-dtypes-in-pytorch-vasiliy-kuznetsov-meta.json index 579d0fb20..6379c868b 100644 --- a/pytorchconf-2024/videos/lightning-talk-low-precision-dtypes-in-pytorch-vasiliy-kuznetsov-meta.json +++ b/pytorchconf-2024/videos/lightning-talk-low-precision-dtypes-in-pytorch-vasiliy-kuznetsov-meta.json @@ -12,7 +12,9 @@ "speakers": [ "Vasiliy Kuznetsov" ], - "tags": [], + "tags": [ + "Lightning Talk" + ], "thumbnail_url": "https://i.ytimg.com/vi_webp/xcKwEZ77Cps/maxresdefault.webp", "title": "Low Precision Dtypes in PyTorch", "videos": [ diff --git a/pytorchconf-2024/videos/lightning-talk-making-the-most-of-heterogeneous-memory-capacity-using-pytorch-syed-ahmed-nvidia.json b/pytorchconf-2024/videos/lightning-talk-making-the-most-of-heterogeneous-memory-capacity-using-pytorch-syed-ahmed-nvidia.json index b9ff20708..86afbddeb 100644 --- a/pytorchconf-2024/videos/lightning-talk-making-the-most-of-heterogeneous-memory-capacity-using-pytorch-syed-ahmed-nvidia.json +++ b/pytorchconf-2024/videos/lightning-talk-making-the-most-of-heterogeneous-memory-capacity-using-pytorch-syed-ahmed-nvidia.json @@ -12,7 +12,9 @@ "speakers": [ "Syed Ahmed" ], - "tags": [], + "tags": [ + "Lightning Talk" + ], "thumbnail_url": "https://i.ytimg.com/vi_webp/srQOK1UusZ4/maxresdefault.webp", "title": "Making the Most of Heterogeneous Memory Capacity Using PyTorch", "videos": [ diff --git a/pytorchconf-2024/videos/lightning-talk-mobile-computational-photography-with-pytorch-low-light-denoising-alexis-baudron.json b/pytorchconf-2024/videos/lightning-talk-mobile-computational-photography-with-pytorch-low-light-denoising-alexis-baudron.json index 6902dd77f..918bfd293 100644 --- a/pytorchconf-2024/videos/lightning-talk-mobile-computational-photography-with-pytorch-low-light-denoising-alexis-baudron.json +++ b/pytorchconf-2024/videos/lightning-talk-mobile-computational-photography-with-pytorch-low-light-denoising-alexis-baudron.json @@ -12,7 +12,9 @@ "speakers": [ "Alexis Baudron" ], - "tags": [], + "tags": [ + "Lightning Talk" + ], "thumbnail_url": "https://i.ytimg.com/vi_webp/hjOtsOqPw3Y/maxresdefault.webp", "title": "Mobile Computational Photography with PyTorch: Low-Light Denoising", "videos": [ diff --git a/pytorchconf-2024/videos/lightning-talk-new-activation-checkpointing-apis-in-pytorch-jeffrey-wan-horace-he-meta.json b/pytorchconf-2024/videos/lightning-talk-new-activation-checkpointing-apis-in-pytorch-jeffrey-wan-horace-he-meta.json index dd57c5387..7e6ecd65b 100644 --- a/pytorchconf-2024/videos/lightning-talk-new-activation-checkpointing-apis-in-pytorch-jeffrey-wan-horace-he-meta.json +++ b/pytorchconf-2024/videos/lightning-talk-new-activation-checkpointing-apis-in-pytorch-jeffrey-wan-horace-he-meta.json @@ -13,7 +13,9 @@ "Jeffrey Wan", "Horace He" ], - "tags": [], + "tags": [ + "Lightning Talk" + ], "thumbnail_url": "https://i.ytimg.com/vi_webp/v3gsrJtGLiA/maxresdefault.webp", "title": "New Activation Checkpointing APIs in PyTorch", "videos": [ diff --git a/pytorchconf-2024/videos/lightning-talk-on-device-profiling-and-debugging-with-executorch-olivia-liu-vaun-puri-meta.json b/pytorchconf-2024/videos/lightning-talk-on-device-profiling-and-debugging-with-executorch-olivia-liu-vaun-puri-meta.json index 36858fe8f..ca5c3d02c 100644 --- a/pytorchconf-2024/videos/lightning-talk-on-device-profiling-and-debugging-with-executorch-olivia-liu-vaun-puri-meta.json +++ b/pytorchconf-2024/videos/lightning-talk-on-device-profiling-and-debugging-with-executorch-olivia-liu-vaun-puri-meta.json @@ -13,7 +13,9 @@ "Olivia Liu", "Vaun Puri" ], - "tags": [], + "tags": [ + "Lightning Talk" + ], "thumbnail_url": "https://i.ytimg.com/vi_webp/_e6E6fnEIDQ/maxresdefault.webp", "title": "On-Device Profiling and Debugging with ExecuTorch", "videos": [ diff --git a/pytorchconf-2024/videos/lightning-talk-optimized-pytorch-inference-on-aarch64-linux-cpus-sunita-nadampalli-amazon-aws.json b/pytorchconf-2024/videos/lightning-talk-optimized-pytorch-inference-on-aarch64-linux-cpus-sunita-nadampalli-amazon-aws.json index 8e57caebf..5dd70a8e1 100644 --- a/pytorchconf-2024/videos/lightning-talk-optimized-pytorch-inference-on-aarch64-linux-cpus-sunita-nadampalli-amazon-aws.json +++ b/pytorchconf-2024/videos/lightning-talk-optimized-pytorch-inference-on-aarch64-linux-cpus-sunita-nadampalli-amazon-aws.json @@ -12,7 +12,9 @@ "speakers": [ "Sunita Nadampalli" ], - "tags": [], + "tags": [ + "Lightning Talk" + ], "thumbnail_url": "https://i.ytimg.com/vi_webp/_9xohmSa2G8/maxresdefault.webp", "title": "Optimized PyTorch Inference on aarch64 Linux CPUs", "videos": [ diff --git a/pytorchconf-2024/videos/lightning-talk-pytorch-release-process-andrey-talman-meta.json b/pytorchconf-2024/videos/lightning-talk-pytorch-release-process-andrey-talman-meta.json index 6e0de648a..e0e730184 100644 --- a/pytorchconf-2024/videos/lightning-talk-pytorch-release-process-andrey-talman-meta.json +++ b/pytorchconf-2024/videos/lightning-talk-pytorch-release-process-andrey-talman-meta.json @@ -12,7 +12,9 @@ "speakers": [ "Andrey Talman" ], - "tags": [], + "tags": [ + "Lightning Talk" + ], "thumbnail_url": "https://i.ytimg.com/vi_webp/4REnTJO3_ps/maxresdefault.webp", "title": "PyTorch Release Process", "videos": [ diff --git a/pytorchconf-2024/videos/lightning-talk-pytorch-xla-auto-sharding-yeounoh-chung-google.json b/pytorchconf-2024/videos/lightning-talk-pytorch-xla-auto-sharding-yeounoh-chung-google.json index 1f16ece47..d195c05f4 100644 --- a/pytorchconf-2024/videos/lightning-talk-pytorch-xla-auto-sharding-yeounoh-chung-google.json +++ b/pytorchconf-2024/videos/lightning-talk-pytorch-xla-auto-sharding-yeounoh-chung-google.json @@ -12,7 +12,9 @@ "speakers": [ "Yeounoh Chung" ], - "tags": [], + "tags": [ + "Lightning Talk" + ], "thumbnail_url": "https://i.ytimg.com/vi_webp/gNfyS9asA7A/maxresdefault.webp", "title": "PyTorch/XLA Auto-Sharding", "videos": [ diff --git a/pytorchconf-2024/videos/lightning-talk-sparsifying-vision-transformers-with-minimal-accuracy-loss-jesse-cai-meta.json b/pytorchconf-2024/videos/lightning-talk-sparsifying-vision-transformers-with-minimal-accuracy-loss-jesse-cai-meta.json index cae9d698c..4eb22749f 100644 --- a/pytorchconf-2024/videos/lightning-talk-sparsifying-vision-transformers-with-minimal-accuracy-loss-jesse-cai-meta.json +++ b/pytorchconf-2024/videos/lightning-talk-sparsifying-vision-transformers-with-minimal-accuracy-loss-jesse-cai-meta.json @@ -12,7 +12,9 @@ "speakers": [ "Jesse Cai" ], - "tags": [], + "tags": [ + "Lightning Talk" + ], "thumbnail_url": "https://i.ytimg.com/vi_webp/mqoIfs5hTxA/maxresdefault.webp", "title": "Sparsifying Vision Transformers with Minimal Accuracy Loss", "videos": [ diff --git a/pytorchconf-2024/videos/lightning-talk-understanding-and-optimizing-pytorch-models-with-thunder-luca-antiga-lightning-ai.json b/pytorchconf-2024/videos/lightning-talk-understanding-and-optimizing-pytorch-models-with-thunder-luca-antiga-lightning-ai.json index b73a0424e..15ac386ee 100644 --- a/pytorchconf-2024/videos/lightning-talk-understanding-and-optimizing-pytorch-models-with-thunder-luca-antiga-lightning-ai.json +++ b/pytorchconf-2024/videos/lightning-talk-understanding-and-optimizing-pytorch-models-with-thunder-luca-antiga-lightning-ai.json @@ -12,7 +12,9 @@ "speakers": [ "Luca Antiga" ], - "tags": [], + "tags": [ + "Lightning Talk" + ], "thumbnail_url": "https://i.ytimg.com/vi_webp/5dQ423cFfLc/maxresdefault.webp", "title": "Understanding and Optimizing PyTorch Models with Thunder", "videos": [ diff --git a/pytorchconf-2024/videos/lightning-talk-what-s-new-for-pytorch-developer-infrastructure-sahan-paliskara-catherine-lee.json b/pytorchconf-2024/videos/lightning-talk-what-s-new-for-pytorch-developer-infrastructure-sahan-paliskara-catherine-lee.json index 50dba65fa..95e1dc968 100644 --- a/pytorchconf-2024/videos/lightning-talk-what-s-new-for-pytorch-developer-infrastructure-sahan-paliskara-catherine-lee.json +++ b/pytorchconf-2024/videos/lightning-talk-what-s-new-for-pytorch-developer-infrastructure-sahan-paliskara-catherine-lee.json @@ -13,7 +13,9 @@ "Sahan Paliskara", "Catherine Lee" ], - "tags": [], + "tags": [ + "Lightning Talk" + ], "thumbnail_url": "https://i.ytimg.com/vi_webp/xlhBDySu7jM/maxresdefault.webp", "title": "What's New for PyTorch Developer Infrastructure", "videos": [ diff --git a/pytorchconf-2024/videos/lightning-talk-whats-new-in-ex-angela-yi-tugsbayasgalan-manlaibaatar-avik-chaudhuri-yidi-wu.json b/pytorchconf-2024/videos/lightning-talk-whats-new-in-ex-angela-yi-tugsbayasgalan-manlaibaatar-avik-chaudhuri-yidi-wu.json index 30b954d24..ab652a0c0 100644 --- a/pytorchconf-2024/videos/lightning-talk-whats-new-in-ex-angela-yi-tugsbayasgalan-manlaibaatar-avik-chaudhuri-yidi-wu.json +++ b/pytorchconf-2024/videos/lightning-talk-whats-new-in-ex-angela-yi-tugsbayasgalan-manlaibaatar-avik-chaudhuri-yidi-wu.json @@ -15,7 +15,9 @@ "Avik Chaudhuri", "Yidi Wu" ], - "tags": [], + "tags": [ + "Lightning Talk" + ], "thumbnail_url": "https://i.ytimg.com/vi_webp/ap1TOf7p0iA/maxresdefault.webp", "title": "What\u2019s New in Export?", "videos": [ From af7be93701eee7e7f37486746f39289319282d10 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Ezequiel=20Leonardo=20Casta=C3=B1o?= <14986783+ELC@users.noreply.github.com> Date: Fri, 21 Nov 2025 21:14:47 -0300 Subject: [PATCH 8/8] Update video metadata for PyTorch Conference 2024 - Refined descriptions by removing speaker names and organization details. - Extracted and added speaker names for several talks. - Cleaned up titles by removing speaker names and unnecessary prefixes. --- ...compile-and-fsdp-l-chu-a-viros-i-martin-b-vaughan.json | 8 +++++--- ...-development-spencer-whitman-vincent-gonguet-meta.json | 7 ++++--- ...n-of-kernels-and-compilers-jacques-pienaar-google.json | 6 +++--- ...ghts-with-max-and-mojo-mikhail-zolotukhin-modular.json | 6 +++--- ...-fine-tuning-mini-summit-kartikay-khandelwal-meta.json | 6 +++--- 5 files changed, 18 insertions(+), 15 deletions(-) diff --git a/pytorchconf-2024/videos/maximizing-training-throughput-using-torch-compile-and-fsdp-l-chu-a-viros-i-martin-b-vaughan.json b/pytorchconf-2024/videos/maximizing-training-throughput-using-torch-compile-and-fsdp-l-chu-a-viros-i-martin-b-vaughan.json index 0eb19b628..cc00fc031 100644 --- a/pytorchconf-2024/videos/maximizing-training-throughput-using-torch-compile-and-fsdp-l-chu-a-viros-i-martin-b-vaughan.json +++ b/pytorchconf-2024/videos/maximizing-training-throughput-using-torch-compile-and-fsdp-l-chu-a-viros-i-martin-b-vaughan.json @@ -1,5 +1,5 @@ { - "description": "Maximizing Training Throughput Using Torch.Compile and FSDP - Linsong Chu & Antoni Viros i Martin, IBM Research; Brian Vaughan, IBM\n\ntorch.compile is a graph compilation technique that improves GPU utilization. A key challenge in getting torch.compile to perform well is to minimize (or eliminate) graph breaks, however, this isn't trivial as even the Llama implementation provided by Meta has many graph breaks resulting in reduced training throughput. In this talk we discuss 1. how we addressed these challenges in order to train a model using torch.compile 2. how we combined torch.compile with FSDP and selective activation checkpointing to achieve the maximum throughput for training 3. model quality comparison between models trained with compile and no-compile, and lastly 4. the best setup we have for different model sizes in the Llama family that achieves the maximum throughput and MFU number (e.g. 68% MFU for the 7B model on A100 GPUs!)", + "description": "torch.compile is a graph compilation technique that improves GPU utilization. A key challenge in getting torch.compile to perform well is to minimize (or eliminate) graph breaks, however, this isn't trivial as even the Llama implementation provided by Meta has many graph breaks resulting in reduced training throughput. In this talk we discuss 1. how we addressed these challenges in order to train a model using torch.compile 2. how we combined torch.compile with FSDP and selective activation checkpointing to achieve the maximum throughput for training 3. model quality comparison between models trained with compile and no-compile, and lastly 4. the best setup we have for different model sizes in the Llama family that achieves the maximum throughput and MFU number (e.g. 68% MFU for the 7B model on A100 GPUs!)", "duration": 220, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,13 @@ } ], "speakers": [ - "TODO" + "Linsong Chu", + "Antoni Viros i Martin", + "Brian Vaughan" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/_CuLeABf_fM/maxresdefault.webp", - "title": "Maximizing Training Throughput Using Torch.Compile and FSDP - L. Chu, A. Viros i Martin, B. Vaughan", + "title": "Maximizing Training Throughput Using Torch.Compile and FSDP", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/meta-llama-3-and-the-future-of-responsible-ai-development-spencer-whitman-vincent-gonguet-meta.json b/pytorchconf-2024/videos/meta-llama-3-and-the-future-of-responsible-ai-development-spencer-whitman-vincent-gonguet-meta.json index 7bdd120de..82eb196ac 100644 --- a/pytorchconf-2024/videos/meta-llama-3-and-the-future-of-responsible-ai-development-spencer-whitman-vincent-gonguet-meta.json +++ b/pytorchconf-2024/videos/meta-llama-3-and-the-future-of-responsible-ai-development-spencer-whitman-vincent-gonguet-meta.json @@ -1,5 +1,5 @@ { - "description": "Meta Llama 3 and the Future of Responsible AI Development - Spencer Whitman & Vincent Gonguet, Meta\n\nAs AI models become increasingly powerful and pervasive, trust and safety have become top priorities. Join us for a timely talk on Llama 3, our latest foundation model, and the cutting-edge trust and safety models and tools we've developed to ensure responsible AI development. In this talk, we'll dive into: \u2022The advancements of Llama 3 and its applications \u2022Our innovative trust and safety approaches, including toxicity detection and mitigation \u2022The open-source tools and resources we're sharing to empower the community Discover how Meta is pushing the boundaries of trust and safety and learn how you can integrate these solutions into your own projects. Let's build a safer, more responsible AI future together!", + "description": "As AI models become increasingly powerful and pervasive, trust and safety have become top priorities. Join us for a timely talk on Llama 3, our latest foundation model, and the cutting-edge trust and safety models and tools we've developed to ensure responsible AI development. In this talk, we'll dive into: \u2022The advancements of Llama 3 and its applications \u2022Our innovative trust and safety approaches, including toxicity detection and mitigation \u2022The open-source tools and resources we're sharing to empower the community Discover how Meta is pushing the boundaries of trust and safety and learn how you can integrate these solutions into your own projects. Let's build a safer, more responsible AI future together!", "duration": 1251, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,12 @@ } ], "speakers": [ - "TODO" + "Spencer Whitman", + "Vincent Gonguet" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/XOIuFIl2-Ao/maxresdefault.webp", - "title": "Meta Llama 3 and the Future of Responsible AI Development - Spencer Whitman & Vincent Gonguet, Meta", + "title": "Meta Llama 3 and the Future of Responsible AI Development", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/mlir-enabling-composition-of-kernels-and-compilers-jacques-pienaar-google.json b/pytorchconf-2024/videos/mlir-enabling-composition-of-kernels-and-compilers-jacques-pienaar-google.json index 98597e37b..b214dbc96 100644 --- a/pytorchconf-2024/videos/mlir-enabling-composition-of-kernels-and-compilers-jacques-pienaar-google.json +++ b/pytorchconf-2024/videos/mlir-enabling-composition-of-kernels-and-compilers-jacques-pienaar-google.json @@ -1,5 +1,5 @@ { - "description": "[MLIR] Enabling Composition of Kernels and Compilers - Jacques Pienaar, Google\n\nHand written kernels and compilers have been part of the toolbox to provide efficient and broad coverage. These approaches have often been positioned as being at odds with one another - and indeed the software solutions either side have sometimes made it such. MLIR, since inception, aimed to enable general, beneficial composition instead. Rather than treating kernels as a black box escape hatch, treat it as a peer in solving the serving needs. This is not magic and requires consideration of how best to combine. In this talk I'll present the approach and effect of this both in IREE and OpenXLA.", + "description": "Hand written kernels and compilers have been part of the toolbox to provide efficient and broad coverage. These approaches have often been positioned as being at odds with one another - and indeed the software solutions either side have sometimes made it such. MLIR, since inception, aimed to enable general, beneficial composition instead. Rather than treating kernels as a black box escape hatch, treat it as a peer in solving the serving needs. This is not magic and requires consideration of how best to combine. In this talk I'll present the approach and effect of this both in IREE and OpenXLA.", "duration": 672, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Jacques Pienaar" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/Dx1fAE9fk8s/maxresdefault.webp", - "title": "[MLIR] Enabling Composition of Kernels and Compilers - Jacques Pienaar, Google", + "title": "[MLIR] Enabling Composition of Kernels and Compilers", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/mojo-lifting-pt-to-new-heights-with-max-and-mojo-mikhail-zolotukhin-modular.json b/pytorchconf-2024/videos/mojo-lifting-pt-to-new-heights-with-max-and-mojo-mikhail-zolotukhin-modular.json index e1d1de8eb..caac5fc1d 100644 --- a/pytorchconf-2024/videos/mojo-lifting-pt-to-new-heights-with-max-and-mojo-mikhail-zolotukhin-modular.json +++ b/pytorchconf-2024/videos/mojo-lifting-pt-to-new-heights-with-max-and-mojo-mikhail-zolotukhin-modular.json @@ -1,5 +1,5 @@ { - "description": "[MOJO] Lifting PT to New Heights with MAX and Mojo - Mikhail Zolotukhin, Modular\n\nIn this talk we'll peek into Modular's inference engine: how it builds on and works with PyTorch and what is unique about it. We will look into how Mojo language can be used to define performant kernels and what optimizations the inference engine can perform. We will also talk briefly about our experience of developing a third party backend for torch.compile.", + "description": "In this talk we'll peek into Modular's inference engine: how it builds on and works with PyTorch and what is unique about it. We will look into how Mojo language can be used to define performant kernels and what optimizations the inference engine can perform. We will also talk briefly about our experience of developing a third party backend for torch.compile.", "duration": 572, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Mikhail Zolotukhin" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/JmHKhc6EGpg/maxresdefault.webp", - "title": "[MOJO] Lifting PT to New Heights with MAX and Mojo - Mikhail Zolotukhin, Modular", + "title": "[MOJO] Lifting PT to New Heights with MAX and Mojo", "videos": [ { "type": "youtube", diff --git a/pytorchconf-2024/videos/welcome-to-the-pytorch-ecosystem-for-llm-fine-tuning-mini-summit-kartikay-khandelwal-meta.json b/pytorchconf-2024/videos/welcome-to-the-pytorch-ecosystem-for-llm-fine-tuning-mini-summit-kartikay-khandelwal-meta.json index 378111517..393385327 100644 --- a/pytorchconf-2024/videos/welcome-to-the-pytorch-ecosystem-for-llm-fine-tuning-mini-summit-kartikay-khandelwal-meta.json +++ b/pytorchconf-2024/videos/welcome-to-the-pytorch-ecosystem-for-llm-fine-tuning-mini-summit-kartikay-khandelwal-meta.json @@ -1,5 +1,5 @@ { - "description": "Welcome to the PyTorch Ecosystem for LLM Fine-tuning Mini Summit - Kartikay Khandelwal, Meta\n\nAs open-source LLMs have become more capable, a substantial ecosystem has developed around the fine-tuning of these models. A thriving community of researchers, developers, practitioners and hobbyists has emerged which focuses on topics ranging from memory efficiency, parameter-efficient fine-tuning and quantization to performance at scale and reproducible evaluations. The goal of this mini-summit is to bring this community together to discuss ideas, share knowledge and build connections.\n\nThe agenda features a keynote from Joe Spisak on the state of the Llama ecosystem followed by invited talks from the founders of Axolotl, Unsloth and torchtune. We conclude the summit with a riveting discussion on what\u2019s next for LLMs, fine-tuning and the PyTorch ecosystem with a fabulous panel of experts - Tim Dettmers (author of bitsandbytes and QLoRA), Hailey Schoelkopf (maintainer of LM Eval Harness at EleutherAI), Aakanksha Chowdhery (Lead author on PaLM and Gemini) and Alexis Conneau (Research Lead at OpenAI)", + "description": "As open-source LLMs have become more capable, a substantial ecosystem has developed around the fine-tuning of these models. A thriving community of researchers, developers, practitioners and hobbyists has emerged which focuses on topics ranging from memory efficiency, parameter-efficient fine-tuning and quantization to performance at scale and reproducible evaluations. The goal of this mini-summit is to bring this community together to discuss ideas, share knowledge and build connections.\n\nThe agenda features a keynote from Joe Spisak on the state of the Llama ecosystem followed by invited talks from the founders of Axolotl, Unsloth and torchtune. We conclude the summit with a riveting discussion on what\u2019s next for LLMs, fine-tuning and the PyTorch ecosystem with a fabulous panel of experts - Tim Dettmers (author of bitsandbytes and QLoRA), Hailey Schoelkopf (maintainer of LM Eval Harness at EleutherAI), Aakanksha Chowdhery (Lead author on PaLM and Gemini) and Alexis Conneau (Research Lead at OpenAI)", "duration": 81, "language": "eng", "recorded": "2024-09-18", @@ -10,11 +10,11 @@ } ], "speakers": [ - "TODO" + "Kartikay Khandelwal" ], "tags": [], "thumbnail_url": "https://i.ytimg.com/vi_webp/Pe_VT5ReB3U/maxresdefault.webp", - "title": "Welcome to the PyTorch Ecosystem for LLM Fine-tuning Mini Summit - Kartikay Khandelwal, Meta", + "title": "Welcome to the PyTorch Ecosystem for LLM Fine-tuning Mini Summit", "videos": [ { "type": "youtube",