APS Logo

Particle Jet Representations via a Joint Embedding Predictive Architecture

ORAL

Abstract

In high energy physics, self-supervised learning methods have the potential to aid in the creation of machine learning models without the need for labeled datasets for a variety of tasks, including those related to jets---narrow sprays of particles produced by quarks and gluons in high energy particle collisions. This study introduces an approach to learning augmentation-independent jet representations using a Jet-based Joint Embedding Predictive Architecture (J-JEPA). This approach aims to predict various physical targets from an informative context, using target positions as joint information. As an augmentation-free method, J-JEPA avoids introducing biases that could harm downstream tasks, which often require invariance under augmentations different from those used in pretraining. This augmentation-independent training enables versatile applications, offering a pathway toward a cross-task foundation model. We fine-tuned the representations learned by J-JEPA for jet tagging and benchmarked them against task-specific representations.

Publication: Accepted by the Machine Learning and the Physical Sciences workshop at NeurIPS 2024.

Presenters

  • Zihan Zhao

    University of California, San Diego

Authors

  • Zihan Zhao

    University of California, San Diego

  • Haoyang Li

    University of California, San Diego

  • Subash Katel

    University of California, San Diego

  • Raghav Kansal

    California Institute of Technology

  • Farouk Mokhtar

    University of California, San Diego

  • Javier M Duarte

    University of California, San Diego