EFMxDM @ PAKDD 2026

9 June 2026, Hong Kong
Efficient Foundation Models for Large-Scale Data Mining

About EFMxDM

This workshop aims to advance the development of efficient foundation models (FM) that can support large-scale data mining tasks. It focuses on techniques that reduce training and inference costs, enable dynamic architectural adaptation, and improve foundation model usability in domain-specific or data-constrained scenarios. The objective is to bring together researchers from machine learning, data mining, and systems communities to explore practical and theoretically grounded methods for scaling foundation models efficiently. The workshop also seeks to stimulate discussions on next-generation architectures that balance performance with deployability. The workshop will cover techniques, systems, and applications centered on making foundation models scalable, efficient, and adaptable for large-scale data mining.
We welcome researchers from machine learning, data mining, distributed systems, and applied AI to collaborate and contribute.

Important Dates

Submission Deadline

February 22
2026

Acceptance Notification

March 15
2026

Camera Ready

March 29
2026

*All deadlines are 23:59 Pacific Standard Time (PST)

Scope & Topics

The workshop will cover techniques, systems, and applications centered on making foundation models scalable, efficient, and adaptable for large-scale data mining.

Efficient Training & Adaptation
  • Low-rank approximation, pruning, distillation, quantization
  • Adaptive and dynamic architectural reconfiguration
  • Lightweight fine-tuning (LoRA, prefix tuning, adapters)
  • Continual, incremental, and federated adaptation strategies
  • Transfer learning under limited data or compute
Scalable Foundation Model Design
  • Efficient LLMs, vision transformers, multimodal models
  • Memory-efficient representation learning
  • Sparse, modular, or expert-based architectures
  • Foundation models for edge or distributed settings
Foundation Model–Enabled Data Mining
  • Integration of FM for classical data mining tasks
  • Representation learning for structured & multimodal data
  • Large-scale pattern discovery using FM
  • Foundation models for graph/time-series mining
  • Cross-domain knowledge transfer
Systems & Deployment
  • Distributed training & inference optimizations
  • Energy-efficient and cost-aware execution
  • Hardware–software co-design for scalable FM
  • Resource-aware serving, caching, compression systems
Applications
  • Healthcare analytics, fraud detection, finance, recommendation
  • Manufacturing, IoT and cyber-physical systems
  • Smart cities, transportation, environmental monitoring
  • Social media mining, language analytics, multimodal KDD applications

Paper Submission

  • Paper Submission Portal (CMT): https://cmt3.research.microsoft.com/EFMxDM2026
  • To create a CMT account, please visit:
    https://cmt3.research.microsoft.com/docs/help/general/account-creation.html
  • For a step-by-step guide on how to submit a paper as an author, see:
    https://cmt3.research.microsoft.com/docs/help/author/author-submission-form.html
  • Submitted papers must be in English.
  • Maximum length: 12 pages (including references and appendices).
  • Abstract limited to 200 words.
  • Formatting must follow Springer LNCS/LNAI guidelines (10pt font).
    Springer LNCS Formatting Instructions
  • Submissions will undergo a double-blind review process.
  • The Microsoft CMT service was used for managing the peer-reviewing process for this conference. This service was provided for free by Microsoft and they bore all expenses, including costs for Azure cloud services as well as for software development and support.
  • All papers must be submitted electronically via CMT in PDF format.
  • Supplementary materials (PDF) may be included, but reviewers are not obliged to consider them.
  • Submissions violating the policy will be rejected without review.
  • Submissions must be original and not under review elsewhere.
  • Blind review requirement: remove author names/affiliations from the manuscript and supplementary files.
  • Self-citations must be written in the third person.
  • Accepted papers will be made available on the PAKDD 2026 website.
  • At least one author of each accepted paper must register and present in person at PAKDD 2026.

Formatting Guidelines



Organizing Committee

Divya Saxena

Dr. Divya Saxena

Assistant Professor
Indian Institute of Technology (IIT) Jodhpur, India


Website
Jiannong Cao

Prof. Jiannong Cao

Vice President (Education)
Otto Poon Charitable Foundation Professor in Data Science
Chair Professor
The Hong Kong Polytechnic University, Hong Kong
Website

Program Committee

(Listed in alphabetical order)

Prof. Anil Kumar Tiwari
Indian Institute of Technology Jodhpur, India
Dr. Ankur Gupta
Netaji Subhas University of Technology (NSUT), Delhi, India
Dr. Gaurav Kumar Nayak
Indian Institute of Technology Roorkee, India
Jialun Zheng
The Hong Kong Polytechnic University, Hong Kong
Prof. Kamalakar Karlapalem
International Institute of Information Technology Hyderabad, India
Dr. Koteswar Rao Jerripothula
Indian Institute of Technology Kanpur, India
Prof. Mukesh Mohania
Indraprastha Institute of Information Technology Delhi, India
Prof. R. Balasubramanian
Indian Institute of Technology Roorkee, India
Prof. Vaskar Raychoudhury
Miami University, Ohio, USA
Xiaoyun Liu
The Hong Kong Polytechnic University, Hong Kong
Dr. Yuqing Zhao
The Hong Kong Polytechnic University, Hong Kong

Contact Information

For any queries, contact:
divyasaxena@iitj.ac.in

Program

Schedule will be published after acceptance notification (15 March 2026).