Wadhwani School of Data Science and Artificial Intelligence
  • People
    • Faculty
    • Researchers
    • Management
    • Staff
    • Alumni
  • Academics
    • UG Programs
    • PG Programs
    • Research
    • Online Certificate Programs
    • Online Courses
    • Training Programs
  • Opportunities
    • Internships
    • Fellowships
    • Industry Collaborations
    • Faculty Careers
  • News & Events
    • Events
    • News
    • Newsletter
  • Research
    • Overview
    • Themes
    • Projects
    • Research Centres
    • Collaborations
  • Outcomes
    • Publications
    • Preprints
    • Whitepapers
    • Software & Datasets
    • Blogs
  • PhD Candidacy Exam Format
  • Open Positions
  • Grievances
  • For Current Students
  • For Prospective Industry
  • For Prospective Faculty
  • For Prospective Student
  • Upcoming Events
  • Contact

Menu

  • People
    • Faculty
    • Researchers
    • Management
    • Staff
    • Alumni
  • Academics
    • UG Programs
    • PG Programs
    • Research
    • Online Certificate Programs
    • Online Courses
    • Training Programs
  • Opportunities
    • Internships
    • Fellowships
    • Industry Collaborations
    • Faculty Careers
  • News & Events
    • Events
    • News
    • Newsletter
  • Research
    • Overview
    • Themes
    • Projects
    • Research Centres
    • Collaborations
  • Outcomes
    • Publications
    • Preprints
    • Whitepapers
    • Software & Datasets
    • Blogs

Quick Links

  • PhD Candidacy Exam Format
  • Open Positions
  • Grievances
  • For Current Students
  • For Prospective Industry
  • For Prospective Faculty
  • For Prospective Student
  • Upcoming Events
  • Contact
  • Home
  • Tags
  • Attention

Attention

On the Importance of Local Information in Transformer Based Models

Publications

The self-attention module is a key component of Transformer-based models, wherein each token pays attention to every other token. Recent studies have shown that these heads exhibit syntactic, semantic, or local …

Tags: NLP, attention, transformer model

Towards Transparent and Explainable Attention Models

Publications

Recent studies on interpretability of attention distributions have led to notions of faithful and plausible explanations for a model’s predictions. Attention distributions can be considered a faithful explanation …

Tags: NLP, LSTM, Attention

On the weak link between importance and prunability of attention heads

Publications

Given the success of Transformer-based models, two directions of study have emerged: interpreting role of individual attention heads and down-sizing the models for efficiency. Our work straddles these two streams: We …

Tags: NLP, Attention, BERT, Pruning

Attend, Adapt and Transfer: Attentive Deep Architecture for Adaptive Transfer from multiple sources

Publications

Transferring knowledge from prior source tasks in solving a new target task can be useful in several learning applications. The application of transfer poses two serious challenges which have not been adequately …

Tags: Transfer Learning, Deep neural network architecture, Attention

logo footer
IITM Logo

Founded in 2024, the School brings together several faculty with expertise in various areas of Data Science and AI, to work together on impactful problems of direct relevance to the society.

Contact Us

044 2257 8980
office@dsai.iitm.ac.in
6th Floor, New Academic Complex 2,
Indian Institute of Technology Madras,
Chennai-600036, India

Quick Links

  1. For Current Students
  2. Grievances
  3. PhD Candidacy Exam Format
  4. Faculty
  5. UG Programs
  6. PG Programs
  7. Research
  8. Online Certificate Programs
  9. Online Courses
  10. Training Programs
  11. Internships
  12. Fellowships
  13. Industry Collaborations
  14. Upcoming Events
  15. Contact

Wadhwani School of Data Science and Artificial Intelligence | IIT Madras © 2026