Detecting Out-of-Distribution Text Using Topological Features of Transformer-Based Language Models

πŸŽ‰ Exciting News! πŸŽ‰

I'm thrilled to share that our paper has been accepted for presentation at IJCAI 2024 in the AI Safety workshop!

πŸ“„ Paper Title: Detecting Out-of-Distribution Text Using Topological Features of Transformer-Based Language Models

πŸ” What’s It About?

Our research focuses on improving how AI models, like BERT, detect when they encounter text that is very different from what they were trained on (known as out-of-distribution or OOD text). This is important for making AI systems safer and more reliable.

πŸ”§ How Did We Do It?

We used a technique called Topological Data Analysis (TDA) to analyze the attention patterns within BERT, a popular AI language model. Think of TDA as a way to look at the "shape" of data to find patterns.

πŸ†š What Did We Compare?

We compared our TDA approach with a more traditional method that looks at BERT's summary of text (CLS embeddings).

πŸ’‘ Key Findings:

This research could help improve AI's ability to recognize when it's facing unfamiliar or unexpected text, making AI applications more robust and trustworthy.

Looking forward to sharing more insights at IJCAI 2024!

You can follow the slide to get more out of the paper. I gave this talk at Sikkim university.