BERT Model Architectures For Semantic Similarity

  Рет қаралды 23,645

Abhishek Thakur

Abhishek Thakur

Күн бұрын

In this video, I discuss different kinds of model architectures that you can use for #SemanticSimilarity using #BERT or any other #Transformers based model.
Please subscribe and like the video to help me keep motivated to make awesome videos like this one. :)
To buy my book, Approaching (Almost) Any Machine Learning problem, please visit: bit.ly/buyaaml
Follow me on:
Twitter: / abhi1thakur
LinkedIn: / abhi1thakur
Kaggle: kaggle.com/abhishek

Пікірлер: 31
@username-notfound9841
@username-notfound9841 3 жыл бұрын
I came here after reading the SBERT paper, and now it starts making sense. Good one.
@rbhambriiit
@rbhambriiit 2 жыл бұрын
Happy to see how you have evolved since the 2017 Berlin talk on Quora duplicate question talk.
@shreyasdantkale
@shreyasdantkale 3 жыл бұрын
I had directly used SBERT in my project(assumed it will be very complex to fine tune 2 BERT models). Here this all made sense with your such easy understandable explanation.
@sushasuresh1687
@sushasuresh1687 3 жыл бұрын
How are the results ? I mean is avert better than the Bert ?
@SurjeetSingh-mz4ne
@SurjeetSingh-mz4ne 3 жыл бұрын
thank you, Abhishek for creating this video... I am working on that model and will let you know the results when it is done...
@maunishdave2573
@maunishdave2573 3 жыл бұрын
I was trying bert on kaggle's new contradictory watson competition but was having little trouble with how to format the data and you uploaded this video. Thank you.
@AlexeyMatushevsky
@AlexeyMatushevsky 3 жыл бұрын
You improved your results?
@maunishdave2573
@maunishdave2573 3 жыл бұрын
@@AlexeyMatushevsky Actually I am trying various thing for many days. I tried both the format shown in this video. with every possible combination of learning_rate and epochs but my training_loss is always around 1.05 (crossentropyloss) it's because it is giving same output every time. I tired same architecture in TF and it worked but I don't know what is wrong with my pytorch model still trying to find out.
@shaurabhsaha2082
@shaurabhsaha2082 3 жыл бұрын
Hi @abhishek Just a small request from your side can you please make a playlist for all your videos uploaded till now it will be really helpful and informative if you align them it will be really easy to follow the right sequence to watch. Appreciate your content really grateful to see your videos ❤️
@ayoubbariki7951
@ayoubbariki7951 3 жыл бұрын
Very informative ! Thanks u
@adityakumarmishra6931
@adityakumarmishra6931 3 жыл бұрын
Thanks for the video sir. very informative. 😁
@shishirkumar1450
@shishirkumar1450 3 жыл бұрын
Hey @abhishek, Do you have a tutorial on the NLP Question Answering system for any custom domain?
@raghavavinay8660
@raghavavinay8660 3 жыл бұрын
Thanks
@luis96xd
@luis96xd Жыл бұрын
Excellent video, great explanation, you taught me a lot of things, thanks 😁💯
@bryancc2012
@bryancc2012 3 жыл бұрын
great video. thanks. is there any way to train a document similarity without using those dataset ? like an autoencoder? any way to create an bert based autoencoder ? thanks
@depenz
@depenz 3 жыл бұрын
When adding two bert heads as bert_1 and bert_2, is that the same thing as a siamese network? :)
@oltionpreka3542
@oltionpreka3542 3 жыл бұрын
What if one would like to create a similarity model that compares the order of similarity among, say, 1000 questions? Is there any alternative than training the model on 1 by 1 similarity as it would grow more than exponentially with the number of questions of the dataset.
@depenz
@depenz 3 жыл бұрын
Is the pooled output ”o2” the same as the mean of each word token. Or if not, how is it computed and what information gain does it give thats better than the mean of tokens / or simply using the CLS token. Thanks :)
@boscojay1381
@boscojay1381 3 жыл бұрын
I trained SBert from transformers library on allNLI data and used to build sentence embeddings for a custom dataset. I cluster those embeddings and find that the representations are of poor quality. InferSent model from face book gave better results. What could be the problem?
@shaikrasool1316
@shaikrasool1316 3 жыл бұрын
Please create a book or video course on nlp
@hardikpachgade6082
@hardikpachgade6082 3 жыл бұрын
Can it be used on Hnglish based data?
@Love_and_wisdom
@Love_and_wisdom 3 жыл бұрын
How do you find this application for specific biomedical applications??
@user-qs3ck1dp7g
@user-qs3ck1dp7g 3 жыл бұрын
Good day, could you give any advices about denoising plroblem of the audio. Any Architectures ?
@raghavavinay8660
@raghavavinay8660 3 жыл бұрын
Sir can you suggest how to send input for lstm for time series data. I've data of patients for 10 days. How to send in the data patient wise..
@shreyasbs2861
@shreyasbs2861 3 жыл бұрын
Where is the code repo ?
@devswarup40
@devswarup40 3 жыл бұрын
🤟🤟🤟
@gtmpai
@gtmpai 3 жыл бұрын
I guess in the self.bert we get 3 outputs rather than o1,o2. I got too many values to unpack error
@abhishekkrthakur
@abhishekkrthakur 3 жыл бұрын
you must have changed your config :)
@gtmpai
@gtmpai 3 жыл бұрын
@@abhishekkrthakur seems like I have not inherited nn.Module but Bertmodel.
@puneetsingh5219
@puneetsingh5219 3 жыл бұрын
Binod?
@abhishekkrthakur
@abhishekkrthakur 3 жыл бұрын
🤦🏽‍♂️
How to use BERTopic - Machine Learning Assisted Topic Modeling in Python
15:04
Python Tutorials for Digital Humanities
Рет қаралды 31 М.
Получилось у Вики?😂 #хабибка
00:14
ХАБИБ
Рет қаралды 7 МЛН
Каха ограбил банк
01:00
К-Media
Рет қаралды 11 МЛН
The child was abused by the clown#Short #Officer Rabbit #angel
00:55
兔子警官
Рет қаралды 24 МЛН
Training BERT Language Model From Scratch On TPUs
34:20
Abhishek Thakur
Рет қаралды 30 М.
STOP Learning These Programming Languages (for Beginners)
5:25
Andy Sterkowitz
Рет қаралды 660 М.
SPLADE: the first search model to beat BM25
28:52
James Briggs
Рет қаралды 18 М.
3 Vector-based Methods for Similarity Search (TF-IDF, BM25, SBERT)
29:24
Получилось у Вики?😂 #хабибка
00:14
ХАБИБ
Рет қаралды 7 МЛН