VATA: Video Attention, Text Attention This is the implementation of our final model for the QIA 2020 Challendge on Kaggle. How to run: Download the dataset from Kaggle. Use MTCNN and OpenCV to crop out the faces from the video. Transform frames into PyTorch tensors. Run the training file: python fe_train.py Structure