Name: Fan Qian
Type: User
Company: Harbin Institute of Technology, HIT
Bio: CS PhD at Harbin Institute of Technology (HIT). My main research interests include speech/audio processing, affective computing, multimodal machine learning.
Location: Harbin, China
Fan Qian's Projects
This code has been developed for detecting sentiment in videos using Convolutional Neural Network and Multiple Kernel Learning.
[ACL'19] [PyTorch] Multimodal Transformer
Multimodal Sarcasm Detection Dataset
Face recognition with deep neural networks.
A script for audio/transcript alignment. Fork of p2fa.
Implementation of a hierarchical CNN based model to detect Big Five personality traits
pytorch implementation for "PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation" https://arxiv.org/abs/1612.00593
Pytorch implementation of R-Transformer. Some parts of the code are adapted from the implementation of TCN and Transformer.
Voice data <= 10 mins can also be used to train a good VC model!
Sentiment Knowledge Enhanced Self-supervised Learning
Implementation of Speech Emotion Recognition using DNN-ELM
Speech recognition module for Python, supporting several engines and APIs, online and offline.
Heart sound segmentation code based on duration-dependant HMM
Stable Diffusion web UI
手写实现李航《统计学习方法》书中全部算法
Text classification using deep learning models in Pytorch
Multimodal Sentiment Analysis with Temporal Modality Attention(Interspeech2021)
🤗Transformers: State-of-the-art Natural Language Processing for Pytorch, TensorFlow, and JAX.
:hibiscus: Minimalist Vim Plugin Manager
Word-wise Sparse Attention for Multimodal Sentiment Analysis