V
主页
十分钟看懂微软大力金刚掌WavLM: Large-Scale Self-Supervised Pre-Training for Full Stack
发布人
WavLM: Large-Scale Self-Supervised Pre-Training for Full Stack Speech Processing
打开封面
下载高清视频
观看高清视频
视频下载器
十分钟告诉你为什么OpenAI的Whisper语音识别没ChatGPT那么好用 [语音语言论文阅读]
三分钟搞定微软零样本语音合成VALL-E
太厉害了 已跪!终于有人能把知识图谱讲的这么通俗易懂了,浙大知识图谱入门及实战公开课分享!-深度学习丨NLP丨知识图谱
[Long Review] GLaM: Efficient Scaling of Language Models with Mixture-of-Experts
十分钟看懂脸书太极拳法Wav2Vec2.0 -- 语音预训练模型就像绝命毒师老白教杰西
[Long Review] Cascaded Diffusion Models for High Fidelity Image Generation
十分钟看懂谷歌易筋经BERT
解锁天顶星科技ChatGPT
[Long Review] Axial Attention in Multidimensional Transformers
十分钟看懂谷歌金钟罩Transformer以及语音的LAS模型
从入门到提示词工程师:全网最通俗易懂Prompt-Learning提示词学习教程!草履虫都学的会!
[Long Review]Switch Transformers: Scaling to Trillion Parameter Models with
[Long Review] Transfer Learning from Speaker Verification to Multispeaker TTS
[Long Review] Deduplicating Training Data Makes Language Models Better
详解OpenAI GPT-3: Language Models are Few-Shot Learners(2/3)
详解OpenAI GPT-3: Language Models are Few-Shot Learners(1/3)
语音文本技术论文阅读 RefineGAN - Universally Generating Waveform Better than Ground ...
[Long Review] Wav2Seq: Pre-training Speech-to-Text Encoder-Decoder Models Using
语音文本技术论文阅读 Joint Unsupervised and Supervised Training for Multilingual ASR
[Long Review]Kullback-Leibler Divergence: Listen, Attend, Spell and Adapt ASR
[Long Review] Conformer: Convolution-augmented Transformer for Speech Recogniti
【黑马】AI大模型训练营2期
[Long Review] Towards Zero-Label Language Learning
语音文本技术论文阅读 Exploring Wav2vec 2.0 fine-tuning for improved speech emotion recogni
语音文本技术论文阅读 SNRi Target Training for Joint Speech Enhancement and Recognition
语音文本技术论文阅读 Improving Speech Recognition Accuracy of Local POI Using Geographical
【零基础学AI】清华大佬200集讲完的AI人工智能从入门到精通全套教程,全程干货无废话!看完绝对让你少走弯路!研一研零必看!机器学习-深度学习-opencv
语音文本技术论文阅读 XLS-R: Self-supervised Cross-lingual Speech Representation Learning a
[Short Review] Fully Sharded Data Parallel: faster AI training with fewer GPUs
(强推!)BERT模型项目实战教程!基于BERT模型的文本分类、情感分析及中文命名实体识别实战教程!真的通俗易懂!(人工智能、深度学习、机器学习算法、神经网络)
[Short Review] Deduplicating Training Data Makes Language Models Better
详解AudioLM: a Language Modeling Approach to Audio Generation
十分钟看懂谷歌W2v-BERT: Combining Contrastive Learning and Masked Language Modeling
[Long Review] CLAS: Deep context: end-to-end contextual speech recognition
语音文本技术论文阅读 UniSpeech-SAT - Universal Speech Representation Learning with Speaker
语音NLP论文阅读 Token-level Sequence Labeling for SLU using Compositional E2E Models
语音合成速度最快的开源TTS | 基于MIT协议可商用的开源TTS | 这是我用过合成速度最快的开源TTS
强推!全网最系统的自然语言处理全套教程分享,学完可少走99%的弯路! NLP基础/NLP实战项目
[Long Review] Xception: Deep Learning with Depthwise Separable Convolution
语音文本技术论文阅读 RNN-T: Sequence Transduction with Recurrent Neural Networks