🤖 audio-classification

emotion-recognition-wav2vec2-IEMOCAP

speechbrain/emotion-recognition-wav2vec2-IEMOCAP

Get AI Model →
530.2K
Downloads
❤️
184
Likes
🏷️
11
Tags
📦
speechbrain
Library
Model Details
Full Model IDspeechbrain/emotion-recognition-wav2vec2-IEMOCAP
Pipeline / Taskaudio-classification
Libraryspeechbrain
Downloads (all-time)530.2K
Likes184
Last Modified7/23/2024
Author / Orgspeechbrain
PrivateNo — public
⚡ Quick Usage (Python)

Using the 🤗 Transformers library. Install with pip install transformers

from transformers import pipeline

# Load the model
pipe = pipeline("audio-classification", model="speechbrain/emotion-recognition-wav2vec2-IEMOCAP")

# Run inference
result = pipe("Your input here")
print(result)
🏷️ Tags
speechbrainaudio-classificationEmotionRecognitionwav2vec2pytorchendataset:iemocaparxiv:2106.04624license:apache-2.0region:us
More audio-classification Models
See all →
clap-htsat-fused

laion/clap-htsat-fused

17.0M❤️ 77
Get AI Model →
wav2vec-vm-finetune

jakeBland/wav2vec-vm-finetune

905.0K❤️ 11
Get AI Model →
wav2vec2-large-robust-12-ft-emotion-msp-dim

audeering/wav2vec2-large-robust-12-ft-emotion-msp-dim

821.5K❤️ 158
Get AI Model →
🚀 Use This Model

Access model files, inference API, and full documentation on Hugging Face.

Open on Hugging Face →Browse Model Files ↗← Browse All Models
🤖 Task: audio-classification

This model is designed for the audio-classification task. Explore more models for this use case.

All audio-classification Models →
📊 Popularity
Downloads530.2K
❤️ Community Likes184
🛠️ Requirements
  • Install: pip install speechbrain
  • Python 3.8+ recommended for Transformers.
  • GPU (CUDA) speeds up inference significantly.
  • Use model.half() for fp16 on limited VRAM.
👋 Need help with code?