DeepAffects is a speech analysis platform for Developers. We offer a number of speech analysis apis like, Speech Enhancement, Multi-Speaker Diarization, Emotion Recognition, Voice-prints, Conversation Metrics etc. For more information, checkout ourdeveloper portal...
- **情感识别** (`src.emotion_recognition`): 基于emotion2vec实现,用于把音频按9种情感分类。 - **多说话人分类(在做)** (`src.speaker_classifier`): 基于3D-Speaker实现,可以将list文件分成不同的说话人。 - **多说话人分类** (`src.speaker_classifier`): 基于funasr/cam++/3d_speaker实现,可以分...
Emotion Recognition From Audio Signal This is a set of soultions of Emotion Recognition From Audio Signal from Surrey Audio-Visual Expressed Emotion (SAVEE) Database using OpenSmile, Principal Components Analysis and set of classifiers from Scikit-learn library. The SAVEE database was recorded from...
git clone https://github.com/stepfun-ai/Step-Audio.git conda create -n stepaudio python=3.10 conda activate stepaudio cd Step-Audio pip install -r requirements.txt git lfs install git clone https://huggingface.co/stepfun-ai/Step-Audio-Tokenizer git clone https://huggingface.co/stepfun-ai...
These features make Audino v2.0 an ideal choice for various tasks, such as Voice Activity Detection (VAD), Diarization, Speaker Identification, Automated Speech Recognition, Emotion Recognition, and more.🚀 Note: Audino v2.0 is actively under development. In the future, we plan to migrate from...
git clone https://github.com/stepfun-ai/Step-Audio.git conda create -n stepaudio python=3.10 conda activate stepaudio cd Step-Audio pip install -r requirements.txt git lfs install git clone https://huggingface.co/stepfun-ai/Step-Audio-Tokenizer git clone https://huggingface.co/stepfun-ai...
Audio-visual Recognition Speech Recognition Speaker Recognition Action Recognition Emotion Recognition Uni-modal Enhancement Speech Enhancement and Separation Object Sound Separation Face Super-resolution and Reconstruction Cross-modal Perception Cross-modal Generation Mono Sound Generation Speech Music Natura...
curl http://0.0.0.0:8080/predict?id=10b09Lb Credits https://github.com/ritikraj660/Speech-emotion-recognition/ https://librosa.github.io/librosa/index.html https://www.datasciencelearner.com/ https://towardsdatascience.com/ 1star 1watching 0forks...
Merge branch 'master' of github.com:david-yoon/multimodal-speech-emotion Nov 18, 2019 LICENSE Initial commit Jan 14, 2019 README.md Update README.md Mar 26, 2024 requirements.txt nltk patch Aug 24, 2019 multimodal-speech-emotion This repository contains the source code used in the following...
Speaker gender recognition82.5%67.2%66.5%35.5%57.0%33.2%29.2%21.9% Emotion recognition60.0%43.2%26.0%29.9%25.7%27.4%37.6%59.5% Speaker age prediction58.8%36.0%42.5%48.7%62.4%51.2%20.4%41.1% Speech entity recognition48.1%71.2%34.0%51.7%26.1%37.2%35.9%69.8% ...