Product Updates

Announcing our Custom Model API

Published on Dec 14, 2023

Meet our Custom Model API — a cutting edge AI tool that integrates language, voice, and/or facial movement to predict human preferences and needs more accurately than any LLM

You can now use our Custom Model API to predict well-being, satisfaction, mental health, and more. Using a few labeled examples, our API integrates dynamic patterns of language, vocal expression, and/or facial expression into a custom multimodal model. 

Leveraging Hume’s AI models pretrained on millions of videos and audio files, our API can usually predict your labels accurately after seeing just a few dozen examples. That means that with just a few labeled examples and a few clicks, you can deploy powerful AI models that predict the outcomes your users care about most. Of course, the models you train using our API are yours alone to deploy and share.

Visit dev.hume.ai/docs/custom-models or login to beta.hume.ai to get started!

Our new API translates nuanced multimodal measures into personalized insights more accurately than any LLM

image

For instance, we partnered with Lawyer.com to predict the quality of customer support calls. Using just 73 calls, we were able to train a model that predicted expert ratings of whether a call went well or poorly with 97.3% accuracy. By contrast, using language models alone—including one of the world’s most capable language models along with our in-house language emotion model—resulted in a 3x higher error rate.

Leveraging dynamic patterns of language, vocal expression, and/or facial movement

Our Custom Model API works by integrating complex patterns of language, vocal expression, and/or facial movement captured using Hume’s expression AI models.

emoclouds for dimensions

To combine these signals with language, we inserted the expression measures extracted by each of our expression models along with transcribed language into a novel empathic large language model (eLLM). We then pretrained our eLLM on millions of human interactions.

highres flow

When you train a custom model using our Custom Model API, you are leveraging the joint language-expression embeddings extracted by our eLLM to predict your own labels.

Pricing

There are two steps to using our Custom Model API, (1) Training and (2) Inference:

1. Training: During our beta release, the model training process is completely free. This includes uploading data, training, evaluating results, and retraining.

2. Inference: When deploying your custom model in your application, a fee is charged for each file processed by your model. Detailed pricing can be found on our pricing page.

Hope you enjoy using our Custom Model API! If you have any questions, you can post them on our Discord channel. We look forward to seeing what you build.

Subscribe

Sign up now to get notified of any updates or new articles.

Recent articles

blogtile
Science

What is semantic space theory?

Our models and products are built on a cutting-edge approach to understanding emotion: semantic space theory (SST), which uses computational methods and data-driven approaches to map the full spectrum of our feelings

Blog - 2.14.24
Science

Publication in iScience: Understanding what facial expressions mean in different cultures

How many different facial expressions do people form? How do they differ in meaning across cultures? Can AI capture these nuances? Our new paper provides new in-depth answers to these questions with the help of machine learning.

Blog Card 2.9.24
Science

Introducing a new evaluation for creative ability in Large Language Models

Introducing HumE-1 (Human Evaluation 1), our new evaluation for large language models (LLMs) that uses human ratings to evaluate LLMs for their ability to perform creative tasks in the ways that matter to us, evoking the feelings we want to feel.