Get the (almost) daily changelog

New Azure OpenAI GPT 4.1 Models

  1. Access to New Azure OpenAI Models: You can now use new GPT 4.1 models in Azure OpenAI such as gpt-4.1-2025-04-14, gpt-4.1-mini-2025-04-14, and gpt-4.1-nano-2025-04-14.
The above models will be available to configure through the console at a later date. For now, configure your assistant to use these models through the API.
  1. Specify Start Node in Workflows with isStart Property: You can now explicitly define the starting point of your workflow by setting the isStart property to true on any node like Say, Gather, or Hangup.

  2. Updated Model Options in GroqModel: You can now use the following new Assistant modles with Groq:

    • meta-llama/llama-4-maverick-17b-128e-instruct
    • meta-llama/llama-4-scout-17b-16e-instruct
    • mistral-saba-24b
    • compound-beta
    • compound-beta-mini
New Groq Models
New Groq Models
Note that some older models have been removed, including llama-3.1-70b-versatile and mixtral-8x7b-32768.
  1. New Kylie Voice Available in Vapi: You can now use the new Kylie voice when using Vapi as your voice provider. You can learn more in the Vapi voices documentation.
New Kylie Voice
New Vapi Voice: Kylie

GoHighLevel Tools for Calendar and Contact Management

You can now use new GoHighLevel tools in all models, templates, and workflows directly through the /tool and /tool/{id} endpoints with the following capabilities:

  1. Configure Conversation Nodes with OpenAI Models: You can now set up your assistant’s workflow conversation nodes to use OpenAI models by specifying WorkflowOpenAIModel. Choose from a range of OpenAI models and customize parameters like maxTokens and temperature to control responses.

  2. Configure Conversation Nodes with Anthropic Models, Including Thinking Feature: Your assistant’s conversation nodes can now use Anthropic models by specifying WorkflowAnthropicModel. Select from various Anthropic models and, for claude-3-7-sonnet-20250219, enable the optional thinking feature for advanced reasoning capabilities.

  1. Workflows Now Marked as Beta Features: The workflow endpoints and related properties have now moved to [BETA], indicating they’re slightly more stable but still in active development. Refer to the Workflows documentation and API documentation for more information.

  2. New {{endedReason}} Variable in Templates: You can now include the {{endedReason}} variable in your post-call analysis templates to access why a call ended. This helps generate more insightful summaries and evaluations based on the call’s outcome.

  3. Introduction of SayAssistantHookAction Schema: A new action, SayAssistantHookAction, allows the assistant to say specific messages during calls. Use this by adding it to call.squad.members.assistant.hooks.do[type=say] to enhance call interactions.

  1. New ‘Conversation’ Node in Workflows: You can now use the Conversation node in your workflows to create conversation tasks, enhancing how assistants interact during calls.

  2. Integration with GoHighLevel via OAuth2 Credentials: You can now connect with GoHighLevel services using new GoHighLevelMCPCredential credentials in the Provider Keys section of the Vapi Dashboard.

  3. Standardized Message Types for clientMessages and serverMessages: When configuring assistants, you now specify Client Messages and Server Messages using predefined message types, ensuring consistency and preventing invalid message configurations.

  1. ClientMessage Additions: Several new client message schemas have been added with additional information about call, customer, assistant, timestamp, and phoneNumber. This includes:
  1. New Hooks for Speech Interruption Events: Two new hooks, Speech Interrupted Assistant Hook and Speech Interrupted Customer Hook, enable you to define actions when speech is interrupted during a call.

  2. Call Schema Updates: There are several notable updates to how Call is structured:

  • costs array now includes a new cost type: KnowledgeBaseCost
  • phoneCallProvider and phoneCallProviderId are now deprecated.
  • waitFunction in LivekitSmartEndpointingPlan has been updated to improve how long the assistant waits before speaking, enhancing call flow responsiveness.
  1. Use Workflows as Call Entry Points: You can now start calls or configure phone numbers using a workflow or workflowId, just like you would with assistant, assistantId, squad, or squadId. This provides more flexibility in defining how calls are initiated and allows direct use of workflows. Refer to the Workflows documentation and API documentation for more information.

  2. New Warm Transfer Mode and Hold Music in TransferPlan: There’s a new transfer mode warm-transfer-experimental in call.squad.members.assistant.hooks.do[type=transfer].destination.transferPlanthat enhances call transfer capabilities, including voicemail detection and customer hold experience. You can also customize the hold music by specifying a holdAudioUrl.

  3. Simplified clientMessages Configuration: The clientMessages property has been updated and is now required in AssistantOverrides, CreateAssistantDTO, and UpdateAssistantDTO. This change simplifies how you specify which messages are sent to your Client SDKs.

  1. New KnowledgeBaseCost in Call Costs:: You can now access detailed costs related to knowledge base queries in a call through the new KnowledgeBaseCost type in call.costs[type=knowledge-base]. This helps in tracking expenses when using knowledge base features during calls.

  2. Deprecated smartEndpointingEnabled Property: The smartEndpointingEnabled property in StartSpeakingPlan is now deprecated. Developers should update their applications to use the new smartEndpointingPlan or customEndpointingRules for controlling endpointing behavior.

  3. Advanced Endpointing with smartEndpointingPlan and customEndpointingRules: The StartSpeakingPlan now includes smartEndpointingPlan and customEndpointingRules properties, providing enhanced control over speech endpointing. Developers can specify endpointing methods or define custom rules to improve conversational interactions.

The smartEndpointingEnabled property in StartSpeakingPlan is now deprecated. Developers should update their applications to use the new smartEndpointingPlan or customEndpointingRules for controlling endpointing behavior.