LLM Models

Language Models (LLMs) in Flyflow

Flyflow supports various Language Models (LLMs) that power the conversational abilities of your agents. LLMs are advanced machine learning models trained on vast amounts of textual data, enabling them to understand and generate human-like responses. Flyflow provides access to both OpenAI's state-of-the-art models and our own custom-developed model optimized for low-latency voice interactions.

GPT-4o: OpenAI's Most Capable Model

GPT-4o is OpenAI's most advanced and capable language model currently available. It represents the cutting edge of natural language processing and understanding, delivering exceptional performance across a wide range of conversational tasks.

Key features of GPT-4o include:

  • Extensive knowledge spanning a broad range of topics
  • Ability to engage in complex and nuanced conversations
  • Strong reasoning and analytical capabilities
  • Contextual understanding and coherence in responses
  • High-quality language generation

By leveraging GPT-4o as the underlying LLM for your agents, you can create highly intelligent and sophisticated conversational experiences. Your agents will be able to handle complex user queries, provide insightful and well-informed responses, and maintain a natural and coherent dialogue flow.

To use GPT-4o for an agent, specify "gpt-4o" as the llm_model property when creating or updating an agent through the Flyflow API.

Example

{
  "name": "My Agent",
  "system_prompt": "You are a highly intelligent assistant powered by GPT-4o.",
  "llm_model": "gpt-4o"
}

Flyflow-Voice: Custom Fine-Tuned Model for Low Latency and High Intelligence

In addition to GPT-4o, Flyflow offers our own custom-developed language model called flyflow-voice. This model is specifically fine-tuned for low-latency voice interactions while maintaining a high level of intelligence and conversational quality.

Key features of Flyflow-Voice include:

  • Optimized for fast response generation
  • Fine-tuned on a large corpus of conversational data
  • Balanced trade-off between latency and intelligence
  • Tailored for voice-based interactions
  • Delivers natural and coherent responses

Flyflow-Voice is designed to provide the lowest possible latency without compromising on the quality and intelligence of the generated responses. By leveraging advanced techniques such as model distillation and architecture optimizations, Flyflow-Voice achieves a remarkable balance between speed and performance.

When using Flyflow-Voice, you can expect response times as low as 300 milliseconds while still maintaining a high level of conversational intelligence. This enables your agents to provide quick, accurate, and natural responses to user queries, creating a smooth and engaging voice interaction experience.

To use Flyflow-Voice for an agent, specify "flyflow-voice" as the llm_model property when creating or updating an agent through the Flyflow API.

Example

{
  "name": "My Agent",
  "system_prompt": "You are a fast and intelligent assistant powered by Flyflow-Voice.",
  "llm_model": "flyflow-voice"
}

Conclusion

Flyflow provides access to powerful language models that empower your agents with advanced conversational capabilities. With GPT-4o, you can leverage OpenAI's most capable model to create highly intelligent and sophisticated conversational experiences. On the other hand, Flyflow-Voice offers a custom fine-tuned model optimized for low-latency voice interactions while maintaining a high level of intelligence.

By choosing the appropriate LLM based on your specific requirements, you can create agents that deliver exceptional conversational quality, responsiveness, and user engagement. Whether you prioritize the highest level of intelligence with GPT-4o or the lowest possible latency with Flyflow-Voice, Flyflow provides the tools and models to power your voice-based conversational applications.