Mastering Generative Answers in Copilot Studio

Mastering Generative Answers in Copilot Studio

Getting Started with Co-Pilot Studio

  • Co-Pilot Studio connects to websites, SharePoint, and custom data sources for AI-based chat.
  • The experience can be improved; guidance is provided to enhance user satisfaction.
  • Traditional chatbots require predefined topics, limiting awareness of user questions.

Generative Answers and Data Sources

  • Generative answers utilize various data sources to provide comprehensive responses.
  • Currently using GPT 3.5 turbo model hosted on Azure OpenAI service.
  • The service is included in the Co-Pilot Studio license costing about $200/month.

Retrieval Augmented Generation (RAG)

  • RAG supplements a generic model with business data without extensive training.
  • Users can ground the co-pilot experience by pointing it to public websites or documents.
  • This method is resource-efficient compared to training a custom model.

Conversation Context and Boosting Topics

  • The system retains context from the last 10 conversation turns for better interaction.
  • Conversation boosting topics enable generative answers as fallback options when needed.
  • Users can build chatbots without predefined topics or mix them with specific intents.

Limitations and Advanced Options

  • Generative answers can be used within specific topics or as overall fallback solutions.
  • Limitations exist; advanced tools are available for more complex setups at higher costs.

Understanding Document Upload Limitations

  • Maximum of four sites can be pointed to for document uploads, with a size limit of 3 MB per file.
  • Dataverse search indexes documents to help find context but cannot provide direct links to uploaded documents.
  • SharePoint or OneDrive offers a better experience for document management compared to direct uploads.

Using Public Data for Co-Pilot

  • Example used is The Better Health Channel website, which provides health information.
  • The bot retrieves results from the specified website without needing predefined topics.
  • URL depth limitation is two levels; however, it can retrieve results from deeper levels.

Generative AI Capabilities

  • Users can request content in different formats, such as simplifying explanations for children.
  • Content moderation settings (high, medium, low) affect answer accuracy and creativity.
  • Low moderation may yield creative but less accurate responses; high moderation ensures more reliable answers.

Common Issues with Bot Responses

  • Users may encounter "I can't help with that" errors due to various reasons including setup time.
  • If no results are found, it could be due to the data source lacking relevant information.

Understanding Course Information Retrieval

  • Discusses issues with retrieving course information from a university website due to different top-level domains.
  • Emphasizes the importance of website structure for successful information retrieval; adding relevant domains can yield results.
  • Notes that if a website has multiple top-level domains, it may complicate search results.

Document Uploading and Citation Challenges

  • Explains uploading policy documents into co-pilot for quick answers about parental leave.
  • Highlights limitations in formatting when documents are uploaded, affecting user experience.
  • Suggests that while document uploads are simple, they may not be practical for real-life applications.

Comparing SharePoint and OneDrive Experiences

  • Contrasts experiences between direct document uploads and using SharePoint for better user interaction.
  • Points out that linking to documents in SharePoint enhances accessibility compared to plain text responses.
  • Mentions the need for authentication when accessing documents on SharePoint.

Enhancing Search Experience with Microsoft 365

  • Introduces Microsoft 365 semantic search as a tool for improved indexing and search capabilities within co-pilot.
  • References a video explaining how this feature works to enhance user experience across various applications.

How to Use Custom Instructions for Generative AI

Understanding Custom Instructions

  • Custom instructions allow users to set a persona for the AI, enhancing interaction quality.
  • Providing meta prompts influences the AI's responses, leading to better user experiences.
  • Example: Acting as an HR assistant with specific parameters improves response tone and relevance.

Tone and Context in Responses

  • Adding warmth and positivity (e.g., emojis) can change the tone of responses significantly.
  • Adjusting instructions based on context (e.g., manager vs. staff interactions) is crucial for appropriateness.
  • Removing casual elements like emojis may yield more professional responses in serious contexts.

Length and Brevity of Responses

  • Custom instructions can specify response length, aiding clarity and conciseness.
  • Users should experiment with character limits to achieve desired brevity in answers.
  • Effective prompting is essential for generating concise and relevant outputs from the AI.

Importance of User Prompting

  • Educating users on effective prompting enhances overall experience with generative AI tools.
  • Clear guidelines help users formulate questions that lead to useful responses from the AI.
  • Specific phrasing in prompts can improve the accuracy of information retrieved by the AI.

Enhancing Output Quality

  • Providing detailed prompts allows the AI to generate more targeted responses related to policies or issues.
  • Using structured requests (e.g., bullet points for meeting preparation) yields clearer outputs from the AI.

Understanding Retrieval Augmented Generation

Key Insights

  • Emphasizes the importance of simplifying complex information for better understanding, akin to explaining it to a child.
  • Highlights the capability of handling multi-turn conversations and the need for improvement in rewording requests.
  • Discusses challenges in getting concise scripts for meetings, indicating limitations in current capabilities.

Performance Improvement Plans

Documenting Progress

  • Focuses on documenting milestones within performance improvement plans, leveraging previous conversation threads effectively.
  • Stresses understanding data sources and use cases to optimize retrieval augmented generation outcomes.
Channel: Lisa Crosbie
Video description

Generative answers is one of the most powerful features available in Copilot Studio, enabling you to create an instant AI bot to search and access to your business data - but so many users get stuck with less than impressive results when they plug in a website, document or SharePoint site for the first time. This video full of real life hit and miss examples will get you over that hurdle to help you understand what you can achieve with generative answers, as well as what's not possible, and why, and equip you with the skills to get the most out of it. 0:00 - Generative Answers in Copilot Studio 1:29 - How generative answers works 3:49 - Conversation boosting topic 5:31 - Limitations and options 7:13 - Public Data - Website 8:52 - Content moderation 9:54 - Error message - I'm sorry I can't help with that 12:36 - Uploading documents 14:18 - SharePoint and OneDrive 17:05 - Custom instructions 20:55 - Prompting 23:38 - Multi turn conversations 24:46 - What matters the most to optimize generative answers ----------------------------------------------------- Connect with me: ☕ Buy me a coffee: https://www.buymeacoffee.com/lisacrosbie 🦉 Learn more about AI: https://aka.ms/learnwithlisa 🖇 LinkedIn: https://www.linkedin.com/in/lisa-crosbie/ 📼 TikTok: https://www.tiktok.com/@lisa.crosbie 🐦 X (Twitter): https://twitter.com/LisaCrosbie 📚Take my LinkedIn Learning Course: Microsoft Power Platform Fundamentals (PL-900) Exam: Power Apps https://www.linkedin.com/learning/microsoft-power-platform-fundamentals-pl-900-cert-prep-power-apps/