Upgrade to Pro — share decks privately, control downloads, hide ads and more …

ABCS25: Enhancing Legal Document Analysis with ...

ABCS25: Enhancing Legal Document Analysis with Reflection Agents, Semantic Kernel, and Azure AI Search by Cédric Mendelin

⭐️ Enhancing Legal Document Analysis with Reflection Agents, Semantic Kernel, and Azure AI Search#
In this session, we will explore the implementation of Retrieval-Augmented Generation (RAG) on Swiss law documents using Semantic Kernel and Azure AI Search. We will delve into the step-by-step optimization process that enhanced the solution’s efficiency and accuracy. Finally, show the architecture using Reflection Agents and advanced Azure AI Search capabilities. Attendees will gain insights into the challenges faced, the strategies employed to overcome them, and the significant improvements achieved in legal document analysis. This session is ideal for developers, data scientists, and legal tech enthusiasts looking to leverage advanced AI techniques for document processing and analysis. Join us to discover how cutting-edge technology can transform the way we interact with complex legal texts.
🙂 CÉDRIC MENDELIN ⚡️ Senior Software Developer @ isolutions

Tweet

More Decks by Azure Zurich User Group

Other Decks in Technology

Transcript

  1. AGENDA LLM Basics Customer Project - FTA The Microsoft Way

    - Azure AI Services Step-by-Step Improvements - FTA Conclusion
  2. AZURE AI SERVICES OpenAI Vision Speech Language Content Safety Face

    Document Intelligence Azure AI services AI Search AI Agent Service AI Model Inference AI Foundry
  3. LOW-CODE VS. PRO-CODE Source: Guest Blog: Semantic Kernel and Copilot

    Studio Usage Series - Part 1 | Semantic Kernel
  4. AI SEARCH • Index • Fields • Chunk • Vector

    Store • Query Processing • Reranking + other advanced features • Indexer & Skills
  5. 1ST ITERATION Focus on 50 documents Index Laws per article

    Index PDFs per page Azure Open AI – on your data Vector Search
  6. EVALUATION TYPES LLM evaluation How good the foundation models performs

    on a certain task. LLM system evaluation How good the LLM performs in your specific use case, on your data, in your domain.
  7. MEAI.EVALUATION OVERVIEW • Open-source • Predefined LLM-based evaluators • Interface

    for custom-evaluators • Local and Azure Storage Account • In Preview
  8. EVALUATION RESULTS 40 45 37 43 0 10 20 30

    40 50 60 70 80 90 100 Vector Search Hybrid Search % LLM system evaluation – Retrieval Step Provided source Applied source
  9. EVALUATION RESULTS 0.8 0.85 0.948 0.948 0.914 0.917 0.41 0.4

    0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Vector Search Hybrid Search LLM system evaluation Relevance Groundedness Cosine Sim Embedding Euclidean Distance
  10. EVALUATION RESULTS 40 45 54 79 37 43 49 75

    0 10 20 30 40 50 60 70 80 90 100 Vector Search Hybrid Search Hybrid Search with Summary Hybrid Search with Reranking LLM system evaluation – Retrieval Step Provided source Applied source
  11. EVALUATION RESULT 0.8 0.85 0.9 0.948 0.948 0.988 0.914 0.917

    0.927 0.41 0.4 0.37 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Vector Search Hybrid Search Hybrid Search with Reranking LLM system evaluation Relevance Groundedness Cosine Sim Embedding Euclidean Distance
  12. IMPLEMENTATION • Not supported by Azure OpenAI - On your

    data • Derive Text Query • Using Azure AI Search SDK + Autogen
  13. EVALUATION – EXECUTION TIME 2889 2925 1999 2170 2966 0

    500 1000 1500 2000 2500 3000 3500 OYOD - Hybrid OYOD - Hybrid + SR Custom - Hybrid Custom - Hybrid + SR Custom - Hybrid + SR + Reflection ms LLM system evaluation
  14. EVALUATION RESULTS 0.9 0.9 0.98 0.97 0.926 0.925 0.41 0.4

    0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Single Agent Multi Agent LLM system evaluation Relevance Groundedness Cosine Sim Embedding Euclidean Distance
  15. CONCLUSION • Step-by-Step Improvements • SK is your SDK of

    choice • Azure AI Search for unstructured data • Use Advanced Search Capabilities • Start Evaluating early • What is your Use case • Business Value & Innovation
  16. REFERENCES • Flaticon.com (for used icons) • Snappify (for code

    snippets) • The Microsoft.Extensions.AI.Evaluation libraries (Preview) • Evaluating Large Language Model (LLM) systems: Metrics, challenges, and best practices • Evaluation and monitoring metrics for generative AI • LLM Evaluation Metrics: The Ultimate LLM Evaluation Guide • A list of metrics for evaluation LLM-generated content • https://github.com/joslat/AgenticAIAutoGen