OpenAI and ChatGPT repo
- GenAI. Where could be applied. Post 1.pdf
- GenAI in Application Refactoring field, Slides.pdf
- Legal problems with AI.pdf
- Paradigms: Rag, Self-RAG, Re-Ranking RAG, FLARE v.2.pdf
- Working with opinionated requests. S2A, RLHF, RLAIF.pdf
- Multi-Modal RAG and its features.pdf
- Measuring the GenAI Quality.pdf
- LLM leveraging RLHF in code review
- Everything of Thoughts (XoT). All modern techniques in one place
- Non deterministic embedding results
- AI Search vs PostgreSQL with pgvector in PROD
- Prod-Ready LLM Solutions. Cook Book.
- Quality Framework For RAG Applications.pdf
- Crew.AI. Agents in LLM Applications (In Progress)
- Pydantic data classes and how to manage the output format (In Progress)
- XML vs Markdown vs Json for tagging in prompting and metaprompting (In Progress)
- Crawlers for LLMs:
- https://python.langchain.com/v0.1/docs/use_cases/web_scraping/ ,
- https://ai.gopubby.com/use-ai-to-scrape-almost-all-websites-easily-in-2025-f868adc41e0f,
- https://github.com/Skyvern-AI/skyvern,
- https://gotenberg.dev/docs/routes,
- https://jina.ai/reader,
- https://github.com/unclecode/crawl4ai,
- https://crawlee.dev/,
- https://github.com/bracesproul/site-rag/,
- https://www.firecrawl.dev,
- https://github.com/mishushakov/llm-scraper
- Table extraction in RAG systems (In Progress)
- Choosing the right programming language for your next AI LLM project
- Misjudgements using LogProbs (In Progress)
- June 2023. My Workshop Presentation. Run 1.pptx
- Online Workshop. ChatGPT -> Azure Function -> PowerAutomate. Run 2.pptx
- Online Workshop. Run 3. Deep Learning -> Prompting -> ChatGPT -> Azure Function -> PowerAutomate
- Online+Offline Workshop for EHU University
- Talk #3. RAG, FLARE, S2A, RLHF, RLAIF, Self-RAG, Re-Ranking. Common approaches and their pros & cons
- Six Principles of responsible AI
- Responsible AI. Trusted AI Framework. Content Filters. Harmful Content. Prerelease Reviews
- What is ChatGPT Doing. and why does it work
- LLM UseCase in Google. Sorting Optimization
- Embeddings. Words to Vector. Useful in Search Scenarios and for Cognitive Search
- Cognitive Search. Video
- Cognitive Search. From Zero to Hero
- Cognitive Search. Indexers. AI Enrichment. Build-in Skills
- Transformers. Embeddings. Foundational Model
- Computer Vision. Cognitive. AI Face. Custom Vision
- Document Intelligence
- Azure AI Speech. Speech To Text. Text To Speech. Azure Services
- Natural Language Processing(NLP). Text Meaning and analysis. General ways how to
- Azure Language Service. Commands interpretation
- Azure Language Service. Question-Answer Knowledge base for bots. Question Answering service.
- Regression. Logistic and Linear Regression. Multiclass regression
- AI Search. Debug Search Issues
- AI Search. Performance and Monitoring
- AI Search. Search and Scoring
- AI Search. Implement Advanced Search Features. Scoring Profiles, Fuzzy Search, Term Boosting, Term Proximity
- AI Search. Scoring profile lab. Add Different Language descriptions
- AI Search. Enchance Index by translation using skills
- AI Search. Custom Skill using Azure Function
- AI Search. Use Custom Analyzers (not default Microsoft Lucene)
- AI Search. Geo-spatial functions
- AI Search. Knowledge Mining. Lab
- AI Search -> PowerBI Table Projection from OCR Document Intelligence
- Composed Document Intelligense Models. Case if you need to analyze several doc types
- Vision. Train a Custom Model using COCO
- Containers. AI Services in Containers, in AKS, ACI, or even locally
- Containers. Run services in Isolated Environment disconnected from the internet
- Analyze Video Indexer. Widgets Integration and API
- Semantic Ranking configuration in AI Search Index
- Knowledge Store & Knowledge Mining with AI Search
- Integrate OpenAI into App. Useful Lab
- Host Mistral and other models in AI Hub
- AI Language. Multi-turn multi-step conversation
- AI Language. Conversation Language understanding. Classical way to build AI-assistant. Utterances: Turn-on Turn-off & Smart home
- AI Language. Custom Named Entities Recognition. Laws, Business Cases
- Key Phrases Extraction from text, Sentiment Analysis, Linked Entities
- Translate speech to text. Materials
- Translate speech to text and synthesize the output if needed. Example
- AI Speech. Speech Synthesis
- Run Cognitive Services in Docker
- Custom Vision. Deploy Custom Vision on the edge devices (phones) using compact models
- Custom Vision. Recognize issues in factory. Upload images, Tag Images, and Train the Model
- Anomaly Detector for IoT. Univariate Detector (multi-stream)
- Cognitive Search. Video
- Cognitive Search. From Zero to Hero
- Cognitive Search. Indexers. AI Enrichment. Build-in Skills
- Document Intelligence
- Vector Database selection & comparison. VectorDB
- Transformer Explainer. Transformer Explainer is an interactive visualization tool designed to help anyone learn how Transformer-based models like GPT work
- Table extraction in RAG systems
- Example:ConsoleApp CommandGuess
- Example: Azure Function with ChatGPT (completion and chat-completion)
- Example: Integration with PowerAutomate
- Example: Integration with PowerApp
- Integration with Outlook (In progress)
- OpenAI + PowerAutomate Workshop by me.pptx
- Example: OpenAI + Redis
- BMW Dealer assistant. ChatGPT Chat + Startup + Redis + Context
- Get Embedding
- Form Recognizer Cognitive Service
- Content Filters (in progress)
- OpenAI straightforward examples
- Azure Bot Service & Chatbot Framework
- LangChain meets Go
- TenzorZero Framework (In progress)
- Key Phrases Extraction. AI Language. Sentiment Analysis. Extracted Linked Entities
- AI Search and Custom Skill using Azure Function
- Document Intelligence, Best Practices (In progress)
- MCP Server example using FastMCP
- Document Retrieval Metrics
a. NDCG@K (Normalized Discounted Cumulative Gain) - Ranking quality with relevance grades
b. Mean Reciprocal Rank (MRR) - First relevant document positioning. How quickly users find their first relevant result. Critical for RAG user experience.
c. Contextual Relevancy - How relevant is the retrieved context to the user's question?
c. Expected Reciprocal Rank (ERR) - User behavior modeling with graded relevance
d. Rank-Biased Precision (RBP) - Early result weighting strategies
e. Embedding Quality Metrics - Intra-cluster vs inter-cluster distance analysis. Quality of your vector space - are similar documents close together - Document Retrieval Metrics 2
a. Fidelity - Measures recall quality - what percentage of all relevant documents in your dataset were actually retrieved in the top-n results.
b. XDCG - Ranking quality within your retrieved top-k chunks, ignoring the rest of your document collection
c. XDCG vs NDCG
d. Max Relevance N - highest relevance score among your top-k retrieved chunks
e. Holes - Counts missing ground truth data - Response Quality Metrics
a. F1, Recall, Precision. Fundamental metrics
b. BLEU Score - N-gram overlap evaluation
c. ROUGE (L/1/2) - Recall-oriented summarization metrics
d. G-Eval (LLM as a judge) - Sophisticated evaluation framework that uses LLMs themselves to evaluate outputs based on detailed criteria.
d. BERTScore - Semantic similarity using contextualized embeddings
e. BLEURT - BERT-based learned evaluation metric
f. SacreBLEU - Standardized BLEU with proper tokenization
g. METEOR - Synonym and paraphrase consideration
h. CIDEr - Consensus-based evaluation
i. CHRF - Character-level F-score for multilingual evaluation - Human-Correlation Metrics
a. Preference-Based Ranking - Win/loss ratios in A/B testing
b. Pearson/Spearman Correlation - Human judge alignment
c. Likert Scale Rating Systems - Multi-point evaluation frameworks
- RAG System assessment and quality control:
a. LangWatch
b. LangFuse
b. Galileo
c. Ragas
d. DeepEval
e. TrueLens
f. HuggingFace (NEW, in progress)
g. AI Foundry - Agentic Application Evaluation
a. General Agentic Application Evaluations
b. Monitoring
c. Trajectory Evaluation
d. Structure of the Evaluation
e. Application Improvements using G-Eval (LLM-as-a-Judge)
- Neural Network Fundamentals
a. ReLU vs Advanced Activations (GELU, Swish/SiLU) b. Layer Normalization vs Batch Normalization - Training stability techniques
c. Gradient Clipping - Exploding gradient prevention
d. Mixed Precision Training - FP16/BF16 memory optimization - CNN Advanced Concepts
a. Kernel Size Impact - Local vs global feature extraction (3x3 vs 7x7)
b. Parameter Sharing Benefits - Translation invariance principles
c. Hierarchical Feature Learning - Low-level to high-level progression
d. CNN vs MLP Scalability - O(k×c×f) vs O(n×m) parameter complexity - Advanced Training Techniques
a. Learning Rate Scheduling - Cosine annealing, linear decay
b. Warmup Steps - Training stability (10% of total steps)
c. Checkpoint Averaging - Model stability improvement
d. Gradient Accumulation - Simulating larger batch sizes
- LoRA (Low-Rank Adaptation)
a. Rank Parameter (r) - 8-64 range, efficiency vs capacity trade-off
b. Alpha Scaling Factor - Typically 16-32
c. Target Module Selection - Query, value, key, output projections
d. AdaLoRA - Adaptive rank allocation
e. QLoRA - 4-bit quantized LoRA for memory efficiency - Training Parameters
a. Learning Rate Ranges - 1e-5 to 5e-4 for LLMs with warmup
b. Batch Size Optimization - 8-32 full fine-tuning, 64-128 LoRA
c. Sequence Length Limits - 512-4096 tokens task dependency
d. Weight Decay (L2 Regularization) - λ||w||² with λ = 1e-4 to 1e-2
-
Re-ranking Algorithms
a. Reciprocal Rank Fusion (RRF) - RRF_score = Σ(1/(k + rank_i))
b. Cross-encoder vs Bi-encoder - Accuracy vs speed trade-offs
c. Neural Re-rankers - BERT/T5-based cross-attention models
d. Learning to Rank (LTR) - ML-based ranking optimization
e. Score Normalization Techniques - Min-max, z-score, sigmoid -
Advanced Retrieval Concepts
a. Semantic Similarity Scoring - Cosine similarity between embeddings
b. Context Preservation - Chunk coherence maintenance
c. Window Size Optimization - Re-ranking candidate selection (100-1000)
-
Evaluation Platforms
a. AI Foundry (Microsoft) - Model testing and evaluation
b. Weights & Biases (W&B) - Experiment tracking
c. Neptune.ai - MLOps platform capabilities
d. LangSmith (LangChain) - LLM application testing
e. Phoenix (Arize AI) - LLM observability and evaluation -
Model Management
a. MLFlow - Model lifecycle management
b. DVC (Data Version Control) - Data and model versioning
c. BentoML - Model serving framework architecture
- LangGraph Basics
a. When to Use What (Decision Framework)
b. Core LangGraph Primitives: StateGraph & MessageGraph, Compilation model, Checkpointers, Thread/Run concepts
c. Graph Execution Model: how LangGraph executes iteratively. StateGraph & MessageGraph
d. LangGraph Checkpointers: MemorySaver, SqliteSaver, PostgresSaver
e. LangGraph composition: START, END, Conditional Edge. Parallel node execution. Cycle limit (recursion_limit), infinite loops
e. Subgraphs & Composition: when to use subgraphs vs separate graphs
f. Error Handling & Interrupts (Critical for production) - LangGraph Advanced Topics
a. State Management - Persistent conversation state
b. Graph Architecture - Nodes and edges for complex workflows
c. Conditional Routing - Dynamic flow based on LLM decisions
d. Human-in-the-Loop - Approval gates and manual interventions
e. Parallel Processing - Concurrent graph branch execution - LangGraph Examples and Prototypes
- LangGraph System Prompt Techniques
a. Decision-Tree Prompts & Pattern
b. Multi-Agent Prompt & Pattern. Primitive Version
c. Plan-Execute Prompt & Pattern
d. ReAct. Prompts & Ideas
e. Prompt-Reflection Pattern. Idea - Semantic Kernel (Microsoft)
a. Kernel Architecture - Central orchestration engine
b. Plugin System - Reusable functions (native C# or prompt-based)
c. Planners - Automatic workflow generation
d. Memory Management - Vector-based semantic memory patterns - Magentic One + Semantic Kernel
- Advanced Framework Concepts
a. Multi-Agent Systems - Collaborative AI agent coordination
b. Error Recovery Strategies - Retry logic, fallback mechanisms
c. Async Execution - Resource management at scale
- Pydantic Advanced Usage
a. Field Validation - Custom validators, constraints (min/max, regex)
b. JSON Schema Generation - Automatic API documentation
c. Error Handling - Detailed validation error message design
d. Schema Compliance Monitoring - Production tracking metrics - Best Practices
a. Schema Complexity vs Success Rates - Optimization strategies
b. Retry Logic Implementation - Parse failure handling
c. Validation Feedback Loops - Error correction workflows
- Distributed Training Strategies
a. Data Parallelism - Batch distribution across GPUs
b. Model Parallelism - Layer splitting across devices
c. Pipeline Parallelism - Sequential processing stages
d. Gradient Synchronization - AllReduce, parameter servers
e. Mixed Precision Training - Memory efficiency optimization
- Regularization Techniques
a. Early Stopping - Validation loss plateau detection
b. Dropout Rates - 0.1-0.3 optimal ranges
c. Training/Validation Loss Curves - Overfitting gap analysis
d. Cross-validation Strategies - 5-10 fold robust evaluation
- Main Topics. What, Why, How
- Q&A. Quick references
a. Whether or not. Use Cases
b. How to Start
c. Parameters
d. Performance and Trade-offs - Examples
a. TenSeal, Concrete-ML, Microsoft Seal
- Initial Example
- Interactive Chat with Chat History
- Model Switching. Hugging Face
- Semantic Function for Conversational Chat
- Semantic Kernel Pipeline
Table of Content:
- LangChain using Golang (In Progress)
- LangChain. Demo examples wiht pipelines
- LangGraph. Patterns. Examples
- ReACT. Pre-coded loop + LLM to calculate the total weight of dogs
- React. Using LangGraph. In Progress
- React. Simple LangGraph Prototype



