Custom Vision Service as part of Azure Cognitive Services landscape of pretrained API services, provides you an ability to customize the state-of-the-art Computer Vision models for your specific use case. Using custom vision service you can upload set of images of your choice and categorize them accordingly using tags/categories and automatically train the image recognition… Continue reading
Category: Emerging Technologies
Emerging technologies include a variety of technologies such as educational technology, information technology, nanotechnology, biotechnology, cognitive science, psychotechnology, robotics, and artificial intelligence.
LLM Caching Strategies: Reducing Costs and Latency at Scale
Introduction: LLM API calls are expensive and slow. A single GPT-4 request can cost cents and take seconds—multiply that by thousands of users and costs spiral quickly. Caching is the most effective way to reduce both cost and latency. But LLM caching is different from traditional caching: exact string matches are rare, and semantically similar… Continue reading
Prompt Compression Techniques: Fitting More Context in Less Tokens
Introduction: Context windows are limited and tokens are expensive. Long prompts with extensive context, examples, or retrieved documents quickly hit limits and drive up costs. Prompt compression techniques reduce token count while preserving the information LLMs need to generate quality responses. This guide covers practical compression strategies: token pruning to remove low-information tokens, extractive summarization… Continue reading
Azure Cosmos DB – TTL (Time to Live) – Reference Usecase
TTL capability within Azure Cosmos DB is a live saver, as it would take necessary steps to purge redudent data based on the configurations you may. Let us think in terms of an Industrial IoT scenario, devices can produce vast amounts of telemetry information, logs and user session information that is only useful until we… Continue reading
Azure Database for MariaDB: Public Preview
During Ignite 2018, Microsoft has announced the availability of Maria DB support in Azure Database services. Today it has been opened for Public Preview for all Azure customers. What is MariaDB? MariaDB is a community-developed fork of the MySQL relational database management system intended to remain free under the GNU GPL.Development is led by some… Continue reading
Document Chunking Strategies: Optimizing RAG Retrieval Quality
Introduction: RAG systems live or die by their chunking strategy. Chunk too large and you waste context window space with irrelevant content. Chunk too small and you lose semantic coherence, making it hard for the LLM to understand context. The right chunking strategy depends on your document types, query patterns, and retrieval approach. This guide… Continue reading
NDepend–VSTS/Azure DevOps Integration–Part 01
In my previous article I wrote an introductory about NDepend and how it will be useful for Agile Team to ensure code quality. In that article we found how we can use NDepend in a developer machine. Now with this article we will familiarize ourselves in using NDepend in your build automation pipeline in your… Continue reading
New Microsoft Azure Certifications
Microsoft has recently announced new certification exam tracks for Azure Administrators, Developers and Architects. Here are the line ups that should help you move your career with right certifications. The three new Microsoft Azure Certifications are: Microsoft Certified Azure Developer Microsoft Certified Azure Administrator Microsoft Certified Azure Architect These certifications would essentially split the previous… Continue reading
Async LLM Patterns: Building High-Throughput AI Applications
Introduction: LLM APIs are inherently slow—even fast models take hundreds of milliseconds per request. When you need to process multiple prompts, make parallel API calls, or handle high-throughput workloads, synchronous code becomes a bottleneck. Async patterns let you overlap I/O wait times, dramatically improving throughput without adding complexity. This guide covers practical async patterns for… Continue reading
LLM Rate Limiting: Maximizing API Throughput Without Getting Throttled
Introduction: LLM APIs have strict rate limits—requests per minute, tokens per minute, and concurrent request limits. Hit these limits and your application grinds to a halt with 429 errors. Effective rate limiting isn’t just about staying under limits; it’s about maximizing throughput while maintaining reliability. This guide covers practical rate limiting patterns: token bucket algorithms… Continue reading