Category: Emerging Technologies

Emerging technologies include a variety of technologies such as educational technology, information technology, nanotechnology, biotechnology, cognitive science, psychotechnology, robotics, and artificial intelligence.

Azure Cognitive Services–Experience Image Recognition using Custom Vision (Build an Harrison Ford Classifier)

Posted on 7 min read

Custom Vision Service as part of Azure Cognitive Services landscape of pretrained API services, provides you an ability to customize the state-of-the-art Computer Vision models for your specific use case. Using custom vision service you can upload set of images of your choice and categorize them accordingly using tags/categories and automatically train the image recognition… Continue reading

LLM Caching Strategies: Reducing Costs and Latency at Scale

Posted on 19 min read

Introduction: LLM API calls are expensive and slow. A single GPT-4 request can cost cents and take seconds—multiply that by thousands of users and costs spiral quickly. Caching is the most effective way to reduce both cost and latency. But LLM caching is different from traditional caching: exact string matches are rare, and semantically similar… Continue reading

Prompt Compression Techniques: Fitting More Context in Less Tokens

Posted on 4 min read

Introduction: Context windows are limited and tokens are expensive. Long prompts with extensive context, examples, or retrieved documents quickly hit limits and drive up costs. Prompt compression techniques reduce token count while preserving the information LLMs need to generate quality responses. This guide covers practical compression strategies: token pruning to remove low-information tokens, extractive summarization… Continue reading

Azure Cosmos DB – TTL (Time to Live) – Reference Usecase

Posted on 2 min read

TTL capability within Azure Cosmos DB is a live saver, as it would take necessary steps to purge redudent data based on the configurations you may.  Let us think in terms of an Industrial IoT scenario, devices can produce vast amounts of telemetry information, logs and user session information that is only useful until we… Continue reading

Azure Database for MariaDB: Public Preview

Posted on 1 min read

During Ignite 2018, Microsoft has announced the availability of Maria DB support in Azure Database services. Today it has been opened for Public Preview for all Azure customers. What is MariaDB? MariaDB is a community-developed fork of the MySQL relational database management system intended to remain free under the GNU GPL.Development is led by some… Continue reading

Document Chunking Strategies: Optimizing RAG Retrieval Quality

Posted on 7 min read

Introduction: RAG systems live or die by their chunking strategy. Chunk too large and you waste context window space with irrelevant content. Chunk too small and you lose semantic coherence, making it hard for the LLM to understand context. The right chunking strategy depends on your document types, query patterns, and retrieval approach. This guide… Continue reading

NDepend–VSTS/Azure DevOps Integration–Part 01

Posted on 2 min read

In my previous article I wrote an introductory about NDepend and how it will be useful for Agile Team to ensure code quality. In that article we found how we can use NDepend in a developer machine. Now with this article we will familiarize ourselves in using NDepend in your build automation pipeline in your… Continue reading

New Microsoft Azure Certifications

Posted on 2 min read

Microsoft has recently announced new certification exam tracks for Azure Administrators, Developers and Architects. Here are the line ups that should help you move your career with right certifications.  The three new Microsoft Azure Certifications are: Microsoft Certified Azure Developer Microsoft Certified Azure Administrator Microsoft Certified Azure Architect These certifications would essentially split the previous… Continue reading

Async LLM Patterns: Building High-Throughput AI Applications

Posted on 17 min read

Introduction: LLM APIs are inherently slow—even fast models take hundreds of milliseconds per request. When you need to process multiple prompts, make parallel API calls, or handle high-throughput workloads, synchronous code becomes a bottleneck. Async patterns let you overlap I/O wait times, dramatically improving throughput without adding complexity. This guide covers practical async patterns for… Continue reading

LLM Rate Limiting: Maximizing API Throughput Without Getting Throttled

Posted on 16 min read

Introduction: LLM APIs have strict rate limits—requests per minute, tokens per minute, and concurrent request limits. Hit these limits and your application grinds to a halt with 429 errors. Effective rate limiting isn’t just about staying under limits; it’s about maximizing throughput while maintaining reliability. This guide covers practical rate limiting patterns: token bucket algorithms… Continue reading

Showing 341-350 of 442 posts
per page