Page Not Found
Page not found. Your pixels are in another canvas.
A list of all the posts and pages found on the site. For you robots out there is an XML version available for digesting as well.
Page not found. Your pixels are in another canvas.
May not be a full list.
“I don’t know”. It is the acceptance of our limited knowledge.
What does it mean to win a game? Many people try to win every game. But at what cost, what was the point of winning?
Are we intelligent? Will any intelligent person think themself of intelligent?
Are we entitled to have something for sure? Are we even entitled to think about our entitlement?
The dictionary meaning “the principles or practice of passive submission to constituted authority even when unjust or oppressive” doesn’t capture the eastern...
Leverage is something you can use to get maximum advantage of something. It is a tool. For you to use a tool, first you have to understand the fundamental pr...
Is someone separate from their perspective? How can we disentangle these two?
It is a standard consensus that experience is the best teacher. How does just the experience can be the best teacher?
What if you could get rid of the idea that you always need to catch-up with your peers status wise? What changes will you make in your life? What stops you f...
If everything you believe is something that you are supposed to believe, what are the odds that it is really a coincidence?
Education is the industrial process of making people compliant. Command and control is the backbone of it. While learning is unleashing of a curious mind ag...
The former is about taking, whereas the latter is all about giving. One is short-lived and the other is long-lived.
Hello world, and everyone.
“From Passive Tools to Active Assistants: The Cognitive Revolution in Software.”
related_dsa_day: 2 related_ml_day: 2 related_speech_day: 2
related_dsa_day: 3 related_ml_day: 3 related_speech_day: 3
related_dsa_day: 4 related_ml_day: 4 related_speech_day: 4
related_dsa_day: 5 related_ml_day: 5 related_speech_day: 5
“To Framework or Not to Framework? Navigating the Agent Ecosystem.”
related_dsa_day: 7 related_ml_day: 7 related_speech_day: 7
“Better workflows beat better models.” — Dr. Andrew Ng
“Giving the Brain a Library: The Foundation of Knowledge-Intensive Agents.”
“Garbage In, Garbage Out. The Art of Reading Messy Data.”
“Finding a Needle in a High-Dimensional Haystack: The Mathematics of Recall.”
“The Finite Canvas of Intelligence: Managing the Agent’s RAM.”
“Thinking Fast and Slow: How to make LLMs stop guessing and start solving.”
“Reason + Act: The Loop that Changed Everything.”
“If you fail to plan, you are planning to fail (and burn tokens).”
“Speed is not a feature. Speed is the product.”
“Talking to machines: The end of the Keyboard.”
“Don’t build the phone network. Just build the app.”
“The art of knowing when to shut up.”
“Removing the Text Bottleneck: The Omni Future.”
“Giving eyes to the brain: How Agents see the world.”
“One model to rule them all: The Unification of Text and Vision.”
“The ultimate API: The User Interface.”
“Knowing WHAT is where, and precisely WHERE it acts.”
“Time: The 4th Dimension of Vision.”
The hash table trick that makes O(n²) become O(n) and why this pattern appears everywhere from feature stores to embedding lookups.
Why a simple stack solves bracket matching, expression parsing, and even neural network depth management in one elegant pattern.
The pointer manipulation pattern that powers merge sort, data pipeline merging, and multi-source stream processing.
The single-pass pattern that powers streaming analytics, online algorithms, and real-time decision making in production systems.
Master the pattern behind online algorithms, streaming analytics, and dynamic programming, a single elegant idea powering countless production systems.
The Fibonacci problem in disguise, teaching the fundamental transition from recursion to dynamic programming to space optimization.
Master the fundamental patterns of tree traversal: the gateway to solving hundreds of tree problems in interviews.
Master BST validation to understand data integrity in tree structures, critical for indexing and search systems.
Master binary search to understand logarithmic algorithms and efficient searching, foundational for optimization and search systems.
Master linked list manipulation through reversal - a fundamental pattern for understanding pointer logic and in-place algorithms.
Master LRU cache design: O(1) get/put with hash map + doubly linked list. Critical for interviews and production caching systems.
Master digit-by-digit addition with linked lists: Handle carry propagation elegantly. Classic problem teaching pointer manipulation and edge cases.
Master the two-pointer greedy technique that powers resource optimization in production ML systems.
Master backtracking to generate all valid combinations—the foundation of ensemble model selection and multi-model systems.
Master hash-based grouping to solve anagrams—the foundation of clustering systems and speaker diarization in production ML.
Master interval processing to handle overlapping ranges—the foundation of event streams and temporal reasoning in production systems.
Simulate arbitrary-precision addition on linked lists—the same sequential pattern used in large-scale distributed training and streaming pipelines.
Master in-place matrix rotation—the same 2D transformation pattern that powers image and spectrogram augmentations in modern ML systems.
Master systematic matrix traversal—the same pattern used for tracking experiments, processing logs, and managing state in ML systems.
Master greedy decision-making to determine reachability—the same adaptive strategy used in online learning and real-time speech systems.
Master grid path counting with dynamic programming—the same optimization technique used in neural architecture search and speech model design.
The classic grid optimization problem that bridges the gap between simple recursion and 2D Dynamic Programming.
A deceptive counting problem that teaches the fundamentals of state transitions and connects directly to Beam Search.
The fundamental string segmentation problem that powers spell checkers, search engines, and tokenizers.
The gatekeeper of data integrity. How do we ensure our sorted structures are actually sorted?
How do you print a corporate hierarchy level by level? CEO first, then VPs, then Managers…
Given two arrays, can you rebuild the original tree? It’s like solving a jigsaw puzzle where the pieces are numbers.
Finding the median or the 99th percentile is easy in a sorted array. Can we do it in a tree?
“Find the point where two paths in a tree first meet.”
“Counting connected components in a 2D grid.”
“Can you finish all courses given their prerequisites?”
“Transforming ‘cold’ to ‘warm’ one letter at a time.”
“Creating a deep copy of a graph structure.”
“Modeling algebraic equations as graph path problems.”
“Capturing regions by identifying safe boundaries.”
“Can you split the treasure evenly?”
“Finding the longest upward trend in chaos.”
“Making change with the fewest coins.”
“Making sense of a stream of characters.”
“Calculating capacity in a fragmented landscape.”
“Finding the optimal path through a sequence of choices.”
“Combining order from chaos, one element at a time.”
“Finding the middle ground between two ordered worlds.”
“Finding the maximum hidden in the valleys and peaks.”
“Finding the king of every window.”
How do you narrow down 10 million items to 1000 candidates in under 50ms? The art of fast retrieval at scale.
From raw data to production predictions: building a classification pipeline that handles millions of requests with 99.9% uptime.
How to build production-grade pipelines that clean, transform, and validate billions of data points before training.
How to design experimentation platforms that enable rapid iteration while maintaining statistical rigor at scale.
How to choose between batch and real-time inference, the architectural decision that shapes your entire ML serving infrastructure.
How to measure if your ML model is actually good, choosing the right metrics is as important as building the model itself.
Feature engineering makes or breaks ML models, learn how to build scalable, production-ready feature pipelines that power real-world systems.
Design production-grade model serving systems that deliver predictions at scale with low latency and high reliability.
Design systems that learn continuously from streaming data, adapting to changing patterns without full retraining.
Design efficient caching layers for ML systems to reduce latency, save compute costs, and improve user experience at scale.
Design a global CDN for ML systems: Edge caching reduces latency from 500ms to 50ms. Critical for real-time predictions worldwide.
Design distributed ML systems that scale to billions of predictions: Master replication, sharding, consensus, and fault tolerance for production ML.
Build production ML infrastructure that dynamically allocates resources using greedy optimization to maximize throughput and minimize costs.
Build production ensemble systems that combine multiple models using backtracking strategies to explore optimal combinations.
Design production clustering systems that group similar items using hash-based and distance-based approaches for recommendations, search, and analytics.
Build production event stream processing systems that handle millions of events per second using windowing and temporal aggregation—applying the same interva...
Design distributed training architectures that can efficiently process massive sequential datasets and train billion-parameter models across thousands of GPUs.
Design a robust data augmentation pipeline that applies rich transformations to large-scale datasets without becoming the training bottleneck.
Design robust experiment tracking systems that enable systematic exploration, reproducibility, and collaboration across large ML teams.
Design online learning systems that adapt models in real-time using greedy updates—the same adaptive decision-making pattern from Jump Game applied to stream...
Design neural architecture search systems that automatically discover optimal model architectures using dynamic programming and path optimization—the same pr...
A comprehensive guide to FinOps for Machine Learning: reducing TCO without compromising accuracy or latency.
The industry-standard algorithm for converting probabilistic model outputs into coherent text sequences.
The critical preprocessing step that defines the vocabulary and capabilities of Large Language Models.
The silent killer of ML models is not a bug in the code, but a change in the world.
Not everything needs to be real-time. Sometimes, “tomorrow morning” is fast enough.
Architecture is destiny. The difference between 50% accuracy and 90% accuracy is often just a skip connection.
How does Google search 50 billion pages in 0.1 seconds? The answer is the “Ranking Funnel”.
“Organizing the world’s information into a structured hierarchy.”
“Leveraging the connection structure to predict what users will love.”
“Managing complex ML workflows with thousands of interdependent tasks.”
“Moving beyond keywords to understand the meaning of a query.”
“Ensuring your ML models are available everywhere, all the time.”
“Structuring the world’s information into connected entities and relationships.”
“Defining where one object ends and another begins.”
“How to share a supercomputer without stepping on each other’s toes.”
“Predicting the next word, the next stock price, the next frame.”
“Finding the perfect knobs to turn.”
“Trust, but verify. Why did the model say No?”
“Scaling from one GPU to thousands.”
“Fitting billion-parameter models into megabytes.”
“The centralized truth for machine learning features.”
“The infrastructure for semantic search and AI-native applications.”
“Serving models that think at human scale.”
“Grounding LLMs in facts, not hallucinations.”
[Challenge] Blizzard Challenge 2015
[Conference] Community-based Building of Language Resources(CBBLR), Brno, Czech Republic, September 2016
[Conference] International Conference on Text, Speech, and Dialogue(TSD), Brno, Czech Republic, September 2016
[Conference] INTERSPEECH 2017 (Show and Tell), Stockholm, Sweden, August 2017
[Conference] INTERSPEECH 2017, Stockholm, Sweden, August 2017
[Conference] Global Conference on Cyberspace (GCCS), New Delhi, India, November 2017
[Conference] Frontiers of Research in Speech and Music (FRSM), Rourkela, India, December 2017
[Conference] The 6th Intl Workshop on Spoken Language Technologies for Under Resourced Languages, Gurugram, India, August 2018
[Conference] INTERSPEECH 2018, Hyderabad, India, September 2018
[MS Thesis] IIT Madras: February 2019; Supervised by Prof. Hema A Murthy
[arXiv] arXiv, May 2020
[Journal] Speech Communication: Volume 123, October 2020, Pages 10-25
[Conference] Speech Synthesis Workshop (SSW), Hungary, Aug 2021
[Conference] National Conference on Communications (NCC 2024), February 2024
[Conference] International Conference on Acoustics, Speech, and Signal Processing ( ICASSP), April 2024
Why batch ASR won’t work for voice assistants, and how streaming models transcribe speech as you speak in under 200ms.
How voice assistants recognize “turn on the lights” from raw audio in under 100ms without full ASR transcription.
How to transform raw audio waveforms into ML-ready features that capture speech characteristics for robust model training.
How voice assistants and video conferencing apps detect when you’re speaking vs silence, the critical first step in every speech pipeline.
How voice assistants recognize who’s speaking, the biometric authentication powering “Hey Alexa” and personalized experiences.
From text to natural speech: understanding modern neural TTS architectures that power Alexa, Google Assistant, and Siri.
Clean audio is the foundation of robust speech systems, master preprocessing pipelines that handle real-world noise and variability.
Build real-time speech processing pipelines that handle audio streams with minimal latency for live transcription and voice interfaces.
Build lightweight models that detect specific keywords in audio streams with minimal latency and power consumption for voice interfaces.
Build systems that enhance voice quality by removing noise, improving intelligibility, and optimizing audio for speech applications.
Separate overlapping speakers with 99%+ accuracy: Deep learning solves the cocktail party problem for meeting transcription and voice assistants.
Build production multi-speaker ASR systems: Combine speech recognition, speaker diarization, and overlap handling for real-world conversations.
Optimize speech pipeline throughput by allocating compute to bottleneck stages using greedy resource management.
Build production speech systems that combine multiple ASR/TTS models using backtracking-based selection strategies to achieve state-of-the-art accuracy.
Build production speaker diarization systems that cluster audio segments by speaker using embedding-based similarity and hash-based grouping.
Build production audio segmentation systems that detect boundaries in real-time using interval merging and temporal processing—the same principles from merge...
Design distributed training pipelines for large-scale speech models that efficiently handle hundreds of thousands of hours of sequential audio data.
Use audio augmentation techniques to make speech models robust to noise, accents, channels, and real-world conditions—built on the same matrix/tensor transfo...
Design experiment management systems tailored for speech research—tracking audio data, models, metrics, and multi-dimensional experiments at scale.
Design adaptive speech models that adjust in real-time to speakers, accents, noise, and domains—using the same greedy adaptation strategy as Jump Game and on...
Design neural architecture search systems for speech models that automatically discover optimal ASR/TTS architectures—using dynamic programming and path opti...
Strategies for building profitable speech recognition systems by optimizing the entire pipeline from signal processing to hardware.
Implementing the core decoding logic of modern Speech Recognition systems, handling alignment, blanks, and language models.
The breakthrough that allows us to treat audio like text, enabling GPT-style models for speech.
How do we know if the audio sounds “good” without asking a human?
Real-time ASR is hard. Offline ASR is big.
Goodbye HMMs. Goodbye Phonemes. Goodbye Lexicons. We are teaching the machine to Listen, Attend, and Spell.
“Play Call Me Maybe”. Did you mean the song, the video, or the contact named ‘Maybe’?
“From broad categories to fine-grained speech understanding.”
“Building recommendation and moderation systems for voice-based social platforms.”
“Orchestrating complex speech processing pipelines from audio ingestion to final output.”
“Finding ‘Jon’ when the user types ‘John’, or ‘Symphony’ when they say ‘Simfoni’.”
“Deploying speech models close to users for low-latency voice experiences.”
“The brain of a task-oriented dialogue system: remembering what the user wants.”
“Knowing when to listen and when to stop.”
“One model to rule them all: ASR, Translation, and Understanding.”
“From waveforms to words, and back again.”
“Tuning speech models for peak performance.”
“Giving machines a voice.”
“Hey Siri, Alexa, OK Google: The gateway to voice AI.”
“Who spoke when? The art of untangling voices.”
“Turning acoustic probabilities into coherent text.”
“Extracting clear speech from the noise of the real world.”
“Speaking with someone else’s voice.”
“Teaching machines to hear feelings.”
While it’s not a principle, I often think of the parable of the Taoist farmer. The Taoist farmer has one horse, and the horse runs off. The villagers lame...
I am so firmly determined, however, to test the constancy of your mind that, drawing from the teachings of great men, I shall give you also a lesson: Set ...
“A fit body, a calm mind, a house full of love. These things cannot be bought—they must be earned.”
“If you ever want to have peace in your life, you have to move beyond good and evil.” “Nature has no concept of happiness or unhappiness. Nature follow...
“Reading is to the mind what exercise is to the body, ”- Richard Steele.
Happiness is not a consumable product. It is not something you find by searching for it. It is a naturally arising byproduct of a fulfilling, well-lived l...
When you care more about getting things right than being right, you get better outcomes and you save time and energy.
The ceramics teacher announced on opening day that he was dividing the class into two groups. All those on the left side of the studio, he said, would be ...
The best way to improve your ability to think is to spend time thinking. Most of us are too busy to think. We have too many meetings. Too many calls. Too ...
We rarely do or say something intentionally that surprises us. That’s because we are in intimate contact with the noise in our heads–we spend our days loo...
Nothing will change your future trajectory like your habits. While goals rely on extrinsic motivation, habits, once formed, are automatic. They literally ...
“How we spend our time is how we spend our days. How we spend our days is how our life goes. How our life goes determines whether we thought it was worth ...
While we tell ourselves that the next level is enough, it never is. The next zero in your bank account won’t satisfy you any more than you are now. The ne...
“Expectation is the grandfather of disappointment. The world can never own a man who wants nothing.” — Aphorisms for Thirsty Fish
One simple way to unlock your best self is to shape your environment so that your desired behavior is the path of least resistance.
“The nature of illusion is that it’s designed to make you feel good. About yourself, about your country, about where you’re going – in that sense it funct...
People are much more honest with their actions than their words.
In turning education into a system of mass production we created a superbly democratic system that made the majority of people, and the world as a whole, ...
“He who knows only his own side of the case, knows little of that.” — John Stuart Mill
Say no (a lot).
To improve your outcomes in life, respond to the world as it is, not as you wish it would be.
Sturgeon’s law states that 90% of everything is crap. If you dislike poetry, or fine art, or anything, it’s possible you’ve only ever seen the crap. Go lo...
“It’s time you realized that you have something in you more powerful and miraculous than the things that affect you and make you dance like a puppet.” — M...
The person who is consistent outperforms the person who is intermittent every time. While inconsistent effort works for some things, for the things that r...
“One day, you will wake up and there won’t be any more time to do the things you’ve always wanted. Do it now.” - Paulo Coelho
New year, new me? Nah, I’m just going to keep on being fabulous and making mistakes like I always do 😜 Happy New Year everyone!
Most people spend the first half of their lives collecting and the second half choosing what to keep. Which lessons learned and pieces of advice do you...
Don’t believe everything you think.
A simple and easy approach to decision-making that prevents us from manipulating ourselves. First, understand the forces at play. Then, understand how you...
Productivity is often a distraction. Don’t aim for better ways to get through your tasks as quickly as possib`le. Instead aim for better tasks that you ne...
Are those things that keep you busy truly important in your life and career?
Don’t define your identity by your beliefs. Define your identity by your willingness to learn.
No one is thinking about you very much. So don’t worry about looking stupid or embarrassing yourself or whatever. No one cares.
Worrying is praying for what you dont want.
We are who we are when nobody else is watching.
Those who cannot live in harmony with the world are fools though they may be highly educated.
The work you do while you procrastinate is probably the work you should be doing for the rest of your life.
“To travel means, ultimately, nothing more than coming back home a different person from the one who left.” — PICO IYER
Try to define yourself by what you love and embrace, rather than what you hate and refuse.
The price you pay for doing what everyone else does is getting what everyone else gets.