We have seen recently a surge in vector databases in this era of generative AI. The idea behind vector databases is to index the data with vectors that relate to that data. Hierarchical Navigable Small World (HNSW) is one of the most efficient ways to build indexes for vector databases. The idea is to build a similarity graph and traverse that graph to find the nodes that are the closest to a query vector. Navigable Small World (NSW) is a process to build efficient graphs for search. We build a graph by adding vectors one after the others and connecting each new node to the most similar neighbors. When building the graph, we need to decide on a metric for similarity such that the search is optimized for the specific metric used to query items. Initially, when adding nodes, the density is low and the edges will tend to capture nodes that are far apart in similarity. Little by little, the density increases and the edges start to be shorter and shorter. As a consequence the graph is composed of long edges that allow us to traverse long distances in the graph, and short edges that capture closer neighbors. Because of it, we can quickly traverse the graph from one side to the other and look for nodes at a specific location in the vector space. When we want to find the nearest neighbor to a query vector, we initiate the search by starting at one node (i.e. node A in that case). Among its neighbors (D, G, C), we look for the closest node to the query (D). We iterate over that process until there are no closer neighbors to the query. Once we cannot move anymore, we found a close neighbor to the query. The search is approximate and the found node may not be the closest as the algorithm may be stuck in a local minima. The problem with NSW, is we spend a lot of iterations traversing the graph to arrive at the right node. The idea for Hierarchical Navigable Small World is to build multiple graph layers where each layer is less dense compared to the next. Each layer represents the same vector space, but not all vectors are added to the graph. Basically, we include a node in the graph at layer L with a probability P(L). We include all the nodes in the final layer (if we have N layers, we have P(N) = 1) and the probability gets smaller as we get toward the first layers. We have a higher chance of including a node in the following layer and we have P(L) < P(L + 1). The first layer allows us to traverse longer distances at each iteration where in the last layer, each iteration will tend to capture shorter distances. When we search for a node, we start first in layer 1 and go to the next layer if the NSW algorithm finds the closest neighbor in that layer. This allows us to find the approximate nearest neighbor in less iterations in average. ---- Find more similar content in my newsletter: TheAiEdge.io Next ML engineering Masterclass starting July 29th: MasterClass.TheAiEdge.io #machinelearning #datascience #artificialintelligence
Scalable System Design
Explore top LinkedIn content from expert professionals.
-
-
Load Balancing: Beyond the Basics - 5 Methods Every Architect Should Consider The backbone of scalable systems isn't just about adding more servers - it's about intelligently directing traffic between them. After years of implementing different approaches, here are the key load balancing methods that consistently prove their worth: 1. Round Robin Simple doesn't mean ineffective. It's like a traffic cop giving equal time to each lane - predictable and fair. While great for identical servers, it needs tweaking when your infrastructure varies in capacity. 2. Least Connection Method This one's my favorite for dynamic workloads. It's like a smart queuing system that always points users to the least busy server. Perfect for when your user sessions vary significantly in duration and resource usage. 3. Weighted Response Time Think of it as your most responsive waiter getting more tables. By factoring in actual server performance rather than just connection counts, you get better real-world performance. Great for heterogeneous environments. 4. Resource-Based Distribution The new kid on the block, but gaining traction fast. By monitoring CPU, memory, and network load in real-time, it makes smarter decisions than traditional methods. Especially valuable in cloud environments where resources can vary. 5. Source IP Hash When session persistence matters, this is your go-to. Perfect for applications where maintaining user context is crucial, like e-commerce platforms or banking applications. The real art isn't in picking one method, but in knowing when to use each. Sometimes, the best approach is a hybrid solution that adapts to your traffic patterns. What challenges have you faced with load balancing in production? Would love to hear your real-world experiences!
-
I love it when AI works out, because when it does - itâs magic. Here is my personal 5-step readiness checklist so you succeed with it. ð¦ðð²ð½ ð: ððð±ð¶ð ð¬ð¼ðð¿ ðð®ðð® ðð»ð³ð¿ð®ððð¿ðð°ððð¿ð² Before any AI conversation, ask: "Is our data clean, accessible, and flowing properly?" - Map your current data sources and quality. - Identify gaps between systems. - Ensure data governance policies are in place ð¦ðð²ð½ ð®: ðððð²ðð ð¬ð¼ðð¿ ð§ð²ð®ðº'ð ð§ð²ð°ðµð»ð¼ð¹ð¼ð´ð ðð¼ðºð³ð¼ð¿ð ðð¼ð»ð² Meet your people where they are, not where you want them to be. - Evaluate current tool proficiency (Are they Excel natives? Advanced analytics users?) - Identify the skills gap between current state and AI requirements. - Plan bridge training programs. ð¦ðð²ð½ ð¯: ððð¶ð¹ð± ðð ðð¶ðð²ð¿ð®ð°ð ðð°ð¿ð¼ðð ð¬ð¼ðð¿ ð¢ð¿ð´ð®ð»ð¶ðð®ðð¶ð¼ð» Create understanding before implementation. - Run AI awareness sessions for leadership and end-users. - Define AI terminology and use cases relevant to your industry. - Address concerns and misconceptions upfront. ð¦ðð²ð½ ð°: ð¦ðð®ð¿ð ð¦ðºð®ð¹ð¹ ðð¶ððµ ð£ð¶ð¹ð¼ð ð£ð¿ð¼ð´ð¿ð®ðºð Test the waters before diving in. - Choose one high-impact, low-risk use case. - Select a team that's excited about innovation. - Measure adoption rates, not just performance metrics ð¦ðð²ð½ ð±: ðððð®ð¯ð¹ð¶ððµ ð¦ðð°ð°ð²ðð ð ð²ðð¿ð¶ð°ð ð®ð»ð± ðð²ð²ð±ð¯ð®ð°ð¸ ðð¼ð¼ð½ð Define what winning looks like. - Set clear ROI expectations. - Create channels for user feedback and iteration. - Plan for scaling successful pilots Organizations that complete this readiness checklist see 3x higher adoption rates and significantly better long-term ROI. AI implementation isn't a sprint, it's a strategic marathon. Where is your organization in this readiness journey? What step are you focusing on right now?
-
â¼ï¸Ever wonder how data flows from collection to intelligent action? Hereâs a clear breakdown of the full Data & AI Tech Stack from raw input to insight-driven automation. Whether you're a data engineer, analyst, or AI builder, understanding each layer is key to creating scalable, intelligent systems. Letâs walk through the stack step by step: 1. ð¹Data Sources Everything begins with data. Pull it from apps, sensors, APIs, CRMs, or logs. This raw data is the fuel of every AI system. 2. ð¹Ingestion Layer Tools like Kafka, Flume, or Fivetran collect and move data into your system in real time or batches. 3. ð¹Storage Layer Store structured and unstructured data using data lakes (e.g., S3, HDFS) or warehouses (e.g., Snowflake, BigQuery). 4. ð¹Processing Layer Use Spark, DBT, or Airflow to clean, transform, and prepare data for analysis and AI. 5. ð¹Data Orchestration Schedule, monitor, and manage pipelines. Tools like Prefect and Dagster ensure your workflows run reliably and on time. 6. ð¹Feature Store Reusable, real-time features are managed here. Tecton or Feast allows consistency between training and production. 7. ð¹AI/ML Layer Train and deploy models using platforms like SageMaker, Vertex AI, or open-source libraries like PyTorch and TensorFlow. 8. ð¹Vector DB + RAG Store embeddings and retrieve relevant chunks with tools like Pinecone or Weaviate for smart assistant queries using Retrieval-Augmented Generation (RAG). 9. ð¹AI Agents & Workflows Put it all together. Tools like LangChain, AutoGen, and Flowise help you build agents that reason, decide, and act autonomously. ð Highly recommend becoming familiar this stack to help you go from data to decisions with confidence. ð Save this post as your go-to guide for designing modern, intelligent AI systems. #data #technology #artificialintelligence
-
ðð©ðªð´ ð¸ð¢ð´ ð´ð°ð®ð¦ðµð©ðªð¯ð¨ ðâð·ð¦ ð£ð¦ð¦ð¯ ð±ð¶ðµðµðªð¯ð¨ ðµð°ð¨ð¦ðµð©ð¦ð³ ðµð©ðªð´ ð¸ð¦ð¦ð¬. ðð¨ð ðð¥ð¥ ðð ð©ð¢ð©ðð¥ð¢ð§ðð¬ ðð«ð ðð«ððððð ððªð®ðð¥. Hereâs how I integrate Microsoft Azure services to create AI that works for just about any business not the other way around. Want to know the secret sauce? ð 7 Lessons from Building Scalable AI Solutions Customers Love: ðððð«ð ð°ð¢ðð¡ ðð¥ððð§ ðððð. â³ Use ðð³ð®ð«ð ðð¨ðð®ð¦ðð§ð ðð§ðð¥ð²ð³ðð« for structured ingestion. â³ Automate preprocessing with ðð³ð®ð«ð ð ð®ð§ððð¢ð¨ð§ ðð©ð©ð¬. â³ Store data securely in ðð³ð®ð«ð ðð¥ð¨ð ððð¨ð«ðð ð. ðð§ð ð¢ð§ððð« ððððð®ð«ðð¬ ðð®ð¬ðð¨ð¦ðð«ð¬ ð¯ðð¥ð®ð. â³ Focus on actionable insights, not noise. â³ Leverage ðð³ð®ð«ð ðððð¡ð¢ð§ð ðððð«ð§ð¢ð§ð for advanced prep. â³ Collaborate with end users for relevance. ðð«ðð¢ð§ ð¦ð¨ððð¥ð¬ ðð¡ðð ðð¥ð¢ð ð§ ð°ð¢ðð¡ ðð®ð¬ð¢ð§ðð¬ð¬ ð ð¨ðð¥ð¬. â³ Test multiple architectures, like custom LLMs. â³ Use ðð³ð®ð«ð ðð and Azure OpenAI to streamline experimentation. â³ Optimize for speed and scalability. ððð©ð¥ð¨ð² ð°ð¢ðð¡ð¨ð®ð ðð¢ð¬ð«ð®ð©ðð¢ð§ð ð°ð¨ð«ð¤ðð¥ð¨ð°ð¬. â³ Host on ðð³ð®ð«ð ðð®ððð«ð§ðððð¬ for reliability. â³ Use ðð³ð®ð«ð ð ð®ð§ððð¢ð¨ð§ð¬ for seamless integration. â³ Monitor deployment with feedback loops. ððð¤ð ðððð ð«ððð«ð¢ðð¯ððð¥ð ðð§ð ðððð¢ð¨ð§ððð¥ð. â³ Index with ðð³ð®ð«ð ðð¨ð ð§ð¢ðð¢ð¯ð Search. â³ Store outputs in ðð¨ð¬ð¦ð¨ð¬ ðð for scalability. â³ Ensure query optimization for real-time use. ðð«ð¢ðð ð ðð ð°ð¢ðð¡ ðð®ð¬ð¢ð§ðð¬ð¬ ð¥ð¨ð ð¢ð. â³ Use ðð³ð®ð«ð ð ð®ð§ððð¢ð¨ð§ð¬ to support decisions. â³ Automate workflows for better efficiency. â³ Integrate insights directly into operations. ðð¨ð¯ðð«ð§ ð°ð¢ðð¡ ð¬ððð®ð«ð¢ðð² ðð§ð ðð ð¢ð¥ð¢ðð² ð¢ð§ ð¦ð¢ð§ð. â³ Use ðð¢ð ð ð¥ð¨ð° for version control. â³ Secure pipelines with ðð¡ððð¤ð¦ðð«ð±. â³ Automate infrastructure with ððð«ð«ððð¨ð«ð¦. Which step will move your business forward today? â»ï¸ Repost to your LinkedIn followers and follow Timothy Goebel for more actionable insights on AI and innovation. #ArtificialIntelligence #AzureCloud #InnovationInTech #AITransformation #MachineLearningPipeline
-
Weâre entering an era where AI isnât just answering questions â itâs starting to take action. From booking meetings to writing reports to managing systems, AI agents are slowly becoming the digital coworkers of tomorrow!!!! But building an AI agent thatâs actually helpful â and scalable â is a whole different challenge. Thatâs why I created this 10-step roadmap for building scalable AI agents (2025 Edition) â to break it down clearly and practically. Hereâs what it covers and why it matters: - Start with the right model Donât just pick the most powerful LLM. Choose one that fits your use case â stable responses, good reasoning, and support for tools and APIs. - Teach the agent how to think Should it act quickly or pause and plan? Should it break tasks into steps? These choices define how reliable your agent will be. - Write clear instructions Just like onboarding a new hire, agents need structured guidance. Define the format, tone, when to use tools, and what to do if something fails. - Give it memory AI models forget â fast. Add memory so your agent remembers what happened in past conversations, knows user preferences, and keeps improving. - Connect it to real tools Want your agent to actually do something? Plug it into tools like CRMs, databases, or email. Otherwise, itâs just chat. - Assign one clear job Vague tasks like âbe helpfulâ lead to messy results. Clear tasks like âsummarize user feedback and suggest improvementsâ lead to real impact. - Use agent teams Sometimes, one agent isnât enough. Use multiple agents with different roles â one gathers info, another interprets it, another delivers output. - Monitor and improve Watch how your agent performs, gather feedback, and tweak as needed. This is how you go from a working demo to something production-ready. - Test and version everything Just like software, agents evolve. Track what works, test different versions, and always have a backup plan. - Deploy and scale smartly From APIs to autoscaling â once your agent works, make sure it can scale without breaking. Why this matters: The AI agent space is moving fast. Companies are using them to improve support, sales, internal workflows, and much more. If you work in tech, data, product, or operations â learning how to build and use agents is quickly becoming a must-have skill. This roadmap is a great place to start or to benchmark your current approach. What step are you on right now?
-
âYour GTM Isnât a ProductâItâs a Platform.â a $21M CEO asked me: âhow did Snowflake grow from zero to $2B+ in revenue in one of the most crowded categories?â my response? âthey didnât just build a product. they built a GTM system that scaled with every stage of growth.â most companies stall after finding early tractionâ ð they scale revenue, but not operations ð they hit product-market fit but donât evolve ð they rely on one channel, one persona, or one hero rep but the best companies donât just grow. they transformâfrom product to platform. and they do it with a go-to-market system. when GTM is a system, it evolves across stages: problem â product â platform so how did Snowflake do it? 1ï¸â£ predictable demand generation â how do we create pipeline at every stage of growth? ð at problem-market fit: â messaging focused on separation of storage & compute â technical founders led early education + sales â first customers were data engineers & architects ð¡ at product-market fit: â launched an enterprise sales engine â paid + partner channels activated â early wins in finance and healthcare verticals ð¢ at platform-market fit: â category creation: âThe Data Cloudâ â multi-cloud strategy + marketplace fueled demand â C-level, IT, and data teams engaged in the same ecosystem ð Snowflake didnât chase channels. they aligned GTM with product maturity. 2ï¸â£ seamless pipeline conversion â how do we turn interest into enterprise deals? â sales process aligned to data transformation roadmap â layered in vertical use cases + security/compliance value â sales + SE + customer success teams worked in pods â weekly forecast + usage reviews to spot and accelerate deals ð every pipeline stage mapped to buyer readiness, not internal quotas. 3ï¸â£ revenue retention & expansion â how do we grow customer value over time? â usage-based pricing â aligned value to cost â net revenue retention (NRR) > 130% â platform expansion: analytics â governance â apps â integrations + marketplace drove stickiness ð they didnât just retain customersâthey expanded into entire ecosystems. final thoughts ð if your GTM strategy doesnât evolve with your productâyouâll stall. ð if you treat GTM as a one-time playâyouâll never become a platform. ð if you donât invest in the system behind the growthâyour wins wonât scale. Snowflake didnât win because of one product. they won because their GTM system evolved at every stage. so iâll ask you: ð is your GTM built to evolveâor are you still selling like itâs day one? letâs discuss ð â love, sangram p.s. follow Sangram Vajre to learn how to scale your GTM from product to platform with GTM O.S. #gotomarket #gtm #growth #b2b #sales #marketing #snowflake #platform #nrr #categorycreation
-
In 2025, deploying GenAI without architecture is like shipping code without CI/CD pipelines. Most companies rush to build AI solutions and create chaos. They deploy bots, copilots, and experiments with no tracking. No controls. No standards. Smart teams build GenAI like infrastructure. They follow a proven four-layer architecture that McKinsey recommends with enterprise clients. Layer 1: Control Portal Track every AI solution from proof of concept to production. Know who owns what. Monitor lifecycle stages. Stop shadow AI before it creates compliance nightmares. Layer 2: Solution Automation Build CI/CD pipelines for AI deployments. Add stage gates for ethics reviews, cost controls, and performance benchmarks. Automate testing before solutions reach users. Layer 3: Shared AI Services Create reusable prompt libraries. Build feedback loops that improve model performance. Maintain LLM audit trails. Deploy hallucination detection that actually works. Layer 4: Governance Framework Skip the policy documents. Build real controls for security, privacy, and cost management. Automate compliance checks. Make governance invisible to developers but bulletproof for auditors. This architecture connects to your existing systems. It works with OpenAI and your internal models. It plugs into Salesforce, Workday and both structured and unstructured data sources. The result? AI that scales without breaking. Solutions that pass compliance reviews. Costs that stay predictable as you grow. Which layer is your biggest gap right now: control, automation, services, or governance?
-
Your GTM strategy can make or break your startup's growth. The challenge? What works changes dramatically between the 3 major growth phases. After seeing hundreds of B2B startups scale (or fail), here's what I think each phase really looks like: PHASE 1: SELL TO LEARN Goal: Find the right offering and audience to focus on. (aka product/market fit) â Audience: Test multiple segments and use cases â Offering: Experiment with different features, pricing, and packaging â Messaging: Test different broad and specific use-case messaging â Channels: High-touch, low-cost (direct outreach, referrals, networking) â Operations: Stay lean, embrace the chaos of learning The truth? Phase 1 feels scattered. That's normal. You're gathering real data about what works. But eventually, you are ready to move to phase 2. PHASE 2: WIN YOUR NICHE Goal: Build a repeatable system to acquire customers in your chosen niche. (1 offering + 1 audience) â Audience: Focus entirely on one segment where you're winning â Offering: Perfect your solution for the core job to be done â Messaging: Hyper-focused on your chosen segment's challenges â Channels: Build a repeatable acquisition system with a few proven channels â Operations: Keep the team laser-focused on mastering one niche Success here means saying no to everything else until you've dominated this audience segment. PHASE 3: EXPAND TO GROW Goal: Build a new repeatable system to acquire a new audience or sell a new offering. â Audience: Add adjacent segments (horizontal growth) â Offering: OR add complementary solutions (vertical growth) â Messaging: Distinct positioning for each market/offering â Channels: Build new acquisition systems for each expansion opp â Operations: Dedicated GTM focus for each audience/offering The key? Choose either horizontal OR vertical expansion. Not both. And only add one new audience or offering at a time. --- The question to ask isn't "What's the perfect GTM strategy?" The real question is, "What's the right GTM strategy for my phase?" Look at your company today: â Are you still searching for what works? â Have you found it but haven't mastered it? â Or are you ready to expand what's already working? Your answer tells you exactly where to focus next. Save this framework for your next strategy session. --- P.S. This framework was inspired by a post by Robert Kaminski ð¯ on startup GTM phases a few months ago. His insights helped me refine my thinking on this. #startupgrowth #gtmstrategy
-
Too many teams accept data chaos as normal. But weâve seen companies like Autodesk, Nasdaq, Porto, and North take a different path - eliminating silos, reducing wasted effort, and unlocking real business value. Hereâs the playbook theyâve used to break down silos and build a scalable data strategy: 1ï¸â£ Empower domain teams - but with a strong foundation. A central data group ensures governance while teams take ownership of their data. 2ï¸â£ Create a clear governance structure. When ownership, documentation, and accountability are defined, teams stop duplicating work. 3ï¸â£ Standardize data practices. Naming conventions, documentation, and validation eliminate confusion and prevent teams from second-guessing reports. 4ï¸â£ Build a unified discovery layer. A single âGoogle for your dataâ ensures teams can find, understand, and use the right datasets instantly. 5ï¸â£ Automate governance. Policies arenât just guidelines - theyâre enforced in real-time, reducing manual effort and ensuring compliance at scale. 6ï¸â£ Integrate tools and workflows. When governance, discovery, and collaboration work together, data flows instead of getting stuck in silos. Weâve seen this shift transform how teams work with data - eliminating friction, increasing trust, and making data truly operational. So if your team still spends more time searching for data than analyzing it, whatâs stopping you from changing that?