Rajat Singhal adlı kullanıcının tam profilini görüntülemek için oturum açın
veya
LinkedIn‘de yeni misiniz? Hemen katılın
Devam Et’i tıklayarak veya oturum açarak LinkedIn Kullanıcı Anlaşması’nı, Gizlilik Politikası’nı ve Çerez Politikası’nı kabul edersiniz.
Rajat Singhal adlı kullanıcının tam profilini görüntülemek için oturum açın
veya
LinkedIn‘de yeni misiniz? Hemen katılın
Devam Et’i tıklayarak veya oturum açarak LinkedIn Kullanıcı Anlaşması’nı, Gizlilik Politikası’nı ve Çerez Politikası’nı kabul edersiniz.
Hindistan
Rajat Singhal adlı kullanıcının tam profili görüntülemek için oturum açın
Rajat sizi Tower Research Capital şirketindeki 10 üzerinde kişiyle tanıştırabilir
veya
LinkedIn‘de yeni misiniz? Hemen katılın
Devam Et’i tıklayarak veya oturum açarak LinkedIn Kullanıcı Anlaşması’nı, Gizlilik Politikası’nı ve Çerez Politikası’nı kabul edersiniz.
1 B takipçi
500+ bağlantı
Rajat Singhal adlı kullanıcının tam profili görüntülemek için oturum açın
veya
LinkedIn‘de yeni misiniz? Hemen katılın
Devam Et’i tıklayarak veya oturum açarak LinkedIn Kullanıcı Anlaşması’nı, Gizlilik Politikası’nı ve Çerez Politikası’nı kabul edersiniz.
Rajat Singhal ile ortal bağlantıları görüntüle
Rajat sizi Tower Research Capital şirketindeki 10 üzerinde kişiyle tanıştırabilir
veya
LinkedIn‘de yeni misiniz? Hemen katılın
Devam Et’i tıklayarak veya oturum açarak LinkedIn Kullanıcı Anlaşması’nı, Gizlilik Politikası’nı ve Çerez Politikası’nı kabul edersiniz.
Rajat Singhal ile ortal bağlantıları görüntüle
veya
LinkedIn‘de yeni misiniz? Hemen katılın
Devam Et’i tıklayarak veya oturum açarak LinkedIn Kullanıcı Anlaşması’nı, Gizlilik Politikası’nı ve Çerez Politikası’nı kabul edersiniz.
Rajat Singhal adlı kullanıcının tam profilini görüntülemek için oturum açın
veya
LinkedIn‘de yeni misiniz? Hemen katılın
Devam Et’i tıklayarak veya oturum açarak LinkedIn Kullanıcı Anlaşması’nı, Gizlilik Politikası’nı ve Çerez Politikası’nı kabul edersiniz.
Hakkında
Tekrar hoş geldiniz
Devam Et’i tıklayarak veya oturum açarak LinkedIn Kullanıcı Anlaşması’nı, Gizlilik Politikası’nı ve Çerez Politikası’nı kabul edersiniz.
LinkedIn‘de yeni misiniz? Hemen katılın
Faaliyet
1 B takipçi
-
Rajat Singhal bunu paylaştıRajat Singhal bunu paylaştıEncouraging message from the best entrepreneurial minds in the world. Huge opportunity for startups to get mentored by well-renowned venture capitalists. See you there at Elevate, Indian Institute of Technology, Madras Participate today: https://lnkd.in/f33---R Deadline for Submission is 20th March #ConnectAscendConquer #Esummit2k18 Kris Gopalakrishnan Desh Deshpande H E A Eddy Campbell Venky Ganesan Prashanth Prakash
-
Rajat Singhal bunu paylaştıRajat Singhal bunu paylaştıAnother flagship event of E-Summit 2018 is here. 72 hours long experience wherein you'll get a taste of entrepreneurship and chance to become an architect of revolutionary products of tomorrow. Register here: http://bit.ly/2EXgu5l #Esummit2k18 #InspireNetworkInnovateUplift #BusinessOfInnovation
-
Rajat Singhal bunu paylaştıRajat Singhal bunu paylaştı"Leadership is about making others better as a result of your presence." Become the leader by being the campus director of ElevateX, preliminary round of pan-India level Pitching Competition Elevate, with a guaranteed total funding opportunity of INR 10 lakhs. Be the one to elevate the network and startup ecosystem of your institute. #BeTheLeader #ConquerAscendConquer #ESummit2018 For more details and registration, visit http://bit.ly/2C51AvW.
-
Rajat Singhal buna tepki verdiRajat Singhal buna tepki verdiHello everyone, I’m excited to share Taskflow-AGENTS, a project I worked on during my final spring semester in Prof.Tsung-Wei Huang's lab. Taskflow-AGENTS is an agent-ready knowledge base for Taskflow (a task-parallel programming system for building and running dependency graphs in modern C++), built on the open AGENTS.md standard. Writing an AGENTS.md file is easy; writing one that captures performance-first engineering judgment is the hard part. This project focuses on making AI coding agents understand not just the Taskflow API, but the decisions behind using it well. It brings together the pieces developers and AI agents need to move faster with Taskflow: practical porting recipes, performance-first design guidance, large-DAG engineering patterns, and real field notes from task-bench and OpenTimer experiments. One part I’m especially excited about is DigestObserver, a human- and LLM-friendly Taskflow profiler that turns massive raw traces into compact, structured diagnostics. Instead of just giving raw data, it helps identify performance patterns and gives insight into what to investigate next. The goal is to make Taskflow adoption faster and more reliable. What often takes months of manual porting, profiling, and trial-and-error can now be guided down to days with existing AI agents and profiler-driven feedback. GitHub: https://lnkd.in/gU_xkfmB I’m grateful to Prof. Tsung-Wei Huang for the mentorship, guidance, and opportunity to work on this. This project helped me bring together parallel programming, performance engineering, and AI-assisted software development in an impactful way. If you are building task-parallel applications and want to accelerate them with Taskflow, this repo is designed to help your AI agent understand the right patterns from the start. #Taskflow #ParallelProgramming #HighPerformanceComputing #Cplusplus #AI #LLM #AGENTSmd #PerformanceEngineering #SoftwareEngineering #Research
-
Rajat Singhal buna tepki verdiRajat Singhal buna tepki verdiAlex Gerko, founder of multi billion dollar quant firm XTX, released a list outlining "what each quant firm really does." Worth a skim if you enjoy market microstructure humor or just want a clearer mental model of who does what in quant land If you want to break into firms like these, we're hosting a free quant recruiting info session this week: https://lnkd.in/e3_nsCpx Disclaimer: WallStreetQuants is not directly affiliated with or endorsed by any of the companies or firms mentioned in this post.
-
Rajat Singhal bunu beğendiRajat Singhal bunu beğendiDuring my PM prep days, the hardest part wasn’t frameworks. It was finding someone good enough to practise with. You’d ask around, try to match schedules, and eventually get on a call where both of you were still guessing what a strong answer even looked like. Feedback stayed surface-level with no real pushback. No clear sense of whether you were improving or just getting comfortable. So I built PM Prep AI. An AI mock interviewer that actually behaves like a real panel: • 🎙️ Voice interviewer that challenges you and asks follow-ups • 🏢 Company-specific questions (Zomato, Swiggy, PhonePe, Razorpay, Meesho, Google Maps, Amazon India, etc.) • 📊 Structured scoring across 6 dimensions + clear hiring verdict • 📈 Track your progress across sessions • 🔍 Precise, actionable feedback from your actual answers 𝐓𝐫𝐲 𝐢𝐭 : www.pmprepai.online #Careers #JobPrep #Upskill #Learning #ProductManagement #PMInterviews #MBAPlacements #PMPrep
-
Rajat Singhal buna tepki verdiRajat Singhal buna tepki verdiRevealing the Mission Patch Every mission carries a story. This represents ours. Introducing the official mission patch for Mission Drishti by GalaxEye - a symbol of the vision, technology, and intent behind the world’s first #OptoSAR imaging satellite. Mission Drishti is our first step toward a brand new way of seeing the Earth. Where clear satellite imagery is reliable, ubiquitous and always actionable. A future where clarity is as accessible as connectivity. Our mission patch reflects that philosophy. Complexity resolving into clarity. And a sharper understanding of the Earth steadily coming into focus. #MissionDrishti #MissionPatch #OptoSAR #EarthObservation #SpaceTech #InnovationFromIndia
-
Rajat Singhal bunu beğendiRajat Singhal bunu beğendiClaude Visuals launched sometime ago and my feed has been losing its mind. And rightfully so. The feature is genuinely incredible. Interactive charts, diagrams, explorable breakdowns, all built from scratch inside your chat. Anthropic absolutely cooked with this one. But I have to share something kind of wild. A few weeks ago Humaira Firdowse Mohammed and I were head down at the Gemini Hackathon building something with a very similar soul. Except we went one dimension further. Meet Thinky3D. You drop in any topic (or upload a PDF) and it generates a full course for you. Slides, voice narration, adaptive quizzes, an AI chatbot, and here's the fun part: Live 3D simulations you can actually play with. - Black holes that bend spacetime as you drag sliders. - DNA helices that unwind in real time. - Pathfinding algorithms racing each other across 3D mazes. - Möbius strips you can twist until your topology intuition finally kicks in. My honest takeaway after building this thing for weeks: 2D is amazing for explaining something. 3D is what makes you actually feel it. Some concepts just click differently when you can rotate them, pull them apart, and watch them respond to your inputs. Here's the bigger thing I keep thinking about. Education still teaches spatial, dynamic concepts using static, flat content. Projectile motion as an arrow on paper. Molecules as 2D diagrams. We've done this for decades because building custom interactive simulations for every concept was simply too expensive. That constraint just quietly died. Every simulation in Thinky3D is fresh code generated on the fly by Gemini 3 Pro. Not a template. The AI writes a new React Three Fiber component for whatever you ask about, wires up the sliders, and hands you a tiny universe to poke at. The flat textbooks are no longer enough, especially when you can grok a concept much easily (and cheaply) with generated simulations. Thinky3D is fully open source and free to use. Devpost writeup: https://lnkd.in/gN8ifrGu GitHub: https://lnkd.in/gEKVPe67 Star us on Github and give some love to the Devpost project if you feel it deserves it 😊 I couldn't resist putting together a side by side comparison. Same topics, same concepts, Claude Visuals on the left, Thinky3D on the right. Full video here: https://lnkd.in/gjRs3h9Y #GeminiHackathon #AI #OpenSource #EdTech #BuildInPublicClaude Visuals vs Thinky3D - Building the future of AI Interactive Education | Gemini HackathonClaude Visuals vs Thinky3D - Building the future of AI Interactive Education | Gemini Hackathon
-
Rajat Singhal buna tepki verdiRajat Singhal buna tepki verdiExcited to start my journey at AMD as an Embedded/Firmware Security Engineer. Looking forward to learning and building impactful systems. Grateful for the support along the way!
-
Rajat Singhal bunu beğendiIn case you missed it, this week we launched our full suite of ERCOT power contracts. There's lots more in store, so stay tuned!Rajat Singhal bunu beğendiNEWS: ElectronX has launched its first product suite of U.S.-regulated power contracts, designed for trading and managing intraday price volatility in the ERCOT market. Hourly bounded futures and binary options for ERCOT are now available on ElectronX’s direct-access platform—the first derivatives tailored for the specific risk transfer and price discovery profiles of distributed energy resources, large load consumers, and energy sector innovators in the rapidly evolving U.S. electricity ecosystem. Sam Tegel, CEO of ElectronX, said, ”By expanding participation and deepening liquidity in power hedging overall, ElectronX can help usher in an era of improved efficiency, reliability and cost optimization for all U.S. electricity markets.” Read the press release, including comments from exchange members Base Power Company, Xcel Energy, and Habitat Energy: https://lnkd.in/gTQWv_bX
Deneyim ve Eğitim
-
Tower Research Capital
*** ******* ******** *********
-
*******
******** *********** ******
-
*********
******* ********* * ****** ****** ** ****
-
****** ********* ** *********** ******
******** ** ********** ******** ******* 8.22/10
-
-
********* ******
**** ****** undefined
-
Rajat Singhal adlı kişinin tam deneyimin görüntüleyin
Unvan, işte kalma süresi ve daha fazlasını görün.
Tekrar hoş geldiniz
Devam Et’i tıklayarak veya oturum açarak LinkedIn Kullanıcı Anlaşması’nı, Gizlilik Politikası’nı ve Çerez Politikası’nı kabul edersiniz.
LinkedIn‘de yeni misiniz? Hemen katılın
veya
Devam Et’i tıklayarak veya oturum açarak LinkedIn Kullanıcı Anlaşması’nı, Gizlilik Politikası’nı ve Çerez Politikası’nı kabul edersiniz.
Gönüllü Deneyimler
Onurlar ve Ödüller
-
Top 20 Innovative Projects - University Challenge
India Innovation Growth Programme (IIGP 2.0)
-
National Winners - Microsoft CodeFunDo++ 2019
Microsoft India
National Winners of Microsoft CodeFunDo ++ Challenge 2018, with a cash prize of Rs. 500,000 plus $5000 Azure Cloud Credits & AI for Earth Grant for the UAVs for Disaster Management project
-
Kishore Vaigyanik Protsahan Yojna (KVPY) Scholar
IISc Bangalore, Department of Science and Technology (DST), India
Diller
-
English
Profesyonel çalışma yetkinliği
-
Hindi
Profesyonel çalışma yetkinliği
-
Gujarati
Başlangıç düzeyinde yetkinlik
Rajat Singhal adlı üyenin tam profilini görüntüleyin
-
Ortak tanıdıklarınızı görün
-
Başka biri aracılığıyla tanış
-
Rajat Singhal ile doğrudan iletişime geçin
Diğer benzer profiller
-
Himanshu Singh
Himanshu Singh
Indian Institute of Technology (Banaras Hindu University), Varanasi
2 B takipçiVaranasi
Diğer gönderileri keşfedin
-
Indrajeet Yadav
ALLEN CAREER INSTITUTE • 3 B takipçi
Dario says models will do full SWE end-to-end in 1-2 years. That means Claude Code, Codex, Cursor — all transitional tools. The real product isn't the code editor. It's the intent translator. We're building scaffolding for a building that will build itself. https://lnkd.in/dyqRmPQm
-
Hitesh Nayak
NA • 8 B takipçi
Agentic Engineering: The Optimization Series (1/50) "How do you design an agentic system that handles 10k RPS without blowing your token budget or hitting 5s P99 latency?" This is the "Senior Architect" interview question for any LLM-native role. If your answer is "use a smaller model," you’ve already lost. The real answer lies in Tiered Intent Routing and Deterministic Sifting. [ I will keep posting set of intriguing topics faced by me ] The Problem: The Monolithic Prompt Trap Most developers cram 20 tool definitions and a 1k-token system prompt into an agent. Every single user "Hello" costs you 1,500 tokens and 2 seconds of LLM "reasoning" just to figure out the user is being polite. The Solution: The "Sifter & Router" Architecture To keep MS (Latency) low and Tokens lean, you must move the decision-making logic out of the LLM and into a high-performance Data Science layer. 1. The L1 Gate (Classic NLP/SLM) Instead of an LLM, use a Bi-Encoder (like all-MiniLM-L6-v2) or a fine-tuned DistilBERT classifier. Mechanism: Map user input to a vector space and compare it against "Gold Standard" intent embeddings using Cosine Similarity. Result: Simple intents (e.g., "Order Status," "Refund Policy") are routed to a Deterministic Script or a Small Language Model (SLM) like Phi-3. Performance: Latency drops from ~2000ms to <40ms. 2. The Dynamic Context Injector For the remaining "High-Reasoning" queries, don't pass all 20 tools. Use the L1 Gate's output to inject only the relevant tool schemas into the prompt. Constraint: Prompt_Tokens Proportional to Relevant_Tools Impact: You reduce the "distraction" for the LLM, increasing tool-calling accuracy while slashing the Input Token bill by 70%. 3. The "Stateful" Fallback If the L1 confidence is $< 0.7$, only then do you engage the "Frontier Model" (GPT-4o/Claude 3.5) to perform an "Ambiguity Resolution" step. Practical Systems Impact By implementing a Router-Worker pattern: Cost: You decouple cost from volume. 80% of traffic hits the $0.00/1k$ token local classifier. Throughput: Your system handles massive spikes because the "Heavy" LLM is only touched for 20% of requests. Reliability: You eliminate "Tool Hallucination" because the LLM only sees the 2 tools it actually needs. Resources [Got this from my fellow students] To master this architecture, I recommend studying these core concepts: RouteLLM (UC Berkeley): An open-source framework specifically for learning how to route queries between "strong" and "weak" models to optimize cost/performance. Semantic Router (Aurelio AI): A great library for implementing high-speed schema routing using embeddings rather than LLM reasoning. Classification vs. Generation: Read up on Natural Language Understanding (NLU) fundamentals specifically "Intent Classification and Slot Filling" to see how classic DS solves 80% of agentic "logic" faster than a Generative model. #AgenticAI #SystemsDesign #LLMOps #DataScience #RetailFoundry #ai #genai
28
3 Yorum -
Rishit Patel
Southern East Inc. • 2 B takipçi
🔥 Real Madrid spent $80M on Ronaldo, and Meta spent $100M to land Jiahui Yu from OpenAI. But the real game-changer? Trapit Bansal, an IIT Kanpur grad who went from Accenture analyst to IISc researcher to UMass Amherst PhD. He’s worked at Google, Facebook, Microsoft, and OpenAI, helped build the o1 reasoning model, and now joins Meta’s Superintelligence Labs with a rumored $100M payday. Meta didn’t stop there. They’ve poached a wave of core OpenAI talent: - Shengjia Zhao (ChatGPT and GPT-4 co-creator) - Jiahui Yu (o3, o4 mini, GPT-4.1) = Shuchao Bi and Hongyu Ren (GPT-4o creators) Meta is building its own $100M dream team. This isn’t just a hiring spree. It’s a strategic knockout move to drain OpenAI of its key people and leap ahead in AGI. 🧠 What this means: Meta is rewriting the roster with superstar engineers OpenAI risks serious talent loss even as it focuses on retention The AI battlefield now includes billion-dollar sign-on bonuses Bottom line: From an Accenture cubicle to Meta’s Superintelligence Labs, this high-stakes transfer saga could reshape the future of AI. Who’s on your fantasy team?
22
2 Yorum -
James Webb
CoinDesk • 995 takipçi
🎩 The Case for LLM Gallantry Posts have gone viral recently citing that manners create "noisy context" and that your "please" is costing millions in compute. I respectfully disagree. 1. Prime for 'Expert Clusters'—Sophisticated reasoning usually inhabits the same domain as polite, professional, discourse. 🔷 Bluntness or rudeness can direct a model toward “internet argument” weighting, encouraging it to cite equally toned sources. Gallantry primes the model to navigate toward ‘Expert Clusters’ such as academic journals and expert forums. 2. Reduce Entropy—Bluntness does not equal directness. Abrasive or chaotic syntax introduces irrelevant emotional tokens that compete for compute attention. 🔷 Clear gallant framing acts as a high-fidelity signal. It minimises entropy, ensuring the model's $K, V$ matrices are focused on your logic problem, and not your emotional tone. 3. Use the Reward Model—LLMs are trained using RLHF, harsh prompting risks triggering safety guardrails or defensive hedging. 🔷 Collaborative, chivalrous tone aligns with model's reward model, creating a fast-lane to complex reasoning capabilities. 4. Tonal Drift & The Preservation of Excellence—The period (.) is 1500 years old, but modern texting culture has caused its meaning to morph into a symbol of passive-aggression. 🔷 Cultural preservation of respectful communication, is not a burden, but a necessary exercise of selflessness, lest your uncouth LLM-speak drift into your real-world communication. Manners maketh man. #AI #PromptEngineering #GenerativeAI #TechTips #LLM #Etiquette I used em dashes before they were cool.
4
-
Shivank Goel
Amazon Web Services (AWS) • 5 B takipçi
Was reading a bit on serving LLMs and space is rapidly evolving. Found vLLM (Liang et al., 2023) quite interesting. It introduced a mechanism called PagedAttention, which virtualizes the KV (key-value) cache by managing it like a paged memory system. In transformer models, the KV cache stores intermediate attention values for previously seen tokens—essential for generating the next token without recomputing everything from scratch. Instead of keeping this entire cache in GPU memory, which grows linearly with prompt and response length, PagedAttention evicts and loads cache blocks on demand. This enables long-context inference without exhausting GPU memory and is a foundational step toward memory-aware serving. Google’s Pathways architecture takes a related route, by conditionally activating only the parts of a model relevant to a given input—known as expert routing. This sparsely activates subsets of model parameters (called "experts") rather than loading the full model, saving compute and memory. Although these architectures focus on model modularity rather than serving infra modularity. Microsoft’s DeepSpeed Inference and MII optimize serving by reducing the resource load of each inference phase and employ techniques such as KV cache sharing, where multiple concurrent requests can reuse the same cache entries instead of duplicating memory; and activation offloading, which temporarily moves intermediate computations (activations) from GPU memory to CPU or host memory during inference to avoid exceeding memory limits. These optimizations allow each GPU to serve more concurrent decode requests without degrading throughput or accuracy. NVIDIA’s Dynamo is currently one of the most complete implementations of fully disaggregated inference. It explicitly splits prefill and decode phases into different GPU pools. A Smart Router component directs decode requests to the GPUs where the relevant KV cache already resides, minimizing cache transfers or recomputation. A separate KV Cache Manager coordinates cache eviction and loading across a hierarchy of memory layers—including GPU VRAM, CPU RAM, SSD, and even networked object stores—similar to how virtual memory systems operate. Meta’s LlamaServe, built on Triton Inference Server, is still in early stages but is beginning to explore smarter scheduling and batching of requests. Across all of these systems, a common trend is taking shape. Inference is no longer treated as a uniform, atomic task. If you're working on inference infrastructure, I’d love to hear more on this from you! #inference
8
-
David Eberle
Typewise • 13 B takipçi
Getting into YC is easy. Only ~2% of applicants get in. Indian Railways? Try 0.3%. In the last drive, ~30 million people applied for 90,000 jobs. Odds can be 1,800 to 1. Six years of study is not unusual. Here’s the key difference: A YC application takes days. The building takes years. Years in which you’re creating something of value, whether YC accepts you or not. Exam prep takes years of cramming obscure facts. If you miss the cut‑off, you might emerge a decade later with no job and no work experience. Invest your best years in building something that compounds: a product, a company, and a skill set no one can take from you.
16
7 Yorum -
Shubham Vedi
Build Fast with AI • 3 B takipçi
An Indian startup just open-sourced a 105B model that outperforms DeepSeek R1 on math and agentic benchmarks. Trained entirely in India. On Indian government GPUs. With $53M in funding. Not $5B. And the weights are live on Hugging face right now: https://lnkd.in/djPhwvEB Here's why Sarvam 105B deserves your attention: The benchmarks that matter: → 88.3 on AIME 2025 (DeepSeek R1: 87.5) → 96.7 on AIME with tools → 49.5 on BrowseComp (DeepSeek R1: 3.2 — not a typo) → 68.3 on Tau2 agentic benchmark (beats o4-mini, DeepSeek R1, and Claude Sonnet 4) → 98.6 on Math500 These aren't cherry-picked comparisons against small models. They're head-to-head with DeepSeek R1, Gemini 2.5 Flash, o4-mini, and Claude Sonnet 4. The backstory makes this even more impressive. Last year, Sarvam caught heavy criticism for fine-tuning Mistral's model and calling it sovereign AI. Downloads were underwhelming. People questioned whether India could build foundational models at all. So they went back and trained from scratch. 12 trillion tokens for the 105B. 16 trillion for the 30B. 128 experts in a Mixture-of-Experts architecture. Custom tokenizer, custom kernels, custom RL pipeline. They threw out KL-divergence regularization entirely — broke from the textbook approach and it worked. The 30B model is equally wild — only 2.4B active parameters. It matches models 10x its active size on coding and reasoning benchmarks. This is what efficient scaling actually looks like. What makes this a signal, not just a model drop: ✅ Trained on IndiaAI Mission compute — sovereign infrastructure, not rented from US hyperscalers ✅ Supports all 22 official Indian languages natively ✅ Already in production — Samvaad (conversational agents) and Indus (reasoning/agentic workflows) ✅ Open-sourced under Apache License on Hugging face ✅ Demoed running on a feature phone with physical buttons That last one hit differently. A 105B-parameter model. On a dumbphone. The AI race isn't just US vs China anymore. A team of ~114 people in Bengaluru just shipped a globally competitive reasoning model on government GPUs, at a fraction of frontier lab costs. And they gave the weights away for free. Download it. Fine-tune it. Ship something with it. Hugging face link: https://lnkd.in/djPhwvEB What's your take — can sovereign AI models compete long-term with frontier labs? 👇 🔄 Repost if your network needs to see what India is building. #AI #OpenSource #SarvamAI #indi
189
4 Yorum -
Tanmay Goel
Qualcomm • 131 B takipçi
Most SDEs assume they’ve already done enough to level up. → They’ve worked on production systems. Solved real engineering problems. Operated in fast-moving teams. On paper, that should qualify them for the next step. Sounds sufficient, right? NOT REALLY. Because SDE-2/3 interviews don’t evaluate your past work the way you expect. They evaluate how you think, live, under constraints. And that changes everything. DSA still shows up first. And for many, that’s where the process ends, before System Design is even discussed. What actually makes the difference 👇 🔸 Stop treating DSA like a numbers game: Volume doesn’t translate to readiness. Pattern recognition does. Focus on how problems are structured, not how many you’ve solved. 🔸 Don’t postpone system design: Waiting until later is the usual mistake. Strong candidates build problem-solving and design intuition together. 🔸 Practice in interview conditions: Mocks expose hesitation, gaps, and communication issues that solo prep hides. The transition from SDE-1 to SDE-2/3 roles isn’t about writing cleaner code alone. It’s about demonstrating structured thinking, articulating decisions clearly, and breaking down unfamiliar problems quickly. That’s what companies are really screening for. So, when you’re managing prep alongside a full-time job, random effort won’t make any difference, you need direction. And that’s where Bosscoder’s Transformer program comes in, built to help you focus on what actually moves the needle, without wasting time on the wrong things. 🔗 Explore Transformer here: bcalinks.com/COZj4LK They’ve helped 2200+ engineers crack top PBC roles through: 💡 Structured curriculum to help you master DSA, System Design, and AI 💡 Real world, AI-powered projects & 24/7 doubt-solving 💡 Leadership skills for Senior roles with 1:1 mentorship from industry experts 💡 Resume reviews, mock interviews, and 100% job switch support. #dsa #sdeprep #pbc #tech #collab
428
54 Yorum -
Alan Kochukalam George
Myovine • 590 takipçi
⚙️ One library, real DSP speed. There’s a gap in the developer ecosystem. Teams doing DSP/ML often migrate to Python, because that’s where most tooling lives. But Python wasn’t built for high-throughput I/O — under load, it spawns new interpreters/processes, creating duplicated memory, slow cold starts, and more containers. Compute stays fast, but orchestration cost explodes. Meanwhile, Node.js/TS scale I/O well, but were never meant for serious compute. Pure-JS DSP hits performance walls. Even worse, serializing data between Redis ↔ DSP/ML adds latency and forces batching — slowing real-time pipelines. So devs are stuck: Python → strong math, weak concurrency Node → strong networking, weak compute dspx closes that gap — TypeScript DX + native C++/SIMD (AVX2/SSE3/SSE2/NEON) under the hood. 🧩 How dspx works FIR / FFT / Conv1D run in optimized C++/SIMD Memory-safe circular buffers → O(1) throughput TS handles Kafka / Redis / WebSockets Redis persistence < 0.5 ms Batched logging avoids I/O stalls At tiny batch sizes, N-API overhead can make naive JS look competitive. At realistic scale, batched native pipelines win. ⚡ Benchmarks Dell OptiPlex 3000 Micro (i5-12600T · AVX2 · Node 22) FFT: 2.6× faster than fft.js, ~9× faster than tfjs-node FIR (51-tap): ~4× faster than fili / naive JS Conv1D (128-kernel): ~3.5× faster than tfjs-node Moving Avg (O(1)): 559× more throughput/sec than naive JS Redis save/load: sub-ms Logging: <3% overhead 📊 Full tables + charts in carousel 💬 Observations SIMD FMA + contiguous memory → big FIR/Conv wins O(1) moving-average design → massive throughput gain Sub-ms Redis + low I/O overhead → real-time persistence in pure Node 🧠 Open Source dspx is Apache 2.0, free for commercial/academic use. Looking for contributors interested in: ARM NEON / Apple M / Graviton tuning Audio / sensor / biomedical DSP Visualization + benchmark tooling 📦 npm i dspx 🔗 https://lnkd.in/e-tWxAgu 🧠 Real-time DSP for Node, TypeScript & Redis 💼 Note I’m exploring opportunities in full-stack, real-time systems, performance engineering, and DSP. If your team works in this space, I’d love to connect. My next post will cover why sub-ms Redis latency isn’t just technical — it’s economic. Lower serialization + compute overhead → lower infra + energy cost. 🔖 Tags & Mentions #NodeJS #TypeScript #DSP #PerformanceEngineering #EdgeComputing #OpenSource #SIMD #Redis #Cplusplus #RealTime NodeJS Developer TensorFlow Google Microsoft JavaScript Developer Amazon Web Services (AWS) Vercel
4
1 Yorum -
Dhairya Purohit
Ekyon • 3 B takipçi
We don't need that much focus on building LLMs, There are far more optimal things happening instead of just diving into a money-burning race. People don't realize it, but India has the largest surplus of software engineering talent. For the first time, the risk of starting something and the resources required to start have gone down so much so that people can finally build. In recent days, I have seen significantly more Indians, yes, us, building something new than anybody else in the world. It's happening.
114
7 Yorum -
Umesh Malik
Expedia Group • 2 B takipçi
RAG vs Fine-Tuning for LLMs (2026): What Actually Works in Production RAG vs fine-tuning in 2026 explained with real tradeoffs, latest trends, and a practical decision framework for production LLM systems. What you'll find inside: - RAG - Fine-Tuning - LLM Engineering Curious how you'd approach this. Drop your take in the comments. Read the full post: https://lnkd.in/gKBrQs2M #RAG #FineTuning #LLMEngineering #AIArchitecture #GenAI2026
16
1 Yorum -
Yashraj Yadav
CreatorJot • 905 takipçi
🤗 Good news for builders in India. Supabase is now fully accessible across the country. The earlier network connectivity issue has been resolved and IP-level restrictions have been removed. All core components are operational: - Compute capacity - ap-south-1 - Auth - Edge Functions If you were affected, try clearing your DNS cache or restarting your network. ISP-side propagation may take up to 24 hours. This incident is also a reminder - infrastructure dependency risk is real. Always think about redundancy, monitoring, and fallback plans. Back to shipping.
5
1 Yorum -
Akash Mahajan
Contextual AI • 819 takipçi
Context engineering → better tools for agents (not just better retrieval/RAG). Traditional retrieval works well on pointed questions over chunks/snippets. But struggles with holistic cross-document questions, forcing you to stuff entire docs into context. We need llms.txt for documents 🤖🗺️. Here's a demo inspired by talks and conversations at @aiDotEngineer: A "document navigator" AI agent (with Cursor + MCP) browsing a 250-page US Govt. document to answer: "summarize all parts of the document about US government debt". (skim to 1:10 and 2:10 for just the demo) Under the hood: - Contextual AI /parse API + auto-generated document_metadata.hierarchy (as llms.txt) - Cursor's agent loop + navigation tools via MCP - Verifiable attribution via interpretable tool call traces with rationales This simple demo uses purely “navigation” tools on one 250 page doc. But combining with solid traditional retrieval can scale context for agents to 10-100x more than can fit in context - while keeping all the agentic goodness intact 🚀. The bigger picture: LLMs can read, summarize, and index various cross-document metadata. The pattern emerging is: Index-time compute → smarter tools → more capable agents. Chat with the US Govt FY 24 Financial report in Cursor yourself: https://lnkd.in/gpvqUbqP
66
4 Yorum -
Garima Shrivastava
Lixta Network • 2 B takipçi
The most ironic part of all this is, the subject they deem only theoretical often define our real world. AI functioning on matrices and theorems? That's not something which you can build or realise. It's a concept, a method, a thought process. And that very thought process now impacts the whole world. Pure Science, Mechanical Engineering, Aeronautics, Food Technology, Philosophy, Psychology, all these supposed niche fields require huge brain power, intellectual capability and creativity. The applied fields and concepts are nothing without their theoretical part. We must remember that.
6
-
Abhishek Kumar Mahto
NOI Software Pvt. Ltd. • 1 B takipçi
🚨 If developer infrastructure can be blocked overnight… what does that mean for builders? Many users across India are currently unable to access Supabase. No notice. No migration window. Apps suddenly break. When infrastructure platforms get blocked at ISP/DNS level: → Student projects break → Startups lose customers → Production systems get affected → Trust in the ecosystem shakes This isn’t about one company. It’s about stability, transparency, and protecting the developer ecosystem. For now, some possible workarounds 👇 • Switch DNS to Cloudflare (1.1.1.1) or Google (8.8.8.8) • Use a VPN • Enable DNS proxying via a custom domain India wants to lead in tech. Let’s protect the builders powering it. #Supabase #IndianDevelopers #StartupIndia #BuildInIndia #DeveloperCommunity #TechPolicy
8
-
Andrew Ng
DeepLearning.AI • 2 Mn takipçi
There is significant unmet demand for developers who understand AI. At the same time, because most universities have not yet adapted their curricula to the new reality of programming jobs being much more productive with AI tools, there is also an uptick in unemployment of recent CS graduates. When I interview AI engineers — people skilled at building AI applications — I look for people who can: - Use AI assistance to rapidly engineer software systems - Use AI building blocks like prompting, RAG, evals, agentic workflows, and machine learning to build applications - Prototype and iterate rapidly Someone with these skills can get a massively greater amount done than someone who writes code the way we did in 2022, before the advent of Generative AI. I talk to large businesses every week that would love to hire hundreds or more people with these skills, as well as startups that have great ideas but not enough engineers to build them. As more businesses adopt AI, I expect this talent shortage only to grow! At the same time, recent CS graduates face an increased unemployment rate, though the underemployment rate — of graduates doing work that doesn’t require a degree — is still lower than for most other majors. This is why we hear simultaneously anecdotes of unemployed CS graduates and also of rising salaries for in-demand AI engineers. When programming evolved from punchcards to keyboard and terminal, employers continued to hire punchcard programmers for a while. But eventually, all developers had to switch to the new way of coding. AI engineering is similarly creating a huge wave of change. There is a stereotype of “AI Native” fresh college graduates who outperform experienced developers. There is some truth to this. Multiple times, I have hired, for full-stack software engineering, a new grad who really knows AI over an experienced developer who still works 2022-style. But the best developers I know aren’t recent graduates (no offense to the fresh grads!). They are experienced developers who have been on top of changes in AI. The most productive programmers today deeply understand computers, how to architect software, and how to make complex tradeoffs — and who additionally are familiar with cutting-edge AI tools. Sure, some skills from 2022 are becoming obsolete. For example, a lot of coding syntax that we had to memorize back then is no longer important, since we no longer need to code by hand as much. But even if, say, 30% of CS knowledge is obsolete, the remaining 70% — complemented with modern AI knowledge — is what makes really productive developers. Without understanding how computers work, you can’t just “vibe code” your way to greatness. Fundamentals are still important, and for those who additionally understand AI, job opportunities are numerous! [Original text: https://lnkd.in/gHSWcKhr ]
10.863
641 Yorum -
Sanyam Sareen
Sareen Career Coaching • 25 B takipçi
Here’s exactly how I would crack a $150K+ SWE job at Microsoft in 6 months. (A real strategy my client used to land interviews at Microsoft, Meta, and Stripe) Too many engineers prepare hard to crack MAANG. But they lack a solid strategy. Here’s the exact roadmap I'll follow if I wanted to land a SWE job at Microsoft. Step 1: Resume + Role Clarity → Reverse-engineer the JD Study at least 10+ Microsoft SWE job listings. Highlight recurring keywords, must-haves, and preferred tools. → Rewrite your resume like a product pitch Show measurable impact: Good: “Improved load time by 43%” Bad: “Worked on performance optimization.” → Make it keyword-optimized for ATS without sounding robotic. Step 2: Master DSA the Microsoft Way → Focus on patterns, not problems Microsoft LOVES: Sliding window Trees (esp. DFS/BFS) Graphs Dynamic programming → Suggested Platforms: Leetcode (Microsoft tag), Neetcode(dot)io roadmap, Grokking series on Educative → 2 problems/day + 1 mock interview/week = compounding prep Step 3: System Design (yes, even for junior roles) → Start with High-Level Design (HLD): Learn how to design APIs, caching, rate limiting, and DB scaling. → Build a project where you actually implement what you learn. Step 4: Microsoft-Specific Behavioral Prep → Microsoft uses structured behavioral interviews Focus on “3As” framework: Action → Approach → Aftermath → Use real projects to show: Collaboration Adaptability Customer focus Engineering rigor → Prep 8–10 STAR stories mapped to their values Step 5: Mock + Real-World Practice → Do 4–5 peer or mentor-led mock interviews (especially for behavioral + design) → Record yourself. Watch for filler words, unstructured answers, lack of metrics. → Apply for 5 roles/week—Microsoft + similar-sized companies (to get into interview flow) Step 6: Apply Strategically + Use Referrals → Connect with 3–5 engineers/recruiters/week on LinkedIn. Don’t ask for a referral right away - engage first. → Reach out using tailored messages like: “Hi [Name], I’ve been preparing for an SWE role at Microsoft. Your journey from [X] to [Microsoft] really stood out. If you're open, I’d love to learn more about your experience.” → Submit applications using referrals whenever possible + 1-click apply where relevant The client who followed this playbook now works at Microsoft Azure and had 3 competing offers before accepting. It’s not about doing everything. It’s about doing the right things in the right order. Give yourself 6 months and follow this roadmap to make it a reality. Share this with someone who dreams of working at Microsoft. P.S. Follow me if you are a tech job seeker in the U.S. I share practical advice that gets you hired. — Additional resources: https://lnkd.in/eKxQmYtP. https://lnkd.in/gwRWpXR9. https://lnkd.in/g94_Cziv
100
4 Yorum -
Vatsal Sanghvi
1811 Labs • 17 B takipçi
There’s something broken in how we train engineers in India Especially in Computer Science And it shows Most fresh CS grads today are unemployable - not because they’re not smart, but because they’ve been poorly trained and terribly advised At 1811 Labs, we have worked with, interviewed and reviewed assignments of many young engineers. Here’s what I’ve seen up close: - They jump from tech to tech: React today, Go tomorrow, maybe a bit of AI on the side. No depth, just surface-level noise - Fundamentals? Weak. Code quality? Worse. Best practices? Never heard of them - AI: Blindly use AI without a clear understanding of the problem/use-case making things even worse - Communication? Most can’t explain what they built, let alone collaborate on a team You might argue, “An engineer just needs to build” But how do you build anything meaningful if you can’t explain your thoughts, write clean code, or understand a problem deeply? The saddest part? They still expect a job just because they have a degree And honestly, I don’t blame them That’s what the system promised But today, skills > degrees The market doesn’t care where you studied It cares what you can do So if you’re a junior dev, here’s the roadmap I wish more people followed: - Build rock-solid fundamentals - Pick one stack. Get decent at it before you jump around - Write clean, readable code. Follow best practices - Do real projects, not just for resumes, but to understand what “good” actually looks like - And please, don’t sleep on communication. It’s your second most important skill after coding Trust me, without solid understanding and skill, AI Agents will eat away at most low and mid-level dev jobs. We are slowly transitioning into an era where engineers will engineer and architect systems supervise progress and fix issues, while AI will do the heavy lifting. A tier-1 degree might fast-track access But a tier-1 skillset is what’ll fast-track growth We don’t need more engineers We need better ones
82
12 Yorum -
Sri Harsha Mudumba
PROCAL TECHNOLOGIES • 996 takipçi
⚡ ML inference continues to fascinate me, and I've been diving deep into LLM inference optimization techniques - something I genuinely enjoy working on! Recently, I've been exploring the intricate challenges of serving large language models efficiently, and it led me to build TorchWeave LLM - an open-source inference compiler that achieves 2-5x throughput improvements through some compelling optimization strategies. What makes ML inference so interesting is how it sits at the intersection of systems engineering and machine learning theory. Every optimization decision has cascading effects on memory, compute, and user experience. 🔄 Continuous Batching: One technique I found particularly rewarding to implement - merging concurrent requests into shared decode steps to maximize GPU utilization. The scheduling algorithms here are genuinely fascinating. ⚡ Per-Request KV-Cache Management: Exploring how attention masks can enable efficient memory management for variable sequence lengths opened up entirely new optimization possibilities. 📡 Real-time SSE Streaming: Building Server-Sent Events with time-to-first-token metrics taught me so much about balancing performance with user experience. The exploration paid off: 98% throughput improvement with 4 concurrent requests Average TTFT of 0.860s Production-ready architecture that scales What I love about working on inference optimization is that every bottleneck you solve reveals three new interesting problems. The rabbit holes of memory management, scheduling algorithms, and distributed systems design never get old. The technical implementation spans everything from PyTorch model interfaces to FastAPI async patterns - each component was a learning opportunity that deepened my appreciation for the complexity of production ML systems. 🔗 Check it out on GitHub: https://lnkd.in/gCZdZ5Pm I'm curious - what aspects of ML inference optimization do you find most compelling? Always eager to learn from others tackling similar challenges! #MachineLearning #LLM #MLInference #PyTorch #Optimization #OpenSource #MLOps #AI #SoftwareEngineering #TechnicalExploration Built with Python 3.12, FastAPI, PyTorch, and genuine curiosity about making models faster ⚡
49
6 Yorum