Category: Technology

  • The Real Reason Startups Are Firing Engineers and Hiring PMs (Or Vice Versa)

    If you’ve been paying attention to tech job postings lately, you’ve noticed a strange pattern. Some startups are quietly trimming their engineering teams — not the dramatic headlines of 30,000 cuts at Oracle, but slow, deliberate reductions. And at the same time, they’re hiring aggressively in product management, developer relations, and customer success.

    The obvious explanation is “AI will replace engineers.” It makes for a good tweet. But the reality is more interesting and more nuanced.

    The Cost-to-Value Equation Has Flipped

    Two years ago, a startup’s competitive advantage was its engineering velocity. If you could ship faster, iterate quicker, and build a more polished product than your competitors, you won. So startups hired engineers — lots of them. Every additional engineer meant more features, more experiments, more shipped code.

    AI has compressed that advantage. What used to take a team of three engineers a week now takes one engineer an afternoon with a capable AI coding assistant. The marginal value of each additional engineer has dropped, dramatically.

    But here’s the thing nobody talks about: building the product was always the easy part. Finding product-market fit, understanding what customers actually want, pricing it right, communicating it effectively, keeping customers happy — those things haven’t gotten any easier. If anything, AI has made them more important, because now everyone can build.

    The Real Bottleneck Moved

    In 2023, the bottleneck was engineering capacity. In 2026, it’s strategic clarity.

    A startup can now build a functioning MVP in a weekend. Three founders with AI assistants, no dedicated engineering team, and a clear vision can ship something that would’ve required six months and a $2M seed round two years ago. The barrier to building has collapsed.

    But the barrier to knowing what to build? That’s still incredibly hard.

    This is where the shift in hiring comes from. Startups are realizing that their scarcest resource isn’t coding capacity anymore — it’s product insight. They need people who can:

    • Talk to customers and translate messy, contradictory feedback into clear feature priorities
    • Define a positioning strategy that cuts through the noise of a thousand AI-wrapped competitors
    • Write PRDs that actually constrain AI behavior instead of vague wishlists
    • Design go-to-market motions that don’t rely on “build it and they will come”

    That’s a product manager’s job. It always has been. It just got way more valuable relative to everything else.

    But Here’s the Twist: It Goes Both Ways

    Not every startup is the same, and the reverse trend is equally real: engineering-heavy startups are finding they don’t need traditional PMs anymore.

    Why? Because a good engineer with an AI assistant can now do most of what a PM used to do. Draft a PRD? AI can help. Analyze user feedback? AI can summarize thousands of reviews in seconds. Create user personas? AI can do it from your existing customer data. Write a competitive analysis? Ten minutes with an LLM and a clear prompt.

    The PM role is getting squeezed from both sides. On one end, AI-augmented engineers are absorbing the tactical PM work (writing specs, prioritizing backlogs, analyzing data). On the other end, PMs who learn to use AI are becoming so efficient at their core work that fewer of them are needed.

    The surviving PMs are the ones who’ve moved up the value chain — from writing tickets to shaping strategy, from backlog management to market positioning, from feature spec to business model.

    What This Means for You

    If you’re an engineer: your coding skills are table stakes now. The engineers who thrive in 2026 are the ones who combine technical depth with product instinct. You need to be able to talk to users, understand market dynamics, and make judgment calls about what to build — not just how to build it.

    If you’re a PM: stop being a ticket factory. If your job is just writing user stories and grooming backlogs, you are one AI prompt away from obsolescence. Move toward strategy, toward user research, toward the parts of the job that require actual human judgment about what the market wants and why.

    The startups that will win in this environment are the ones that figure out the right ratio. Too many engineers without product direction means you’re building efficiently in the wrong direction. Too many PMs without building capacity means you’re strategizing with nothing to ship.

    The sweet spot is a small, sharp team of T-shaped people — engineers who understand their customers, and PMs who understand the technical tradeoffs — all operating at maximum leverage with AI doing the heavy lifting on execution.

    The org chart is flattening. The roles are blurring. And the people who’ll thrive are the ones who stop thinking about what their title is and start thinking about what the product needs.

    What do you think? Has your team’s ratio shifted, or are you seeing the opposite trend? I’m genuinely curious what the data looks like on the ground.

  • AI Agent Weekend Chronicles: 5.73 Million Tokens, Zero Grass Touched

    What do you do on a long weekend? Some people touch grass. I decided to dive headfirst into the glorious chaos of AI agents. Naturally.

    First things first: I spun up an Ubuntu VM. Why? Because I’ve been around the internet long enough to know that letting an autonomous AI agent loose on my personal machine is like giving a toddler a loaded smartphone. The VM had internet access, zero personal data, and enough leeway to make mistakes I wouldn’t have to explain to anyone. Safety first.

    Agent #1: #PaperclipAI. I hooked it up to my #OpenAI Codex subscription, created a company, hired a virtual content development team, and let them loose. Before I knew it, they were cranking out posts and articles of surprisingly decent quality. I even got the agent to publish directly to my self-hosted WordPress site. At this point, I was basically a media mogul who hadn’t left the couch.

    Next up: #OpenClaw, the crowd favourite. Installed it, pointed it at qwen/qwen3.6-plus:free on #OpenRouter, and asked it to blog about Oracle layoffs and shiny new AI models. It did a solid job. Grammarly’s AI detector gave it a clean 0% AI-generated bill of health. Take that, detectors. Either the AI is getting scary good at sounding human, or the detector is just vibing.

    Then came #Hermes. And wow, what a tool. It practically deserves its own podcast. This thing can run the entire show solo. I fed it my resume PDF for a review. It said, “There’s potential here,” which is polite code for “this needs work.” Then it handed me a questionnaire like a career counsellor at a crossroads. I filled it out, fed it back, and Hermes promptly realized it didn’t have PDF creation tools. No panic. It made a .md file instead, told me to install the missing tools, and ten minutes later I had a freshly polished resume. Ten minutes. My last resume update took a procrastination cycle measured in seasons.

    The plot twist: all of this is glorious, but these agents are absolute token guzzlers. They eat through tokens like I eat through snacks on a movie night. If you’re billing your corporate AmEx, sure, party on. If you’re like me and riding the free-model wave, you’re essentially paying with your data. The age-old bargain: convenience for surveillance.

    Oh, and I almost forgot #Claude Code. I paired it with stepfun/step-3.5-flash:free on OpenRouter and asked it to build a WebUI so I could chat with it from a browser. Two hours. 5.73 million tokens. Endless questions and approvals later… I got a codebase that doesn’t work. Five point seven three million tokens. I could’ve written War and Peace in fewer tokens. Or at least a working to-do app.

    All in all, the long weekend was a blast. I built companies, reviewed resumes, published blogs, and burned through tokens like a dragon with a credit card. Would I do it again? Absolutely. Would I do it on my main machine? …Let’s not get crazy.

  • Beyond the Hype: A Technical Deep Dive into Qwen 3.6’s ‘1M Context’

    In the race for AI supremacy, “context window” has become the new battleground. With Qwen 3.6-Plus boasting a massive 1 million token context window, Alibaba is claiming it can process entire codebases or technical manuals in a single pass. But what does that actually mean, and how do they keep the model from “forgetting” the first page by the time it reaches the last?

    The “Lost in the Middle” Problem

    For a long time, Large Language Models (LLMs) suffered from a phenomenon researchers call “Lost in the Middle.” If you fed a model 100 pages of text, it would remember the beginning and the end but would struggle to recall specific details buried in the 50th page. This was a fundamental limitation of how “attention mechanisms”—the core of a transformer model—process data.

    Qwen 3.6-Plus addresses this through architectural advancements in RoPE (Rotary Positional Embeddings) and specialized attention span optimizations. Essentially, the model has been trained to maintain a “sharp focus” regardless of where the information sits in a massive document.

    How It Handles the Load: KV Caching

    Processing 1 million tokens isn’t just about memory; it’s about speed. If the model had to re-read everything every time it generated a new word, it would be incredibly slow. Qwen 3.6 uses a technique called KV Caching (Key-Value Caching).

    Think of it like a student taking notes during a lecture. Instead of re-reading their entire textbook for every new question, they keep a “cache” of the most important information (the keys and values) ready for immediate access. This allows Qwen to scale to huge contexts without a massive drop in inference speed.

    Why This Changes Everything for Developers

    For developers, a 1M context window means you can stop “chunking” your code. You no longer have to write complex scripts to break your repository into small pieces and hope the AI picks the right ones. You can simply feed the entire project structure to Qwen 3.6 and say, “Refactor this,” and it will understand the dependencies across different files.

    While the hype around “1M tokens” can feel like a marketing number, the engineering required to make it actually useful is a massive leap forward. It’s not just about how much the model can read; it’s about how well it understands what it has read.

    Have you tested Qwen 3.6 with large codebases yet? Did you notice a difference in its ability to connect distant parts of your project? Share your experiences below.

  • Qwen 3.6-Plus: A New Era for AI Agents

    Following the successful launch of the Qwen 3.5 series earlier this year, Alibaba has just dropped its latest powerhouse: Qwen 3.6-Plus. If you’ve been following the AI space, you know that each incremental update brings something new, but this one feels like a genuine leap forward—especially if you’re into building AI agents or doing complex coding tasks.

    What’s New in Qwen 3.6-Plus?

    Available right now via the Alibaba Cloud Model Studio API, Qwen 3.6-Plus isn’t just a minor tweak. It’s designed to be the engine behind “real-world agents.” Here are the big-ticket items that have the community buzzing:

    • Agentic Coding on Steroids: Whether you’re fixing a frontend bug or tackling a massive, repository-level architectural change, Qwen 3.6-Plus has been tuned to handle it with impressive accuracy. It’s built to “vibe code” alongside you, handling terminal operations and automated tasks like a seasoned engineer.
    • 1 Million Token Context Window: Yes, you read that right. By default, the model can process a massive amount of information at once. This is a game-changer for developers who need to feed entire codebases or massive technical manuals into the AI without losing the thread.
    • Sharper Multimodal Reasoning: It doesn’t just “see” images or charts; it understands them with much higher accuracy. This makes it incredibly reliable for tasks that involve interpreting complex diagrams or scientific data.

    Why It Matters for Developers

    The biggest hurdle with previous AI models was often their ability to stay on track during long, multi-step tasks. Qwen 3.6-Plus addresses this by deeply integrating reasoning, memory, and execution. In benchmarks like SWE-bench and Terminal-Bench 2.0, it’s matching or even surpassing industry leaders.

    For the average developer, this means less time babysitting the AI and more time seeing it actually do the work. It’s a move toward “highly autonomous super-agents” that can handle cross-domain planning and complex code management without constant human hand-holding.

    The “Vibe Coding” Experience

    Alibaba explicitly mentions that this release is designed to deliver a transformative “vibe coding” experience. It’s about making the interaction with AI feel more natural, stable, and reliable. By addressing feedback from the Qwen 3.5-Plus deployment, they’ve smoothed out the rough edges, making it a solid foundation for the next generation of AI-powered apps.

    Final Thoughts

    With Qwen 3.6-Plus, Alibaba is making a clear statement: the future of AI isn’t just about chatbots; it’s about agents that can actively participate in the development process. If you’re a developer looking to speed up your workflow or just curious about the cutting edge of open-weight models, Qwen 3.6-Plus is definitely worth a spin.

    Have you tried it out yet? Let me know how it handles your latest coding challenges!