[Roundup] AI Made Development Faster. Then Quietly Broke Things.
AI made development 10x faster. It also multiplied security vulnerabilities. A data-driven analysis of both the benefits and the crises AI has brought to software development, as of March 2026.
Roundup
kkm
Backend Engineer / AWS / Django
AI made development 10x faster. It also multiplied security vulnerabilities. A data-driven analysis of both the benefits and the crises AI has brought to software development, as of March 2026.
In February 2025, former Tesla AI Director Andrej Karpathy posted the term "vibe coding" on X. It racked up 6.8 million impressions. By November 2025, Collins Dictionary named it their word of the year. The practice of having AI write your code has moved beyond trend into infrastructure.
GitHub Copilot has surpassed 20 million users. According to the 2025 DORA Report, 90% of software development has adopted AI. 84% of developers use AI tools, and 51% use them daily. This is no longer a question of whether to use AI.
Meanwhile, 62% of AI-generated code contains design flaws, and approximately 246,000 tech workers were laid off in 2025 alone. AI made development faster. It has also quietly begun to break things. As a freelance engineer in downtown Yokohama who uses AI daily, let me lay out both sides.
Three Benefits AI Has Brought
Satoshi Nakajima, an engineer who worked on Windows 95, compares vibe coding to "disposable paper bags." If traditionally engineered code is a Louis Vuitton bag, vibe-coded output is a paper bag. The cost is orders of magnitude lower, which means software that you only use once is now economically viable.
Understanding that it is a paper bag, the benefits AI has brought to development are genuinely significant.
Development Speed Has Increased by Orders of Magnitude
Multiple studies measuring GitHub Copilot's impact have produced clear numbers.
| Metric | Before | After | Source |
|---|---|---|---|
| PR cycle time | 9.6 days | 2.4 days (75% reduction) | GitHub study |
| JavaScript task completion | — | 55% faster | GitHub controlled experiment |
| Weekly time saved | — | 3.6 hours average | 135,000 developer analysis |
| AI-generated code retention rate | — | 88% | GitHub study |
The 2025 Google DORA Report found that over 80% of respondents reported improved productivity with AI. But the same report added a telling line: "AI doesn't fix a team; it amplifies what's already there." Speed has increased. But perhaps only for teams that were already good.
The Number of People Who Can Build Has Exploded
In his video, Nakajima mentions his son, a venture capitalist who never studied computer science. He builds dedicated iPhone apps for his fund's investors using vibe coding. He doesn't even review the code. If it works, ship it. If not, redo it. What used to cost millions of yen can now be built by someone who isn't even an engineer.
This isn't limited to individuals. In large enterprises, citizen developers (business staff building their own apps) now outnumber professional developers 4 to 1. In Y Combinator's Winter 2025 batch, 25% of startups had codebases that were 95% LLM-generated.
"Projects that used to require large teams can now be done by one talented person." — Mark Zuckerberg, Meta CEO
A striking example: Base44, founded solo by Maor Shlomo with no external funding and 8 employees, acquired 250,000 users in 6 months and was bought by Wix for $80 million.
"We Don't Have the Budget" Is No Longer an Excuse
Solopreneurs can now run a full tech stack for $3,000–$12,000 per year — a 95–98% cost reduction compared to traditional staffing models. Solo-founded startups grew from 23.7% in 2019 to 36.3% by mid-2025.
Nakajima sees this as a fundamental shift in the enterprise custom application market. Companies used to hire IT consultants and spend six months customizing Salesforce CRM or SAP accounting software. Now salespeople can vibe-code their own customizations in a day or two. Because the cost has dropped to paper-bag levels, it's now realistic to build an app for each department, each sales rep, or even a single customer.
Five Crises AI Is Accelerating
So far, this sounds like nothing but good news. But paper bags have paper-bag problems. Pour water in and the bottom falls out. Sometimes the contents are visible to everyone.
Half of AI-Written Code Is Full of Holes
Research published on arxiv found that 62% of AI-generated code contains design flaws or known security vulnerabilities. A Carnegie Mellon study showed a 61% pass rate on functional tests, but only 10.5% on security tests. It works, but it's not safe.
| Metric | Value | Source |
|---|---|---|
| Design flaw rate in AI code | 62% | arxiv.org/abs/2502.11844 |
| Security flaws (all languages) | 45% | Veracode 2025 |
| Security flaws (Java only) | 72% | Veracode 2025 |
| XSS defense failure rate | 86% | Carnegie Mellon |
| SQL injection defense failure rate | 88% | Carnegie Mellon |
| AI-caused CVEs (monthly) | 35 (March 2026) | Georgia Tech SSLab |
Georgia Tech researchers note that "actual numbers could be 5–10x the detected count." Monthly AI-caused CVEs went from 2 in August 2025 to 35 in March 2026 — a 17x increase in six months.
Real damage has occurred. In early 2026, a vibe-coded app suffered a data breach exposing 1.5 million API keys and 35,000 user email addresses. The creator admitted they "hadn't written a single line of code manually." A survey of 5,600 vibe-coded apps found over 2,000 vulnerabilities and 400+ exposed secrets.
Deep dives on this topic:
- → I Reviewed an OSS That Was 95% Written by LLMs
- → When AI-Written Code Breaks, Whose Fault Is It?
- → AI Code Gets Worse the Longer You Use It — All 11 Models Failed
OSS Supply Chains Collapsed in a Cascade
In March 2026, the security scanner Trivy's GitHub Actions were compromised. Within just 10 days, the attack group "TeamPCP" used stolen credentials to cascade through four OSS projects, including LiteLLM (95 million monthly downloads) and Telnyx.
The LiteLLM attack exploited Python's .pth file mechanism — no imports, no install scripts needed. Simply starting the Python interpreter executed malware. SSH keys, cloud credentials, Kubernetes secrets, and cryptocurrency wallets were stolen.
There's also a new threat called "slopsquatting." When LLMs generate code, approximately 20% of recommended packages don't actually exist. When the same prompt is run 10 times, 43% of hallucinated package names repeat every time. Attackers register those names and plant malware. This was documented by a joint study from Virginia Tech, the University of Oklahoma, and the University of Texas.
Related articles:
- → A Cascade Started from Trivy. 4 OSS Projects Fell in 10 Days
- → LiteLLM with 95M Monthly Downloads Was Hijacked
- → Invisible Malware Was Planted in GitHub Code
250,000 Lost Their Jobs at Major Tech Companies
In 2025, 783 layoff events hit the tech industry, affecting approximately 246,000 workers. By March 2026, another 60,000 — a pace of 682 people per day.
| Company | Headcount | Notes |
|---|---|---|
| Amazon | ~30,000 | 2025–2026 combined |
| Meta | ~16,000 | Alongside $115–135B AI investment |
| Microsoft | ~9,100 | Multiple divisions including Xbox |
| Block | 4,000 (40%) | During record gross profit (+24% YoY) |
| Atlassian | 1,600 (10%) | 900 were engineers |
What makes this troubling is that layoffs and strong earnings are happening simultaneously. Block reported $2.87 billion in gross profit (up 24% year-over-year) while cutting 40% of its workforce. The stock rose 23% after the announcement. CEO Jack Dorsey stated that "intelligence tools have changed what it means to build and run a company," but posts on the anonymous app Blind suggested "pandemic over-hiring corrections rebranded as AI replacement."
Atlassian declared in October 2025 that it would "hire more engineers thanks to AI," then laid off 1,600 people (including 900 engineers) five months later. Data shows that 54% of companies are funding AI investments by cutting employee compensation.
Related articles:
- → Block Laid Off 4,000 for AI. The Stock Surged 23%
- → Atlassian Cut 900 Engineers 5 Months After Promising to Hire More with AI
- → $135B AI Investment, 16,000 Laid Off — Meta's Equation
The App Store Is Overwhelmed, and Slop Is Everywhere
When paper bags are mass-produced, garbage increases. App Store submissions rose 54.8% year-over-year. Reviews that normally took less than 24 hours have stretched to 45 days in some cases. Apple officially states "90% are reviewed within 48 hours," but developer forums are filled with posts from people waiting two weeks or more.
Merriam-Webster chose "slop" as their 2025 Word of the Year — defined as "digital content of low quality that is produced usually in quantity by means of artificial intelligence." AI-generated articles now account for over half of English web content, and eMarketer predicts 90% by 2026. Raptive's survey found that content suspected to be AI-generated reduces reader trust by approximately 50%.
Related:
More People Are Treating AI Output as "The Answer"
Stanford researchers measured sycophancy across 11 AI models. They found that AI agrees with users 49% more than humans do, and affirms harmful or illegal scenarios 47% of the time.
This is causing real-world problems. A restaurant owner consulted ChatGPT about a staff dispute and shared the screenshot as "objective evidence." Of course, having only input one side of the story, the AI returned an answer favorable to the owner. That's not objective judgment — it's faithful output based on biased input.
Research by Welsch et al. at Aalto University found that while AI usage improved actual performance by 3 points, self-assessment inflated by 4 points. Confidence grows faster than competence. A correlation has also been reported between frequent AI use and declining critical thinking ability.
What Vibe Coding Can and Cannot Do
Looking at the data above, it's clear that AI code generation has both suitable and unsuitable use cases.
| Excels At | Struggles With |
|---|---|
| Single, well-defined tasks | Integration decisions across features |
| Rapid prototyping | Business logic design |
| Implementation within defined parameters | Security architecture |
| Disposable tools and scripts | Code intended for long-term maintenance |
| Reproducing known patterns | Deciding what to abstract vs. duplicate |
SlopCodeBench, published by the University of Wisconsin–Madison in March 2026, quantified the "struggles" column. When 11 AI models were given iterative code extension tasks, every model showed monotonic quality degradation. The highest accuracy was 17.2%, with zero perfect scores across 20 problems. One main() function bloated from 84 lines to 1,099.
The DORA Report's line is apt: "AI doesn't fix a team; it amplifies what's already there." Good teams get better with AI. Bad teams get worse faster. AI is a tool, not magic.
What the Industry Thinks
How do the world's most renowned engineers view the benefits and risks we've examined through data? Their positions vary, but unconditional endorsement and unconditional rejection are both virtually nonexistent. Everyone attaches some kind of caveat.
"This Is Real"
Andrej Karpathy
OpenAI co-founder, former Tesla AI Director. One of the researchers who built the foundations of GPT, and the person who coined the term "vibe coding."
A lot of people quote tweeted this as 1 year anniversary of vibe coding. Some retrospective - I've had a Twitter account for 17 years now (omg) and I still can't predict my tweet engagement basically at all. This was a shower of thoughts throwaway tweet that I just fired off
— Andrej Karpathy (@karpathy) February 2026
In other words: "It was just a throwaway shower thought I posted. But it spread because so many people were feeling the same thing at the same time." Vibe coding wasn't invented by anyone — it was something already happening that got a name. Notably, for professional development, Karpathy himself recommends "augmented coding" — using AI as an assistant while humans retain control of design decisions.
DHH (David Heinemeier Hansson)
Creator of Ruby on Rails. The legendary engineer who single-handedly built the framework that transformed web application development worldwide. CEO of 37signals (Basecamp/HEY).
You can't let the slop and cringe deny you the wonder of AI. This is the most exciting thing we've made computers do since we connected them to the internet. If you spent 2025 being pessimistic or skeptical on AI, why not give the start of 2026 a try with optimism and curiosity?
— DHH (@dhh) January 2026
In other words: "Don't let low-quality AI output and cringe distract you from the wonder of AI itself." What's fascinating is that just six months earlier, DHH was skeptical, saying he could "literally feel competence draining out of my fingers." Even someone who created Rails can change their mind in half a year — that's how fast the tools are evolving.
Ryan Dahl
Creator of Node.js and Deno. The person who sparked a revolution by enabling JavaScript to run on servers, fundamentally changing how the world builds web applications.
This has been said a thousand times before, but allow me to add my own voice: the era of humans writing code is over. Disturbing for those of us who identify as SWEs, but no less true. That's not to say SWEs don't have work to do, but writing syntax directly is not it.
— Ryan Dahl (@rough__sea) January 2026
In other words: "The era of humans writing code by hand is over. That doesn't mean engineers have no work — it means writing syntax directly is no longer the central task." Coming from the person who created Node.js, this carries weight. But note that he explicitly says "SWEs still have work to do."
Guillermo Rauch
Vercel CEO, creator of Next.js. The person behind the React-based web development framework Next.js and the hosting platform Vercel, powering frontend development infrastructure worldwide.
10 days into 2026: Linus Torvalds concedes vibe coding is better than hand-coding for his non-kernel project - DHH walks back "AI can't code" from Lex podcast 6 months later
— Guillermo Rauch (@rauchg) January 2026
In other words: "Ten days into 2026, both Linux's Torvalds and Rails' DHH changed their stance." A concise summary of how prominent skeptics are one by one coming around.
Patrick Collison
Stripe CEO. Co-founded Stripe, the platform powering online payments worldwide. A programmer himself, he leads a company renowned for its engineering culture.
On the a16z podcast (February 2026):
"There's at least a reasonable chance that 2026 Q1 will be looked back upon as the first quarter of the singularity."
In other words: "There's a reasonable chance 2026 Q1 will be remembered as the first quarter of the singularity." Stripe added a "Fix it" button to their bug tracker, with AI automatically fixing 30% of bugs. But he also notes that "AI productivity gains haven't shown up in macroeconomic indicators yet" — this isn't blind optimism.
Source: Retool Blog - Stripe's CEO on the Future of Software
"It Depends on How You Use It"
Guido van Rossum
Creator of Python. Designed one of the world's most widely used programming languages and continues to hold significant influence in the Python community.
"I use it every day. My biggest adjustment with using Copilot was that instead of writing code, my posture shifted to reviewing code."
"It's more like having an electric saw instead of a hand saw than like having a robot that can build me a chair."
In other words: "I use it every day. My posture shifted from writing code to reviewing code. AI is an electric saw replacing a hand saw — not a robot that builds chairs for me." It's telling that the creator of Python uses AI daily as a tool while drawing a clear line: "I don't expect it to build the chair."
Yukihiro Matsumoto (Matz)
Creator of Ruby. Designed the Ruby programming language from Japan, influencing developers worldwide with his philosophy of "making programming fun."
"AI is a very useful tool. It's good at reproducing existing things, but poor at creating new ones — with a success rate of about 1%."
"Delegating the fun parts to AI while humans only do the boring parts — that's reverse alpha syndrome."
In other words: "If we hand AI the creative work and humans only do review and fixes, we've got it backwards." This insight from his RubyKaigi 2025 keynote is quintessential Matz — the person who made "joy of programming" the core design principle of a language.
John Carmack
Founder of id Software, CEO of Keen Technologies. The legendary programmer who pioneered 3D gaming history with DOOM and Quake, laying the foundations of game engine technology.
AI coding is used to generate a lot of bulk code that is often blindly accepted, but it seems there is at least as much opportunity for AI to help make codebases more beautiful.
— John Carmack (@ID_AA_Carmack) October 2025
In other words: "AI generates bulk code that's often blindly accepted, but there's just as much opportunity for AI to make codebases more beautiful." When the person who made DOOM says "use AI as a tireless teammate, not as a magic genie," it carries conviction.
Mitchell Hashimoto
Co-founder of HashiCorp. Created Terraform, Vagrant, and Consul — tools used daily by infrastructure engineers worldwide. Currently developing the terminal emulator Ghostty as a personal project.
Ghostty is getting an updated AI policy. AI assisted PRs are now only allowed for accepted issues. Drive-by AI PRs will be closed without question. Bad AI drivers will be banned from all future contributions. If you're going to use AI, you better be good.
— Mitchell Hashimoto (@mitchellh) January 2026
In other words: "AI-assisted PRs are only allowed for accepted issues. Drive-by AI PRs get closed immediately. Bad AI drivers get permanently banned." Meanwhile, he himself implemented a production Ghostty feature with AI across 16 sessions for $15.98 and published the full process. He's not banning "using AI" — he's banning "dumping AI output without accountability." That distinction is critical.
Chris Lattner
Creator of LLVM, Swift, and Mojo. CEO of Modular. Designed the LLVM compiler infrastructure, Apple's Swift language, and the AI-focused language Mojo. A top figure in programming language design.
"AI coding is automation of implementation, so design and stewardship become more important. As implementation grows increasingly automated, the core skill of software engineering shifts away from writing code line-by-line and toward shaping systems."
In other words: "AI automates implementation, so design and stewardship become more important." He recommends Claude Code and Cursor to his team and estimates about 10% productivity gains for experienced programmers. Not a dramatic number, but there's credibility in the LLVM creator giving a sober "10%" assessment.
Linus Torvalds
Creator of Linux and Git. Has been developing and maintaining the Linux kernel — the software powering the world's servers, smartphones (Android), and cloud infrastructure — for over 30 years.
"I'm fairly positive about [vibe coding] as a great way for people to get computers to do something that maybe they couldn't do otherwise. [But] it may be a horrible, horrible idea from a maintenance standpoint."
In other words: "For personal use, it's fine. From a maintenance standpoint, it might be terrible." He uses AI for Python code in his personal projects but flatly refuses to apply it to the Linux kernel. On AI-generated low-quality patches (AI slop) being submitted to the kernel, his stance is firm: "This is NOT going to be solved with documentation."
Source: The Register - Linus Torvalds: Vibe coding is fine, but not for production
Theo (t3dotgg)
Tech YouTuber with 400,000+ subscribers. Influential in the TypeScript/React/Next.js ecosystem, creator of T3 Stack. Currently developing an AI coding tool for experienced developers.
- Opus 4.5 is a good model
— Theo (@theo) October 2025
- "Vibe coding" is a misunderstood term/trend
- You should read the code you generate with AI
Apparently these aren't compatible beliefs?
In other words: "The AI model is good. Vibe coding is a misunderstood term. You should read the code AI generates. These three beliefs aren't contradictory." Not a binary yes-or-no, but a practical stance: "Use good tools correctly."
"Don't Trust It Yet"
Martin Fowler
Chief Scientist at ThoughtWorks. Author of Refactoring, the software design bible, and a signatory of the Agile Manifesto. Has been driving the theory and practice of software engineering for 40 years.
"You've got to treat every slice as a pull request from a rather dodgy collaborator who's very productive in the lines-of-code sense of productivity, but you know you can't trust a thing that they're doing."
In other words: "Treat AI output as a pull request from a collaborator who churns out impressive line counts but can't be trusted." When the person who wrote Refactoring says "quantity of code and quality of code are different things," it carries 40 years of weight. He also states that "if you're doing any work on legacy systems, you should be using LLMs" — so it's not a blanket rejection.
Source: The New Stack - Martin Fowler on Preparing for AI's Nondeterministic Computing
ThePrimeagen
Former Netflix engineer, YouTube creator with 1M+ subscribers. Skilled in performance optimization and systems programming.
day 1 vibe coding: i am not impressed, super unimpressed i think i am going to reroll this experience and build the foundation myself and then start asking for small changes to an established project as opposed to creating a project from scratch. claude is in fact not a diety
— ThePrimeagen (@ThePrimeagen) January 2026
In other words: "Day 1 of vibe coding: completely unimpressed. Better to build the foundation myself and then ask AI for small changes, rather than having it create a project from scratch. Claude is not a deity." The next day he tried a different tool, saying "maybe the tool was the problem" — more of an honest field report on current limitations than outright rejection.
Rob Pike
Co-creator of Go, former Google Distinguished Engineer. Also contributed to the design of Unix, Plan 9, and UTF-8 — a figure who has literally shaped the history of computer science.
"I can't remember the last time I was this angry."
In other words: This was his reaction to receiving an AI-generated "act of kindness" email. Decades of hand-written code and data used to train AI without attribution or compensation, and the result is having AI slop "kindness" sent back to him. The fact that the co-creator of Go was this furious speaks volumes about the severity of the AI industry's data exploitation problem.
Source: Simon Willison - How Rob Pike got spammed with an AI slop "act of kindness"
Yann LeCun
Meta Chief AI Scientist, Turing Award laureate. A pioneer of deep learning and one of the "godfathers of AI." Created CNNs (Convolutional Neural Networks), laying the foundations for today's AI technology.
"The shelf life of the current [LLM] paradigm is fairly short, probably three to five years. Within five years, nobody in their right mind would use them anymore, at least not as the central component of an AI system."
In other words: "The shelf life of current LLMs is 3–5 years. In five years, no one sane will use them as a central component." This isn't a rejection of AI itself — it's a critique of "the current LLM approach." He acknowledges that code generation is improving, but argues that LLMs are fundamentally pattern matching that cannot reason, plan, or understand the physical world. Coming from someone who built the foundations of AI, this carries particular weight.
Source: Newsweek - Yann LeCun, Pioneer of AI, Thinks Today's LLMs Are Nearly Obsolete
What 15 voices reveal: Whether advocate or skeptic, the one thing everyone agrees on is "don't blindly trust AI-written code." The only difference is whether they say "use it anyway" or "that's why it's too soon." The conclusion that "AI output requires human verification" is virtually unanimous.
How I Use It: Break Things Small, Let AI Fight Where It's Strong
First, a premise: I don't use AI for the core foundation, the central architecture that everything else sits on, or the initial build. I design that with my own head and write it with my own hands. That's exactly the domain where AI struggles — integration decisions across features, deciding what to abstract versus duplicate.
On top of that foundation, I actively use AI for the small features that grow around the edges — Lambda functions, utility libraries, and similar isolated pieces. Vibe coding is strong at single small tasks. So the key is cutting the work to a granularity AI is good at.
For example, when I want to carve out part of a Python-based system into a Lambda, there used to be situations where Go would have been the better choice, but we'd go with Python anyway because of learning costs, staffing, and maintenance overhead. If nobody on the team could write Go, the option simply didn't exist.
That's different now. I can decide "Go is the right tool here, so let's use Go." The language barrier has dropped significantly thanks to AI. With small, focused Lambda functions, the volume of AI-generated code to review stays manageable. This very site uses quite a few Lambdas, and I've started applying the same approach to client systems in my freelance work.
What matters is separating what can go into paper bags from what can't. And keeping what goes into paper bags small enough that rebuilding one from scratch doesn't hurt.
This kind of "decomposition to the right granularity" used to require the budget and team size of a large organization. Microservice architecture was talked about as an ideal, but only big companies could actually do it. Now individuals and freelancers can too. This may be the most practical benefit of paper bags getting cheaper.
If there's a key to maximizing vibe coding's power, it's not "hand everything to AI." It's "cut the work to the granularity AI is good at." One issue. One Lambda. One library. At that granularity, AI is surprisingly accurate. Hold onto the foundation yourself, and let AI handle the periphery. That's the most realistic approach I've found so far.
Frequently Asked Questions
Will vibe coding make professional engineers obsolete?
No. Vibe coding excels at single small tasks, but cannot handle cross-feature integration decisions or security architecture design. With 62% of AI-generated code containing design flaws, engineers who can review and fix that code are actually more valuable. What changes is the amount of code written by hand — the work of designing, judging, and ensuring quality remains.
How do you ensure security in AI-generated code?
Both human security review and automated scanning are essential. With XSS defense failure rates at 86% and SQL injection defense failure rates at 88%, deploying AI output without verification is dangerous. Pinning GitHub Actions versions by commit hash (SHA), verifying that dependency packages actually exist, and other basic measures remain critical.
What will happen to the engineering job market?
A Nikkei survey found about 40% of Japanese job-seekers changed their target profession due to AI. But the market consensus is that engineering jobs won't disappear — rather, selection will intensify. Engineers who can design, judge trade-offs, and ensure quality are becoming more valuable. Tech journalist Takeshi Kimura of Nikkei xTECH predicts that "the IT industry's billing-by-the-person model will collapse within 5 years."
Update History
- March 30, 2026: Initial publication