Hiring for Startups: Takeaways from Brian Chesky
Chesky's hiring framework is direct, no-nonsense, and more relevant to startup founders than anything I've read in an HR playbook.
thoughts on software, systems, and good practice
Chesky's hiring framework is direct, no-nonsense, and more relevant to startup founders than anything I've read in an HR playbook.
Compliance requirements land on engineering teams whether they asked for them or not. Here's what that actually looks like.
What changed. What being CTO actually means at this stage of the company.
Specific decisions. Not principles.
Why this combination. What it costs. What breaks. What I'd change.
I've seen OKRs done badly. Here's the version that worked under real pressure.
Not the research version. The version that has to not fail inside a real product.
The engineering decisions that scaled. The ones that didn't.
Waste monitoring in the field. Cameras, weather, variable light. The real world is messier than the dataset.
Product directing tech. No ownership. Three-hour meetings. The first 30 days.
Not the architecture. The judgment. The things you can only learn by being on call when it breaks.
I've been diligenced and I've diligenced others. What the process actually surfaces.
GPT-4V launched. After three years building computer vision products, here's what I think it actually means.
The dismissal is wrong. Here's why prompt engineering is a real engineering discipline.
MMLU scores don't tell you what happens when your prompt is ambiguous at 11pm and the model confidently gives you the wrong answer.
Not a hot take. A practical decision made after running both in production for the same tasks.
Our enterprise clients can't send data to OpenAI. For several of them, it's a legal constraint. Here's how we solved it.
Six months of AI-assisted code review in production. The architecture, the results, and the things it still misses.
Bolting an LLM onto a product as a feature and announcing 'now with AI' is not a strategy.
Early API access to GPT-4. Six weeks in, here's the honest account.
Having run GPT-3 in production since mid-2020, here's an honest read on what ChatGPT actually is.
0.5 seconds average response time in logistics. Not a benchmark — a business requirement.
What the patent covers, why we filed, and what the process actually looks like from inside a startup.
Closed pilots with LVMH, Lego, Moncler, Puma, and Hugo Boss. What we got wrong before we got it right.
Enterprise computer vision at production scale does not require enterprise infrastructure budgets. What we built and what it cost.
In a warehouse, under fluorescent lights, with a conveyor belt at operational speed. What it took.
The wrong choice costs you six months. How to not make it.
When request volume stopped being a number we celebrated and started being a number we engineered for.
The papers show clean benchmark datasets. The logistics environment is not that.
What closing a seed round looks like from inside it, as a technical founder who had never raised before.
The architecture decision that comes up in every investor conversation, written down clearly once.
Three months in. What actually happened, before the story gets cleaned up in retrospect.
Two years watching the Hero Mindset framework applied by founders. What it means in practice.
How we validated Countercheck before building anything — and what 50 customer conversations actually produces.
Six months as an Entrepreneur in Residence. What the model actually is, from someone who's done it.
Early API access. Here's what it actually does.
COVID hit. Three weeks to move a residential global entrepreneurship program fully online without losing quality.
Mapping whether computer vision is the right fit for automating product authentication at scale in logistics and retail.
The hype arc of the 2010s, what actually changed, and what the predictions got badly wrong.
Two years, several hundred pitches analyzed. What we actually learned about building AI products.
Aaron Levie on why distribution is not a post-product problem — and how Box figured out enterprise go-to-market.
AI theater: when a product uses AI to perform intelligence rather than to produce it.
Culture is what you tolerate, not what you declare. Notes from Tony Hsieh's session with founders.
Phil Libin on asymmetric retention — why some products get more valuable the longer you use them.
OpenAI didn't release the full model. The stated reason was misuse. Here's what I think it actually means.
A year of running STT on founder pitch recordings. What the benchmark numbers don't tell you.
BERT dropped in October. What it changes for practitioners doing NLP in production.
Biz Stone on taste as a product skill — and why the discipline of removal is harder than addition.
Six weeks, 60 founders from 40 countries, and what actually changes people — from the program side.
The paper says 92% accuracy. The production system says something different.
Several hundred pitches analyzed. What NLP actually surfaces — and where the ceiling is.
Three months building on Watson for Zillion Pitches. The gap between the demo and the integration is real, and nobody talks about it.
The ambient conversations alone changed how I thought. You can't get that remotely.
When a technology becomes boring, it becomes useful. Notes on the maturation of the stack.
The pull of the AI wave. The decision to go be closer to where things are happening.
What I built, what I broke, and the one decision I'd reverse.
Turkey to Silicon Valley. What the ecosystem actually looks like when you're an outsider.
2017. Every startup deck says AI. Almost none of them mean it.
We used them in production. Here's what the docs don't prepare you for.
Not a tutorial. A survival guide for doing both platforms at once with no dedicated mobile team.
The gap between writing code and leading technical direction is wider than it looks.
Free cloud infrastructure is a gift. It's also a trap if you're not careful.
We won and immediately started making expensive decisions.
One engineer. Two platforms. One backend. The decisions that kept me sane.
Google open-sourced it two months ago. I've been poking at it since. Here's where I am.
A note from someone who was convinced ML was for PhDs. It isn't.
Automated testing for mobile apps that generate sensory data. The problem kept reshaping itself.
I spent three weeks on this. Here's the version I wish someone had written for me.
Built a text classifier for my MSc thesis. Got 94% accuracy. Shipped garbage.