> But if you take the point of view of a customer, it might not matter as much 'which' part is broken. To use a bad analogy, if my car is in the shop 10% of the time, it's not much comfort if each individual component is only broken 0.1% of the time.
Not to go too out of my way to defend GH's uptime because it's obviously pretty patchy, but I think this is a bad analogy. Most customers won't have a hard reliability on every user-facing gh feature. Or to put it another way there's only going to be a tiny fraction of users who actually experienced something like the 90% uptime reported by the site. Most people are in practice are probably experienceing something like 97-98%.
Sorry, by 'customer' I meant to say something like a large corporate customer - you're buying the whole package, and across your org, you're likely to be a little affected by even minor outages of niche services.
But yeah, totally agree that at the individual level, the observed reliability is between 90% and 99%, and probably toward the upper end of that range.
I think the parent is just pointing out that these things lie on a spectrum. I have a website that consists largely of static content and the (significant) scraping which occurs doesn't impact the site for general users so I don't mind (and means I get good, up to date answers from LLMs on the niche topic my site covers). If it did have an impact on real users, or cost me significant money, I would feel pretty differently.
Putting everything on a spectrum is what got us into this mess of zero regulation and moving goal posts. It's slippery slope thinking no matter which way we cut it, because every time someone calls for a stop sign to be put up after giving an inch, the very people who would have to stop will argue tirelessly for the extra mile.
What mess are you talking about? The existence of LLMs? I think it's pretty neat that I can now get answers to questions I have.
This is something I couldn't have done before, because people very often don't have the patience to answer questions. Even Google ended up in loops of "just use Google" or "closed. This is a duplicate of X, but X doesn't actually answer the question" or references to dead links.
Are there downsides to this? Sure, but imo AI is useful.
It's just repackaged Google results masquerading as an 'answer.' PageRank pulled results and displayed the first 10 relevant links and the LLM pulls tokens and displays the first relevant tokens to the query.
1. LLMs can translate text far better than any previous machine translation system. They can even do so for relatively small languages that typically had poor translation support. We all remember how funny text would get when you did English -> Japanese -> English. With LLMs you can do that (and even use a different LLM for the second step) and the texts remain very close.
2. Audio-input capable LLMs can transcribe audio far better than any previous system I've used. They easily understood my speech without problems. Youtube's old closed captioning system want anywhere close to as good and Microsoft's was unusable for me. LLMs have no such problems (makes me wonder if my speech patterns are in the training data since I've made a lot of YouTube videos and that's why they work so well for me).
3. You can feed LLMs local files (and run the LLM locally). Even if it is "just" pagerank, it's local pagerank now.
4. I can ask an LLM questions and then clarify what I wanted in natural language. You can't really refine a Google search in such a way. Trying to explain a Google search with more details usually doesn't help.
5. Iye mkx kcu kx VVW dy nomszrob dohd. Qyyqvo nyocx'd ny drkd pyb iye. - Google won't tell you what this means without you knowing what it is.
LLMs aren't magic, but I think they can do a whole bunch of things we couldn't really do before. Or at least we couldn't have a machine do those things well.
Generalizing with "everything", "all", etc exclusive markers is exactly the kind of black/white divide you're arguing against. What happened to your nuanced reality within a single sentence? Not everything is black and white, but some situations are.
The person he's replying to argued against putting things on a spectrum. Does that not imply painting everything in black and white? Thus his response seems perfectly sensible to me.
He argued against putting things in a spectrum in many instances where that would be wrong, including the case under the question. What's your argument against that idea? LLM'ed too much lately?
In a funny way it reminds me of writing survey questions. You have to be so careful not to introduce some bias just with the wording, as you can basically nudge the LLM to the answer you want with some little hints in the q e.g. "is it right that..."
That's mostly because of London's financial center, where a lot of foreign money is laundered; the city's GDP is comparable in size with small EU countries like Belgium or Ireland. If you take London out of the equation, what's left has an average GDP of only 30k per capita [0].
A quick comparison with [1] (using 1 GBP ~ 1.30 USD) shows that London would rank #8 in Europe (between Denmark and Norway), while the rest of the UK would come in somewhere around #25, between Spain and Italy.
Sure, if you exclude the wealthiest parts of the country then it does look significantly poorer. Just as if you exclude California then the GDP of the US drops significantly. The point was whether the UK is economically relevant, not whether the economy is ethically sound (which is quite a nebulous question I'm sure you'll agree).
> You're only missing out if that's what you want to do.
Who writes software and doesn't have a list of "I'll fix this one day" issues as long as their arm?
This is honestly one of the things I enjoy most at the moment. There's whole classes of issues where I know the fix is probably pretty simple but I wouldn't have had time to sort it previously. Now I can just point claude at it and have a PR 5mins later. It's really nice when you can tell users "just deployed a fix for your thing" rather than "I've made a ticket for your request" your issue is on the never-ending backlog pile and might get fixed in 5 years time if you're lucky.
Claude code makes it so easy to do things the "right way" that it also makes it really easy for you to let scope creep get out of hand. I have a personal project that I haven't deployed yet that in some ways is way overengineered for its purpose. It's hard to blame the tool though, it's always telling me I'm making it more complicated than it needs to be but I don't listen
I've felt this recently. I've often been bad about scope creep. CC makes it so easy.
On the other hand, I can see these tools getting good enough that scope creep doesn't even matter.
ATM I usually get stuck around the review/verification stage. As in, my code works, I have tested that it works, but it is failing CI or someone left a PR comment. And for each comment I'll have to make sure it makes sense, make the change, test again, and get CI passing again.
In my team we have strict rules for scope creep in pull request. Each one needs to introduce a single thing, not a dozen little refactorings. This helps, but not when you're working alone in a personal project. Maybe you can setup your review agent to help with scope creep?
Many people don't. You can write a ticket and the PM can deal with it. Not everyone is intimately involved in their job enough to care about stuff like that. And some projects might not last long enough for you to care. You should project your dev experience on everyone, specially as a software development enthusiast.
I haven't used it so just spit balling, but surely it depends on the quality of the review? If it picks up lots of issues and prevents downtime then it could work out as worthwhile. What would it cost an engineer with deep knowledge of the codebase to do a similar job? You could spend an hour really digging into a PR, poking around, testing stuff out etc. Im guessing most engineers are paid more than $15-25/hr, not to mention the opportunity cost.
Nothing revolutionary, but there is a small organisation called The Himalayan Database [1] who have recorded (in great detail) climbing expeditions to peaks in the area. The data is available in a downloadable format, but it is a little awkward to browse. I have been working in collaboration with them to build a website for making the database more accessible https://himalayan-database.climbing-history.org/
This is just the reality of the power asymmetry and is exactly the same for small company Vs big company. As a small company your business is just not worth that much to big company, so you choices are accept the terms offered or go elsewhere. Or, in an ideal world, there is a competitor who's found a space in the market offering better terms than big company.
Not to go too out of my way to defend GH's uptime because it's obviously pretty patchy, but I think this is a bad analogy. Most customers won't have a hard reliability on every user-facing gh feature. Or to put it another way there's only going to be a tiny fraction of users who actually experienced something like the 90% uptime reported by the site. Most people are in practice are probably experienceing something like 97-98%.
reply