“Strategy” implies he actually thinks about it. I think it’s just a reflex; fault belongs elsewhere, always. The man is incapable of critical thought, especially inward.
“Strategy” implies he actually thinks about it. I think it’s just a reflex; fault belongs elsewhere, always. The man is incapable of critical thought, especially inward.
Machine learning has many valid applications, and there are some fields genuinely utilizing ML tools to make leaps and bounds in advancements.
LLMs, aka bullshit generators, which is where a huge majority of corporate AI investment has gone in this latest craze, is one of the poorest. Not to mention the steaming pile of ethical issues with training data.
Very nice writeup. My only critique is the need to “lay off workers to stop inflation.” I have no doubt that some (many?) managers etc… believed that to be the case, but there’s rampant evidence that the spike of inflation we’ve seen over this period was largely due to corporate greed hiking prices, not due to increased costs from hiring too many workers.
Agreed. The solution to this is to stop using LLMs to present info authoritatively, especially when facing directly at the general public. The average person has no idea how an LLM works, and therefore no idea why they shouldn’t trust it.
My guess is that your name is so poorly represented in the training data that it just picked the most common kind of job history that is represented.
Yeah, exactly. The issue is precisely that it’s NOT just showing search results. MS’s software is generating libelous material and presenting it as fact.
Air Canada was forced to give a customer the compensation its chat bot made up. Germany/Europe in general is a bit stronger on public protections than Canada, so I’d expect MS would be held liable if this journalist decides to press a suit.
Bullshit generator generating bullshit, news at 11.
There is no way that arming Taiwan results in Taiwan starting a war of aggression against China.
What arming Taiwan does is make it an increasingly bad idea for China to invade Taiwan. It’s a deterrent to make sure the nuclear power who constantly threatens Taiwan (read: China) doesn’t think they can just go and take what they want without consequence, and probably commit a little genocide on the side.
You know, like Ukraine.
This. Satire would be writing the article in the voice of the most vapid executive saying they need to abandon fundamentals and turn exclusively to AI.
However, that would be indistinguishable from our current reality, which would make it poor satire.
The issue with “Human jobs will be replaced” is that society still requires humans to have a paying job to survive.
I would love a world where nobody had to do dumb labour anymore, and everyone’s needs are still met.
What part of “we paid these guys and they said we’re fine” do you not? Why would they choose and pay and release the results from a company they didn’t trust to clear them?
I’m not saying it’s rotten, but the fact that the third party was unilaterally chosen by and paid for LMG makes all the results pretty questionable.
It’s hard to trust a firm that is explicitly being paid by the company they’re investigating. I could be convinced that they are actually a neutral third party and that their investigation was unbiased if they had a track record of finding fault with their clients a significant portion of the time. (I haven’t done the research to see if that’s the case.)
However, you have to ask yourself - how many companies would choose to hire a firm which has that track record? Wouldn’t you pick one more likely to side with you?
The way to restore credibility is to have an actually independent third party investigation. Firm chosen by the accuser, perhaps. Or maybe something like binding arbitration. Even better, a union that can fight for the employees on somewhat even footing with the company.
The fundamental difference is that the AI doesn’t know anything. It isn’t capable of understanding, it doesn’t learn in the same sense that humans learn. A LLM is a (complex!) digital machine that guesses the next most likely word based on essentially statistics, nothing more, nothing less.
It doesn’t know what it’s saying, nor does it understand the subject matter, or what a human is, or what a hallucination is or why it has them. They are fundamentally incapable of even perceiving the problem, because they do not perceive anything aside from text in and text out.
If an LLM had to say “I don’t know” when it doesn’t know, that’s all it would be allowed to say! They literally don’t know anything. They don’t even know what knowing means. They are complex (and impressive, admittedly) text generators.
Best I can find in Canada is in BC. I think you could get longer distances in a few other provinces, but the issue is a lack of roads/destinations in the northern corners, haha.
I don’t know about the regulatory side, but Boeing gutted their experienced engineering corps starting about 10 years ago. In the pursuit of profit of course. I think we’re seeing the effects of that finally coming to the fore.
My understanding of the role of the regulatory agencies for stuff like this is that they can ground a model of plane if they believe there’s a systemic issue. Like we saw with the MAX.
NFTs do not solve the problem of proof of ownership. Nor can they. If someone steals it from you - whether by trickery, force, or any other means - it’s just as lost to you as any other stolen thing, digital or physical. (Not to touch on the fact that NFTs to date have just been URLs to web hosted media, i.e. ridiculously non-unique and insecure.)
Also, your whole paragraph about theoretical NFT replacement for DRM is just describing a different kind of DRM.
In any industrial context, a “robot” is short for robotic arm. Those things you see in footage of automotive factories.
They also don’t have any kind of AI. It’s just a regular (if specialized) computer in control.
Your former coworkers are incredibly lucky that after showing management how to turn 80h/week into 4h/week, they didn’t keep the automation train going. Because the very next thing they would do is lay off 90% of the staff and make the remainder still work full time.
Automation should do what you did - give people more time off. Just about every corp uses it to minimize labour costs, though.
Did you tell him you guess you have to stop doing non-web development then? Clearly you’re not qualified if you can’t have the corresponding title.