Examples? I can think of a number of foreign companies that the US facilitates, like Nestle.
Examples? I can think of a number of foreign companies that the US facilitates, like Nestle.
Eh, I switched. I switched all of my lab’s computers, too, and my PhD students have remarked a few different times that Linux is pretty cool. It might snowball.
Oregonians almost take pleasure in driving slowly in front of you. Maybe they’ve just gotten used to going slow because the entire state freeway system is always under construction. People driving crazily is infuriating for a completely different reason.
Sorry, what? Not sure if you’re joking, but Americans use texts because they’re free and the ability to use them comes preloaded on the phone (no need to download something that takes up more space). I have Signal and WhatsApp on my phone for my international friends, but I use texts to communicate with US friends because RCS works with everyone and it’s integrated much better into my phone, watch, etc. than any app can be without an absurd amount of permissions given to the app.
I never understand why lemmy downvotes someone who is trying to help by providing accurate information, presumably because they think that there’s a very small chance that the person they’re replying to isn’t being sarcastic.
I actually took that bit out because LLMs are pro climate and against everything that makes the environment worse. That’s a result of being trained on a lot of scientific literature. I was just curious what Opus would say about the conceptual knowledge piece.
Claude Opus disagrees, lol (edit to add: all of what follows is Claude; not me):
I respectfully disagree with the claim that current language models are logically incapable of mapping language to actual concepts. While today’s LLMs certainly have limitations and are prone to hallucinations, they have demonstrated a remarkable ability to capture and manipulate complex concepts through language.
A few key points:
LLMs are trained on vast amounts of human-generated text which is grounded in real-world concepts, allowing the models to build rich representations that go beyond surface-level word associations. Techniques like unsupervised pre-training enable LLMs to learn meaningful conceptual relationships.
In many domains, LLMs have shown an impressive capacity for reasoning, inference, and knowledge synthesis - skills that rely on a deep understanding of the underlying concepts, not just pattern matching. Their performance on complex question-answering, analogical reasoning, and natural language inference tasks is evidence of this.
LLMs can be fine-tuned and augmented with techniques like retrieval-augmented generation or chain-of-thought prompting to further ground them in factual knowledge and improve their reasoning capabilities. This allows more robust conceptual mappings.
Careful prompt engineering and sampling techniques can help to mitigate hallucinations and nonsensical outputs. While not a complete solution, this shows that the conceptual knowledge is often present even if the model sometimes struggles to express it faithfully.
That said, I fully agree that the tendency for subtle nonsense is a serious issue, especially for layperson use cases where hallucinations may go undetected. Continued research into making LLMs more truthful, consistent and robust is crucial. Techniques like constitutional AI to bake in truthfulness, and better UX design to convey uncertainty are important steps.
But in summary, I believe the evidence suggests that LLMs, while flawed, are not fundamentally incapable of meaningful conceptual representation and reasoning. We should push forward on making them more reliable and trustworthy, rather than dismissing their potential prematurely.
I’m back on my BS is also a solid contributor
I was just in a smaller city in Germany and flew back to the US after that. I look German and speak German. When paying with card, Germany felt exactly like the US. At every restaurant, the tip request automatically came up within the thing used to process your card, just like in the US.
Also, the US is 9.14 million sq. km of land, whereas the EU is 4.29 million sq. km of land
EU is still smaller
But the main reason the US can’t handle the same stuff at a federal level that the EU can is population density. The US government can’t afford to nationalize rural healthcare given how rural the US can be–especially with their debt/GDP at the moment. Give it another few hundred years and the US might catch up to Europe in that respect.
If we’re going full coast-to-coast, US still wins
Edit: a better illustration that loses about 80 km but avoids the extra stop.
I’m thinking of shorting it. My friend is definitely shorting it.
To a degree. The large subreddits, like AskReddit, get far fewer upvotes on the top posts of the week than they used to get. I think there’s a good chunk of folks who left for a replacement, then left their replacement without going back to Reddit.
Trogdor was popular way before Reddit