AndrasKrigare
@AndrasKrigare@beehaw.org
- Comment on Sony shuts down Concord developer Firewalk Studios, game will remain permanently offline 3 weeks ago:
Oh shit, I’m sorry. I misunderstood what you were saying, I thought you were referring to them purchasing and running their own physical server hardware as opposed to running their servers off of a cloud platform.
- Comment on Sony shuts down Concord developer Firewalk Studios, game will remain permanently offline 3 weeks ago:
That’s kinda a weird take, since the private server model was the only model until 10 years ago or so. Companies definitely know it. It’s just not financially efficient comparing to benefiting from economies of scale with hosting. Plus you don’t lose a ton of money or piss of players if you over or under estimate how popular the game will be.
Had they gone with private servers here, they would have lost even more money than they already have. The problem here is they spent too much money on a game no one wanted to play, chasing a fad that ended before it launched.
- Comment on [deleted] 3 weeks ago:
One note on “sick” being slang for “good”: that particular slang started in the 80s, and some of the younger generation consider it to be old person slang.
- Comment on Integer addition algorithm could reduce energy needs of AI by 95% 5 weeks ago:
I’d say it’s not just misleading but incorrect if it says “integer” but it’s actually floats.
- Comment on Thank you Skövde 1 month ago:
I actually looked into this, part of the explanation is that in the 80s, Sweden entered a public/private partnership to subsidize the purchase of home computers, which otherwise would have been prohibitively expensive. This helped create a relatively wide local consumer base for software entertainment as well as have a jump start on computer literacy and software development.
- Comment on AI hallucinations are impossible to eradicate — but a recent, embarrassing malfunction from one of China’s biggest tech firms shows how they can be much more damaging there than in other countries 2 months ago:
I think to some extent it’s a matter of scale, though. If I advertise something as a calculator capable of doing all math, and it can only do one problem, it is so drastically far away from its intended purpose that the meaning kinda breaks down. I don’t think it would be wrong to say “it malfunctions in 99.999999% of use cases” but it would be easier to say that it just doesn’t work.
Continuing (and torturing) that analogy, if we did the disgusting work of precomputing all 2 number math problems for integers from -1,000,000 to 1,000,000 and I think you could say you had a (really shitty and slow) calculator, which “malfunctions” for numbers outside that range if you don’t specify the limitation ahead of time. Not crazy different from software which has issues with max_int or small buffers.
If it were the case that there had only been one case of a hallucination with LLMs, I think we could pretty safely call that a malfunction (and we wouldn’t be having this conversation). If it happens 0.000001% of the time, I think we could still call it a malfunction and that it performs better than a lot of software. 99.999% of the time, it’d be better to say that it just doesn’t work. I don’t think there is, or even needs to be, some unified understanding of where the line is between them.
Really my point is there are enough things to criticize about LLMs and people’s use of them, this seems like a really silly one to try and push.
- Comment on AI hallucinations are impossible to eradicate — but a recent, embarrassing malfunction from one of China’s biggest tech firms shows how they can be much more damaging there than in other countries 2 months ago:
We’re talking about the meaning of “malfunction” here, we don’t need to overthink it and construct a rigorous proof or anything. The creator of the thing can decide what the thing they’re creating is supposed to do. You can say
hey, it did X, was that supposed to happen?
no, it was not supposed to do that, that’s a malfunction.
We don’t need to go to
Actually you never sufficiently defined its function to cover all cases in an objective manner, so ACTUALLY it’s not a malfunction!
Whatever, it still wasn’t supposed to do that
- Comment on AI hallucinations are impossible to eradicate — but a recent, embarrassing malfunction from one of China’s biggest tech firms shows how they can be much more damaging there than in other countries 2 months ago:
The purpose of an LLM, at a fundamental level, is to approximate text it was trained on.
I’d argue that’s what an LLM is, not its purpose. Continuing the car analogy, that’s like saying a car’s purpose is to burn gasoline to spin its wheels. That’s what a car does, the purpose of my car is to get me from place to place. The purpose of my friend’s car is to look cool and go fast. The purpose of my uncle’s car is to carry lumber.
I think we more or less agree on the fundamentals and it’s just differences between whether they are referring to a malfunction in the system they are trying to create, in which an LLM is a key tool/component, or a malfunction in the LLM itself. At the end of the day, I think we can all agree that it did a thing they didn’t want it to do, and that an LLM by itself may not be the correct tool for the job.
- Comment on AI hallucinations are impossible to eradicate — but a recent, embarrassing malfunction from one of China’s biggest tech firms shows how they can be much more damaging there than in other countries 2 months ago:
Where I don’t think your argument fits is that it could be applied to things LLMs can currently do. If I have an insufficiently trained model which produces a word salad to every prompt, one could say “that’s not a malfunction, it’s still applying weights.”
The malfunction is in having a system that produces useful results. An LLM is just the means for achieving that result, and you could argue it’s the wrong tool for the job and that’s fine. If I put gasoline in my diesel car and the engine dies, I can still say the car is malfunctioning. It’s my fault, and the engine wasn’t ever supposed to have gas in it, but the car is now “failing to function in a normal or satisfactory manner,” the definition of malfunction.
- Comment on AI hallucinations are impossible to eradicate — but a recent, embarrassing malfunction from one of China’s biggest tech firms shows how they can be much more damaging there than in other countries 2 months ago:
It implies that, under the hood, the LLM is “malfunctioning”. It is not - it’s doing what it is supposed to do, to chain tokens through weighted probabilities.
I don’t really agree with that argument. By that logic, there’s really no such thing as a software bug, since the software is always doing what it’s supposed to be doing: giving predefined instructions to a processor that performs some action. It’s “supposed to” provide a useful response to prompts, anything other than is it not what it should be and could be fairly called a malfunction.
- Comment on I don't hate Body Type replacing Gender, I hate laziness 2 months ago:
To be clear, your stance is it’s such a small step in the right direction, you’d prefer no step at all? Keep it cis-only or invest time/money in extra character models?
- Comment on Walmart's use of digital price tags signal the future of retail shopping, but consumers are worried 2 months ago:
- Comment on No More Room In Hell: How Tech & Games Are Desperately Rotting (The Jimquisition) 3 months ago:
I think that’s been a fair description of the AAS space for a long time, which is fine. If you want innovation, go indie, if you want big budget, go AAA
- Comment on 0.0.0.0 Day - 18 Yr Old Vulnerability Let Attackers Bypass All Browser Security 3 months ago:
Yeah, I just did a quick test in Python to do a tcp connection to “0.0.0.0” and it made a loopback connection, instead of returning an error as I would have expected.
- Comment on Astronomers discover technique to spot AI fakes using galaxy-measurement tools 3 months ago:
Looking completely realistic and being able to discern between real and fake are competing goals. If you can discern the difference, then it does not look completely realistic.
I think what they’re alluding to is generative adversarial networks …wikipedia.org/…/Generative_adversarial_network where creating a better discriminator that can detect a good image from bad is how you get a better image.
- Comment on The Indie Chat & Recommendation Thread 5 months ago:
To help give love to some games I think are underrated, here’s a list of my favorite games with 4,000 reviews or less on steam under $25 ranked by my personal play time.
Neo scavenger $15
Post apocalyptic survival sim, that reminds me a tiny bit of Oregon Trail. There’s a good chance a scratch will kill you, and finding a plastic bag so you can carry more than what you hold in your two hands makes you feel OP. I’ve put 74 hours into this game, have died and restarted countless times, and have hardly gotten anywhere in it, but it’s exactly my kind of survival game
Fae tactics $20
Turn-based grid combat reminiscent of Final Fantasy Tactics, with just a splash of pokemon. The mechanics and setting I found really fun, although the difficulty can fluctuate a good bit at times.
Xenonauts $25
If OG XCOM went more crunchy than streamlined, it’d be Xenonauts instead of Firaxis’s Enemy Unknown. The combat gives you a ton of control during combat, specifying how much time they should spend aiming before shooting, specific hours of overwatch, crouching, etc.
Star Renegades $25 (currently $5)
Roguelike turn based party RPG. It doesn’t do a crazy amount that’s new or novel, but it executes very well, and lining up a good combo with your build feels amazing.
Rogue Book $25
Slay the Spire with some smart additions. Instead of one hero, you play two, which gives some extra possibilities to mix and match between runs. Instead of an overmap with a couple branching paths, there’s a hex overworld where you can use resources to reveal tiles.
Wildfire $15
Avatar the Last Airbender as a 2d stealth action game. The level layouts are great, and the ability upgrades strike a good balance between being impactful and not trivializing encounters.
Don’t Escape: 4 Days to survive $15
A classic point and click adventure, except using human logic instead of insane Game Logic. Reminds me of a bunch of similar games I played at the height of Newgrounds. It’s a tight, solid experience that doesn’t over stay its welcome.
Alina of the Arena $15
What if Slay the Spire had a hex grid system? I’ve seen other games ask this question, but Alina is the best I’ve played. There are some really clever design decisions they’ve made where certain builds very intuitively form some classic archetypes.
Shardpunk $14 (currently $10)
Roguelike XCOM themed as a crystalpunk version of Vermintide. Combat is solid, but the theme of running to the exit while shooting rats on the way with crystal powered machine guns sets it apart for me.
The Case of the Golden Idol $18
This one breaks my “4,000 or less” review rule by a little bit, so I’m putting it at the bottom, but it is one of my favorite games. I understand the love for Obra Dinn, but Golden Idol is better in my opinion. Each myself is a scene more or less frozen in time, which you can click on things for clues as what’s happening. What sets it apart is how you really do need to solve the mystery to progress; the game doesn’t walk you into it nor really lets you brute force it. Hands down the best mystery game I’ve ever played.
- Comment on The Indie Chat & Recommendation Thread 5 months ago:
Really fun GDC talk by spiderweb games youtu.be/stxVBJem3Rs?si=mZdu6eyyWD4OEWGw
- Comment on Former Square Enix exec on why Final Fantasy sales don’t meet expectations and chances of recouping insane AAA budgets | Game World Observer 5 months ago:
I think it highlights how perverse the stock market itself is. It doesn’t really seem like it functions much as a way for riskier ventures to raise capital outside of a bank, but a giant casino that gives the illusion of not being a zero sum game.
It’s hypothetically possible for a company to make more money in the stock market by investing in themselves than by creating anything (see Tesla). And if all companies could behave this way and somehow knew what the stock market would do for 5 years, I’d wager a TON of companies wouldn’t meet it, invest in the stock market, drive up the “value,” more don’t meet it, etc. etc. until no one is making anything, and everyone is happy with their paper fortunes and try to sell.
- Comment on [deleted] 5 months ago:
Huh… Will this message then get re-ingested by chatgpt? Did it just poison itself?
- Comment on Steam is a ticking time bomb 7 months ago:
A monopoly with checks notes 30% market share. It has a plurality, but not a majority.
- Submitted 7 months ago to gaming@beehaw.org | 24 comments
- Comment on WWII first person shooters 9 months ago:
I’d just throw out that my recollection is that it was really more of a mid-to-late 2000’s thing for the oversaturation of WW2 games, if you’re willing to move your window forward a bit. That and there weren’t nearly as many games being released at that time period, so it didn’t take much to saturate the market; there were roughly 1/50th the number of released in 2008 as today (www.statista.com/…/number-games-released-steam/ using steam releases as a rough approximation of total).
In terms of specific games, I don’t have any that aren’t already mentioned elsewhere. The Battlefield, Band of Brothers, and Call of Duty recurring releases are really the big ones. …wikipedia.org/…/List_of_World_War_II_video_games has a good list if you want to browse more.
- Comment on ‘Significant security loophole’ found in Google software container system 9 months ago:
misconfigured
Makes me skeptical this is a real “loophole”
The issue revolves around permissions, with GKE allowing users access to the system with any valid Google account. Orca Security said this creates a “significant security loophole when administrators decide to bind this group with overly permissive roles.”
Orca Security noted that Google considers this to be “intended behavior” because in the end, this is an assigned permission vulnerability that can be prevented by the user. Customers are responsible for the access controls they configure.
The researchers backed Google’s assessment that organizations should “take responsibility and not deploy their assets and permissions in a way that carries security risks and vulnerabilities.”
Yeah, PEBKAC
- Comment on Hertz 180: Rental giant to sell 20,000 EVs and replace them with gas-powered vehicles 10 months ago:
What maintenance?
- Comment on OpenAI says it’s “impossible” to create useful AI models without copyrighted material 10 months ago:
I don’t see why it wouldn’t be able to. That’s a Big Data problem, but we’ve gotten very very good at searches. Bing, for instance, conducts a web search on each prompt in order to give you a citation for what it says, which is pretty close to what I’m suggesting.
As far as comparing to see if the text is too similar, I’m not suggesting a simple comparison or even an Expert Machine; I believe that’s something that can be trained. GANs already have a discriminator that’s essentially measuring how close to generated content is to “truth.” This is extremely similar to that.
I completely agree that categorizing input training data by whether or not it is copyrighted is not easy, but it is possible, and I think something that could be legislated. The AI you would have as a result would inherently not be as good as it is in the current unregulated form, but that’s not necessarily a worse situation given the controversies.
On top of that, one of the common defenses for AI is that it is learning from material just as humans do, but humans also can differentiate between copyrighted and public works. For the defense to be properly analogous, it would make sense to me that it would need some notion of that as well.
- Comment on OpenAI says it’s “impossible” to create useful AI models without copyrighted material 10 months ago:
I know it inherently seems like a bad idea to fix an AI problem with more AI, but it seems applicable to me here. I believe it should be technically feasible to incorporate into the model something which checks if the result is too similar to source content as part of the regression.
My gut would be that this would, at least in the short term, make responses worse on the whole, so would probably require legal action or pressure to have it implemented.
- Comment on Duolingo Fires Translators in Favor of AI 10 months ago:
Yup, that’s the one
- Comment on 5+ man group games 10 months ago:
Ones I’ve played, in no particular order
- Rust
- Civilization
- Valheim
- Project Cars
- Battlebit
- Stolen Realm
- Hell Let Loose
- Comment on Duolingo Fires Translators in Favor of AI 10 months ago:
I’ve found Babble to be okay
- Comment on OLED is nice and all, but the upgrade I really want is this. 11 months ago: