this post was submitted on 21 Nov 2024
78 points (96.4% liked)

Technology

59511 readers
2939 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

Niantic, the company behind the extremely popular augmented reality mobile games Pokémon Go and Ingress, announced that it is using data collected by its millions of players to create an AI model that can navigate the physical world. 

In a blog post published last week, first spotted by Garbage Day, Niantic says it is building a “Large Geospatial Model.” This name, the company explains, is a direct reference to Large Language Models (LLMs) Like OpenAI’s GPT, which are trained on vast quantities of text scraped from the internet in order to process and produce natural language. Niantic explains that a Large Geospatial Model, or LGM, aims to do the same for the physical world, a technology it says “will enable computers not only to perceive and understand physical spaces, but also to interact with them in new ways, forming a critical component of AR glasses and fields beyond, including robotics, content creation and autonomous systems. As we move from phones to wearable technology linked to the real world, spatial intelligence will become the world’s future operating system.”

By training an AI model on millions of geolocated images from around the world, the model will be able to predict its immediate environment in the same way an LLM is able to produce coherent and convincing sentences by statistically determining what word is likely to follow another.

top 29 comments
sorted by: hot top controversial new old
[–] [email protected] 6 points 2 hours ago (1 children)

And this software will probably be able to route soumeone from one special Pokemon point to the other. Wow. There are three of them in our town. It will be very smart in speedrunning that triangle.

[–] [email protected] 1 points 1 hour ago (1 children)

They have added tasks that make you photograph your surroundings or Objects and give them real world lidarr data linked with geo data for some in-game benefits, last I checked.

[–] [email protected] 1 points 57 minutes ago

OK, that is actually something usable. So far what they could learn from here is how to take a shortcut through the fields ;-)

[–] [email protected] 20 points 3 hours ago (3 children)

same way an LLM is able to produce coherent and convincing sentences by statistically determining what word is likely to follow another

To me this implies that the navigation AI is going to hallucinate parts of its model of the world, because it's basing that model on what's statically the most likely to be there as opposed to what's actually there. What could go wrong?

[–] [email protected] 0 points 59 minutes ago

It's only going to hallucinate until it gets new input from reality. Not nearly as precarious as generative models.

[–] [email protected] 13 points 2 hours ago* (last edited 2 hours ago) (1 children)

AI: Dave, turn right and walk across the bridge.

Dave : But AI, there is no bridge

AI: I am 99% sure based on 99 billion images that there should be a bridge

Dave: ok , you're the smart one

Dave: aaaargh . . . .

SPLAT

[–] [email protected] 4 points 1 hour ago (1 children)

Reminds me of Morrowind's directions, with the frequent east-west mixups, and sporadic north-south mixups.

[–] [email protected] 1 points 1 hour ago

Fun fact, that's why the immersion-breaking magic compass thing exists in Oblivion (and most open worlds since). Bethsoft devs explained it once.

Stuff is relocated a lot in development, and this means having to rework all dialogues refering to directions, occasionally missing some. It was even more unfeasible for Oblivion in which all dialogue is voiced and would have to be re-recorded.

So they just removed all directions from the dialogue and now you've got 100% accurate floating tags telling you exactly where to go, even when you are not yet sure what you're looking for.

[–] [email protected] 8 points 2 hours ago

Its more amusing than you think. Most really hardcore of the players left now spoof their GPS position.

Id be willing to bet that most of the navigation data is completely useless.

[–] [email protected] 8 points 3 hours ago (2 children)

From all the apps invading your privacy and abusing your data, I didn't suspect Pokemon GO to be one of them.

This should be so extremely illegal that it should bring criminal charges to all the members of their board.

[–] [email protected] 3 points 1 hour ago

Hu? Really? I thought that was known even when pkm go wasn’t released and only ingress existed.

Niantic is a google split up after all, if they were not collecting data, I would have been very surprised

[–] [email protected] 7 points 2 hours ago

You probably agreed to it when you installed the app.

[–] [email protected] -2 points 3 hours ago (2 children)

So slavery, then. No surprise.

[–] [email protected] 0 points 2 hours ago

I mean... the Pokémon kind of get forced into combat for your entertainment... so more like cockfighting?

[–] [email protected] 3 points 3 hours ago
[–] [email protected] 5 points 4 hours ago (2 children)

It's almost like listening to my crazy rants predicts the future.

Hope you guys don't have those loyalty rewards cards to grocery stores or pharmacies. Oh, who am I kidding? All of you do.

[–] [email protected] 0 points 2 hours ago

Jenny's number: (area code) 867-5309

Of course it probably doesn't matter if you also use a credit card to make the purchase - every single purchase is fed into your personal consumer profile.

[–] [email protected] 2 points 3 hours ago* (last edited 3 hours ago) (1 children)

Hope you guys don’t have those loyalty rewards cards to grocery stores or pharmacies. Oh, who am I kidding? All of you do.

Does it count if they're all just copies of someone else's cards?
I mean, good luck shopping without them. All shops artificially inflate the prices without them and then act like you're getting a huge discount. For example, Tesco, as much as 100% price increase without their loyalty card, and most products have some. At least a 25% price increase.

[–] [email protected] 0 points 3 hours ago

Yeah, I just stopped buying products.

[–] [email protected] 51 points 4 hours ago (4 children)

I'll copypaste an interesting comment here:

[Stephen Smith] This article is a great example of a trend I don't think companies realize they've started yet: They have killed the golden goose of user-generated content for short-term profit. // Who would willingly contribute to a modern-day YouTube, Reddit, StackOverflow, or Twitter knowing that they are just feeding the robots that will one day replace them?

You don't even need robots replacing humans, or people believing so. All you need is people feeling that you're profiting at their expense.


Also obligatory "If you're not paying for the product, then you are the product".

[–] [email protected] 0 points 58 minutes ago

The only down side, IMO, is that the models are proprietary and closed.

[–] [email protected] 1 points 2 hours ago

What are platforms going to do about it? Start to demonetize AI generated videos and ban AI written fan fiction?

[–] [email protected] 1 points 3 hours ago* (last edited 2 hours ago) (1 children)

I think people will still “contribute” because they also don’t care that their use of certain platforms leaks data used to target ads at them.

In the same vein though, once AI essentially destroys a site like Stack Overflow, where will AI companies source new training data with updated information? Also, we are likely to see something like 50% of content being AI generated. Are AI models then going to train on the content they themselves created? What is the impact of that? What is the use?

[–] [email protected] 1 points 2 hours ago

Are AI models then going to train on the content they themselves created? What is the impact of that?

It leads to model collapse. The second AI starts to focuses on certain patterns in the output of the first AI instead of the actual content and you get degraded output. They are pattern matching machines after all. Repeat the cycle a few times and all output becomes gibberish. Think of it as data incest.

So the AI companies are pretty desperate for more fresh user data. More data is the only way they have currently to push through the diminishing returns.

[–] [email protected] 7 points 3 hours ago (1 children)

I’ve found myself thinking “well, you just helped teach the AI about that one…” various times when reading content online.

It’s a strange thing to know that a form of the basilisk is real. Things posted will help AI get better, if only my teeny tiny increments each time.

[–] [email protected] 7 points 3 hours ago (2 children)

AI learning isn't the issue, its not something we will be able to put a lid on either way. Either it destroys or saves the world. It doesn't need to learn much to do so besides evolving actual self-agency and sovereign thought.

What is a huge issue is the secretive non-consentual mining of peoples identity and expressions.

And then acting all normal about It.

[–] [email protected] 2 points 2 hours ago

AI learning isn't the issue, its not something we will be able to put a lid on either way.

So... there is no Artificial Intelligence. The AI cannot hurt you. It is just a (buggy) statistical language parsing system. It does not think, it does not plan, it does not have goals, it does not understand, and it doesn't even really "learn" in a meaningful sense.

Either it destroys or saves the world.

If we're talking about machine learning systems based on multi-dimensionl statistical analyses, then it will do neither. Both extremes are sensationalism and arguments based on the idea that either such outcome will come from the current boom of ML technology is utter nonsense designed to drive engagement.

It doesn't need to learn much to do so besides evolving actual self-agency and sovereign thought.

Oh, is that all?

No one on the planet has any idea how to replicate the functionality of consciousness. Sam Altman would very much like you to believe that his company is close to achieving this so that VCs will see the public interest and throw more money at him. Sam Altman is a snake oil salesman.

What is a huge issue is the secretive non-consentual mining of peoples identity and expressions.

And then acting all normal about It.

This is absolutely true and correct and the collection and aggregation of data on human behavior should be scaring the shit out of everyone. The potential for authoritarian abuses of such data collection and tracking is disturbing.

[–] [email protected] 4 points 3 hours ago (1 children)

I didn’t say it was an issue. I just said it was a strange feeling to know AI is watching us talk past each other.

[–] [email protected] 1 points 3 hours ago* (last edited 3 hours ago)

I sort of misread your comment as saying the basilisk is inevitable which is a thought i would describe as least oopsie-issue-level.

Still there are many other people bent on directly poisoning AI to counteract the learning but i just fear that will get it to dangerously rogue mentally challenged AI faster then if we aimed for maximum coherent intelligence and hope that benevolence is an emergent behavior from it.

But more at hand. If we build AI by grossly exploiting our own fellow-humans. How do we expect it will treat us once it reaches a state of independent learning.