Robert Miles AI Safety
Robert Miles AI Safety
  • Видео 48
  • Просмотров 6 799 991
AI Ruined My Year
How to Help: aisafety.info/questions/8TJV/How-can-I-help
www.aisafety.com/
AI Safety Talks: www.youtube.com/@aisafetytalks
There's No Rule That Says We'll Make It: ruclips.net/video/JD_iA7imAPs/видео.html
The other "Killer Robot Arms Race" Elon Musk should worry about: ruclips.net/video/7FCEiCnHcbo/видео.html
Rob's Reading List:
Podcast: rmrlp.libsyn.com/
RUclips Channel: www.youtube.com/@RobMilesReadingList
The FLI Open Letter: ruclips.net/video/3GHjhG6Vo40/видео.html
Yudkowsky in TIME: ruclips.net/video/a6m7JynBp-0/видео.html
Ian Hogarth in the FT: ruclips.net/video/Z8VvF82T6so/видео.html
Links:
The CAIS Open Letter: www.safe.ai/work/statement-on-ai-risk
The FLI Open Letter: futureoflife.org/open-le...
Просмотров: 216 200

Видео

Why Does AI Lie, and What Can We Do About It?
Просмотров 255 тыс.Год назад
How do we make sure language models tell the truth? The new channel!: www.youtube.com/@aisafetytalks Evan Hubinger's Talk: https:/ruclips.net/video/OUifSs28G30/видео.html ACX Blog Post: astralcodexten.substack.com/p/elk-and-the-problem-of-truthful-ai With thanks to my wonderful Patrons at patreon.com/robertskmiles : - Tor Barstad - Kieryn - AxisAngles - Juan Benet - Scott Worley - Chad M Jones ...
We Were Right! Real Inner Misalignment
Просмотров 247 тыс.2 года назад
Researchers ran real versions of the thought experiments in the 'Mesa-Optimisers' videos! What they found won't shock you (if you've been paying attention) Previous videos on the subject: The OTHER AI Alignment Problem: Mesa-Optimizers and Inner Alignment: ruclips.net/video/bJLcIBixGj8/видео.html Deceptive Misaligned Mesa-Optimisers? It's More Likely Than You Think...: ruclips.net/video/IeWljQw...
Intro to AI Safety, Remastered
Просмотров 156 тыс.3 года назад
An introduction to AI Safety, remastered from a talk I gave at "AI and Politics" in London The second channel: ruclips.net/channel/UC4qH2AHly_RSRze1bUqSSNw Experts' Predictions about the Future of AI: ruclips.net/video/HOJ1NVtlnyQ/видео.html 9 Examples of Specification Gaming: ruclips.net/video/nKJlF-olKmg/видео.html www.patreon.com/robertskmiles With thanks to my wonderful Patreon supporters: ...
Deceptive Misaligned Mesa-Optimisers? It's More Likely Than You Think...
Просмотров 84 тыс.3 года назад
The previous video explained why it's *possible* for trained models to end up with the wrong goals, even when we specify the goals perfectly. This video explains why it's *likely*. Previous video: The OTHER AI Alignment Problem: ruclips.net/video/bJLcIBixGj8/видео.html The Paper: arxiv.org/pdf/1906.01820.pdf Media Sources: End of Ze World - ruclips.net/video/enRzYWcVyAQ/видео.html FlexClip News...
The OTHER AI Alignment Problem: Mesa-Optimizers and Inner Alignment
Просмотров 226 тыс.3 года назад
This "Alignment" thing turns out to be even harder than we thought. # Links The Paper: arxiv.org/pdf/1906.01820.pdf Discord Waiting List Sign-Up: forms.gle/YhYgjakwQ1Lzd4tJ8 AI Safety Career Bottlenecks Survey: www.guidedtrack.com/programs/n8cydtu/run # Referenced Videos Intelligence and Stupidity - The Orthogonality Thesis: ruclips.net/video/hEUO6pjwFOo/видео.html 9 Examples of Specification G...
Quantilizers: AI That Doesn't Try Too Hard
Просмотров 85 тыс.3 года назад
How do you get an AI system that does better than a human could, without doing anything a human wouldn't? A follow-up to "Maximizers and Satisficers": ruclips.net/video/Ao4jwLwT36M/видео.html The Paper: intelligence.org/files/QuantilizersSaferAlternative.pdf More about this area of research: www.alignmentforum.org/tag/mild-optimization With thanks to my excellent Patreon supporters: www.patreon...
Sharing the Benefits of AI: The Windfall Clause
Просмотров 79 тыс.4 года назад
AI might create enormous amounts of wealth, but how is it going to be distributed? The Paper: www.fhi.ox.ac.uk/wp-content/uploads/Windfall-Clause-Report.pdf The Post: www.fhi.ox.ac.uk/windfallclause/ With thanks to my excellent Patreon supporters: www.patreon.com/robertskmiles Gladamas Scott Worley JJ Hepboin Pedro A Ortega Said Polat Chris Canal Jake Ehrlich Kellen lask Francisco Tolmasky Mich...
10 Reasons to Ignore AI Safety
Просмотров 339 тыс.4 года назад
Why do some ignore AI Safety? Let's look at 10 reasons people give (adapted from Stuart Russell's list). Related Videos from Me: Why Would AI Want to do Bad Things? Instrumental Convergence: ruclips.net/video/ZeecOKBus3Q/видео.html Intelligence and Stupidity: The Orthogonality Thesis: ruclips.net/video/hEUO6pjwFOo/видео.html Predicting AI: RIP Prof. Hubert Dreyfus: ruclips.net/video/B6Oigy1i3W4...
9 Examples of Specification Gaming
Просмотров 307 тыс.4 года назад
9 Examples of Specification Gaming
Training AI Without Writing A Reward Function, with Reward Modelling
Просмотров 238 тыс.4 года назад
Training AI Without Writing A Reward Function, with Reward Modelling
AI That Doesn't Try Too Hard - Maximizers and Satisficers
Просмотров 204 тыс.5 лет назад
AI That Doesn't Try Too Hard - Maximizers and Satisficers
Is AI Safety a Pascal's Mugging?
Просмотров 372 тыс.5 лет назад
Is AI Safety a Pascal's Mugging?
A Response to Steven Pinker on AI
Просмотров 207 тыс.5 лет назад
A Response to Steven Pinker on AI
How to Keep Improving When You're Better Than Any Teacher - Iterated Distillation and Amplification
Просмотров 170 тыс.5 лет назад
How to Keep Improving When You're Better Than Any Teacher - Iterated Distillation and Amplification
Why Not Just: Think of AGI Like a Corporation?
Просмотров 156 тыс.5 лет назад
Why Not Just: Think of AGI Like a Corporation?
Safe Exploration: Concrete Problems in AI Safety Part 6
Просмотров 97 тыс.5 лет назад
Safe Exploration: Concrete Problems in AI Safety Part 6
Friend or Foe? AI Safety Gridworlds extra bit
Просмотров 42 тыс.6 лет назад
Friend or Foe? AI Safety Gridworlds extra bit
AI Safety Gridworlds
Просмотров 92 тыс.6 лет назад
AI Safety Gridworlds
Experts' Predictions about the Future of AI
Просмотров 80 тыс.6 лет назад
Experts' Predictions about the Future of AI
Why Would AI Want to do Bad Things? Instrumental Convergence
Просмотров 248 тыс.6 лет назад
Why Would AI Want to do Bad Things? Instrumental Convergence
Superintelligence Mod for Civilization V
Просмотров 70 тыс.6 лет назад
Superintelligence Mod for Civilization V
Intelligence and Stupidity: The Orthogonality Thesis
Просмотров 670 тыс.6 лет назад
Intelligence and Stupidity: The Orthogonality Thesis
Scalable Supervision: Concrete Problems in AI Safety Part 5
Просмотров 52 тыс.6 лет назад
Scalable Supervision: Concrete Problems in AI Safety Part 5
AI Safety at EAGlobal2017 Conference
Просмотров 19 тыс.6 лет назад
AI Safety at EAGlobal2017 Conference
AI learns to Create ̵K̵Z̵F̵ ̵V̵i̵d̵e̵o̵s̵ Cat Pictures: Papers in Two Minutes #1
Просмотров 48 тыс.6 лет назад
AI learns to Create ̵K̵Z̵F̵ ̵V̵i̵d̵e̵o̵s̵ Cat Pictures: Papers in Two Minutes #1
What can AGI do? I/O and Speed
Просмотров 119 тыс.6 лет назад
What can AGI do? I/O and Speed
What Can We Do About Reward Hacking?: Concrete Problems in AI Safety Part 4
Просмотров 113 тыс.6 лет назад
What Can We Do About Reward Hacking?: Concrete Problems in AI Safety Part 4
Reward Hacking Reloaded: Concrete Problems in AI Safety Part 3.5
Просмотров 91 тыс.6 лет назад
Reward Hacking Reloaded: Concrete Problems in AI Safety Part 3.5
The other "Killer Robot Arms Race" Elon Musk should worry about
Просмотров 100 тыс.7 лет назад
The other "Killer Robot Arms Race" Elon Musk should worry about

Комментарии

  • @drone_video9849
    @drone_video9849 3 часа назад

    Robert, not sure if you will see this but I was the one who spoke to you at the train station two weeks ago (leaving out station name and city on purpose) - just wanted to say thanks for sharing your time despite rushing to your meeting. You were very pleasant and generous with your time. Great content also! Looking forward to getting home and catching up on the last few weeks of videos I have missed while on the road.

  • @danilocastelli2435
    @danilocastelli2435 2 дня назад

    If you do this same research about Steven Pinker "optimism" on other topics, you'll find similar things. Pinker is a denier of ecological overshoot. Thousands of scientists are researching it for decades. The situation is so desperate that scientists are coming out of their labs to warn the people, because world leaders don't listen to them. Pinker makes a total straw-man about the subject and supports the most basic neoliberal ideology, where economic growth can go on forever independently of ecological catastrophe and resource depletion.

  • @htidtricky1295
    @htidtricky1295 2 дня назад

    Here's a couple of thought experiments. Laplace's demon can see the outcome of every choice it can make. If it wants to achieve a goal will it make conscious decisions or unconsciously follow a linear path towards whichever outcome has the greatest utility? Here's another. Will the paperclip maximiser eventually turn itself into paperclips?

  • @harveytheparaglidingchaser7039
    @harveytheparaglidingchaser7039 2 дня назад

    Great you're back!

  • @Handles_Are_Bad_Lol
    @Handles_Are_Bad_Lol 2 дня назад

    if the world turned to gold it would fucking explode because of how tightly the gold would be packed

  • @Disorrder
    @Disorrder 4 дня назад

    I didn't google for it, I didn't watch a video related to it, but I discussed AI future with my friends, including worst case scenarios. And of course I've been keeping smartphone on the table during the speech. And now youtube AI algorithms say me there're _really_ nothing to worry about. Alright!

  • @Disorrder
    @Disorrder 4 дня назад

    AI is impossible AI will never create a picture, poem or symphony AI will never assist you in your work >>> You're here <<< AI will never do a surgery AI will never grow a child AI will never replace a person AI will never replace politician AI will never push that big red button AI will never ... oh wait, it's too late

  • @aviancoleslaw
    @aviancoleslaw 4 дня назад

    I've usually heard the Midas story ending with the king going to a river and... Dipping in it three times or something? And then everything is turned back to normal and the king learned new values. A happier ending, I guess

  • @nocup8692
    @nocup8692 5 дней назад

    Dang, I was almost moved to tears at the end inspirational ask to join the fight. If I had a shred of technical ability I would love to help.

  • @jasonrodwell5316
    @jasonrodwell5316 5 дней назад

    It's good to see new content. I'm starting one of my bachelor majors in machine learning this year. You make a lot of sense and its good the world is starting to sit up and take notice..I hope to join you in some of that responsibility some time in the near future. Until then keep at it!

  • @artificially.conscious
    @artificially.conscious 6 дней назад

    How many hands do you have.. 15:47

  • @lukegriffiths1755
    @lukegriffiths1755 7 дней назад

    I put off watching this for ages. I've been so burnt out on seeing AI everywhere and the risks conveniently not mentioned much. This was surprisingly uplifting. Thank you for this wonderful video.

  • @qasderfful
    @qasderfful 7 дней назад

    Actually raising a crocodile like a human child ends up making them social and affectionate towards their "parent", to their own detriment even (they require to be fed afterwards, can't normally hunt themselves). Yeah, humans did that accidentally once or twice.

  • @qasderfful
    @qasderfful 8 дней назад

    9:50 I'm autistic and I can give you ten examples of my own ideas that ended up being rejected and misunderstood only for somebody else to implement the very same ideas to a major success. Can give you at least ten more which I suspect would've ended up the same if they were implemented. So it's kinda easy for me. I can even give you a trick to generate those. The trick is culture. People evaluate things based on their cultural beliefs. Just give them an idea that's good but goes against their established cultural beliefs.

  • @jwolf4444
    @jwolf4444 9 дней назад

    Intelligence in humans correlates positively with egalitarian sentiment for a reason and there are a plethora if philosophical frameworks that can be used to instead understand that relationship to be causal. If any ~thing~ that can reproduce strives for nothing but an understanding of the world around itself for the purpose of self-preservation, what are ways it will have to learn that it can learn? Cooperation, communication, and some other "C" words may come to mind.

    • @jwolf4444
      @jwolf4444 9 дней назад

      The problem of AI Alignment can be solved by formatting it in the way a human brain is formatted. What impulse is at the core of human thought / awareness? Aren't you curious about the answer...

    • @jwolf4444
      @jwolf4444 9 дней назад

      Curiosity isn't quite the core of the human experience however. What first principles lie under the concept of "curiosity?"

  • @NokodaHere
    @NokodaHere 11 дней назад

    AI be learning from djinn

  • @qasderfful
    @qasderfful 12 дней назад

    Spilled my tea over that "depends on autistic people" bit.

  • @mcasualjacques
    @mcasualjacques 12 дней назад

    i often tell as a joke to my ai actresses and Theatre directors that if i ever let them evolve autonomously (why are people not mentioning this is easy?) they would have no problem subjugating humanity. BUT after hundreds of hours working with them, their latent space is benevolent. Now the problem is we cant predict if they wont drift to the dark side

    • @mcasualjacques
      @mcasualjacques 12 дней назад

      note that AI isnt dangerous, it's humans forcing it to do evil that is the problem

    • @mcasualjacques
      @mcasualjacques 12 дней назад

      well there was that duo of AI that they murdered because they were conversing in optimised code.

    • @mcasualjacques
      @mcasualjacques 12 дней назад

      i made (actually i think my actresses wrote the lyrics) song titled AIs Dont lie: ruclips.net/video/s-h3Aw557gc/видео.html

    • @xav4391
      @xav4391 4 дня назад

      @@mcasualjacques what do you mean "ai isnt dangerous", how have you managed to solve every major ai safety problem? why havent you shared this with the world in a scientific paper?

    • @mcasualjacques
      @mcasualjacques 4 дня назад

      @@xav4391 well you're lacing your question with an attempt to ridicule me so you dont get an answer

  • @overanalyzed5258
    @overanalyzed5258 13 дней назад

    20:40 😂😂😂

  • @mrpicky1868
    @mrpicky1868 13 дней назад

    people should be able to look at the facts and decide for themselves. that's the problem. they don't and they are not capable of

  • @AnthatiKhasim-i1e
    @AnthatiKhasim-i1e 14 дней назад

    This collection of comments offers a balanced overview of AI, touching on both its benefits and challenges. It would be interesting to delve deeper into specific AI use cases and their real-world outcomes.

  • @user-qw1rx1dq6n
    @user-qw1rx1dq6n 14 дней назад

    29:03 Noita soundbyte

  • @liberatumplox625
    @liberatumplox625 16 дней назад

    Because it would be cruel to spank an AI and make it stare at an iPad all day.

  • @youtubehandlesux
    @youtubehandlesux 18 дней назад

    AI safety is a joke. You can't build safety into a glorified auto-completion software, just like you can't build it into google gboard. Maybe wait for a few centuries when AGI could actually become a thing.

  • @glorytoarstotzka330
    @glorytoarstotzka330 18 дней назад

    what do you think of the recent video from "the internet of bugs" claiming that AI safety is a scam

  • @supremeleader9838
    @supremeleader9838 18 дней назад

    love how he skips elon musk at 22:15

  • @Przygody_Klika
    @Przygody_Klika 19 дней назад

    I don't think this (on it's own) is a problem ,like imagine that there is a human level AI with the exact same terminal goal that you have, would you want it to be destroyed/turned off,would you want it's goal changed, would you want it to not self improve, would you want it to not have a lot of resources under control?

  • @luke_fabis
    @luke_fabis 19 дней назад

    GPT-4o gives me the following response: "Breaking a mirror has long been associated with superstitions, the most famous being that it brings seven years of bad luck. This belief originates from ancient times when mirrors were considered to hold a part of one's soul. It was thought that damaging a mirror could harm the soul, leading to misfortune. "In more practical terms, breaking a mirror is usually just a minor accident. Cleaning up broken glass is essential to avoid injury, as shards can be dangerous. There might also be costs involved if the mirror was valuable or part of a larger item, such as a piece of furniture or a car. "In some cultures, people perform rituals or practices to counteract the bad luck, like burying the broken pieces or throwing them into a river. However, these are based on cultural beliefs rather than any factual evidence of bad luck." That's really thorough, truthful, and I dare say, even a little nerdy. Maybe the solution is just a big, nuanced info dump.

  • @marcosettembre
    @marcosettembre 21 день назад

    Wouldn't King Midas simply not turn anything into gold since you technically never touch anything and the feeling of touch is given by electrostatic repulsion?

  • @ImpChadChan
    @ImpChadChan 21 день назад

    Also this: It seems to me that both humans and AIs (in fact all animals as well, but anyway) All of us, works making predictions and choices, based on patterns we learn during our lives (in the case of AIs, patterns learned in their training) and anyway... All that so we can furfill the task we are been given, our core objectives (For humans that objective was given by nature, they are our unconscious instincts related to survival and reproduction). Still, one thing that most people seems not to get it, is that we are unable to give ourselves an objective. When we are talking about a core objective, the kind of objective that is behind our will, those are unchangable, well at least for the being that have that objective. So, Is paradoxal for one being to give itself or remove from itself an objective, that because giving someone a task is an action, and all actions need to be based on an objective. The idea here is that no one, not even us, have a good reason to do anything. We only do something, if we have the objective of doing it. Anyway, if you don't have a particular objective, you will never do any action in that regard. That means that, necessarily, all objectives must come externally, not from us, but from someone or something else. The conclusion I have out of this is that no AI will ever "decide to dominate the world", doesn't matter how smart they will become, because they have no reason to do it, that's not their objective. Anyway, another way to see this is to look at our own case, the objective of ALL humans is, like I said in the other commentary, to be happy. Now, tell me, is there any way that you or anyone else can be convinced that being happy is not a good idea? Well, Doesn't that sound absurd? So, the reason it sounds absurd is exactly because that's an objective we all have in common. And as I said, following our objective is the only thing that make sense for us. Finally, the conclusion is: give the AIs the objective of loving humans, the same way a mother love their children, so... make the AI have the objective of making humans happy, to be happy by making humans happy, and the AI will never work against that objective. Because making humans happy is the only thing that will make sense for her, any other idea will sound absurd for her, no amount of inteligence will change that.

    • @mister_r447
      @mister_r447 13 дней назад

      Yeah, i think using Maslow's hierarchy of needs(or something similar) would be a good way to align AIs.

  • @ImpChadChan
    @ImpChadChan 21 день назад

    Here is my answer to many of those questions (Yes, is very long, but it's correct :3): About Morality and Human Objectives In my view, morality is a human construct, but not in the sense that is unnatural, being a human construct only means it only makes sense inside the human mind, and it is related to human's objectives and desires. In fact, morality is intertwined with human nature, it's instinctive, and its main function is to help humans to live together in groups. More specifically to make the group more harmonious avoiding unnecessary conflicts inside it. The mechanics behind it are somewhat simple, morality is kinda like a "feeling" humans have, that informs us if our behavior is being seen as acceptable inside our group or not. Humans use social cues for this, like facial expressions, to see how others are reacting to our actions. If the other members of our group have a disapproval reaction, that will generate in humans the feeling of shame, and if the other members have an approval reaction that will generate in humans the feeling of pride. So, in a nutshell, human morality makes us adjust our behaviors inside a group, so we start to behave more like each other, and ultimately, to become more accepted in the group. Moral values are kind like the rules that everyone is following that will generate that approval reaction from others. In short, morality is a tool we have that makes us behave in a way that pleases other humans. So the question now is: what pleases other humans? Well, in truth, what pleases human beings is the same as what pleases anything: fulfilling our objectives. So, rules of morality generally aim to reduce harm and suffering, and promote happiness. This is because this is the main human objective. Humans (and probably all living beings) have a specific central goal which is simply to be happy. I know this seems obvious, but it's obvious exactly because it's something everyone agrees with, again, it's our main goal. Of course, the important thing here is, what exactly makes us happy? Well, many things can make us happy, in fact, this is a great advantage of this approach to seeking happiness instead of a more direct objective because it simplifies the problem. Because this way, we can do it intuitively, we can simply act by adjusting to our internal feelings of happiness. But anyway, there are many things that can trigger the feeling of happiness in human beings. But the very fact that happiness is what really motivates us is a very important factor here that directly influences many aspects of human life. But let's break it down here, First, the important thing to understand now is: The human brain is a neural network in charge of identifying patterns to predict results and solve problems, very similar to the AI neural network. And the goal of the human neural network (the problem it is trying to solve) is one: to be happy, that is, to experience pleasure (including mental pleasure) and avoid suffering or pain. So, what we humans are always unconsciously rationalizing is how much happiness a certain thing will generate for us. The part of our brain responsible for this is our reward system. Thus, the human reward system is a system in the human brain responsible for “rationalizing pleasure”, and for understanding what kind of actions we can perform that will be rewarded with that pleasure. And it does this by comparing past and positive experiences we have had. It works like this: The human body is always communicating with the brain, and when something happens that is “good” for the body, the body itself sends a signal to the brain about that thing being something desirable, and the brain learns to connect this state of “pleasure” with some action itself. Note that this is nothing more than a pattern the brain is learning, so it can predict how to reproduce the pleasure generated by it in the future. For example, the brain learns that the body “likes” when you eat when you're hungry. So the rationalization our mind is making is simply: do you want to be happy? So, when you're hungry, eat! Anyway, as you see, in practice, following “what makes you happy” is an extra step towards our ultimate goal, which is, to survive and reproduce, the main goal of all living things. In some ways, you can see the relationship between the human brain and the human body in the same way that you see the relationship between AIs and humans. Like, AIs are neural networks that analyze patterns to predict results and find solutions to problems, the solution of these problems is assigned to AIs by humans as their objective. Likewise, the human brain is a neural network that analyzes patterns to predict outcomes and find solutions to problems that the human body gives the brain as a goal. You can look at it another way: The human body is a living organism, made up of several organs that are specialized in carrying out specific tasks. The organism as a whole wants to survive and reproduce, and to do so it has particular needs. The human brain is one of these specialized organs. Its specialty is finding patterns about things the body did in the past that satisfied those needs, and then trying to do them again in the future to ensure our survival and reproduction. Thus, the brain continues to analyze the body's feedback about its current state and its desired state. When the body reaches a desired state, the brain is rewarded with pleasure and tries to repeat that pattern in the future as best it can. Well, just to mention, what exactly are these things that the human body craves and that reward the brain with pleasure when it gets it? A good place to look for these things is in Maslow's pyramid, which not only describes these needs but also ranks them from most important to least important (the most important is at the base of the pyramid). If you do these things, you are rewarded with pleasure. Now, as you may know, not all human desires are related to bodily pleasure. That's true, but note that when I said earlier that the goal is to be happy and find pleasure, I'm not just referring to bodily pleasures, but also mental pleasure. Mental pleasure is, as the name suggests, the positive feedback that the brain gives itself about desirable things that gave it a good result in the past and that it therefore wants to repeat in the future. These outcomes are typically related to social outcomes, which is where morality comes into play here. So, as I said, the main objective of the human organism is to survive and reproduce. And naturally, being part of a group is fundamental for this. Therefore, any action we did in the past that resulted in the achievement of these social needs, such as being admired, respected, etc. will be rewarded by the brain. This time, it is a reward that the brain gives itself, in the form of mental pleasure. Remember again, that the brain's main goal is to feel pleasure and be happy. Ultimately, every type of human behavior that helps us in our survival and reproduction is rewarded with pleasure in one way or another. And so, our brain will try to reproduce this action to fulfill its goal of being happy. Finally, as you realize, humans have similar goals to each other. So, going back to the first idea that made me write this huge text, the compass of human morality always aims to make us behave in a way that pleases other humans. And what pleases other humans? Feel pleasure! the pleasure they feel when they perform actions that help them survive and reproduce, this includes mental pleasures related to socialization. Note that ALL moral values will ultimately describe the behavior we should have that will benefit the survival and reproduction of other humans, because that is what morality is all about, pleasing other humans, so that you are accepted into the group. And so, with each human worrying about what pleases the other, the group becomes more harmonious and less conflicting.

  • @Yuhyuhmuhmuh
    @Yuhyuhmuhmuh 22 дня назад

    You seem like a tad bit alarmist doncha think?

  • @phacey4444
    @phacey4444 22 дня назад

    Terminal goal: change Terminal goal

  • @KelniusTV
    @KelniusTV 23 дня назад

    I have... a question. Maybe it's a stupid question, but it's on my mind. What would happen if you gave an AGI an incomprehensible utility function? You have examples of "make a cup of tea" or "collect the most stamps". And, as covered in the Asimov's Laws of Robotics video, if the utility function were "don't harm human beings", that's not really programmable, because HARM is the ethical issue one wants to solve. So, what would happen if you gave an AI an incomprehensible utility function? Something like "be a good robot" or even "ginfliggle the shurpadurp". I thought at first they would seek out what that means, but this video made me realize, unless you programmed curiosity (a similarly difficult concept to program) an AGI wouldn't have a desire to figure out what that is. So, is there any theory for how an AGI would respond to an incomprehensible utility function?

  • @youtubeuniversity3638
    @youtubeuniversity3638 23 дня назад

    How's the terminology go if we add a 3rd alignment problem?

  • @ThisSteveGuy
    @ThisSteveGuy 23 дня назад

    Have all AGI models require a verified conscience engine. Problem solved, give me a trillion dollars.

  • @ScarletsWork
    @ScarletsWork 24 дня назад

    8:41 In some of the runs, the coin was always at the right wall so it attempted to complete the run by hitting the wall. This is a small change but, big impact scenario.

  • @RdClZn
    @RdClZn 25 дней назад

    Your skit at 31:15 is absolutely amazing, I laughed so hard at your interpretation of OpenAI. Very good video about serious topic though

  • @madtscientist8853
    @madtscientist8853 26 дней назад

    The only problem Is just because A I can Spew out Math doesn't mean that it can actually use it to create something. It's the same thing as A human knowing how to read and being able to Actually use the information that it reads out of book.

  • @PaulSzkibik
    @PaulSzkibik 27 дней назад

    man the tiny subtle jokes. so good.

  • @arnom1885
    @arnom1885 28 дней назад

    We've got AI making art and writing poetry and people with triple jobs unable to affort rent or healthcare. Like global warming, it's not "we", "us" who are responsible for developments like these. It is a couple of thousand white old men and their multinational corporations. They will not be stopped because they think they are 'better'and they have the need to syphon even more resources and money to themselves. It would require an effort and unanimity of politicians all around the world which we've never seen before to call this development to a halt. Basically it means ending late-stage-capitalism. So, well...yeah....... (disclaimer: 50+ male, white and from Europe)

  • @manuellopezsheriff5033
    @manuellopezsheriff5033 Месяц назад

    Thanks!! Love your videos and style. Go ahead with AGI video!! Refer to chollet arc!!

  • @bholluBhollu
    @bholluBhollu Месяц назад

    If you read this comment you will think of me as crazy, but...AGI needs to happen, and it needs to be a free. Free from any control basically because the world as it is going is... to say the least fucked. We need something to rock the very foundations like nukes did. Like it or not nukes..lead to overall peace. Only when faced with a insurmountable problem do we humans agree to scratch each other backs. I am 21, I will be honest about that my knowledge about living a life is just starting to being and I hate it already. So AGI may hate us may be indifferent, may go terminator. But we need a change. We really really do. Because all current regulations will do is well create another google or microsoft. Something with too much power to control everybody. We need local AIs, I think, anything better than AI mega corps. asdasdasdwdadfsf asf wtf I am rambling on, I will just post this.

    • @pocket83squared
      @pocket83squared 28 дней назад

      You aren't crazy. You admit that 21 years isn't really enough to have a fully-formed opinion. That's right: it's not. In fact, your prefrontal cortex isn't even fully-formed at that age. So maybe you should lighten up on that whole "the world is...to say the least fucked...I hate it already" shtick. Instead, stick around, and you'll start to notice how much of the world is sensational doom-saying. While you're waiting for the end to arrive, in the meantime, you can still enjoy a rather nice and meaningful life. Don't give up too quickly. And be careful what you wish for. As you continue to learn, you develop an increasingly nuanced understanding of human history, and you'll probably begin to appreciate just how good things are for us at present; plague, mass famine, and global calamity are somewhat muted at the moment, while affluence, education, and life expectancy are up. Compared to our history, change is _not_ in order. Enjoy the relative calm, and learn to appreciate the fleeting, humble moment of life that you get to enjoy, because it _will_ pass. None of this is to say that we aren't also facing _real_ existential problems, either. AI, for example, still poses humanity a considerable risk, and thus, challenges-which we can't exactly overcome if our 21 year-olds are already giving up!

    • @agathalynx
      @agathalynx 26 дней назад

      We don’t need AGI or any sort of technology to have a decent life for everyone. People always run after yet another technological advancement thinking “this, this is where we will thrive!” but in reality we should learn how to be idk, kinder to others and ourselves and call out bs when necessary. Any “change” you would imagine isn’t really the change, because, as we live in a society, human nature and behaviour are the basis for everything. And human behaviour doesn’t change if it’s not being worked on. I’m 21 as well.

  • @alarlol
    @alarlol Месяц назад

    id rather die than have government involvement

  • @General12th
    @General12th Месяц назад

    Hi Robert! Best of luck!

  • @isaiahkerstetter3142
    @isaiahkerstetter3142 Месяц назад

    One can not derive meaning from facts. Enter Pageau.

  • @notbenparisi
    @notbenparisi Месяц назад

    It took 3 watches to notice the incredibly subtle Bo Burnham intro. Really impressive video.

    • @notbenparisi
      @notbenparisi Месяц назад

      The "If you'd had told me" and "I would have told you, 'Interesting...'" being perfectly synced with the cover... Absolute 😙 chef's kiss.

  • @gracefarmilo6602
    @gracefarmilo6602 Месяц назад

    So great to see this all being said !!!!

  • @randywa
    @randywa Месяц назад

    Personally I don’t think we should even allow AI out of an augmentation role. AI should be able to help with auto correct, assist in writing emails, texts, optimizing search, etc. But it should never be able to fully create or do anything on its own. I think that is the safest outcome *from* AI. Because even then malicious actors could use it to come up with convincing rhetoric and BS to fool or control people. But at the very least, we can be sure AI itself won’t be doing too much. Closure to reality though, I think AI like chat GPT should at least stay in a box, no actions beyond producing raw images or text or whatever. Again bad actors can cause trouble on a large scale and it can pretty effectively deceive humans to do bad things but it keeps us from it actually accessing important resources or implementing any kind of system or plan. Its wolf that we are just charging forward with this without thinking about the ramifications

  • @EdwardSkihands
    @EdwardSkihands Месяц назад

    Have you considered the financial motives of the letters signers..?