• vunderba 20 hours ago

    I'm not sure how this affects the premise of the article, but the "jaw dropping quote created by an LLM about NYC" reads like so much pretentious claptrap:

      Because New York City is the only place where the myth of greatness still 
      feels within reach—where the chaos sharpens your ambition, and every street 
      corner confronts you with a mirror: who are you becoming?
    
      You love NYC because it gives shape to your hunger. It’s a place where 
      anonymity and intimacy coexist; where you can be completely alone and still 
      feel tethered to the pulse of a billion dreams.
    
    If I read this even before ChatGPT was a mote in the eye of Karpathy, my eyes would have rolled so far back that metacognitive mindfulness would have become a permanent passive ability stat.

    The author of Berserk said it so much better: "Looking from up here, it's as if each flame were a small dream, for each person. They look like a bonfire of dreams, don't they? But, there's not flame for me here. I'm just a temporary visitor, taking comfort from the flame."

    • antithesizer 20 hours ago

      >reads like so much pretentious claptrap

      >my eyes would have rolled so far back that metacognitive mindfulness would have become a permanent passive ability stat

      Yes, I agree it is impressively lifelike, just like it was written by a real flesh-and-blood New Yorker.

      • righthand 17 hours ago

        I think it reads like an ad for what people think a real New Yorker sounds like. No one’s sticking around for you to ramble about the social weather.

        • randallsquared 11 hours ago

          I lived in NYC for a few years, and there are definitely lots of New Yorkers who would find themselves described in the quoted passage.

          • keysdev 5 hours ago

            Or it can be a piece from the New Yorker

            • antithesizer 11 hours ago

              Feeling implicated?

          • barrenko 7 hours ago

            Or a Ghost in the Shell perspective.

            • atoav 7 hours ago

              Sounds to me like a description of what it feels like to live in any sort of metropolis. Georg Simmel, one of the first sociologists, wrote similar observations on how city dwellers (in 1900s Berlin) would mentally adapt themselves within the context of this city in his 1903 work »The Metropolis and Mental Life«.

              The anonymity and (from the POV of a country pumpkin: rudeness) is basically a form of self-protection, because a person would go mad if they had to closely interact with each human being they met. Of course this leads to its own follow-up problems of solitude within a mass of people etc.

              But if you e.g. come from a close-knit rural community with strict (religious?) rules, a typical rural chit-chat and surveillance, simething like the anonymity of a city can indeed feel like a revalation, especially if you see yourself as an outcast within the rural community, e.g. because you differ from the expected norms there in some way.

              I am not sure about the English translation of Simmels text as I read it in German, but as I remember it it was quite on point, not too long and made observations that are all still valid today, recommended read.

              • bawolff 12 hours ago

                I'd say it reads like propaganda for a place i do not live.

                Sounds great to the target audience but everyone else just rolls their eyes.

                • geon 10 hours ago

                  Yes. I realize some people like places like that. You couldn’t pay me to live there.

                • echelon 18 hours ago

                  > reads like so much pretentious claptrap

                  Somewhere out there, there's an LLM that reads exactly like us pretentious HN commentators. All at once cynical, obstinate, and dogmatic. Full of the same "old man yells at cloud" energy that only a jaded engineer could have. After years of scar tissue, overzealous managers, non-technical stakeholders, meddling management, being the dumping ground for other teams, mile-high mountains of Jira tickets.

                  Techno-curmudgeonism is just one more flavor that the LLMs will replicate with ease.

                  • evidencetamper 7 hours ago

                    Different from your comment, that comes from your insight from your lived experiences in this community, the quote is a whole bunch of nothing.

                    Maybe the nothingness is because LLM can't reason. Maybe it's because it was trained a bit too much in marketing speak and corporate speak.

                    The thing is, while LLMs can replicate this or that style, they can't actually express opinions. LLM can never not be fluff. Maybe different types of AI can make use of LLM as part of their components, but while we keep putting all of our tokens in LLMs, we will forever get slop.

                    • echelon 3 hours ago

                      > LLM can't reason

                      > we will forever get slop

                      I think the New York quote is a lot better than whatever I wrote. A human saw that diamond in the rough and decided to share it, and that was the transformative moment.

                      A human curator of AI outputs, a tastemaker if you will, can ensure we don't see the sloppy outputs. Only the sublime ones that exceed expectations.

                      I suspect that artists and writers and programmers will pick and choose the best outputs to use and to share. And then they do, that their AI-assisted work will stand above our current expectations of whatever limitations we think AI has.

                    • jsosshbfn 6 hours ago

                      Eh, I'll take it any day over the cocaine-fuled technooptimism

                  • 13years a day ago

                    > AI is not inevitable fate. It is an invitation to wake up. The work is to keep dragging what is singular, poetic, and profoundly alive back into focus, despite all pressures to automate it away.

                    This is the struggle. The race to automate everything. Turn all of our social interactions into algorithmic digital bits. However, I don't think people are just going to wake up from calls to wake up, unfortunately.

                    We typically only wake up to anything once it is broken. Society has to break from the over optimization of attention and engagement. Not sure how that is going to play out, but we certainly aren't slowing down yet.

                    For example, take a look at the short clip I have posted here. It is an example of just how far everyone is scaling bot and content farms. It is an absolute flood of noise into all of our knowledge repositories. https://www.mindprison.cc/p/dead-internet-at-scale

                    • pixl97 21 hours ago

                      The Culture dives into this concept with the idea of hegemonizing swarms, and Bolstrom touches on this with optimizing singletons.

                      Humans are amazing min/maxers, we create vast, and at least temporarily productive mono cultures. At the same time a scarily large portion of humanity will burn and destroy something of beauty if it brings them one cent of profit.

                      Myself I believe technology and eventually AI were our fate once we became intelligence optimizers.

                      • 13years 20 hours ago

                        > Myself I believe technology and eventually AI were our fate once we became intelligence optimizers.

                        Yes, everyone talks about the Singularity, but I see the instrumental point of concern to be something prior which I've called the Event Horizon. We are optimizing, but without any understanding any longer for the outcomes.

                        "The point where we are now blind as to where we are going. The outcomes become increasingly unpredictable, and it becomes less likely that we can find our way back as it becomes a technology trap. Our existence becomes dependent on the very technology that is broken, fragile, unpredictable, and no longer understandable. There is just as much uncertainty in attempting to retrace our steps as there is in going forward."

                        • pixl97 20 hours ago

                          >but without any understanding any longer for the outcomes.

                          A concept in driving where your braking distance exceeds your view/headlight range at any given speed. We've stomped on the accelerator and the next corner is rather sharp.

                          Isaac Asimov did a fictional version of this in the Foundation trilogy.

                          • 13years 20 hours ago

                            Yes, that's an excellent description.

                          • psychoslave 10 hours ago

                            That's a very idealistic view to believe there ever was something as a point were some people had a really more clear and precise representation which was accurate of what was going to come.

                            • gsf_emergency 9 hours ago

                              Iirc Eva's Instrumentality comes from Cordwainer-Smith..

                              Subtlety missed by TFA but not necessarily Eva: Government is meant to be the Instrument (like AI), NOT the people/meat.

                              https://en.wikipedia.org/wiki/Instrumentality_of_Mankind#Cul...

                              Eva's optimization: Human instrumentality=> self governance. Which is not that ambiguous, one could say, less so than Star Child (2001)

                              "Of the people" vs <of the people>

                            • exe34 2 hours ago

                              > We are optimizing, but without any understanding any longer for the outcomes.

                              That's how the Vile Offspring is born.

                            • jsosshbfn 6 hours ago

                              I feel confident humans could not define intelligence if their survival depended on it.

                              Tbh, the only thing I see when looking at Terminator is a metaphor for the market. It makes more sense than any literal interpretation

                            • jfarmer 20 hours ago

                              John Dewey on a similar theme, about the desire to make everything frictionless and the role of friction. The fallacy that because "a thirsty man gets satisfaction in drinking water, bliss consists in being drowned."

                              > The fallacy in these versions of the same idea is perhaps the most pervasive of all fallacies in philosophy. So common is it that one questions whether it might not be called the philosophical fallacy. It consists in the supposition that whatever is found true under certain conditions may forthwith be asserted universally or without limits and conditions.

                              > Because a thirsty man gets satisfaction in drinking water, bliss consists in being drowned. Because the success of any particular struggle is measured by reaching a point of frictionless action, therefore there is such a thing as an all-inclusive end of effortless smooth activity endlessly maintained.

                              > It is forgotten that success is success of a specific effort, and satisfaction the fulfilment of a specific demand, so that success and satisfaction become meaningless when severed from the wants and struggles whose consummations they are, or when taken universally.

                              • dwaltrip 19 hours ago

                                Our societies and our people are overdosing on convenience and efficiency.

                                • the_af 18 hours ago

                                  Agreed.

                                  I remember a few years back, here on HN everyone was obsessed with diets and supplements and optimizing their nutrients.

                                  I remember telling someone that eating is also a cultural and pleasurable activity, that it's not just about nutrients, and that it's not always meant to be optimized.

                                  It wasn't well received.

                                  Thankfully these days that kind of posts are much less common here. That particular fad seems to have lost its appeal.

                              • juenfift 18 hours ago

                                > However, I don't think people are just going to wake up from calls to wake up, unfortunately.

                                > We typically only wake up to anything once it is broken. Society has to break from the over optimization of attention and engagement.

                                > However, I don't think people are just going to wake up from calls to wake up, unfortunately.

                                But here is the thing, we cannot let this bleak possibility occur.

                                It is morally wrong essentially for lack of a better phrase: "sit on our asses and do nothing"

                                Now I am aware that society has hit the accelerator and we are driving to a wall, however call me naive idiotic optimist, in fact call me a fool, and idiot, a "fucking r**ed wanker" as one called me long ago for all I damn care. But I am a fervent believer that we can change, we can stop this.

                                This has to stop because this is morally wrong to just let it happen, we've got to stop this? How I'm not sure, but I'm know for certain we have to start somewhere.

                                Because it's the right option.

                                • verisimi 21 hours ago

                                  > However, I don't think people are just going to wake up from calls to wake up, unfortunately.

                                  > We typically only wake up to anything once it is broken. Society has to break from the over optimization of attention and engagement.

                                  I don't think anyone will be waking up as long as their pronouns are 'we' and 'us' (or 'people', 'society'). Waking up or individuation is a personal, singular endeavour - it isn't a collective activity. If one hasn't even grasped who one is, if one is making a category error and identifies as 'we' rather than 'I', all answers will fail.

                                • keiferski 20 hours ago

                                  A bit of a rambling essay without much depth, but it did make me wonder: if AI tools weren’t wrapped in a chatbot pretending to be a person, would all of this hullabaloo about AI and human nature go away? That seems to be the root of the unease many people have with these tools: they have the veneer of a human chatting but obviously aren’t quite there.

                                  I tend to treat Ai tools as basically just a toolset with an annoying chat interface layered on top, which in my experience leads me to not feel any of the feelings described in the essay and elsewhere. It’s just a tool that makes certain outputs easier to generate, an idea calculator, if you will.

                                  As a result, I’m pretty excited about AI, purely because they are such powerful creative tools - and I’m not fooled into thinking this is some sort of human replacement.

                                  • Avicebron 20 hours ago

                                    I wonder if investors would have dumped the same amount of money in if it was pitched as something like "Semantic encoding and retrieval of data in latent space" vs "hey ex-machina though"

                                    • keiferski 20 hours ago

                                      Definitely, whether they intrinsically believe it or not, the hunt for AGI is driving a lot of funding rounds.

                                      • goatlover 19 hours ago

                                        They think it will eliminate most payroll costs while driving productivity way up, without destroying the economy or causing a revolution. They also don't take worries about misaligned AGI very seriously.

                                        • FridgeSeal 17 hours ago

                                          I have to wonder if they’ve really thought through the consequences of their own intentions.

                                          > eliminate payroll costs

                                          And what do they think everyone who’s been shunted out of a livelihood is going to do? Roll over and die? I don’t see “providing large scale social support” being present in any VC pitch deck. How do they imagine that “without destroying the economy” will happen in this scenario?

                                    • gavmor 20 hours ago

                                      Yes, it would—the dialogic interface is an anchor weighing us down!

                                      Yes, yes, it's an accessible demonstration of the technology's mind-blowing flexibility, but all this "I, you, me, I'm" nonsense clutters the context window and warps the ontology in way that introduces a major epistomological "optical illusion" that exploits (inadvertently?) a pretty fundamental aspect of human cognition—namely our inestimably powerful faculty for "theory of mind."

                                      Install the industrial wordsmithing assembly line behind a brutalist facade—any non-skeumorphic GUI from the past 20 years aughta do.

                                      Check out eg https://n8n.io for a quick way to plug one-shot inference into an ad-hoc pipeline.

                                      • niemandhier 19 hours ago

                                        I fully agree, where do we get the training data to create a base model? Where do we source the terabyte of coherent text that is devoid of ego?

                                      • numpad0 17 hours ago

                                        it's just not there. Whether it's text, chat, images, video - the quality sometimes appear regular to cursory look, but aren't actually up there with that of humans. That's the problem.

                                        ---

                                        This gets into cheap scifi territory but: I think someone should put people in fMRI and make them watch live archive recordings of something consistent and devoid of meanings, like a random twitch channels on same theme, or foreign language sports, and sort results by date relative to time of experiment.

                                        Expected result would be that the data is absolutely random, maybe except for something linked to seasons or weather, and/or microscopic trend shift in human speech, and/or narrator skill developments. But what if it showed something else?

                                        • ambicapter 20 hours ago

                                          > if AI tools weren’t wrapped in a chatbot pretending to be a person

                                          I don't think this is the central issue, considering all the generative AI tools that generates art pieces, including various takes on the cherished styles of still-living artists.

                                          • keiferski 19 hours ago

                                            Right, but then the conversation would mostly be akin to photography replacing portraiture - a huge technological change, but not one that makes people question their humanity.

                                            • numpad0 17 hours ago

                                              It's remarks like these that strengthen my suspicion, that continuous exposure to AI output causes model collapse in humans too.

                                        • jwalton a day ago

                                          It's somehow a little poetic that the author's chatgpt example has already been plagiarized by a realtor blog: https://www.elikarealestate.com/blog/beautiful-suffering-of-...

                                          • Garlef 21 hours ago

                                            This made me happy: It offers an interesting take on AI.

                                            (After reflecting a bit on this I think this is for the following reason: Not only does this take a step back to offer a meta perspective. It also does so without falling into the trap of rooting this perspective in the hegemonic topos of our everyday discourse (economics).

                                            Usually, takes on AI are very economic in nature: "Gen AI is theft", "We/our jobs/our creativity will all be replaced", "The training data is preduced by exploiting cheap labour".

                                            In this sense this perspective avoids the expected in not only in one but two ways.)

                                            • roxolotl 20 hours ago

                                              This piece addresses the major thing that’s been frustrating to me about AI. There’s plenty else to dislike, the provenance, the hype, the potential impacts, but what throws me the most is how willing many people have been to surrender themselves and their work to generative AI.

                                              • HellDunkel 19 hours ago

                                                This is the most perplexing thing about AI. And it is not just their own work but also every product of culture they love, which they are ready to surrender for beeing „ahead of the curve“ in very shallow ways.

                                              • myaccountonhn 20 hours ago

                                                There are quite a few practical problems that bother me with AI: centralization of power, enablement of fake news, AI porn of real women, exploitation of cheap labour to label data, invalid AI responses from tech support, worse quality software, lowered literacy rates, the growing environmental footprint.

                                                The philosophical implications are maybe the least of my worries, and maybe a red herring? It seems like the only thing those in power are interested in discussing while there are very real damages being done.

                                                • 13years 19 hours ago

                                                  A philosophical lens can sometimes help us perceive the root drivers of a set of problems. I sometimes call AI humanity's great hubris experiment.

                                                  AI's disproportionate capability to influence and capture attention versus productive output is a significant part of so many negative outcomes.

                                                • pglevy 20 hours ago

                                                  Just happened to read Heidegger's Memorial Address this morning. Delivered to a general audience in 1955, it is shorter and more accessible. Certainly not as complex as his later works but related.

                                                  > Yet it is not that the world is becoming entirely technical which is really uncanny. Far more uncanny is our being unprepared for this transformation, our inability to confront meditatively what is really dawning in this age.

                                                  https://www.beyng.com/pages/en/DiscourseOnThinking/MemorialA... (p43)

                                                  • TimorousBestie a day ago

                                                    > Instead, Heidegger compels us to do something much harder: to see the world as it is being reframed by technology, and then to consciously reclaim or reweave the strands of meaning that risk being flattened.

                                                    As a call to action this is inadequate. I have no idea what this is persuading me to do.

                                                    If I dig into how Heidegger solved this problem in his own life, well, I don’t think that should be replicated.

                                                    • gchamonlive a day ago

                                                      Leave passivity behind. You should put work into understanding what's required of you. This is why it's so easy to fall back to passivity, but there are things that must be done just because it's the right thing to do. Doing anything other than that is akin to committing philosophical suicide.

                                                      • TimorousBestie a day ago

                                                        I don’t think I have an ethical duty to parse obscurantist nonsense.

                                                        • viccis 21 hours ago

                                                          There's nothing obscure about that. You might be out of the habit of consuming challenging material, but it's definitely not a good response to react reflexively with contempt for something that takes a moment of thought to understand. There's already enough vulgar anti-intellectualism in society right now without adding to it.

                                                          • gchamonlive 21 hours ago

                                                            If that's obscurantist nonsense you aren't going to get far with Heidegger.

                                                            • daseiner1 21 hours ago

                                                              Contempt prior to investigation ought not be a point of pride, I think.

                                                          • daseiner1 a day ago

                                                            "The Question Concerning Technology" [1] mentioned in this piece is dense but can be understood by the literate layman, I think, with patience and a bit of work.

                                                            Re: "call to action", part of Heidegger's project by my read is to interrogate such phrases. I think he would refute that "action" is what we need and that orienting ourselves towards the world in terms of "action" is obscuring the Question of Being. He himself offers no real way out. In his posthumously published Der Spiegel interview [2] he himself says "only a God can save us".

                                                            I assume you're making a snide reference to his involvement with Nazism, which I'm not going to attempt to downplay or respond to here. He himself in his later life, however, went and lived a humble life in the Black Forest. Can or should we all "return to the land"? No. But his writing certainly has expanded my view of the world and my "image of thought". He is a worthwhile study.

                                                            How to Read Heidegger [3] is a great primer for any who may be interested.

                                                            [1] https://www2.hawaii.edu/~freeman/courses/phil394/The%20Quest...

                                                            [2] https://www.ditext.com/heidegger/interview.html

                                                            [3] https://a.co/d/dK5dp2t

                                                            P.S. just noticed/remembered that my username is a Heidegger reference. heh.

                                                          • bombdailer 18 hours ago

                                                            Meanwhile this article is quite clearly written (or cleaned up) by AI. Or perhaps too much dialectic with AI causes one to internalize its tendency to polish turds smooth. It just has the unmistakable look of something where the original writing was "fixed up" and what remains is exactly the thing is warns against. I understand the pull to get an idea across as efficiently as possible, but sucking the life out of it ain't the way.

                                                            • dtagames a day ago

                                                              This is fantastic. Perhaps the best philosophical piece on AI that I've read.

                                                              • DiscourseFan a day ago

                                                                I need to write something, then.

                                                                • smokel 21 hours ago

                                                                  Please do. It is unfortunate that those who shout hardest are not necessarily the smartest. There is a lot of nonsense being put out there, and it would be refreshing to read alternative perspectives.

                                                                  Edit: this was a generic comment, not judging the article. I still have trouble understanding what its premise is.

                                                              • throwawaymaths 19 hours ago

                                                                "When an LLM “describes the rain” or tries to evoke loneliness at a traffic light, it produces language that looks like the real thing but does not originate in lived experience"

                                                                does it not originate in the collective experience ensouled in the corpus it is fed?

                                                                • djoldman 20 hours ago

                                                                  1. I suspect that the vast majority couldn't care less about the philosophical implications. They're just going to try to adapt as best they can and live their lives.

                                                                  2. LLMs are challenging the assumption, often unvoiced, that humans are special, unique even. A good chunk of people out there are starting to feel uncomfortable because of this. That LLMs are essentially a distillation of human-generated text makes this next-level ironic: occasionally people will deride LLM output... In some ways this is just a criticism of human generated text.

                                                                  • Barrin92 20 hours ago

                                                                    > LLMs are challenging the assumption, often unvoiced, that humans are special, unique even. A good chunk of people out there are starting to feel uncomfortable because of this.

                                                                    I'm gonna be honest, after Copernicus, Newton and Darwin it's a bit hilarious to think that this one is finally going to do that worldview in. If you were already willing to ignore that we're upright apes, in fact not in the center of the universe and things just largely buzz around without a telos I'd think you might as well rationalize machine intelligence somehow as well or have given up like 200 years ago

                                                                  • niemandhier 19 hours ago

                                                                    In Heideggers philosophy objects and people are defined by their relations to the real world, he calls it “ in der Welt sein”.

                                                                    Llms pose an interesting challenge to this concept, since they cannot interact with the physical world, but they nevertheless can act.

                                                                    • poopiokaka 18 hours ago

                                                                      Clearly written by AI

                                                                      • bowsamic 20 hours ago

                                                                        This is basically my experience. LLMs have made me deeply appreciative of real human output. The more technology degrades everything, the clearer it shows what matters

                                                                        • 13years 20 hours ago

                                                                          I think it is creating a growing interest in authenticity among some. Although, it still feels like this is a minority opinion. Every content platform is being flooded with AI content. Social media floods it into all of my feeds.

                                                                          I wish I could push a button and filter it all out. But that's the problem we have created. It is nearly impossible to do. If you want to consume truly human authentic content, it is nearly impossible to know. Everyone I interact with now might just be a bot.

                                                                        • alganet 12 hours ago

                                                                          > In Neon Genesis Evangelion, the “Human Instrumentality Project” offers to dissolve all suffering through perfect togetherness.

                                                                          That is what Gendo says, and it is obviously a lie. It's an _unreliable universe_ story: you don't really know anything. Even the most powerful characters lack knowledge of what is going on.

                                                                          All the endings, of all revisions, include Gendo realizing he didn't knew something vital (then, after that, the story becomes even more unreliable). If that's the goal of the story (Gendo's arc of pure loss despite absolute power), it's not ambiguous at all.

                                                                          So, very strange that you used the reference to relate to AI.

                                                                          • aspenmayer 9 hours ago

                                                                            Yeah, I would agree. NGE is like a reverse Garden of Eden story, where the Adam and Eve escape from society to some kind of idealistic paradise because they were the ones chosen by humans who essentially had created a modern day reverse Tower of Babel to become gods (remember Tokyo-3 is underground skyscrapers that can pop above ground?), and created their own fake plagues in the form of the Angels as justification to create the EVAs, which were merely a cover story to fight their fake plagues, but which were actually necessary to trigger the Human Instrumentality Project, so that they could become gods for real.

                                                                            NGE is an allegory for our present, with something like 9/11 truther government coverup false flag attack paranoia, in the form of the one kid who always has a VHS camcorder, who is kind of a stand in for a conspiracy theorist who is accidentally correct, combined with Christian apocalyptic eschatology, combined with Japanese fears about being the only non-nuclear armed modern democracy in some hypothetical future, and some mecha fights and waifus. It’s us, the little people versus industrial gods.

                                                                            Gendo was a true believer, he just became jealous of his own son because Shinji was able to pilot the EVAs, and thus was forced to confront his own poor treatment of Shinji. Once Gendo realized that S.E.E.L.E. (the U.N. group formed to fight the Angels) may not know what they’re doing, before they can trigger Instrumentality with themselves via the synthetic Angels, Gendo triggers Instrumentality with Shinji and Asuka. So in that way, I would say that Gendo was lying because he wanted to trigger Instrumentality himself so he could bring back his dead wife, but had to settle for indoctrinating his son to do it by proxy.

                                                                            Gendo was lying, but not about the fact that the Human Instrumentality Project does what it says on the tin, but about how many eggs he had to break to make that omelet. Rather than trust Instrumentality and thus the literal future of humanity to literal faceless bureaucrats, Gendo put his son on the throne and told him to kill their false gods and become one himself, and trusted that love would conquer all in the end. Gendo lied to Shinji so that he could tell him a deeper truth that he could never say aloud in words, especially after the loss of his wife, that he loved his son, and he did that by allowing Shinji to create a world without pain, whatever that meant to him. Gendo was a flawed man and a genius who was duped to become a useful idiot for the deep state, a true believer of his own bullshit, but he loved his son in his extremely stereotypically warped Japanese way, and because his son was able to accept that love and learn to love himself, Shinji was able to love the world and his place in it, and thus achieved enlightenment via his realization that heaven was on earth all along.

                                                                            “God’s in his heaven, all is right with the world,” indeed.

                                                                            If anything, AI is part of what might one day become our own Human Instrumentality Project, but in and of itself, I don’t think it’s enough. AIs aren’t yet effectively embodied.

                                                                            I think Final Fantasy VII would be a better story/setting to explore for ideas related to AI. Sephiroth is a “perfect” synthetic human, and he basically turns himself into a paperclip maximizer that runs on mako energy by co-opting Shinra Corp via a literal hostile takeover of the parent company of the lab that created him.

                                                                          • layer8 17 hours ago

                                                                            > 2003 Space Odyssey

                                                                            ???

                                                                            • gkanai 3 hours ago

                                                                              yeah that made me stop too. Who, writing on these topics, does not know that the movie is 2001: A Space Odyssey.

                                                                            • johnea 17 hours ago

                                                                              > What unsettles us about AI is not malice, but the vacuum where intention should be. When it tries to write poetry or mimic human tenderness, our collective recoil is less about revulsion and more a last stand, staking a claim on experience, contradiction, and ache as non-negotiably ours.

                                                                              Vacuous BS like this sentence make me think this whole article is LLM generated text.

                                                                              What unsettles me isn't some existential "ache", it isn't even the LLM tech itself (which does have _some_ very useful applications), it's the gushing, unqualified anthropomorphization by people who aren't technically qualified to judge it.

                                                                              The lay populous is all gaga, while technically literate people are by and large the majority of those raising red flags.

                                                                              This topic of the snippet about "life in NYC": is the perfect application for the statistical sampling and reordering of words, already written by OTHER PEOPLE.

                                                                              We can take the vacuous ramblings of every would-be poet who ever moved to NYC, and reorder them into a "new" writing about some subjective topic that can't really be pinned down as correct or not. Of course it sounds "human", it was trained on preexisting human writing. duh...

                                                                              Now, try to apply this to the control system for your local nuclear power plant and you definitely will want a human expert reviewing everything before you put it into production...

                                                                              But does the c-suite understand this? I doubt it...

                                                                              • freen an hour ago

                                                                                Ted Chiang: “I tend to think that most fears about A.I. are best understood as fears about capitalism. And I think that this is actually true of most fears of technology, too. Most of our fears or anxieties about technology are best understood as fears or anxiety about how capitalism will use technology against us. And technology and capitalism have been so closely intertwined that it’s hard to distinguish the two.”

                                                                                https://www.nytimes.com/2021/03/30/podcasts/ezra-klein-podca...

                                                                                • akomtu 21 hours ago

                                                                                  AI will be the great divider of humanity. It's one of those ideas that can't be ignored or waited out, and everyone will need to take a decisive stance on it. Human civilization and machine civilization won't coexist for long.

                                                                                  • httbs a day ago

                                                                                    Great read

                                                                                    • BlueTemplar 13 hours ago

                                                                                      And just how do you know LLMs don't have a soul, hmm ?

                                                                                      "Uncanny valley" is interesting here because I am pretty sure I would have failed this "Turing Test" if stumbling on this text out of context. But yeah, within context, there is something of this rejection indeed... And there would be probably a lot more acceptance of AIs if they were closer to humans in other aspects.

                                                                                      • deadbabe 21 hours ago

                                                                                        We never valued the human element in the work that surrounds us. Do you care that the software engineer who produced the CRUD app you use everyday had a “craftsman mentality” toward code? Do you care about the hours a digital artist spent to render some CGI just right in a commercial? Do you appreciate the time a human took to write some local news article?

                                                                                        Probably not, you probably didn’t even notice, and now it’s over. It’s too late to care. These things will soon be replaced with cheaper AI pipelines and much of what we consume or read digitally will be proudly AI generated or at best only merely suspected of being AI generated. Did you know that soon you’ll even be able to install browser plugins that will automatically pay you to have AI insert ads into comments you write on popular websites? It’s true, and people will do it, because it’s an easy way to make money.

                                                                                        Reversing this AI trend means everyone should just do things the hard way, and that’s just not going to happen. If no one cares about how you do your work (and they really don’t give a fuck) you might as well use AI to do it.

                                                                                        • micromacrofoot 20 hours ago

                                                                                          that's not the point at all, the question is that in the face of this inevitability of slop how do we create meaning for ourselves

                                                                                          • deadbabe 20 hours ago

                                                                                            By choosing to do things the hard way. There is no meaning without struggle. And keep in mind some of that struggle will mean accepting the fact that others using AI will surpass you, and even be praised.

                                                                                            • codr7 2 hours ago

                                                                                              Surpass on a path that leads in circles, sooner or later nothing will work and few will remember how to use their brains for anything beyond slop prompting.

                                                                                        • neuroelectron a day ago

                                                                                          My chatgpt doesn't like nyc that much:

                                                                                          New York City, as a global symbol, exports the myth of America—its exceptionalism, hustle culture, capitalism-as-dream, fashion, Wall Street bravado, media dominance, cultural swagger. NYC has always been a billboard for "brand America," selling a narrative of limitless opportunity, grit-as-glory, and urban sophistication. Think Times Square's overstimulation, Broadway's fantasy, Wall Street's speculation, and how these are consumed worldwide as aspirational content.

                                                                                          But what's exported isn't necessarily real—it’s hype. The marketed dream, not the lived reality.

                                                                                          “...and its biggest import is grime and grief”

                                                                                          In contrast, what flows into NYC is the cost of that image: the labor of the marginalized, the psychological toll, the physical debris. “Grime” evokes literal pollution, overwork, and class stratification; “grief” brings in the emotional fallout—displacement, burnout, violence, economic precarity, and cycles of trauma.

                                                                                          NYC absorbs the despair of a world it pretends to uplift. Refugees, artists, outcasts, and exhausted believers in the American Dream all converge here, only to be consumed by the very machine that exports the myth of hope.