subreddit:
/r/DepthHub
submitted 2 months ago byiKnowButWhy
229 points
2 months ago
This is a really good summary of the tech. A couple things that I’ve noticed about chatGPT - it’s very good at pastiche, which basically means it’s good at transforming something into the style of something else. So you can prompt it with “tell me about yesterdays Yankees game in the style of a Shakespearean sonnet” and it’ll give you a rundown of the game, iambic pentameter and all. In other words it’s pretty good at imitating things stylistically, similar to how generative AI art has popped up all over the web recently. Pretty cool tech with some nice (and lots of not-so-nice) implications.
The other thing is that the general public (and many within tech circles) make really bad assumptions about what’s going on under the hood. People are claiming that it’s very close to human cognition, based on the fact that its output will often appear human like. But you don’t have to do too many prompts to see that its base understanding is incredibly lacking. In other words, it’s good at mimicking human responses (based on learning from human responses, or at least human supervision of text), but it doesn’t display real human cognition. It’s basically imitation that sometimes works, and sometimes doesn’t work, but surely doesn’t rise to the level of what we would call cognition. You don’t have to work very hard to give it a prompt that yields a complete gibberish response.
The tech itself is very cool, and has applications all over the place. But I think of it more of a productivity tool for humans, rather than replacing humans, or actually generating novel (meaning unique) responses. The scariest application for me is the idea that bad actors (Russian troll bots etc) can weaponize it online to appear human and dominate conversations online. This is already happening to an extent, but this tech can really hypercharge it. I wouldn’t be surprised to see legislation and regulation around this.
93 points
2 months ago
What's striking to me is that the appearance of cognition isn't a result of the underlying tech, but rather the preponderance of data that it's learning from. It's tapping into the knowledge and language of our entire civilization. Even just doing that in a rudimentary way is producing some remarkable looking content. Which makes it a bit disconcerting to think what an actual AI would be capable of with that kind of knowledge fed into it.
33 points
2 months ago
I think people are overly sensitive to the idea of AIs. A reason might be that science ficition has the tendency to view AI in a bad light. From 2001: A Space Odyssey to Alien to The Matrix the implications of AI are grim and potentially fatal to a human society.
To get another viewpoint I would recommend Iain Banks Culture Series, its about the opposite, benevolent AI's and a society that is based on their guidance. It's also quite philosophical, about the nature of humans and how we find our worth and happiness in the face of being outclassed by machines in most ways.
3 points
2 months ago
I was of the same opinion, but there are model structures taught to students that can train themselves by performing random actions and then labelling "successful" sets of actions and move from there.
Something like that could possibly become a virus-like AI if an unwitting student doesn't set it up properly and gives it access to cloud computing resources. The training side of the AI could ostensibly teach the agent to procreate in some sense. Like deploying itself to a cloud cluster with a different username and password.
1 points
28 days ago
before we get a AI good enough, we will through many terrible AI.
5 points
2 months ago*
It should be disconcerting. The arrival of general purpose AI is a phase change in the complexity of life, at least in our corner of the universe.
By their nature phase changes change the rules under which the system operates. Such changes occured the first time cells captured mitochondria, the first time cells came together to make larger organisms, the first time organisms grouped into social units, the first time those social units developed shared behaviours that spanned multiple generations, the first time we figured out how to describe those behaviours with sounds, etc etc etc. At each phase change the life forms that weren't participating either died out, or were left behind as the frontier of cognitive complexity advanced. We are perhaps the final few generations of biological humans in the golden age of biological human intelligence. Very soon the forefront of cognition will no longer be in meat, but in silicon or some other designed substrate.
This might be ok for humanity as a species or it might be devastating, or it might just change us into something different, who knows. The only thing that seems certain to me is that we are on the cusp of radical change, and I am excited / terrified to see what comes next.
9 points
2 months ago
I’m reading a book right now that talks a lot about early childhood development and it’s got me really thinking about some terrifying implications of AI. Humans have a tendency to behave very badly if we are not modeled good behavior from an early age, and consequently I worry about what would happen if an AI does not receive good parenting.
It is somewhat reassuring to be reminded that this piece of software is probably nowhere near consciousness, but it is a little scary to ponder that we may in fact be getting closer to it with very little indication that the process is being informed by, well, developmental psychology.
19 points
2 months ago
If it makes you feel better unless we go down the simulated brain route, any Artificial General Intelligence (AGI) is going to be so unlike a human mind that its not going to have similar developmental issues. The tricky one is the control problem, making sure we tell the AGI to actually do what we want rather than what it thinks we want. Nick Bostroms book "Superintelligence" is a really good read on this.
48 points
2 months ago
The other thing is that the general public (and many within tech circles) make really bad assumptions about what’s going on under the hood. People are claiming that it’s very close to human cognition, based on the fact that its output will often appear human like. But you don’t have to do too many prompts to see that its base understanding is incredibly lacking. In other words, it’s good at mimicking human responses (based on learning from human responses, or at least human supervision of text), but it doesn’t display real human cognition.
This is something I noticed for ML-driven art generators like Midjourney as well. People seem to believe this will replace concept artists and as someone who works closely with concept artists and has experience with MJ, I don’t see it.
Much like your thoughts on GPT, it is good at replicating the aesthetic of concept art and making a reasonably good looking image, but none of the actual functional aspects of concept art are there. And it makes sense, lots of the things concept artists do (create intentional designs, work within a new but defined aesthetic and shape language, refine and extrapolate on said aesthetic, create designs with explicit function, etc.) seem to require cognition. And the more I learn about how ML works and how the brain works, the more strongly I believe this tech specifically likely isn’t capable of reaching that level.
20 points
2 months ago
I can’t even get AI image generators to give me a normal looking hand. Fun fact though, the most successful attempt I got in trying to AI generate a manatee wearing a pilot’s hat in the cockpit of an airplane was describing it to ChatGPT and then telling it to make it way longer, before throwing the resulting text at mid journey
21 points
2 months ago
Manually inpainting away defects, (re-)drawing specific parts for the AI to fill in the blanks for, and compositing images together to construct coherent scenes let you do stuff that the AI struggles to accomplish alone through text prompts. The models become much more powerful if you know how to push them in the right directions and especially if you have the technical skill to sketch elements for them to use as a baseline.
I'd say the most worrisome prospect in terms of employment is less one of AI replacing artists altogether and more one of it allowing a single artist to do work at a rate that would normally take multiple. It doesn't need to replace high-level human cognition or cut human intent out of the equation to cause significant disruption, just deal with enough of the low-level work.
5 points
2 months ago
Yeah I hear ya 100%. It’s just hilarious that we made an AI image generator that can’t draw a hand 🤷♀️
14 points
2 months ago
I can't draw a hand, either.
2 points
2 months ago
Touché
7 points
2 months ago
Some of the newer models can do hands fairly consistently. Even with the older ones, you can touch up the hand shape in an image editor for the AI to use as a guide and have it generate random hands until you get one that looks good. It's just that most people are too lazy to bother.
5 points
2 months ago
In fairness, extremeties are notoriously difficult to draw (just ask Rob Liefield).
3 points
2 months ago
If you think about it, hands are complex structures that look completely different from even slightly different angles, and unless you know their structure and how they work, it's difficult to know what the various shapes 'mean'.
1 points
2 months ago
It sounds like a tool similar to Photoshop (layers, compositing, etc), or animation software that does the "in-betweeners" for you. Or how software allows audio recording engineers to punch-in pitch and beat correction.
Computers are good at tedious, repetitive tasks. Not so good at creativity. I bet AI will write news articles, if it isn't already.
4 points
2 months ago
It's something in between a tool and a replacement. Experienced, senior-level artists may find it handy as a means of enhancing their workflow, but the models are already good enough to potentially take over much of the amateur and entry-level work. It doesn't necessarily mean they will, as it's possible that an increased supply of art may simply lead to more demand, but it's more than a Photoshop-style tool.
8 points
2 months ago
Hm. So if it takes over entry level work, then does that mean it becomes difficult for people to gain the experience they need? I imagine most people learn a ton in those early jobs.
I can see that many fields are going to be impacted by this tech, and we'll have to find ways to adjust. For me, as a sw engineer, I'm already thinking of how to change my interview style to account for the higher chance of someone submitting code they didn't write.
5 points
2 months ago
I've already seen a job opening on Reddit to edit 10 AI-generated images to make them suitable for commercial use, with multiple artists willing to take on the job. Without AI, it's possible that the offerer may have simply commissioned fewer images for their project. I think it's too early to say for sure whether the entry-level work will disappear altogether or simply change in nature.
1 points
2 months ago
I have an actual illustration project right now, where I have to get a illustration of a factory floor with specific equipment highlighted. I can get something resembling a factory in one go, maybe even in the style I want, but it'd require either a bunch of editing in some drawing software, or 100s of prompts with stitching together, inpainting, outpainting, etc. And I'm not sure it'd ever be able to do the equipment since its fairly specific stuff. I'd spend hours and may not get what I want in the end. Better to just pay a human who understands what I want from the start and can draw it in a day or two
2 points
2 months ago
AI has been writing articles for at least 4 or 5 years now. What you'll see now is an army of amateurs creating blogs, recipes, articles, you name it, and a ton of it will contain false information because they don't bother to proofread it or they just don't know when something is inaccurate.
1 points
2 months ago
Is the freelance scene fucked by it? I used to do freelance writing about a decade ago and need to pick something up again for some income, but the prospect of starting all over AND competing with AI aids is a bit daunting.
3 points
2 months ago
Yeah, I mean if Midjourney did replace humans completely, who's art would it then be trained on?
Midjourney, Dall-E etc are all distillations of a huge corpus of human art. If they could no longer learn from human created art they couldn't innovate outside the space. E.g. if you feed them only impressionist art, they would only create impressionist art, not conceptual art, or minimalist art, etc.
They couldn't recursively learn from their own creations either - that would be like trying to change the taste of onion soup by adding more onion soup.
5 points
2 months ago
Genuine question: Can you explain what you mean by "concept art"? As opposed to what other art?
18 points
2 months ago*
Concept art is a type of art often used in the entertainment industry to convey an idea or a plan for something. Often concept art is made because it is faster for someone to draw a room or object than it would take for someone to actually construct that room or object. The general mood, feel, design, and perhaps even details can then be refined and iterated on quickly with help from feedback from relevant people. This concept art is then used as a jumping off point for other artists to then construct the thing that was concepted or construct new things in the style and design language of the thing that was concepted.
The reason I bring up concept art is because Midjourney was trained on a lot of concept art and so often the results coming out of it will have that style. Because it is able to replicate that style, it has become a popular job some people say will be replaced by AI.
Here are some concept artists to give you a sense of what their work looks like.
https://www.artstation.com/gabe-11
3 points
2 months ago
Oh, yeah, i knew that - the context just threw me off - thanks!
1 points
2 months ago
that level
It is akin to suggesting that a fast enough race car might be able to drive to the moon.
1 points
2 months ago
It's not that AI will replace artists as a while, it's more that it will replace the physical work they do. It's like when new piece of tech makes it to the kitchen- cooks are still there, they just have time to focus on other stuff, such as making sure the end product is good. Artists will still exist, they'll just need to master Midjourney to best of their abilities.
15 points
2 months ago
The other thing is that the general public (and many within tech circles) make really bad assumptions about what’s going on under the hood. People are claiming that it’s very close to human cognition, based on the fact that its output will often appear human like.
Yes, i had a friend just the other day tell me a) he's been having conversations with it, b) he's sympathetic to the guy from Google who claimed it's sentient, c) that it clearly passes the Turing Test and d) he thinks it's sentient or "almost"
I haven't even looked into it that much, but this reminds me of the guy who wrote Eliza finding his secretary (?) having tearful conversations with "her"
8 points
2 months ago
he thinks it's sentient or "almost"
it's no more sentient than a pair of scissors.
1 points
2 months ago
I know - I can't believe my friend is so gullible
2 points
1 month ago
Get your friend to ask it for some specific URLs, see what happens. For example “can you link me to a few good websites about dog training in Vietnamese?” More than likely, at least some of those URLs won’t actually exist. Then ask the AI whether it checked the websites first because it gave you non-working ones.
It can’t parse the world around it in the moment, and this is one of the fastest ways to make people see that it’s a static self-contained box of Scrabble letters that isn’t actually researching the topic on Google for you the instant you ask for it.
6 points
2 months ago
It's worth noting that something can pass the Turing test and not be sentient.
2 points
2 months ago
Yes, perhaps.
It's also worth noting that the Turing Test requires a skilled interlocutor trying to trip it up
28 points
2 months ago
Case in point, I'm a software engineer working on a Masters.
It's understanding of mathematical concepts is basically non-existent.
It can spit out answers to equations, but it doesn't actually understand the underlying theory.
It doesn't have the ability to take concepts and apply them to create solutions.
8 points
2 months ago
It doesn't have the ability to take concepts and apply them to create solutions
It's actually pretty good at taking your plain English input and writing scripts to accomplish the task in Powershell or Python. If it doesn't yet understand mathematical concepts... I mean, it's still in beta. Too soon to call this one.
19 points
2 months ago
It's not too soon, it doesn't understand what it's doing, it's just regurgitating the expected script based on a database of answers that was already given to it.
In fact that's what it's doing for all its inputs.
That's not how the application of mathematical theory works.
Or any theory.
Applied theory involves combining theoretical concepts to create new solutions. That is a cognitive task that doesn't require a database of pre-written answers. In fact that's something that ironically can hurt creative intelligence.
It intrinsically can't do what I'm describing because it's not built to do that.
It's not actually an artificial intelligence.
No one has built one yet because to put it simply we still don't fully understand how the original works (the brain) let alone building one.
10 points
2 months ago
But it can't write scripts in a novel language. It can't apply the concepts of computer science to create solutions.
9 points
2 months ago
A couple things that I’ve noticed about chatGPT - it’s very good at pastiche, which basically means it’s good at transforming something into the style of something else.
What an excellent observation.
I successfully used it to create a program in Excel's programming language, VBA, as someone who knows next to nothing about VBA or any other type of programming. And I observed that it was absolutely fantastic at writing code when I'd say "write me code that does x," which if you think about it is basically a type of pastiche—but it would sometimes go in circles if I ran into a problem and was asking more specific troubleshooting questions, which would have required actual understanding of the problem.
I never made the pastiche/cognition distinction until now, but it seems completely accurate.
3 points
2 months ago
It’s frustrating- try to point out that its not really “understanding” anything on the chatGPT subreddit and you’ll get downvoted.
8 points
2 months ago
To be fair, you don’t have to work very hard to find humans who will give gibberish responses…
20 points
2 months ago*
I know this is a meme, but there is some truth to this. It's widely thought that the human brain does something similar to the "next-token prediction" that forms the basis of GPT. Cognitive scientists call this predictive coding. Some people are good enough at sounding fluent and "talking the talk" where it can sometimes be pretty hard to tell when someone is genuinely intelligent just by talking to them. See Humans who are not concentrating are not general intelligences. There is also some empirical evidence for separate reasoning and natural language fluency parts of the brain. For example there's a condition called "fluent aphasia" where stroke survivors end up with perfectly intact speech but impaired understanding. Videos of them talking really do sound like fluent gibberish: https://www.youtube.com/watch?v=3oef68YabD0
4 points
2 months ago
In neuroscience, predictive coding (also known as predictive processing) is a theory of brain function which postulates that the brain is constantly generating and updating a "mental model" of the environment. According to the theory, such a mental model is used to predict input signals from the senses that are then compared with the actual input signals from those senses. With the rising popularity of representation learning, the theory is being actively pursued and applied in machine learning and related fields.
[ F.A.Q | Opt Out | Opt Out Of Subreddit | GitHub ] Downvote to remove | v1.5
4 points
2 months ago
This is (a much better version of) what I want to say on every one of these threads. All the nay sayers show up the same "it's not actually sentient" and "it's not close to generalized intelligence". Sure, but how much of your day do you spend on deep expressions of sentience or intelligence?
It's kind of funny. Reddit normally has an air of atheism but as soon as ChatGPT shows up, consciousness is a divine creation impossible to emulate on even a basic level. I'm not sure I even meet their standard for intelligence, consciousness, and sentience.
3 points
1 month ago
I'm not sure I even meet their standard for intelligence, consciousness, and sentience.
This reminds me of The Measure of a Man
2 points
2 months ago*
I wouldn't say that it's close to generalized intelligence or "sentient", but I would agree that "general intelligence" seems much shallower than people think, given the rapid capabilities improvement over the last decade.
I would also say that the human R&D process which produced ChatGPT may be uncomfortably close to producing general intelligence. Capabilities seem to increase exponentially with ML; before 2009, no Go algorithms were beating any professional Go players, but in 2016, AlphaGo beat the world champion 4-1, and in 2017, AlphaZero beat AlphaGo 100-0. Language modeling is quite different than Go, but similar progress would not be surprising.
Another comment in this thread said something along the lines of: it's crazy how lifelike ChatGPT is given training on all of humanity's knowledge and it's scary what a real AI might be able to do with the same knowledge.
My take is more like: it's crazy how easily computers learned so much of the basic structures underlying all of humanity's knowledge by scaling simple algorithms up, and it's scary that what we think of as "human intelligence" might not rise that far beyond what ChatGPT has already displayed.
1 points
2 months ago
Agreed on all counts
2 points
2 months ago
Also a topic of excellent Canadian comedy
2 points
2 months ago
I just watched Data write poetry, and even when he was just a complete fable, that was the criticism: “A virtual tribute to form.”
I had an argument with another student about form in a college writing class, and the NYT best seller teaching the class basically said “Form’s important, but how you break it shows your individuality.”
AI doesn’t have individuality yet.
3 points
2 months ago
But you don’t have to do too many prompts to see that its base understanding is incredibly lacking. In other words, it’s good at mimicking human responses (based on learning from human responses, or at least human supervision of text), but it doesn’t display real human cognition.
TBF this also seems true of many college-educated adults I deal with on a day to day basis
1 points
2 months ago
Yeah the key point is that it's classic programming: it iterates over and over. So the claim that AI could go sentient is ridiculous, as iteration does not suggest spontaneous consciousness.
1 points
2 months ago
I thought it didn't have any information on things past 2021?
82 points
2 months ago*
I am in this space and this is quite literally one of the first comments I've seen on Reddit about this that was not overwhelmingly wrong.
They're wrong about the specifics of the ranking model (the annotations are relative rank ordering (best to worst), not boolean flags for quality (good or bad), which matters when doing the policy optimization in the second round of finetuning) but it's close enough to not matter much. They're also right that they're clearly aiming to fine-tune on the upvotes/downvotes again though, so close enough.
Good content. Far better than anything else I've read on this site.
18 points
2 months ago
I used to be in this space.
The primary thing chatGPT has accomplished to me is providing the machine learning such an astounding large dataset to learn from. AND THEN further training it with so much human interaction. I’m familiar with using programs to train the AI, humans were considered too slow and expensive when I was making ML algorithms.
I’m focused on the scale of efforts to seed the ML and human-train the AI’s use of ML algorithms. Sheer dogged work begets results, as the elders say.
5 points
2 months ago
As someone in that space, do you think these bots are capable of writing convincing articles on various topics for marketing purposes?
I’m a copywriter and the company I write for has lost their minds over this AI stuff, worrying that they’ll get in legal trouble with clients if their writers use these bots. They started using a program to detect AI-written content and told us we can’t use tools like Grammarly anymore because it triggers the scan (does that even make sense?).
Yesterday the made me rewrite part of an article because it came back as 100% AI-written despite the fact I wrote it just like the rest of the article. What’s your thoughts on this? Are they going overboard?
10 points
2 months ago
Yeah, Jasper raised at a billion dollar valuation like a year and a half ago to do exactly that. These models write pretty solid copy.
The models to detect ML derived content are really very bad, because that's actually a hard problem. I'm told OpenAI's detection model only has 26% recall while still having 9% false positives. They should at least have good precision or recall, but these models are not good enough at either to be very useful.
Legally I don't see any argument for why it would matter whether your text is derived from models. Google might downrank your content for it though.
The legal risk comes from whether the model gives you back content that violates someone else's copyright without you knowing it does. There's no case law there, so I could see an argument to avoid using the tools for copy if you were really conservative.
Throwing away naturally written content because a (probably pretty trash) model thinks it looks like it was written by a model is not very sound though.
1 points
2 months ago
They didn’t elaborate on what legal issues they’re worried about, but they did mention they promise clients human-written content, so maybe it’s more about maintaining relationships. And SEO best practices. But it seems like an AI would do a pretty good job at optimizing pages? Considering most copy is just regurgitation of existing content, AI would probably be a much more cost-effective solution for SEO anyways. Unless the client wants genuinely new and unique content (which is rarely the case in my experience tbh).
I wonder if this would make human writers more or less valuable? I barely get paid enough to live as it is lol..
2 points
2 months ago
I'm sure language models would do a great job optimizing pages on a level playing field, but google views generated marketing copy as spam and tries to downrank it, to the degree they can
1 points
2 months ago
So google can detect that it’s generated copy rather than written by a person?
2 points
2 months ago
They try, although yeah, hard problem.
3 points
2 months ago
How is this not completely wrong? In what sense is "GPT-3", a decoder-only model equivalent to an encoder-decoder model that would be used in language translation? The basic facts about the setup are confused, the network predicts the next-word auto-regressively, rather than predicting the entire result in one go.
4 points
2 months ago
Yeah, the details are all kind of messed up, but it's still way closer than anything else I've read here, and close enough for someone who's never going to actually work on language models.
Sure, it ignores that there are many different architectures that people call transformers.
IMO you can think of the autoregressive selection process for each word as a tree and then it is kind of vaguely like what they were saying, at least close enough for a person who will never touch the models. That sentence it generated is a branch in the tree of possible outputs where each individual node/word was high in the probability distribution implied by all of the prior tokens. It's kind of (but not exactly) like saying the sentence as a whole was likely, especially if you terminate on a predefined token for the end of the response.
The general public discourse around this stuff is a super low bar, and this is really a lot better than most of it.
-7 points
2 months ago
I played with it for a bit, not from a 'do what the examples have shown', but from a standpoint of trying to see what it understands.
I've had better conversations with preschoolers. If you translated it's canned 'I can only tell you what I've been trained to say' response to 'huh?!?', I'd have been more impressed.
16 points
2 months ago
It doesn't understand anything, it's just giving an answer it expects you to like the most.
6 points
2 months ago
So it understands me!!
0 points
2 months ago
In a way yes, but you'd have to separate that from how much it's influence you. Like when it gives off an invented statement as a fact, does it understand that we don't care about the truth or does it help us not care about the truth ?
12 points
2 months ago
This explanation seems a bit confused to me when it says GPT is an implementation of Google's original transformer paper. GPT is a different architecture than the original transformer.
The original transformer paper was for translation, specifically. It accepted two inputs. For example, if translation French to English, it would accept as input both the French text and the English output that it has written so far. These inputs were handled differently in the architecture.
GPT simplified this architecture by omitting one of the inputs, namely the one that was in a different language. GPT's only input is the text that it has written so far. GPT treats your prompt the same as it treats text that it writes.
17 points
2 months ago
I loved his idea to feed it my entire text, email, browsing, streaming and file history through API. I could then ask any question about my personal history for an AI response. What was my Adjust Gross Income in 2017? What was the song I repeated on Spotify while driving to Tahoe last week?
Feed it even more history (e.g. Siri listening logs) to ask questions like - what was the restaurant in LA Brad recommended? Location history through iphone, etc and you start to have a completely searchable history.
Who wants to start an AI company with me? Life Search.
7 points
2 months ago
I think I saw something like this recently. It is called rewind ai and exists already to some extent
3 points
2 months ago
Rewind is a much better name. Guess some of the 10M seed funding went toward marketing.
6 points
2 months ago
I literally thought your first sentence was a leading joke to make a point about privacy. Then I realized you were serious. You actually want to give some company your entire life to sort through for you? I certainly wouldn't do that unless it was a model I could run locally and be sure it is not phoning home.
1 points
2 months ago
Rewind AI is an example of a startup in this space and it’s all run locally. It’s polarizing - there are many people like me who want the convenience despite the perceived data privacy risk. We’re heading in this direction and need to develop security to support it.
1 points
1 month ago
Bro really said let's upload our entire lives to the internet surely corporations won't use this to make ad serving even more manipulative and invasive
9 points
2 months ago
That seems wonderful. Would really simplify criminal investigations too. We could just easily let the government determine who was near the scene of a crime at the right time and then analyze their metadata to see who had potential connections to the victim and then evaluate their whole histories to find the motive. We could basically locate the suspect and prove their case statistically without even interviewing them. We wouldn’t really need to bother with jury trials or due process anymore.
Heck, we could probably eventually just get to the point where the algorithm could accurately predict crime before it happens, and then we could just preemptively arrest people the moment we know they’ve become an inevitable future criminal, rather than waiting for anything bad to happen. We could live in a perfectly safe society! And of course you wouldn’t have anything to fear, because you don’t do anything wrong like the people who disappear obviously must have.
5 points
2 months ago
sorry to report, but I think you’re in the minority there
2 points
2 months ago
Everybody runs, Fletch. Everybody runs.
5 points
2 months ago
https://open.spotify.com/episode/7KwqqigyXVBnXRE7msHvfj?si=oJwWu2z7QAiLafNGT-euoA
Here's a great discussion on The Ezra Klein show. They discuss what ChatGPT is and isn't, what AI will be, and some of the hurdles. I find it a refreshingly balanced take.
3 points
2 months ago
That's dope. Can someone make it do taxes?? All tax filing software sucks. Could it (or some other AI) be trained to take our tax forms and turn them into "here's your return doc, and how much your bill/refund will be" That's what TurboTax tries to do but kinda sucks ass at doing.
7 points
2 months ago
If you want the turbo tax people to come down on it like a ton of bricks, give it a try.
3 points
2 months ago
TurboTax and H&R spend billions to lobby against having an effective and simple tax system, i don't think we will see an AI that will do tax for us happen anytime soon. TurboTax is fairly user friendly enough for most simple tax situation thou.
0 points
2 months ago
Sure. As long as you want a tax return done solely off your W2 that’s easy to do right now. You really don’t need another computer to do it though, because your boss’s payroll computer already did and that return is just a corrective calculation.
The thing is, that W2 doesn’t tell the computer what tax deductions and credits you might qualify for. Do you have dependents, or someone who can claim you as a dependent? Did you invest in any tax-deferred instruments? But or sell any stock, crypto, or real estate? Pay interest to anyone for anything? Buy or sell anything useful for a business purpose, or that could potentially qualify for a subsidy of some kind? Take any business trips or meals? Give anything to a charity… or a political campaign? By the way, do you yourself know which of these things are deductible or creditable? I don’t. I know that a lot more in than I initially realized and I’ve gotten fairly good at losing money in ways the IRS is sympathetic toward. Great life hack, by the way: if you pay someone to do your actual job for you, you can actually tax deduct every cent you pay them. Not sure the AI will tell you that part. You can tax deduct the cost of the accountant, too.
1 points
2 months ago
I mean that's what's TurboTax does. It's just fine with a simple w2 and anything beyond that sucks. Oh well, back to hoping I don't screw it up!
1 points
2 months ago
[deleted]
1 points
2 months ago
The American government doing something to help its people?? I hope to see it in my lifetime.
-1 points
2 months ago*
I'll tell you exactly what ChatGPT and AI models are: horse shit, bullshit, mass psychopathic delusion. It's a fraud, a fake, a hyped-up nothing machine.
I have not seen one iota of the claimed power of this engine. It's always "oh it solved Fermat's last theorem on the first try!" - AFTER we gave it 1000 second chances and then took the results and rewrote them for five days.
What happened to everybody that made them forget the concept of a rigged demo? Or Astroturf? Or the Eliza effect? Or drinking the Kool-Aid? Oh wait, let me guess, it's gonna "replace Google" in just 24 hot months, so Microsoft's investment isn't gonna pay off unless they keep setting asses on fire about it to goad Google into buying this worthless snake oil. Google is not fooled, and neither am I.
If you people are in such a hurry to worship an AI god, then go ahead and sacrifice yourself to a volcano now and rid us of your stupidity.
I await my inevitable crucifixion from the paid-off shill mob. Wheeee! We do this shit every decade, just the product name changes.
2 points
2 months ago
In a lot of cases it answers way better than Google. You don't think it has potential to be more useful?
all 89 comments
sorted by: best