4/23/2025 at 5:24:03 PM
I could not agree more with this. 90% of AI features feel tacked on and useless and that’s before you get to the price. Some of the services out here are wanting to charge 50% to 100% more for their sass just to enable “AI features”.I’m actually having a really hard time thinking of an AI feature other than coding AI feature that I actually enjoy. Copilot/Aider/Claude Code are awesome but I’m struggling to think of another tool I use where LLMs have improved it. Auto completing a sentence for the next word in Gmail/iMessage is one example, but that existed before LLMs.
I have not once used the features in Gmail to rewrite my email to sound more professional or anything like that. If I need help writing an email, I’m going to do that using Claude or ChatGPT directly before I even open Gmail.
by joshstrange
4/23/2025 at 10:22:04 PM
One of the interesting things I've noticed is that the best experiences I've had with AI are with simple applications that don't do much to get in the way of the model, e.g. chatgpt and cursor/windsurf.I'm hopeful that as devs figure out how to build better apps with AI we'll have have more and more "cursor moments" in other areas in our lives
by petekoomen
4/24/2025 at 4:12:19 AM
Perhaps the real takeaway is that there really is only one product, two if you count image generation.Perhaps the only reason Cursor is so good is because editing code is so similar to the basic function of an LLM without anything wrapped around it.
Like, someone prove me wrong by linking 3 transformative AI products that:
1. Have nothing to do with "chatting" to a thin wrapper (couldn't just be done inside a plain LLM with a couple of file uploads added for additional context)
2. Don't involve traditional ML that has existed for years and isn't part of the LLM "revolution."
3. Has nothing to do with writing code
For example, I recently used an AI chatbot that was supposed to help me troubleshoot a consumer IoT device. It basically regurgitated steps from the manual and started running around in circles because my issue was simply not covered by documentation. I then had to tell it to send me to a human. The human had more suggestions that the AI couldn't think of but still couldn't help because the product was a piece of shit.
Or just look at Amazon Q. Ask it a basic AWS question and it'll just give you a bogus "sorry I can't help with that" answer where you just know that running over to chatgpt.com will actually give you a legitimate answer. Most AI "products" seem to be castrated versions of ChatGPT/Claude/Gemini.
That sort of overall garbage experience seems to be what is most frequently associated with AI. Basically, a futile attempt to replace low-wage employees that didn't end up delivering any value to anyone, especially since any company interested in eliminating employees just because "fuck it why not" without any real strategy probably has a busted-ass product to begin with.
Putting me on hold for 15 minutes would have been more effective at getting me to go away and no compute cycles would have been necessary.
by dangus
4/24/2025 at 9:57:59 AM
Outside of coding, Google's NotebookLM is quite useful for analysing complex documentation - things like standards and complicated API specs.But yes, an AI chatbot that can't actually take any actions is effectively just regurgitating documentation. I normally contact support because the thing I need help with is either not covered in documentation, or requires an intervention. If AI can't make interventions, it's just a fancy kind of search with an annoying interface.
by leoedin
4/24/2025 at 12:31:10 PM
I don’t deny that LLMs are useful, merely that they only represent one product that does a small handful of things well, where the industry-specific applications don’t really involve a whole lot of extra features besides just “feed in data then chat with the LLM and get stuff back.”Imagine if during the SaaS or big data or containerizaiton technology “revolutions” the application being run just didn’t matter at all. That’s kind of what’s going on with LLMs. Almost none of the products are all that much better than going to ChatGPT.com and dumping your data into the text box/file uploader and seeing what you get back.
Perhaps an analogy to describe what I mean would be if you were comparing two SaaS apps, like let’s say YNAB and the Simplifi budget app. In the world of the SaaS revolution, the capabilities of each application would be competitive advantages. I am choosing one over the other for the UX and feature list.
But in the AI LLM world, the difference between competing products is minimal. Whether you choose Cursor or Copilot or Firebase Studio you’re getting the same results because you’re feeding the same data to the same AI models. The companies that make the AI technologies basically don’t have a moat themselves, they’re basically just PaaS data center operators.
by dangus
4/24/2025 at 11:14:01 AM
Everything where structured output is involved, from filling in forms based on medical interview transcripts / court proceedings / calls, to an augmented chatbot that can do things for you (think hotel reservations over the phone), to directly generating forms / dashboards / pages in your system.by miki123211
4/24/2025 at 11:39:24 AM
If thats the best current llms can do, my job is secured till retirementby jajko
4/24/2025 at 4:48:14 PM
The best that current LLMs can do is PhD-level science questions and getting high scores in coding contests.Your job? Might be secure for a lifetime, might be gone next week. No way to tell — "intelligence" isn't yet so well understood to just be an engineering challenge, but it is so well understood that the effect on jobs may be the same.
by ben_w
4/25/2025 at 10:26:47 AM
Tbh, your job is secured as long as your boss is not brainwashed and FOMO by the infinite cash burning and lobbying to sell LLM as the all solving holy grail. Depending on how long it takes to brainwash all leadership into thinking, LLMs can do your job(despite reality, I mean it takes some time to fail and then additional to deal with sunk cost fallacy). One of my colleague went to work for a startup that promises to replace ALL general practitioner/family doctor within next 2-3 years by LLMs. While I hope they badly burn and go bankrupt, I am also worried about the future as this particular startup is also getting major backing from certain big deep pockets.by n_ary
4/24/2025 at 9:22:02 AM
LLMs make all sorts of classification problems vastly easier and cheaper to solve.Of course, that isn't a "transformative AI product", just a regular old product that improves your boring old business metrics. Nothing to base a hype cycle on, sadly.
by otabdeveloper4
4/24/2025 at 1:05:29 PM
Agree 100%.We built a very niche business around data extraction & classification of a particular type of documents. We did not have access to a lot of sample data. Traditional ML/AI failed spectacularly.
LLMs have made this super easy and the product is very successful thanks to it. Customers love it. It is definitely transformative for them.
by molf
4/24/2025 at 4:39:26 AM
I have used LLMs for some simple text generation for what I’m going to call boilerplate, eg why $X is important at the start of a reference architecture. But maybe it saved me an hour or two in a topic I was already fairly familiar with. Not something I would have paid a meaningful sum for. I’m sure I could have searched and found an article on the topic.by ghaff
4/24/2025 at 12:04:13 PM
Two off the top of my head:There are a lot of tools in the sales space which fit your criteria.
by ZephyrBlu
4/24/2025 at 12:18:26 PM
Granola is the exact kind of product I’m criticizing as being extremely basic and barely more than a wrapper. It’s just a meeting transcriber/summarizer, barely provides more functionality than leaving the OpenAI voice mode on during a call and then copying and pasting your written notes into ChatGPT at the end.Clay was founded 3 years before GPT 3 hit the market so I highly doubt that the majority of their core product runs on LLM-based AI. It is probably built on traditional machine learning.
by dangus
4/24/2025 at 5:05:19 AM
LLMs in data pipelines enable all sorts of “before impossible” stuff. For example, this creates an event calendar for you based on emails you have received:https://www.indexself.com/events/molly-pepper
(that’s mine, and is due a bugfix/update this week. message me if you want to try it with your own emails)
I have a couple more LLM-powered apps in the works, like next few weeks, that aren’t chat or code. I wouldn’t call them transformative, but they meet your other criteria, I think.
by whiddershins
4/24/2025 at 5:53:06 AM
What part of this can't be done by a novice programmer who knows a little pattern matching and has enough patience to write down a hundred patterns to match?by semi-extrinsic
4/24/2025 at 10:05:11 AM
Long tail, coping with typos, and understanding negation.If natural language was as easy as "enough patience to write down a hundred patterns to match", we'd have had useful natural language interfaces in the early 90s — or even late 80s, if it was really only "a hundred".
by ben_w
4/24/2025 at 9:12:20 PM
For narrow use cases we did have natural language interfaces in the 90s, yes. See e.g. IRC bots.Or to take a local example, for more than 20 years my city has had a web service where you can type "When is the next bus from Street A to Road B", and you get a detailed response including any transfers between lines. They even had a voice recognition version decades ago that you could call, which worked well.
From GP post, I was replying specifically to
> LLMs in data pipelines enable all sorts of “before impossible” stuff. > For example, this creates an event calendar for you based on emails you have received
That exact thing has been a feature of Gmail for over a decade. Remember the 2018 GCal spam?
https://null-byte.wonderhowto.com/how-to/advanced-phishing-i...
by semi-extrinsic
4/24/2025 at 10:35:07 PM
> For narrow use cases we did have natural language interfaces in the 90s, yes. See e.g. IRC bots."Narrow" being the key word. Thing is, even in the 2010s, we were doing sentiment analysis by counting the number of positive words and negative words, because it doesn't go past "narrow".
Likewise, "A to B" is great… when it's narrow. I grew up on "Southbrook Road" — not the one in London, not the one in Southampton, not the one in Exeter, …
And then there's where I went to university. Ond mae hynny'n twyllo braidd, oherwydd y Gymraeg. But not cheating very much, because of bilingual rules and because of the large number of people with multi-lingual email content. Cinco de mayo etc.
I also grew up with text adventures, which don't work if you miss the expected keyword, or mis-spell it too hard. (And auto-correction has its own problems, as anyone who really wants to search for "adsorption" not "absorption" will tell you).
> That exact thing has been a feature of Gmail for over a decade. Remember the 2018 GCal spam?
Siri has something similar. It misses a lot and makes up a lot. Sometimes it sets the title to be the date and makes up a date.
These are examples of not doing things successfully with just a hundred hard-coded rules.
by ben_w
4/24/2025 at 6:18:44 AM
> Perhaps the only reason Cursor is so good is because editing code is so similar to the basic function of an LLM without anything wrapped around it.I think this is an illusion. Firstly, code generation is a big field - it includes code completion, generating entire functions, and even agenting coding and the newer vibe-coding tools which are mixes of all of these. Which of these is "the natural way LLMs work"?
Secondly, a ton of work goes into making LLMs good for programming. Lots of RLHF on it, lots of work on extracting code structure / RAG on codebases, many tools.
So, I think there are a few reasons that LLMs seem to work better on code:
1. A lot for work on it has been done, for many reasons, mostly monetary potential and that the people who build these systems are programmers.
2. We here tend to have a lot more familiarity with these tools (and this goes to your request above which I'll get to).
3. There are indeed many ways in which LLMs are a good fit for programming. This is a valid point, though I think it's dwarfed by the above.
Having said all that, to your request, I think there are a few products and/or areas that we can point to that are transformative:
1. Deep Research. I don't use it a lot personally (yet) - I have far more familiarity with the software tools, because I'm also a software developer. But I've heard from many people now that these are exceptional. And they are not just "thing wrappers on chat", IMO.
2. Anything to do with image/video creation and editing. It's arguable how much these count as part of the LLM revolution - the models that do these are often similar-ish in nature but geared towards images/videos. Still, the interaction with them often goes through natural language, so I definitely think these count. These are a huge category all on their own.
3. Again, not sure if these "count" in your estimate, but AlphaFold is, as I understand it, quite revolutionary. I don't know much about the model or the biology, so I'm trusting others that it's actually interesting. It is some of the same underlying architecture that makes up LLMs so I do think it counts, but again, maybe you want to only look at language-generating things specifically.
by edanm
4/24/2025 at 12:37:55 PM
1. Deep Research (if you are talking about the OpenAI product) is part of the base AI product. So that means that everything building on top of that is still a wrapper. In other words, nobody besides the people making base AI technology is adding any value. An analogy to how pathetic the AI market is would be if during the SaaS revolution everyone just didn’t need to buy any applications and directly used AWS PaaS products like RDS directly with very similar results compared to buying SaaS software. OpenAI/Gemini/Claude/etc are basically as good as a full blown application that leverage their technology and there’s very limited need to buy wrappers that go around them.2. Image/video creation is cool but what value is it delivering so far? Saving me a couple of bucks that I would be spending on Fiverr for a rough and dirty logo that isn’t suitable for professional use? Graphic designers are already some of the lowest paid employees at your company so “almost replacing them but not really” isn’t a very exciting business case to me. I would also argue that image generation isn’t even as valuable as the preceding technology, image recognition. The biggest positive impact I’ve seen involves GPU performance for video games (DLSS/FSR upscaling and frame generation).
3. Medical applications are the most exciting application of AI and ML. This example is something that demonstrates what I mean with my argument: the normal steady pace of AI innovation has been “disrupted” by LLMs that have added unjustified hype and investment to the space. Nobody was so unreasonably hyped up about AI until it was packaged as something you can chat with since finance bro investors can understand that, but medical applications of neural networks have been developing since long before ChatGPT hit the scene. The current market is just a fever dream of crappy LLM wrappers getting outsized attention.
by dangus
4/24/2025 at 3:28:41 PM
Is Cursor actually good though? I get so frustrated at how confidently it spews out the completely wrong approach.When I ask it to spit out Svelte config files or something like that, I end up having to read the docs myself anyway because it can’t be trusted, for instance it will spew out tons of lines to configure every parameter as something that looks like the default when all it needs to do is follow the documentation that just uses defaults()
And it goes out of its way to “optimise” things that actually picks the wrong options versus the defaults which are fine.
by aetherspawn
4/24/2025 at 6:36:22 AM
This challenge is a little unfair. Chat is an interface not an application.by kybernetikos
4/24/2025 at 7:00:37 AM
Generating a useful sequence of words or word-like tokens is an application.by RedNifre
4/24/2025 at 7:12:21 AM
I would describe that as a method or implementation, not as an application.Almost all knowledge work can be described as "generating a useful sequence of words or word like tokens", but I wouldn't hire a screen writer to do the job of a lawyer or a copy editor to do the job of a concierge or an HR director to do the job of an advertising consultant.
by kybernetikos
4/24/2025 at 12:53:37 PM
So then the challenge is valid but you just can’t think of any ways to satisfy it. You said yourself that chat is just the interface.That means you should be able to find many popular applications that leverage LLM APIs that are a lot different than the interface of ChatGPT.
But in reality, they’re all just moving the chat window somewhere else and streamlining the data input/output process (e.g., exactly what Cursor is doing).
I can even think of one product that is a decent example of LLMs in action without a chat window. Someone on HN posted a little demo website they made that takes SEC filings and summarizes them to make automatic investor analysis of public companies.
But it’s kind of surprising to me how that little project seems to be in the minority of LLM applications and I can’t think of two more decent examples especially when it comes to big successful products.
by dangus
4/23/2025 at 6:21:27 PM
> This demo uses AI to read emails instead of write themLLMs are so good at summarizing that I should basically only ever read one email—from the AI:
You received 2 emails today that need your direct reply from X and Y. 1 is still outstanding from two days ago, _would you like to send an acknowledgment_? You received 6 emails from newsletters you didn’t sign up for but were enrolled after you bought something _do you want to unsubscribe from all of them_ (_make this a permanent rule_).
by teeray
4/23/2025 at 6:49:43 PM
I have fed LLMs PDF files, asked about the content and gotten nonsense. I would be very hesitant to trust them to give me an accurate summary of my emails.by namaria
4/23/2025 at 7:13:03 PM
One of our managers uses Ai to summarize everything. Too bad it missed important caveats for an offer. Well, we burned an all nighters to correct the offer, but he did not read twenty pages but one...by HdS84
4/23/2025 at 7:31:57 PM
I don't know if this is the case but be careful about shielding management from the consequences of their bad choices at your expense. It all but guarantees it will get worse.by namaria
4/23/2025 at 8:22:55 PM
Letting a thing implode that you could prevent is a missed opportunity for advancement and a risk to your career because you will be on a failing team.The smarter move is to figure out how to fix it for the company while getting visibility for it.
by kortilla
4/23/2025 at 9:39:28 PM
> Letting a thing implode that you could prevent is a missed opportunity for advancementNo matter how many times I bail out my managers it seems that my career has never really benefit from it
I've only ever received significant bumps to salary or job title by changing jobs
by bluefirebrand
4/24/2025 at 9:40:01 AM
I don't know what your experience is, but mine is the opposite. Nobody ever notices people who put out fires, and it's hard to should "hey guys! There's a fire here that John started, I'm putting it out!" without looking like a jerk for outing John.by stavros
4/24/2025 at 1:13:03 PM
Fewer still notice the fire-preventer.by HPsquared
4/24/2025 at 1:32:37 PM
Oh, no, neither prevent the fires not put them out. Instead, predict them, and then say "see?" when they break out.by stavros
4/24/2025 at 2:12:42 PM
That's a risky business, you can get the blame if you're not careful. "Why didn't you try harder if you knew this would happen" etc.by HPsquared
4/24/2025 at 2:23:01 PM
If you say "look, the stuff they're doing there is risky, you should <do thing>", and they don't do it, how can they blame you? If they do do it, then mission accomplished, no?E.g. "the way that team builds software isn't robust enough, you should replace the leader or we'll have an incident", how can you be blamed for the incident when it happens?
by stavros
4/23/2025 at 8:52:40 PM
You are right. I don't think the only alternative to shielding management from the consequences of their bad choices is letting things implode and going down with the ship.by namaria
4/24/2025 at 6:50:44 AM
yup, an employee is more than just a gear, better keep the motor running than explode along with the other parts.by spookie
4/23/2025 at 9:09:36 PM
Management should be hiring lawyers for those details anyway...by bluGill
4/24/2025 at 1:50:07 PM
Wasn't even legal but concerned the scope of the offer. Nuance, but nuance can be important. Like "rework the service and add minor festures" VS "slightly rework and do major features" - this affected the direction of our offer a lot.by HdS84
4/23/2025 at 9:33:13 PM
Yes. Reliable domain experts are very important.by namaria
4/24/2025 at 3:59:55 AM
Should I mention that yesterday I just saw a diagram with a box that said “Legal Review LLM”?by shermantanktop
4/24/2025 at 7:04:11 AM
Maybe you should point them to the news stories about that sort of thing blowing up spectacularly in court. Or maybe you could just let them learn that by themselves.by namaria
4/23/2025 at 7:35:17 PM
Did he pull all nighters to fix it? If not, it wasn't "too bad" for him. I doubt he'll change his behavior.by BeetleB
4/24/2025 at 9:00:09 AM
Where's the IBM slide about "a machine cannot be held accountable, therefore a machine should never make a management decision"?Of course, often it's quite hard to hold management accountable either.
by pjc50
4/24/2025 at 12:31:25 PM
Isn't a solution to assign vicarious liability to whomever approves the use of the decision-making machine?by checkyoursudo
4/23/2025 at 7:19:29 PM
LLMs are terrible at summarizing technical emails where the details matter. But you might get away with it, at least for a while, in low performing organizations that tolerate preventable errors.by nradov
4/23/2025 at 7:27:28 PM
This. LLMs seem to be great for 90+% of stuff, but sometimes, they just spew weird stuff.by imp0cat
4/24/2025 at 11:50:50 AM
If I get a technical email I read it myself. The summary just needs to say technical email from X with priority Y about problem Zby HDThoreaun
4/24/2025 at 4:10:37 AM
I got an email from the restaurant saying "We will confirm your dinner reservation as soon as we can", and Apple Intelligence summarizing it as "Dinner reservation confirmed." Maybe it can not only summarize, but also see the future??by FabHK
4/24/2025 at 7:28:27 AM
Well, at least it doesn’t make up words. The Portuguese version of Apple Intelligence made up “Invitaçāo” (think “invitashion”) and other idiocies the very first day it started working in the EU.by rcarmo
4/23/2025 at 7:48:38 PM
> LLMs are so good at summarizing that I should basically only ever read one email—from the AIThis could get really fun with some hidden text prompt injection. Just match the font and background color.
Maybe these tools should be doing the classic air gap approach of taking a picture of the rendered content and analyzing that.
by koolba
4/23/2025 at 6:47:15 PM
What system are you using to do this? I do think that this would provide value for me. Currently, I barely read my emails, which I'm not exactly proud of, but it's just the reality. So something that summarized the important things every day would be nice.by joshstrange
4/24/2025 at 3:53:08 AM
Do you ever check its work?by meroes
4/24/2025 at 6:40:45 AM
What is the reason to unsub ever in that world? Are you saying the LLM can't skip emails? Seems like an arbitrary ruleby throwaway290
4/24/2025 at 5:13:02 PM
I fed an LLM the record of a chat between me and a friend, and asked it to summarize the times that we met in the past 3 months.Every time it gave me different results, and not once did it actually get it all right.
LLMs are horrible for summarizing things. Summarizing is the art of turning low information density text into high information density text. LLMs can’t deal in details, so they can never accurately summarize anything.
by amrocha
4/25/2025 at 10:37:53 AM
You should try with lowered temperature(0.1 or even 0.0). Usually accuracy improves dramatically.by n_ary
4/23/2025 at 6:10:36 PM
I enjoy Claude as a general purpose "let's talk about this niche thing" chat bot, or for general ideation. Extracting structured data from videos (via Gemini) is quite useful as well, though to be fair it's not a super frequent use case for me.That said, coding and engineering is by far the most common usecase I have for gen AI.
by danielbln
4/23/2025 at 6:48:16 PM
Oh, I'm sorry if it wasn't clear. I use Claude and ChatGPT to talk to about a ton of topics. I'm mostly referring to AI features being added to existing SaaS or software products. I regularly find that moving the conversation to ChatGPT or Claude is much better than trying to use anything that they may have built into their existing product.by joshstrange
4/23/2025 at 6:27:50 PM
I think the other application besides code copiloting that is already extremely useful is RAG-based information discovery a la Notion AI. This is already a giant improvement over "search google docs, and slack, and confluence, and jira, and ...".Just integrated search over all the various systems at a company was an improvement that did not require LLMs, but I also really like the back and forth chat interface for this.
by sanderjd
4/24/2025 at 7:25:49 AM
The e-mail agent example is so good that it makes everything else I’ve seen and used pointless by comparison. I wonder why nobody’s done it that way yet.by rcarmo
4/23/2025 at 10:44:50 PM
I wonder sometime if this is why there is such an enthusiasm gap over AI between tech people and the general public. It's not just that your average person can't program; it's that they don't even conceptually understand why programming could unlock.by knightscoop
4/24/2025 at 4:49:50 AM
Have you ever been cooking and asked Siri to set a timer? That's basically the most used AI feature outside of "coding" I can think of.by bamboozled
4/24/2025 at 1:40:48 PM
Setting a timer and setting a reminder. Occasionally converting units of measure. That's all I can rely on Siri (or Alexa) for and even then sometimes Siri doesn't make it clear if it did the thing. Most importantly, "set a reminder", it shows the text, and then the UI disappears, sometimes the reminder was created, sometimes not. It's maddening since I'm normally asking to be reminded about something important that I need to get recorded/tracked so I can "forget" it.The number of times I've had 2 reminders fire back-to-back because I asked Siri again to create one since I was _sure_ it didn't create the first one.
Siri is so dumb and it's insane that more heads have not rolled at Apple because of it (I'm aware of the recent shakeup, it's about a decade too late). Lastly, whoever decided to ship the new Siri UI without any of the new features should lose their job. What a squandered opportunity and effectively fraud IMHO.
More and more it's clear that Tim Cook is not the person that Apple needs at the helm. My mom knows Siri sucks, why doesn't the CEO and/or why is he incapable of doing anything to fix it. Get off your Trump-kissing, over-relying-on-China ass and fix your software! (Siri is not the only thing rotten)
by joshstrange
4/24/2025 at 8:00:40 AM
I find that ChatGPT o3 (and the other advanced reasoning models) are decently good at answering questions with a "but".Google is great at things like "Top 10 best rated movies of 2024", because people make lists of that sort of thing obsessively.
But Google is far less good at queries like "Which movies look visually beautiful but have been critically panned?". For that sort of thing I have far more luck with chatgpt because it's much less of a standard "top 10" list.
by dale_glass
4/24/2025 at 1:32:13 PM
o3 has been a big improvement on Deep Research IMHO. o1 (or whatever model I originally used with it) was interesting but the results weren't always great. o3 has done some impressive research tasks for me and, unlike the last model I used, when I "check its work" it has always been correct.by joshstrange
4/23/2025 at 6:13:22 PM
I like perplexity when I need a quick overview of a topic with references to relevant published studies. I often use it when researching what the current research says on parenting questions or education. It's not perfect but because the answers link to the relevant studies it's a good way to get a quick overview of research on a given topicby nicolas_t
4/23/2025 at 5:28:46 PM
garmin wants me to pay for some gen-ai workout messages on connect plus. Its the most absurd AI slop of all. Same with strava. I workout for mental relaxation and i just hate this AI stuff being crammed in there.Atleast clippy was kind of cute.
by apwell23
4/23/2025 at 7:25:56 PM
Strava employees claim that casual users like the AI activity summaries. Supposedly users who don't know anything about exercise physiology didn't know how to interpret the various metrics and charts. I don't know if I believe that but it's at least plausible.Personally I wish I could turn off the AI features, it's a waste of space.
by nradov
4/23/2025 at 10:15:59 PM
Anytime someone from a company says that users like the super trendy thing they just made I take it with a sizeable grain of salt. Sometimes it's true, and maybe it is true for Strava, but I've seen enough cases where it isn't to discount such claims down to ~0.by rurp
4/24/2025 at 8:41:51 AM
The guy at the Wendy's drive thru has told me repeatedly that most people don't want ketchup so they stopped putting it in bags by default.by genewitch
4/23/2025 at 6:14:23 PM
Strava's integration is just so lackluster. It literally turns four numbers from right above the slop message into free text. Thanks Strava, I'm a pro user for a decade, finally I can read "This was a hard workout" after my run. Such useful, much AI.by danielbln
4/24/2025 at 6:23:06 AM
I use AI chatbots for 2+ hours a day but the Garmin thing was too much for me. The day they released their AI Garmin+ subscription, I took off my Forerunner and put it in a drawer. The whole point of Garmin is that it feels emotionally clean to use. Garmin adding a scammy subscription makes the ecosystem feel icky, and I'm not going to wear a piece of clothing that makes me feel icky. I don't think I'll buy a Garmin watch again.(Since taking off the watch, I miss some of the data but my overall health and sleep haven't changed.)
by sandspar
4/23/2025 at 6:27:52 PM
At this point, "we aren't adding any AI features" is a selling point for me. I've gotten real tired of AI slop and hype.by bigstrat2003
4/23/2025 at 6:22:46 PM
Honestly I don't even enjoy coding AI features. The only value I get out of AI is translation (which I take with a grain of salt because I don't know the other language and can't spot hallucinations, but it's the best tool I have), and shitposting (e.g. having chatGPT write funny stories about my friends and sending it to them for a laugh). I can't say there's an actual productive use case for me personally.by bigstrat2003
4/24/2025 at 8:35:25 AM
I've anecdotally tested translations by ripping the video with subtitles and having whisper subtitle it, and also asking several AI to translate the .srt or .vtt file (subtotext I think does this conversion if you don't wanna waste tokens on the metadata)Whisper large-v3, the largest model I have, is pretty good, getting nearly identical translations to chatgpt or whatever, Google's default speech to text. The fun stuff is when you ask for text to text translations from LLMs.
I did a real small writeup with an example but I don't have a place to publish nor am I really looking for one.
I used whisper to transcribe nearly every "episode" of the Love Line syndicated radio show from 1997-2007 or so. It took, iirc, several days. I use it to grep the audio, as it were. I intend to do the same with my DVDs and such, just so I never have to Google "what movie / tv show is that line from?" I also have a lot of art bell shows, and a few others to transcribe.
by genewitch
4/24/2025 at 8:52:59 AM
> I used whisper to transcribe nearly every "episode" of the Love Line syndicated radio show from 1997-2007 or so.Yes - second this. I found 'Whisper' great for that type of scenario as well.
A local monastery had about 200 audio talks (mp3). Whisper converted them all to text and GPT did a small 'smoothing' of the output to make it readable. It was about half a million words and only took a few hours.
The monks were delighted - they can distribute their talks in small pamplets / PDFs now and is extra income for the community.
Years ago as a student I did some audio transcription manually and something similar would have taken ages...
by farrelle25
4/24/2025 at 4:43:23 PM
I actually was asked by Vermin Supreme to hand-caption some videos, and i instantly regretted besmirching the existing subtitles. I was correct, the subtitles were awful, but boy, the thought of hand-transcribing something with Subtitle Edit had me walking that back pretty quick - and this was for a 4 minute video - however it was lyrical over music, so AI barely gave a starting transcription.by genewitch
4/24/2025 at 9:22:15 AM
I wanted this to work with Whisper, but the language I tried it with was Albanian and the results were absolutely terrible - not even readable English. I'm sure it would be better with Spanish or Japanese.by pjc50
4/24/2025 at 10:21:51 AM
According to the Common Voice 15 graph on OpenAI's github repository, Albanian is the single worst performance you could have had: https://github.com/openai/whisperBut for what it's worth, I tried putting the YouTube video of Tom Scott presenting at the Royal Institute into the model, and even then the results were only "OK" rather than "good". When even a professional presenter and professional sound recording in a quiet environment has errors, the model is not really good enough to bother with.
by ben_w
4/24/2025 at 8:07:37 AM
> Auto completing a sentence for the next word in Gmail/iMessage is one exampleInterestingly, I despise that feature. It breaks the flow of what is actually a very simple task. Now I'm reading, reconsidering if the offered thing is the same thing I wanted over and over again.
The fact that I know this and spend time repeatedly disabling the damned things is awfully tiresome (but my fault for not paying for my own email etc etc)
by Ntrails
4/24/2025 at 8:22:04 AM
I've been using Fastmail in lieu of gmail for ten or eleven years. If you have a domain and control the DNS, I recommend it. At least you're not on Google anymore, and you're paying for fastmail, so it feels better - less like something is reading your emails.by genewitch
4/24/2025 at 4:52:00 PM
I really like my speech-to-text program, and I find using ChatGPT to look up things and answer questions is a much superior experience to Google, but otherwise, I completely agree with you.Companies see that AI is a buzzword that means your stock goes up. So they start looking at it as an answer to the question: "How can I make my stock go up?" instead of "How can I create a better product", and then let the stock go up from creating a better product.
by tomjen3
4/23/2025 at 5:58:03 PM
> I’m actually having a really hard time thinking of an AI feature other than coding AI feature that I actually enjoy.If you attend a lot of meetings, having an AI note-taker take notes for you and generate a structured summary, follow-up email, to-do list, and more will be an absolute game changer.
(Disclaimer, I'm the CTO of Leexi, an AI note-taker)
by Andugal
4/23/2025 at 6:08:22 PM
The catch is: does anyone actually read this stuff? I've been taking meeting notes for meetings I run (without AI) for around 6 months now and I suspect no one other than myself has looked at the notes I've put together. I've only looked back at those notes once or twice.A big part of the problem is even finding this content in a modern corporate intranet (i.e. Confluence) and having a bunch of AI-generated text in there as well isn't going to help.
by AlexandrB
4/23/2025 at 6:55:11 PM
When I was a founding engineer at a(n ill-fated) startup, we used an AI product to transcribe and summarize enterprise sales calls. As a dev it was usually a waste of my time to attend most sales meetings, but it was highly illustrative to read the summaries after the fact. In fact many, many of the features we built were based on these action items.If you're at the scale where you have corporate intranet, like Confluence, then yeah AI note summarizing will feel redundant because you probably have the headcount to transcribe important meetings (e.g. you have a large enough enterprise sales staff that part of their job description is to transcribe notes from meetings rather than a small staff stretched thin because you're on vanishing runway at a small startup.) Then the natural next question arises: do you really need that headcount?
by Karrot_Kream
4/23/2025 at 6:13:46 PM
I thought the point of having a meeting-notes person was so that at least one person would pay attention to details during the meeting.by bee_rider
4/23/2025 at 7:16:51 PM
I thought it was so I could go back 1 year and say, 'I was against this from the beginning and I was quite vocal that if you do this, the result will be the exact mess you're asking me to clean up now.'by jethro_tell
4/23/2025 at 7:48:07 PM
Ah, but a record for CYA and “told you so”, that’s pure cynicism. “At least one person paying attention” at least we can pretend the intent was to pair some potential usefulness with our cynicism.by bee_rider
4/23/2025 at 7:20:02 PM
Also, ensure that if the final decition was to paint the the bike shed green, everyone agree it was the final decitions. (In long discusions, sometimes people misunderstand which was the final decition.)by gus_massa
4/23/2025 at 7:50:15 PM
If they misunderstood they will still disagree so the meeting notes will trigger another mail chain and, you guessed right, another meeting.by soco
4/23/2025 at 6:22:50 PM
What is the problem?Notes are valuable for several reasons.
I sometimes take notes myself just to keep myself from falling asleep in an otherwise boring meeting where I might need to know something shared (but probably not). It doesn't matter if nobody reads these as the purpose wasn't to be read.
I have often wished for notes from some past meeting because I know we had good reasons for our decisions but now when questioned I cannot remember them. Most meetings this doesn't happen, but if there were automatic notes that were easy to search years latter that would be good.
Of course at this point I must remind you that the above may be bad. If there is a record of meeting notes then courts can subpoena them. This means meetings with notes have to be at a higher level were people are not comfortably sharing what every it is they are thinking of - even if a bad idea is rejected the courts still see you as a jerk for coming up with the bad idea.
by bluGill
4/23/2025 at 7:07:43 PM
Accurate notes are valuable for several reasons.Show me an LLM that can reliably produce 100% accurate notes. Alternatively, accept working in a company where some nonsense becomes future reference and subpoenable documentation.
by namaria
4/23/2025 at 8:16:12 PM
Counterpoint: show me a human who can reliably produce 100% accurate notes.Seriously, I wish to hire this person.
by lpapez
4/23/2025 at 8:35:42 PM
Seriously, do people around you not normally double check, proofread, review what they turn in as done work?Maybe I am just very fortunate, but people who are not capable of producing documents that are factually correct do not get to keep producing documents in the organizations I have worked with.
I am not talking about typos, misspelling words, bad formatting. I am talking about factual content. Because LLMs can actually produce 100% correct text but they routinely mangle factual content in a way that I have never had the misfortune of finding in the work of my colleagues and teams around us.
by namaria
4/23/2025 at 8:41:43 PM
A friend of mine asked an AI for a summary of a pending Supreme Court case. It came back with the decision, majority arguments, dissent, the whole deal. Only problem was that the case hadn't happened yet. It had made up the whole thing, and admitted that when called on it.A human law clerk could make a mistake, like "Oh, I thought you said 'US v. Wilson,' not 'US v. Watson.'" But a human wouldn't just make up a case out of whole cloth, complete with pages of details.
So it seems to me that AI mistakes will be unlike the human mistakes that we're accustomed to and good at spotting from eons of practice. That may make them harder to catch.
by aaronbaugher
4/23/2025 at 8:58:27 PM
I think it is more like the clerk would say "There never was a US vs Wilson" (well there probably was given how common that name is, but work with me). The AI doesn't have a concept of maybe I misunderstood the question. AI would likely give you a good summary if the case happened, but if it didn't it makes up a case.by bluGill
4/23/2025 at 9:03:57 PM
Yes. That is precisely the problem with using LLMs. They wantonly make up text that has no basis in reality. That is the one and only problem I have with them.by namaria
4/24/2025 at 4:22:15 AM
It would be kind of funny if we build a space probe with an LLM and shoot it out into space. Many years later intelligent life from far away discovers it and it somehow leads to our demise do to badly hallucinated answers.by Bluecobra
4/24/2025 at 1:39:42 PM
Space is so big and space travel is so slow that our sun will be dead before the probe is found by anyone else out there.And that is assuming there even is someone out there, which isn't a given.
by bluGill
4/24/2025 at 8:15:07 AM
What are the odds that the comment you're responding to was AI-generated?by mangamadaiyan
4/24/2025 at 12:50:14 PM
Good question. So far comments here mostly seem to be human generated, but I would be surprised if there were no AI generated ones. It is also possible to fool me. I'm going with - for now - the default that it was not AI.by bluGill
4/24/2025 at 9:53:48 AM
You are mixing up notes and full blown transcript of the meeting. The latter is impossible to produce by the untrained humans. The former is relatively easy for a person paying attention, because it is usually 5 to 10 short lines per an hour long meeting, with action items or links. Also in a usual work meeting, a person taking notes has possibility to simply say "wait a minute, I will write this down" and this does happens in practice. Short notes made like that usually are accurate in the meaning, with maybe some minor typos not affecting accuracy.by Yizahi
4/23/2025 at 7:16:56 PM
You show me human meeting minutes written by a PM that accurately reflect the engineer discussions first.by Tadpole9181
4/23/2025 at 7:37:06 PM
Has it been your experience? That's unacceptable to me. From people or language models.by namaria
4/23/2025 at 7:35:35 PM
If it is just for people in the meeting we don't need 100%, just close enough that we remember what was discussed.by bluGill
4/23/2025 at 7:41:24 PM
I really don't see the value of records that may be inaccurate as long as I can rely on my memory. Human memory is quite unreliable, the point of the record is the accuracy.by namaria
4/23/2025 at 8:59:43 PM
Written records are only accurate if they are carefully reviewed. Humans make mistakes all the time too. We just are better at correcting them, and if we review the record soon after the meeting there is a chance we remember well enough to make a correction.There is a reason meeting rules (ie Robert's rules of order) have the notes from the previous meeting read and then voted on to accept them - often changes are made before accepting them.
by bluGill
4/23/2025 at 9:02:35 PM
Do just that. Enter an organization that has regular meetings and follows Robert's rules of order. Use an LLM to generate notes. Read the notes and vote on them. See how long the LLM remains in use.by namaria
4/23/2025 at 8:28:51 PM
Meh, show me a human that can reliably produce 100% accurate notes. It seems that the baseline for AI should be human performance rather than perfection. There are very few perfect systems in existence, and humans definitely aren't one of them.by ewhanley
4/23/2025 at 8:46:33 PM
I agree, and my vision of this is that instead of notes, the meeting minutes would be catalogued into a vector store, indexed by all relevant metadata. And then instead of pre-generated notes, you'll get what you want on the fly, with the LLM being the equivalent of chatting with that coworker who's been working there forever and has context on everything.by falcor84
4/24/2025 at 9:46:08 AM
You can probably buy another neural net SAAS subscription to summarize the summaries for you :)by Yizahi
4/23/2025 at 7:48:20 PM
I'm not a CTO so maybe your wold is not my world, but for me the advantage of taking the notes myself is that only I know what's important to me, or what was news to me. Teams Premium - you can argue it's so much worse than your product - takes notes like "they discussed about the advantages of ABC" but maybe exactly those advantages are advantageous to know right? And so on. Then like others said, I will review my notes once to see if there's a followup, or a topic to research, and off they go to the bin. I have yet to need the meeting notes of last year. Shortly put: notes apps are to me a solution in search of a problem.by soco
4/23/2025 at 6:52:12 PM
Is Leexi's AI note-taker able to raise its hand in a meeting (or otherwise interrupt) and ask for clarification?As a human note-taker, I find the most impactful result of real-time synthesis is the ability to identify and address conflicting information in the moment. That ability is reliant on domain knowledge and knowledge of the meeting attendees.
But if the AI could participate in the meeting in real time like I can, it'd be a huge difference.
by yesfitz
4/23/2025 at 7:45:35 PM
If you are attending the meeting as well as using an AI note-taker, then you should be able to ask the clarifying question(s). If you understand the content, then you should understand the AI notes (hopefully), and if you ask for clarification, then the AI should add those notes too.Your problem really only arises if someone is using the AI to stand in for them at the meeting vs. use it to take notes.
by bdavisx
4/23/2025 at 9:35:12 PM
I'll pretend you asked a few questions instead of explaining my work to me without understanding.1. "Why can't you look at the AI notes during the meeting?" The AI note-takers that I've seen summarize the meeting transcript after the meeting. A human note-taker should be synthesizing the information in real-time, allowing them to catch disagreements in real-time. Not creating the notes until after the meeting precludes real-time intervention.
2. "Why not use [AI Note-taker whose notes are available during the meeting]?" Even if there were a real-time synthesis by AI, I would have to keep track of that instead of the meeting in order to catch the same disagreements a human note-taker would catch.
3. "What problem are you trying to solve?" My problem is that misunderstandings are often created or left uncorrected during meetings. I think this is because most people are thinking about the meeting topics from their perspective, not spending time synthesizing what others are saying. My solution to this so far has been human note-taking by a human familiar with the meeting topic. This is hard to scale though, so I'm curious to see if this start-up is working on building a note-taking AI with the benefits I've mentioned seem to be unique to humans (for now).
by yesfitz
4/23/2025 at 6:16:45 PM
But that isn't writing for me, it is taking notes for me. There is a difference. I don't need something to write for me - I know how to write. What I need is someone to clean up grammar, fact check the details, and otherwise clean things up. I have dysgraphia - a writing disorder - so I need help more than most, but I still don't need something to write my drafts for me: I can get that done well enough.by bluGill
4/23/2025 at 8:49:19 PM
We've had the built-in Teams summary AI for a while now and it absolutely misses important details and nuance that causes problems later.by yoyohello13
4/23/2025 at 6:46:06 PM
I've used multiple of these types of services and I'll be honest, I just don't really get the value. I'm in a ton of meetings and I run multiple teams but I just take notes myself in the meetings. Every time I've compared my own notes to the notes that the the AI note taker took, it's missing 0-2 critical things or it focuses on the wrong thing in the meeting. I've even had the note taker say essentially the opposite of what we decided on because we flip-flopped multiple times during the meeting.Every mistake the AI makes is completely understandable, but it's only understandable because I was in the meeting and I am reviewing the notes right after the meeting. A week later, I wouldn't remember it, which is why I still just take my own notes in meetings. That said, having having a recording of the meeting and or some AI summary notes can be very useful. I just have not found that I can replace my note-taking with an AI just yet.
One issue I have is that there doesn't seem to be a great way to "end" the meeting for the note taker. I'm sure this is configurable, but some people at work use Supernormal and I've just taken to kicking it out of of meetings as soon as it tries to join. Mostly this is because I have meetings that run into another meeting, and so I never end the Zoom call between the meetings (I just use my personal Zoom room for all meetings). That means that the AI note taker will listen in on the second meeting and attribute it to the first meeting by accident. That's not the end of the world, but Supernormal, at least by default, will email everyone who was part of the the meeting a rundown of what happened in the meeting. This becomes a problem when you have a meeting with one group of people and then another group of people, and you might be talking about the first group of people in the second meeting ( i.e. management issues). So far I have not been burned badly by this, but I have had meeting notes sent out to to people that covered subjects that weren't really something they needed to know about or shouldn't know about in some cases.
Lastly, I abhor people using an AI notetaker in lieu of joining a meeting. As I said above, I block AI note takers from my zoom calls but it really frustrates me when an AI joins but the person who configured the AI does not. I'm not interested in getting messages "You guys talked about XXX but we want to do YYY" or "We shouldn't do XXX and it looks like you all decided to do that". First, you don't get to weigh in post-discussion, that's incredibly rude and disrespectful of everyone's time IMHO. Second, I'm not going to help explain what your AI note taker got wrong, that's not my job. So yeah, I'm not a huge fan of AI note takers though I do see where they can provide some value.
by joshstrange
4/24/2025 at 9:45:00 AM
In my company have a few "summaries" made by Zoom neural net, which we share for memes on the joke chats, they are so hilariously bad. No one uses that functionality seriously. I don't know about your app, but I've yet to see a working note taker in the wild.by Yizahi
4/24/2025 at 12:57:29 PM
You do you.I attend a lot of meetings and I have reviewed the results of an AI note taker maybe twice ever. Getting an email with a todo-list saves a bit of time of writing down action items during a meeting, but I'd hardly consider it a game changer. "Wait, what'd we talk about in that meeting" is just not a problem I encounter often.
My experience with AI note takers is that they are useful for people who didn't attend the meeting and people who are being onboarded and want to be able to review what somebody was teaching them in the meeting and much much much less useful for other situations.
by UncleMeat