• 0 Posts
  • 43 Comments
Joined 6 months ago
cake
Cake day: August 27th, 2025

help-circle

  • I agree - mostly. But…things online are RADICALLY different now, vs late 90s / early 2000s.

    I’ve outlined some of my media and tech curation for my kids above; I would LOVE for them to stumble across stuff like we did. Hell, in time, I’d even let them grok the edgier stuff (yes, like you, I was there 3000 years ago. I know of the old magics)

    But that internet is long gone…or if not…severely booby trapped. The competence required of (say) a curious 8yr old in 2026 vs 2002 to navigate the online landscape and NOT encounter those booby traps (I feel) is several orders of magnitude higher.

    I don’t think we can just park our kids in front of the 486 and say “here’s Encarta; have at it. Then I’ll show you this cool thing called a BBS”.

    Kinda sucks.

    Still, there are useful funnels / curation pathways. You CAN recreate that experience for your kids…but it’s no longer “are you winning, son?” set it and forget it meme. Now it’s “Daddy needs to be a part time sysadmin and know what’s what, so some pedo doesn’t catfish you for feet pics via ROBLOX”.


  • I have to admit, I’ve gone sort of retro-tech with my kids in the main. Wii over Xbox or PS5, DS-lite over Switches, OLPC over chromebook etc. Social media - ha ha, get fucked :)

    My eldest asked for a phone, so I gave her my old flipfone. She loves it (and I kinda want it back now, lol).

    The HMD Nokia branded Barbie phones with KaiOS is what I’ve promised to get her when she’s earned it. That’s a dumbphone with some smart features…and as bonus, I can probably (with enough effort) create some apps for her directly (KaiOS is basically Firefox OS in a container, which even my dumb-ass can probably grok with some effort)

    I give them access to my tablets…but the tablets have kid mode (app lock) and are hardened with firewall, pi-hole, app timers and curated content (eg: Smart-tube with ONLY the channels I select as safe being visible, no click thrus, comment section blocked etc. ABC kids. Jellyfin pointing to ONLY kids stuff etc etc). I know this is a technical solution to a behavioral problem. OTOH as much as I would LOVE to just yeet all this shit into the sun, the realistic position is kids need to know how to use tech. I even leave little breadcrumbs for my eldest to try and “hack” my systems so she can get access to “hidden” software (which, matrix-in-matrix style, I’ve allowed her access to. Don’t give SUDO equivalent to an 8yr old…once bitten, twice shy. I could tell you a recent horror story that would curl your hair)

    Anyway…it’s not 1983 any more (sadly, in some ways), But I have observed that by curating content like this it FEELS like the kids are interacting with tech like we did back in the day; morning cartoons are once again morning cartoons and not a chance for MrBeast to invite my 8yr old to “comment, like and subscribe”.

    My eldest has ASD and I’ve noticed these small tweaks have had significant improvements on her behaviour / media consumption patterns (eg: she will get bored of media now and self regulate away from it…and…gasp…play).

    I dunno man. I’m trying out here. Shit ain’t easy. Too many plates spinning, not enough hands, and father time is a motherfucker. I’m tired, boss.


  • Possibly…but I think some of that depends too on what is meant by “online.” Obviously, if you frequent questionable sites and install unvetted software, that’s a bad idea. OTOH, having a machine with strict firewall rules (so not everything can just phone home), limited outbound access, no daily browsing/email, and only going online occasionally for specific, known downloads is a different situation than using it as a general-purpose internet PC.

    Even occasional access to a small number of mainstream, HTTPS-authenticated sites (e.g., major services where the browser can verify certificates) isn’t the same exposure as wide-open browsing. (nb: Firefox’s ESR releases have historically helped extend browser security support on older systems for a while, which can reduce risk somewhat - though obviously not indefinitely.)

    Look, I’m not arguing that EOL systems are “safe.” They’re not getting patches. But exposure matters. A mostly appliance-like gaming box that’s segmented and tightly controlled isn’t the same risk profile as someone’s primary web machine.

    ICBW and YMMV.


  • I like linux and I use it (Raspbian, Zorin, Ubuntu, Arch: diff machines). I also enjoy using Win 8.1 on my Lenovo M93p Tiny (8GB ram), as a Playnite appliance / console. This allows me to play emulated games (Wii, Gamecube, PS2, to about 1.5-2x upscale), ~2013ish era AAA titles (Fallout 3, Just Cause 2, Dead Rising 2, GTA IV) and select indy games (like Donut County, Untitled Goose Game, EXO ONE) all from one device.

    Normally, the advice would be to use something like Bazzite or Batocera (and I agree!)…but given the hardware limitations and the “it just runs” nature of these older Window games (under windows) I’ve had better experiences sticking to Win 8.1.

    YMMV but the “switch to linux cause windows too old” thing has some shades of gray.








  • You’re over-egging it a bit. A well written SOAP note, HPI etc should distill to a handful of possibilities, that’s true. That’s the point of them.

    The fact that the llm can interpret those notes 95% as well as a medical trained individual (per the article) to come up with the correct diagnosis is being a little under sold.

    That’s not nothing. Actually, that’s a big fucking deal ™ if you think thru the edge case applications. And remember, these are just general LLMs - and pretty old ones at that (ChatGPT 4 era). Were not even talking medical domain specific LLM.

    Yeah; I think there’s more here to think on.


  • Agreed!

    I think (hope) the next application of this tech is in point of care testing. I recall a story of a someone in Sudan(?) using a small, locally hosted LLM with vision abilities to scan hand written doctor notes and come up with an immunisation plan for their village. I might be misremembering the story, but the anecdote was along those lines.

    We already have PoC testing for things like Ultrasound… but some interpretation workflows rely on strong net connection iirc. It’d be awesome to have something on device that can be used for imaging interpretation where there is no other infra.

    Maybe someone can finally win that $10 million dollar X prize for the first viable tricorder (pretty sure that one wrapped up years ago? Too lazy to look)…one that isn’t smoke and mirror like Theranos.


  • Funny how people over look that bit enroute to dunk on LLMs.

    If anything, that 90% result supports the idea that Garbage In = Garbage Out. I imagine a properly used domain-tuned medical model with structured inputs could exceed those results in some diagnostic settings (task-dependent).

    Iirc, the 2024 Nobel prize in chemistry was won on the basis of using ML expert system to investigate protein folding. ML =! LLM but at the same time, let’s not throw the baby out with the bathwater.

    EDIT: for the lulz, I posted my above comment in my locally hosted bespoke llm. It politely called my bullshit out (Alpha fold is technically not an expert system, I didn’t cite my source for Med-Palm 2 claims). As shown, not all llm are tuned sycophantic yes man; there might be a sliver of hope yet lol.


    The statement contains a mix of plausible claims and minor logical inconsistencies. The core idea—that expert systems using ML can outperform simple LLMs in specific tasks—is reasonable.

    However, the claim that “a properly used expert system LLM (Med-PALM-2) is even better than 90% accurate in differentials” is unsupported by the provided context and overreaches from the general “Garbage In = Garbage Out” principle.

    Additionally, the assertion that the 2024 Nobel Prize in Chemistry was won “on the basis of using ML expert system to investigate protein folding” is factually incorrect; the prize was awarded for AI-assisted protein folding prediction, not an ML expert system per se.

    Confidence: medium | Source: Mixed



  • I don’t think it’s their information per se, so much as how the LLMs tend to use said information.

    LLMs are generally tuned to be expressive and lively. A part of that involves “random” (ie: roll the dice) output based on inputs + training data. (I’m skipping over technical details here for sake of simplicity)

    That’s what the masses have shown they want - friendly, confident sounding, chat bots, that can give plausible answers that are mostly right, sometimes.

    But for certain domains (like med) that shit gets people killed.

    TL;DR: they’re made for chitchat engagement, not high fidelity expert systems. You have to pay $$$$ to access those.



  • Agree.

    I’m sorta kicking myself I didn’t sign up for Google’s MedPALM-2 when I had the chance. Last I checked, it passed the USMLE exam with 96% and 88% on radio interpretation / report writing.

    I remember looking at the sign up and seeing it requested credit card details to verify identity (I didn’t have a google account at the time). I bounced… but gotta admit, it might have been fun to play with.

    Oh well; one door closes another opens.

    In any case, I believe this article confirms GIGO. The LLMs appear to have been vastly more accurate when fed correct inputs by clinicians versus what lay people fed it.



  • So, I can speak to this a little bit, as it touches two domains I’m involved in. TL;DR - LLMs bullshit and are unreliable, but there’s a way to use them in this domain as a force multiplier of sorts.

    In one; I’ve created a python router that takes my (deidentified) clinical notes, extracts and compacts input (user defined rules), creates a summary, then -

    1. benchmarks the summary against my (user defined) gold standard and provides management plan (again, based on user defined database).

    2. this is then dropped into my on device LLM for light editing and polishing to condense, which I then eyeball, correct and then escalate to supervisor for review.

    Additionally, the llm generated note can be approved / denied by the python router, in the first instance, based on certain policy criteria I’ve defined.

    It can also suggest probable DDX based on my database (which are .CSV based)

    Finally, if the llm output fails policy check, the router tells me why it failed and just says “go look at the prior summary and edit it yourself”.

    This three step process takes the tedium of paperwork from 15-20 mins to 1 minute generation, 2 mins manual editing, which is approx a 5-7x speed up.

    The reason why this is interesting:

    All of this runs within the llm (or more accurately, it’s invoked from within the llm. It calls / invokes the python tooling via >> commands, which live outside the LLMs purview) but is 100% deterministic; no llm jazz until the final step, which the router can outright reject and is user auditble anyway.

    Ive found that using a fairly “dumb” llm (Qwen2.5-1.5B), with settings dialed down, produces consistently solid final notes (5 out of 6 are graded as passed on first run by router invoking policy document and checking output). It’s too dumb to jazz, which is useful in this instance.

    Would I trust the LLM, end to end? Well, I’d trust my system, approx 80% of the time. I wouldn’t trust ChatGPT … even though its been more right than wrong in similar tests.