• 0 Posts
  • 14 Comments
Joined 11 months ago
cake
Cake day: January 25th, 2024

help-circle

  • Now would be a good time to look for a .com you like, or one of the more common TLDs. And register it at Namecheap, Porkbun, or Cloudflare. (Cloudflare is cheapest but all-eggs-in-one-basket is a concern for some.)

    Sadly, all the cheap or fun TLDs have a habit of being blocked wholesale, either because the cheap ones are overused by bad actors or because corporate IT just blacklists “abnormal” TLDs (or only whitelists the old ones?) because it’s “easy security”.

    Notably, XYZ also does that 1.111B initiative, selling numbered domains for 99¢, further feeding the affordability for bad actors and justifying a flat out sinkhole of the entire TLD.

    I got a three character XYZ to use as a personal link shortener. Half the people I used it with said it was blocked at school or work. My longer COM poses no issue.


  • Some would think this is horrible, but to me, it would be wholly dependent on the title/what was bought and sold.

    Nothing in this world is free. Development, servers, character licensing, it all costs money and if those costs aren’t passed down, you’ll never afford to continue. So for a game, especially one with online content or continuing content, to be free to play, money has to come from somewhere.

    Where the road splits is what is being sold. Things that give an edge in the game, pay-to-win? Uninstalled. Time limited FOMO triggers? Disgusting. Random loot boxes? Begone foul spirit.

    On the other end, if all that is for sale is shiny baubles and trinkets, things no one needs but can have as a reward for “supporting development”? I’m cool with that. If I feel no requirement to pay up, it’s being handled right, and if I like they game, sure, I can part with a fiver to look like I’m dipped in gold or whatever the supporter pack adds to help them keep the lights on(at least until I get bored of it in a week or two and switch back :P).

    I’d be curious what the divide is between the two kinds of purchases are. I’m sure I’ll be disappointed to find it was mostly P2W scum, though.


  • PassingThrough@lemmy.worldtoMemes@lemmy.mlGet rich quick
    link
    fedilink
    English
    arrow-up
    1
    ·
    6 months ago

    Forgive me, I’m no AI expert to fully compare the needed tokens per second measurement to relate to the average query Siri might handle, but I will say this:

    Even in your article, only the largest model ran at 8/tps, others ran much faster, and none of these were optimized for a task, just benchmarking.

    Would it be impossible for Apple to be running an optimized model specific to expected mobile tasks, and leverage their own hardware more efficiently than we can, to meet their needs?

    I imagine they cut out most worldly knowledge etc/use a lightweight model, which is why there is still a need to link to ChatGPT or Apple for some requests, would this let them trim Siri down to perform well enough on phones for most requests? They also advertised launching AI on M1-2 chip devices, which are not M3-Max either…


  • PassingThrough@lemmy.worldtoMemes@lemmy.mlGet rich quick
    link
    fedilink
    English
    arrow-up
    1
    ·
    edit-2
    6 months ago

    Onboard AI chips will allow this to be local.

    Phones do not have the power to ~~~

    Perhaps this is why these features will only be available on iPhone 15 Pro/Max and newer? Gotta have those latest and greatest chips.

    It will be fun to see how it all shakes out. If the AI can’t run most queries on the phone with all this advertising of local processing…there’ll be one hell of a lawsuit coming up.

    EDIT: Finished looking for what I thought I remembered…

    Additionally, Siri has been locally processed since iOS 15.

    https://www.macrumors.com/how-to/use-on-device-siri-iphone-ipad/


  • PassingThrough@lemmy.worldtoMemes@lemmy.mlGet rich quick
    link
    fedilink
    English
    arrow-up
    12
    arrow-down
    1
    ·
    6 months ago

    I think there’s a larger picture at play here that is being missed.

    Getting the weather is a standard feature for years now. Nothing AI about it.

    What is “AI” is, Hey Siri, what is the weather at my daughter’s recital coming up?

    The AI processing, calculated on-device if what they claim is true, is:

    1. the determination of who your daughter is
    2. What is a recital? An event? Are there any upcoming calendar events that match this concept?
    3. Is the “daughter” associated with this event by description or invitation? Yes? OK, what’s the address?
    4. Submit zip code of recital calendar event involving the kid to the weather API, and churn out a reply that includes all this information…

    Well {Your phone contact name}, it looks like it will {remote weather response} during your {calendar event from phone} with {daughter from contacts} on {event date}.

    That is the idea between on-device and cloud processing. The phone already has your contacts and calendar and does that work offline rather than educating an online server about your family, events and location, and requests the bare minimum from the internet, in this case nothing more than if you opened the weather app yourself and put in a zip code.


  • Plug it into a monitor or TV and keep an eye on the console.

    I have an older NUC that will not cooperate with certain brands of NVMe drive under PVE…the issue sounds like yours where it would work for an arbitrary amount of time before crashing the file system, attempting to remount read-only and rendering the system inert and unable to handle changes like plugging a monitor in later, yet it would still be “on”.




  • I’ll take a compromise where “3.1” is etched in each head end, and I can trust that “3.1” means something, and start with that.

    The real crux of the issue is that there is no way to identify the ability of a port or cable without trying it, and even if labeled there is/was too much freedom to simply deviate and escape spec.

    I grabbed a cable from my box to use with my docking station. Short length, hefty girth, firm head ends, certainly felt like a featured video/data/Dock cable…it did not work. I did work with my numpad/USB-A port bus thing though, so it had some data ability(did not test if it was 2.0 or 3.0). The cable that DID work with my docking station was actually a much thinner, weaker feeling one from a portable monitor I also had. So you can’t even judge by wiring density.

    And now we have companies using the port to deviate from spec completely, like the Raspberry Pi 5 technically using USB-C, but at a power level unsupported by spec. Or my video glasses that use USB-C connections all over, with a proprietary design that ensures only their products work together.

    Universal appearance, non-universal function, universal confusion.

    I hate it. At least with HDMI, RCA, 3.5mm, Micro-USB…I could readily identify what a port and plug was good for, and 99/100 the unknown origin random wires I had in a box worked just fine.


  • Actually, that leads me to another point:

    One upon a time, the concept behind a universal USB-C connector was so we could do exactly that.

    Laptop? Phone? Camera? America? Germany? Japan? Power? Connect the to TV? Internet?

    Wouldn’t matter anymore. USB-C to cover it all. Voltage high for the laptop, low for the camera, all available just the same in every country, universal. So yes, fill the airports and hotels with them. Use them for power and to play videos on the TV. Because we weren’t supposed to have to question the voltage or abilities of the ports and cables in use.

    Did/will that future materialize?


  • I feel the only place for a €1 cable is met by those USB-A to C cables that you get with things for 5V charging. That’s it. And it’s very obvious what the limits on those are by the A plug end.

    Anything that wants to be USB-C on both ends should be fully compatible with a marked spec, not indistinguishable from a 5V junk wire or freely cherry picking what they feel like paying for.

    Simply marking on the cable itself what generation it’s good for would be a nice start, but the real issue is the cherry picking. The generation numbers don’t cover a wire that does maximum everything except video. Or a proprietary setup that does power transfer in excess of spec(Dell, Raspberry Pi 5). But they all have the same ends and lack of demarcation, leading to the confusion.



  • As someone with video glasses like those included here, it might be a step forward but it has a lot of room for improvement before it will survive mass market.

    For starters, unlike a screen, these glasses must be tailored to your eyesight. If you wear prescription, you will need to fit double glasses or have some ability for the video ones to be prescription. And a huge problem in the market right now is pupil distance, or eye spacing/head size. Mass market wants one-size-fits all, but that means those outside the designed size will have difficulty using them if they can at all.

    These are problems currently experienced with the current market like Rokid, XReal, and Viture.

    And then of course there’s power, if we keep to 1080p we’ll need more computing power and battery than a Steam Deck screen, which some handhelds might be able to accommodate, maybe more so depending on the weight and shape trades of the new style. But so far it might be disappointing, especially if it has the appearance of a huge screen and still needs to low-res upscale/FSR to meet performance.

    Just my thoughts. Still cool, but no confidence in it as a winner yet.