They/Them, agender-leaning scalie.

ADHD software developer with far too many hobbies/trades: AI, gamedev, webdev, programming language design, audio/video/data compression, software 3D, mass spectrometry, genomics.

Learning German (B2), Chinese (HSK 3-4ish), French (A2).

  • 0 Posts
  • 19 Comments
Joined 1 year ago
cake
Cake day: June 18th, 2023

help-circle




  • Western companies no longer operating in the Russian market, but still producing desirable content. … Western companies have ‘legalized’ piracy in Russia.

    100% this.

    Media is culture, and IMO people have a right to participate in culture. If it’s excessively difficult or impossible to legitimately access culture, one has the moral right to illegitimately access culture, and share it so others also have access.

    It’s inexcusable to refuse to directly sell media. The internet has made it easier than ever to trade access to media for money. Geo-restricted subscription services should be a nice add-on option for power-consumers, not the only way to get access to something.




  • The website does a bad job explaining what its current state actually is. Here’s the GitHub repo’s explanation:

    Memory Cache is a project that allows you to save a webpage while you’re browsing in Firefox as a PDF, and save it to a synchronized folder that can be used in conjunction with privateGPT to augment a local language model.

    So it’s just a way to get data from browser into privateGPT, which is:

    PrivateGPT is a production-ready AI project that allows you to ask questions about your documents using the power of Large Language Models (LLMs), even in scenarios without an Internet connection. The project provides an API offering all the primitives required to build private, context-aware AI applications.

    So basically something you can ask questions like “how much butter is needed for that recipe I saw last week?” and “what are the big trends across the news sites I’ve looked at recently?”. But eventually it’ll automatically summarize and data mine everything you look at to help you learn/explore.

    Neat.


  • I agree that older commercialized battery types aren’t so interesting, but my point was about all the battery types that haven’t had enough R&D yet to be commercially mass-produced.

    Power grids don’t care much about density - they can build batteries where land is cheap, and for fire control they need to artificially space out higher-density batteries anyway. There are heaps of known chemistries that might be cheaper per unit stored (molten salt batteries, flow batteries, and solid state batteries based on cheaper metals), but many only make sense for energy grid applications because they’re too big/heavy for anything portable.

    I’m saying it’s nuts that lithium ion is being used for cases where energy density isn’t important. It’s a bit like using bottled water on a farm because you don’t want to pay to get the nearby river water tested. It’s great that sodium ion could bring new economics to grid energy storage, but weird that the only reason it got developed in the first place was for a completely different industry.





  • Yeah, I was over-enthusiastic based on their cherry-picked examples. SeamlessExpressive still leaves a lot to be desired.

    It has a limited range of emotions and can’t change emotion in the middle of the clip. It can’t produce the pitch shifts of someone talking excitedly, making the output sound monotonous. Background noise in the input causes a raspy, distorted output voice. Sighs, inter-sentence breaths, etc. aren’t reproduced. Sometimes the sentence pacing is just completely unnatural, with missing pauses or pauses in bad places (e.g. before the sentence-final verb in German).

    IMO their manual dataset creation is holding them back. If I was in this field, I would try to follow the LLM route: Start with a next-token predictor trained indiscriminately on large-scale speech+text data (e.g. TV shows, movies, news radio, all with subtitles even if the subs need to be AI generated), fine-tune it for specific tasks (mainly learning to predict and generate based on “style tokens” (speaker, emotion, accent, pacing)), then generate a massive “textbook” synthetic dataset. The translation aspect could be almost completely outsourced to LLMs or multilingual subtitles.




  • Sadly archive.li seems to be in a broken CAPTCHA loop, so I can’t see the full article. However, I’m struggling to imagine a fundamental universe-spanning interaction that triggers weeks after the big bang, given that the universe has already expanded/cooled enough by 20 minutes to stop fusing nuclei. If there is evidence for a Dark Matter big bang for weeks after the Matter big bang, surely this must have some extreme implications about the possible mass range of DM particles?

    One thing nobody seems to be talking about: Just like String Theory, the more new phenomena are needed to make the Dark Matter model work, the further we stray from the edge of Occam’s Razor. While all the research into detecting hypothetical particles has been fun to follow, I can’t help but feel we’re just a few equations away from discovering that the universe is actually pretty MONDane.



  • Newtra@pawb.socialtoTechnology@lemmy.worldUnity apologises.
    link
    fedilink
    English
    arrow-up
    16
    ·
    1 year ago

    They’ve had days to prepare this response. They didn’t rescind or explain the one thing that people universally hated, which means they’re just stalling and trying to save their reputation without actually changing trajectory.

    We’ve seen this corporate bullshit so much in recent years. No more “benefit of the doubt”.


  • ooo, I love this. It reminds me of how nice C#'s LINQ is…

    “Pipeline style” DB queries have some interesting advantages as well:

    • It’s straightforward to write efficient queries for DBs that don’t include a query optimizer stares at Datomic
    • You can split the pipeline into server-side and client-side steps when working with less capable DBs stares at most of NoSQL
    • It would be much easier to transition from a pipeline API to a non-text-based API so that our ORMs/query builders can directly talk to DBs without the overhead of generating and parsing SQL.