• 2 Posts
  • 15 Comments
Joined 2 years ago
cake
Cake day: December 31st, 2023

help-circle






  • Learned helplessness is an insidious foe, and one that market forces have tended to side with over the past 20 years (probably for far longer than that, but as I was a mere child back then I wouldn’t claim it with as much certainty).

    It’s an “easy way” for those like you and me who have more or less already built up the know-how over countless small steps, but if you’ve never known “life” outside of these corporate surveillance playgrounds I imagine it seems very scary and deserted.


  • I agree that in the short-term, NVidia stand to make more money as a destroyed datacenter prompts it’s owners to rebuild it. However, until the armed conflict stops, there might be a pause on datacenter construction in the region - why pay to (re)build a datacenter if it’s probably going to be blown up again? From my (admittedly limited) reading of history, most reconstruction only happens in the years following an armed conflict’s end.




  • I’m disappointed there’s no threadiverse integration.

    The basic grunt work of bringing together bluesky and mastodon feeds is pretty decently done, if a bit opinionated.

    I like the concept and presentation of their feed “packages”, but for most of the feeds I’ve been proposed in the app I’m not interested in the entire package. It’s nice that you can still directly follow the actors packaged into a feed from that feed’s page.

    It’s very annoying that Surf wants their/my Surf account to act as the intermediary for my mastodon account. From what I can see, this means I need to exit the app to take some action with my mastodon account. Maybe I just haven’t found the option, but if there isn’t one then it’s a hard blocker for me to use it as my main fediverse browser.



  • A small gui to automate generating some pdfs from some CSV files.

    There’s a small non-profit in my area helping people operate localized energy distribution (as producers and consumers). Each month, they receive a zip file containing the raw kiloWatt-hours produced and consumed by each participant over the past month as CSV files. So far the non-profit has been manually importing these CSVs into LibreOffice to generate graphs and tables and export the whole thing as an individualized PDF file for each participant. Now that they’re starting to help more than 2-3 operations, it’s become useful to try to automate that process.

    I’ve been writing it in rust for a few reasons. First of all I wanted cross-compilation to be sure to work and at this point I’m more familiar with rust than go, secondly I read a blog post recently that evaluated rust gui solutions in terms of accessibility and IME-compatibility on windows. I started off looking for a “direct” pdf-writing library but eventually switched to using typst to generate the pdfs from templates I write. typst being written in rust has enabled me to bundle its engine into the program in a pretty-straightforward way.

    I’m currently working on allowing the import of multiple sets of data so that the generated PDFs can show line plots of the electricity production and consumption over several months.


    1. chunk_size := file_size / cpu_cores. Compile regex.

    2. spawn cpu_cores workers:
      2.a. worker #n starts at n * chunk_size bytes. If n > 0, skip bytes until newline encountered.
      2.b worker starts feeding bytes from file/chunk into regex. When match is found, write to output (stdout or file, whichever has better performance). When newline encountered, restart regex state automata.
      2.c after having read chunk_size bytes, continue until encountering a newline to ensure the whole file is covered by the parallel search.

    Optionally, keep track of byte number and attach them to the found matches when outputting, to facilitate eventually de-duplicating and/or navigating to said match in the file.

    To avoid problems, have each worker output to a separate file, and only combine these output files when the workers are all finished.

    As others have said, it’s going to be hard to get more speedup than this, and you will ultimately be limited by your storage’s read speed and throughput if the whole file cannot fit into memory.