Hacker News new | past | comments | ask | show | jobs | submit | faxmeyourcode's comments login

Rock Auto has a fantastic to use website. It uses a very intuitive tree based representation and has search that actually works. It also shows you if a part is going to be shipped from the same warehouse as another part in your cart so you can save on shipping.

rockauto.com


I've got something very similar as a function in my ~/.zshrc, except instead of days separated by lines I start a new file for each topic.

Most of my notes are write-only for a day or so but I keep the old ones around just in case I need to grep through everything (and I've saved myself doing this a few times).

    function notes {
      THIS_MONTH="$(date '+%Y-%m')"

      mkdir -p "$NOTES_FOLDER/$THIS_MONTH" # create folder if it doesn't exist.
      cd "$NOTES_FOLDER/$THIS_MONTH"

      if [ -z "$1" ]; then
        return 0
      fi

      vim $1
      cd -
    }
Running `notes` will take you to today's directory to poke around, and `notes file.md` will open file.md in the appropriate dir.


I do something very similar for my public notes as well! I have `tn` to search for a note and open it in a browser, and `tne foo` to open my `foo.md` note in my editor, which in this case is not vim. I then commit them to git, where a post-commit hook uses mkdocs to build them, and gh-deploy to publish them EG https://danielhoherd.com/tech-notes/exiftool/ is the output of my exiftool.md note.


As always, it's the users who are at fault. Nevermind the thousands of kids and teens who are exposed to this shit day in and day out.


Before social media you had sites like rotten. And if you were a teenager with access to internet you perused some of those.


There's been garbage online forever. But you had to seek it out and if you didn't visit those sites you weren't going to come across this content organically in the same way.

Now you just have an infinite feed that shows you this stuff organically and unless you are diligent in clicking "not interested" over and over, it'll show up in your feed more and more often.


You checked out rotten once in a while

A lot of users spend HOURS each day doom scrolling and are exposed to 1000s of pieces of such content weekly if not daily


You didn't stumble up on Rotten while scrolling past your cousin's prom photos like you do with "suggested reels" on Insta.

It's being deliberately pushed to the kids, that's my problem.


It is the user's fault when the user says in his post that he doesn't interact with the algorithm or give it anything to go on except to click on content he doesn't like and then lo and behold it gives him content he doesn't like.

Children should not be on social media at all.


So the algorithm should show an elephant stomping a man to death to anyone who starts a new account until they "correct" "their algorithm" by interacting with better content?

The issue is that users are being pushed towards violence, gore, and pornography by default.

The bar for these giant multi billion dollar tech companies with hundreds of thousands of employees and the most advanced AI tech in the world is truly as low as it's ever been.


Snowflake and DuckDB are two flavors of SQL that allow things like trailing commas. My personal favorite feature is `GROUP BY ALL`.

    select 
      c1, 
      c2, 
      c3, 
      ...,
      c50,
      sum(c51),
    from 
      table
    group by all


I've been fighting trying to chunk SEC filings properly, specifically surrounding the strange and inconsistent tabular formats present in company filings.

This is giving me hope that it's possible.


(from the gemini team) we're working on it! semantic chunking & extraction will definitely be possible in the coming months.


>>I've been fighting trying to chunk SEC filings properly, specifically surrounding the strange and inconsistent tabular formats present in company filings.

For this specific use case you can also try edgartools[1] which is a library that was relatively recently released that ingests SEC submissions and filings. They don't use OCR but (from what I can tell) directly parse the XBRL documents submitted by companies and stored in EDGAR, if they exist.

[1] https://github.com/dgunning/edgartools


I'll definitely be looking into this, thanks for the recommendation! Been playing around with it this afternoon and it's very promising.


If you'd kindly tl;dr the chunking strategies you have tried and what works best, I'd love to hear.


isn't everyone on iXBRL now? Or are you struggling with historical filings?


XBRL is what I'm using currently, but it's still kind of a mess (maybe I'm just bad at it) for some of the non-standard information that isn't properly tagged.


I've wanted to do something like this for years. I might have to actually stop fiddling with the idea in my head and give it a real shot in 2025.

I'm curious - how does the design process go? Do you propose a design, do they usually have a pretty complete vision or do you have templates that they can take inspiration from?


force = mass * acceleration

work = force * displacement

work = mass * acceleration * displacement

lower mass, less work.


So my username is a little less ridiculous than I originally thought? :)

The fact that this can introduce OCR bugs into your C code is hilarious, and this is diabolical:

    #define one ( 4 - 3 )
    #define eleven ( 3 + 4 + 4 )

Source code is here https://github.com/lexbailey/compilerfax


> OCR bugs

Especially if your fax machine uses JBIG2 compression. See: https://googleprojectzero.blogspot.com/2021/12/a-deep-dive-i...


I think it's appropriate linking directly to Kriesel's blog¹ or his talk, as that's about the scanner creating fake data and not about rce. Though technically it too is not an OCR bug as there's no ocr in JBIG2.

¹: http://www.dkriesel.com/en/blog/2013/0802_xerox-workcentres_...


I wonder if OCR could be improved by adding a "language model" of sorts...

Like, sure, maybe it's hard to tell apart a "1", "i", or "l" purely visually, but if you knew it was supposed to be code, I'd suspect one could significantly improve the recognition accuracy if the system just worked in the probability of each confusable option given the preceding (and following) text.


This would also have a higher risk of introducing some nasty, hard to spot errors.

It's actually better for the compilation to fail than for the Clippy to make up something syntactically and compilation correct, but wrong.


You might be right in a practical sense, but for an art project like this, maybe not?


Need a proper preprocessor to take a code file and make it OCR-safe by substituting for dangerously glyphs.


This might be a good reason to support trigraphs again! https://en.wikipedia.org/wiki/Digraphs_and_trigraphs_(progra...

edit: fixed link, copy paste fail dropped the ++


Amateur! Use a barcode font!


monospace font OCR-B


I would not touch another computer for a very long time.


Yep. I'd cut computer technology out of my life as much as I could.


To me the videos are absolutely mind blowing, congratulations on the launch.

It feels like a leap forward in an area of tech that has totally stagnated (e-ink). I think this device is tackling a few hard problems all at once. Well done and I wish you luck.


> the videos

Please note (you will find info in these very pages) that they are tentative, made on prototypes. And they were apparently meant to be demonstrative of capabilities, not of real-life experience.

The Company says they will be updated.


Consider applying for YC's Summer 2025 batch! Applications are open till May 13

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: