• 2 Posts
  • 350 Comments
Joined 2 years ago
cake
Cake day: June 6th, 2023

help-circle






  • Mi was trying out labplot yesterday, and as far as I can tell it can only really plot, not do any sort of transformation or data analysis. The plotting UI itself is pretty nice and the plots look good, but for most of my use cases its worth it to just spin up a Jupyter notebook and work with MatPlotLib directly.

    If it could become a general-purpose UI for matplotlib, thatd be fantastic, but its pretty limited in actual usability for me at the moment.







  • While I agree that publishers charging high open access fees is a bad practice, the ACS journals aren’t the kind of bottom-of-the-barrel predatory journals you’re describing. ACS nano in particular is a well respected journal for nanochem, with a generally well-respected editorial board, and any suspicions of editorial misconduct of the type you’re describing would be a three-alarm fire in the community.

    I will also note that this article is labelled “free to read” – when the authors have paid an (as you said, exhorbitant) publishing fee to have the paper be open access, the label used by ACS journals is “open access”. The “free to read” label would be an editorial decision, typically because the article is relevant outside the typical readerbase of the journal, and so it makes sense both from a practical perspective (and more cynically for the journal’s PR) to make it available to everyone, not just the community who has institutional access.

    Also, the fact that the authors had a little fun with the title doesn’t mean its low-effort slop – this was actually an important critique at the time, because for years people had been adding different modifications to graphene and making a huge deal about how revolutionary their new magic material was.

    The point this paper was trying to make is that finding modifications to graphene which make it better for electrocatalysis is not some revolutionary thing, because almost any modification works. It was actually a useful recalibration for expectations, as well as a good laugh.

    Edit: typo


  • Not somebody who knows a lot about this stuff, as I’m a bit of an AI Luddite, but I know just enough to answer this!

    “Tokens” are essentially just a unit of work – instead of interacting directly with the user’s input, the model first “tokenizes” the user’s input, simplifying it down into a unit which the actual ML model can process more efficiently. The model then spits out a token or series of tokens as a response, which are then expanded back into text or whatever the output of the model is.

    I think tokens are used because most models use them, and use them in a similar way, so they’re the lowest-level common unit of work where you can compare across devices and models.





  • My favorite overheard undergrad story:

    I was walking past the lecture hall right after an organic chemistry midterm, and there was a cluster of 4-5 students talking about the exam. One asked about question 8b, and another one said “you’re not supposed to mix nitric acid and ethanol, that makes TNT, right?” I had to stifle a chuckle as I walked by.

    So close, and yet so far! Nitrated acetone is explosive, and TNT (trinitrotoluene) is also made with nitric acid, but toluene is a much more complex molecule than acetone. If those undergrads could figure out how to turn acetone into TNT efficiently, they’d get a Nobel!


  • Agreed! I’m just not sure TOPS is the right metric for a CPU, due to how different the CPU data pipeline is than a GPU. Bubbly/clear instruction streams are one thing, but the majority type of instruction in a calculation also effects how many instructions can be run on each clock cycle pretty significantly, whereas in matrix-optimized silicon its a lot more fair to generalize over a bulk workload.

    Generally, I think its fundamentally challenging to generate a generally applicable single number to represent CPU performance across different workloads.