25+ yr Java/JS dev
Linux novice - running Ubuntu (no windows/mac)

  • 0 Posts
  • 306 Comments
Joined 1 year ago
cake
Cake day: June 14th, 2023

help-circle
  • I think job postings are better in indeed, but tbh >75% I’ve gotten in pretty much my whole 25+ year career has been through a recruiter. Dice.com used to be big for tech jobs back in the day but I’m not sure any more.

    As a SSE, mostly I have recruiters hitting me up through linked in. This is also a really bad time. I’ve been back to work for about a month after 5 months of not finding anything. That’s the worst drought I’ve had in almost 15 years. Usually it’s < 1 month.

    Be seriously prepared about cloud. It’s so anyone fucking wants right now. I’m a damn good Java/js developer, but I’m still learning the tech stack and I haven’t touched a line of code yet in this job. Everything has been configuration and pipelines. I feel more like devops than developer.


  • Me and some old guildies have kept in touch off and on over the years. Every once in a while I’d buy a wow expansion and do a couple of dungeons. We were really looking forward to making Diablo 4 our new hang out.

    We played like hell all through the beta. Then like twice in live. Then we all kinda decided it sucked. I think my good friend’s daughter is graduating soon. Or possibly already did. I can’t remember how much older than my own kids she was. I can remember when she was born.

    He’s still like a brother to me, but we’ve got fuck all in common any more and we can’t keep talking about glory days that were damn near 20 years ago.






  • I don’t think I would agree that just because something is public that it’s a public forum. I feel like the public has to own it as well. I looked it up and maybe it’s because I predate social media by rather a lot, but I think of it in the classical sense:

    Public forums are typically categorized into three types:

    1. Traditional Public Forums: Long-established spaces like parks or sidewalks, where people have historically exercised their rights to free speech and assembly.
    2. Designated Public Forums: Areas that the government intentionally opens up for public expression, such as town halls or school meeting rooms.
    3. Limited Public Forums: Spaces opened for specific types of discussions or activities but with certain restrictions on the subject matter or participants.

    The important factor being public ownership of the forum. I will concede that it has colloquially come to include public social media, but I think it’s important to distinguish that it’s not really the same thing at all as has been discussed through most of our history.

    Food for thought. I just think calling them public forums attaches too much importance to a profit seeking endeavor.


  • I would only note that for the vast majority of my experience these streams can only return up to a single match. Determinism isn’t really preserved by findFirst, either, unless the sort order is set up that way.

    Finding the first Jim Jones in a table is no more reliable that finding any Jim Jones. But finding PersonId 13579 is deterministic whether you findFirst or findAny.

    Perhaps you work in a different domain where your experience is different.


  • I try to prefer .findAny() over .findFirst() because it will perform better in some cases (it will have to resolve whether there are other matches and which one is actually first before it can terminate - more relevant for parallel streams I think. findAny short circuits that) but otherwise I like the first. I’d probably go with some sort of composed predicate for the second, to be able to easily add new criteria. But I could be over engineering.

    I mostly just posted because I think not enough people are aware of the reasons to use findAny as a default unless findFirst is needed.



  • I asked ChatGPT for a tldr because same. The result reads like ad copy. Idk, man.


    The memory packaging market is evolving with advancements like flip-chip, wire-bond, and through-silicon via (TSV) technologies. These innovations enable smaller, more powerful, and faster devices, particularly in smartphones, where efficient space use is crucial for sleek designs. DRAM, while still used in PCs, faces declining adoption due to its complexity and the rise of alternatives like 3D TSV, which offer better functionality. The APAC region, especially China, is leading the growth in memory packaging, driven by investments in assembly infrastructure and rising demand for mobile applications using system-in-package (SiP) technologies.







  • You made a lot of points here. Many I agree with, some I don’t, but I specifically want to address this because it seems to be such a common misconception.

    It does and it doesn’t discard the original. It isn’t impossible to recreate the original (since all the data it gobbled up gets stored somewhere in some shape or form and can be truthfully recreated, at least judging by a few comments bellow and news reports). So AI can and does recreate (duplicate or distribute, perhaps) copyrighted works.

    AI stores original works like a dictionary does. All the words are there, but the order and meaning is completely gone. An original work is possible to recreate by randomly selecting words from the dictionary, but it’s unlikely.

    The thing that makes AI useful is that it understands the patterns words are typically used in. It orders words in the right way far more often than random chance. It knows “It was the best of” has a lot of likely options for the next word, but if it selects “times” as the next word, it’s far more likely to continue with, “it was the worst of times.” Because that sequence of words is so ubiquitous due to references to the classic story. But over the course of following these word patterns, it will quickly glom onto a different pattern and create a wholly new work from the original “prompt.”

    There are only two cases in which an original work should be duplicated: either the training data is far too small and the model is overtrained on that particular work, or the work is the most derivative text imaginable lacking any flair or originality.

    Adding more training data makes it less likely to recreate any original works.

    I am aware of examples where it was claimed an LLM reproduced entirely code functions including original comments. That is either a case of overtraining, or far too many people were already copying that code verbatim into their own, thus making that work very over represented in the training data (same thing, but it was infringing developers who poisoned the data, not researchers using bad training data).

    Bottom line: when created with enough data, no original works are stored in any way that allows faithful reproduction other than by chance so random that it’s similar to rolling dice over a dictionary.

    None of this means AI can do no wrong, I just don’t find the copyright claim compelling.