I like your optimism best to look on the bright side and all— curious what do you mean by fabless? Do they not require as complex facilities because they’re a larger process or something? Or for some other reason?
I like your optimism best to look on the bright side and all— curious what do you mean by fabless? Do they not require as complex facilities because they’re a larger process or something? Or for some other reason?
deleted by creator
Hmm I can’t decide if this is a joke or if I’m just very privileged in the internet department
For fucks sake I hope your joking x.x
I remember back in 2013 I picked up the full expansion set at a thrift store for $4 it was the best summer of middle school by far— got the werewolf mod working and went around finding out who was essential to the plot by killing them and then reverting the saves— good times
Can it be running some unix derivative?
Thanks for the tip! https://github.com/jcuberdruid/DeepBlue
I’ll add more to the repo but here it is for now :)
Not just yet but it’s in the works!
Thanks! I made a submarine game called DeepBlue where you explore the ocean floor generated by ocean depth map data:
Millions of hard working Dutch people depend on the tulip system for their retirement. The crash was not due to the management of the tulip market, rather it’s an unavoidable pattern of boom and bust. Given this, it’s only right that the Dutch government to bail out the tulip funds to avoid a breakdown of the liliaceae system.
While the iris study is interesting, looking at their dataset the pictures seem to include the area around the eye a little bit, including eye lashes, so after a cursory glance it seems odd that they even titled it as iris. However I didn’t read the full thing so it cold be that they cropped it somewhere. Although they are using large convolutions so a lot of detail is lost.
I do really like the error system in rust for its descriptions. I guess the difficulty for me, which maybe will go away after writing more rust, is that my intuition for what is efficient and what isn’t totally breaks down.
I find myself passing copies of values around and things like that, it might be that the compiler just takes care of that, or that I just don’t know how to do it well but that’s often the point of friction for me.
Totally agree on the refactor though, most of the time it doesn’t even take that much time since you know the skeleton of what you want at that point!
Maybe it’s just because I haven’t had to deal with the scenario yet but does compile time really matter? I mean for small programs it seems it’s almost instant on modern machines and for large programs I would assume, if it exists, that you would be using the equivalent of make so you would only be recompiling the small changes made.
deleted by creator
Llama2 is pretty good but there are a ton of different models which have different pros and cons, you can see some of them here: https://ollama.com/library . However I would say that as a whole models are generally slightly less polished compared to chat gpt.
To put it another way: when things are good they’re just as good, but when things are bad the AI will start going off the rails, for instance holding both sides on the conversation, refusing to answer, just saying goodbye, etc. More “wild westy” but you can also save the chats and go back to them so there are ways to mitigate, and things are only getting better.
Also Swift!
While I agree wholeheartedly with the idea that we need to emphasize quality over quantity, so long as software pays well there will be people who don’t care. In my university I’ve met a fair few people that complain about having to learn about compilers, assembly, and whatnot because “I’ll never need to know that in my actual job”. While to some extent in the United States you can blame the fact that classes just cost a ton, I think it’s a sad reality that, barring some key change in the way our whole education and economic systems work, there will be unimaginative apathetic people that will ruin things for the rest. Plus people are fallible or something I dunno. But yeah void pointers are my jam because I don’t have to wait precious clock cycles making new ones jk.
Just run ollama locally and download uncensored versions— runs on my m1 MacBook no problem and is at the very least comparable to chatgpt3. Unsure for images though, but there should be some open source options. Data is king here, so the more you use a platform the better its AI gets (generally) so don’t give the corporations the business.
You could just add a small nonvolatile buffer to each camera if it’s not wired, such that if it loses connection with your home assistant server it will start recording. With 720p video and a 64gb flash storage you could, depending on encoding, store well over a day of footage. (Napkin math so could be wrong)
IRC: it’s open source, it’s free, its retro