vampatori

joined 2 years ago
[–] [email protected] 3 points 2 years ago (1 children)

Definitely give Ruthless a go, I love it.. reminds me of early game ARPG's on higher difficulties. Positioning really matters, you have to adapt based on what you get. It seems to have been the proving ground for PoE2's new tempo.

[–] [email protected] 1 points 2 years ago

I was going to do an origin character as a solo play-through and a custom character for a group play-through with my mates, but now I might do it the other way around... which means hours in the character creator! Ha.

[–] [email protected] 8 points 2 years ago

Often the question marked as a duplicate isn't a duplicate, just the person marking it as such didn't spend the time to properly understand the question and realise how it differs. I also see lots of answers to questions mis-understanding the question or trying to force the person asking down their own particular preference, and get tons of votes whilst doing it.

Don't get me wrong, some questions are definitely useful - and some go above-and-beyond - but on average the quality isn't great these days and hasn't been for a while.

[–] [email protected] 19 points 2 years ago (1 children)

Google's first quarter 2023 report shows they made massive profits off vast revenue due to advertising.

It is about control though. The thing that caught my eye is that they're saying that only "approved" browsers will be able to access these WEI sites. So what does that mean for crawlers/scrapers? That the big tech companies on the approval board will be able to lock potential competitors out of accessing the web - new browsers, search engines, etc. but much more importantly... Machine Learning.

Google's biggest fear right now is that ML systems will completely eliminate most people's reason to use Google's search, and therefore their main source of revenue will plummet. And they're right to be scared, it's already starting to happen and it's showing us very quickly just how bad Google's search results are.

So this seems to me like an attempt to control things from that side. It's essentially the "big boys" trying to consolidate and firm-up their hold in the industry and not let newcomers rival them, as with ML the barrier to entry has never been lower.

[–] [email protected] 3 points 2 years ago

NPCs! Definitely not PCs! Ha.

I have seen people that don't track spell slots for NPCs and just have them all at-will, which I think is an interesting idea. But I tend to give players non-combat objectives in their encounters, which prolongs them significantly so spell slot usage can become important for balance for NPCs in those cases.

[–] [email protected] 2 points 2 years ago (2 children)

I just have a series of "pips" that I colour-in when used and erase when claimed back. Super simple, easy to see at-a-glance, and robust so it's not going to get messed up in my bag. Added bonus is that it works when being DM too and you have several casters to track simultaneously.

[–] [email protected] 1 points 2 years ago (1 children)

How do Linux distro's deal with this? I feel like however that's done, I'd like node packages to work in a similar way - "package distro's". You could have rolling-release, long-term service w/security patches, an application and verification process for being included in a distro, etc.

It wouldn't eliminate all problems, of course, but could help with several methods of attack, and also help focus communities and reduce duplication of effort.

[–] [email protected] 3 points 2 years ago* (last edited 2 years ago)

If I’m okay with the software (not just trying it out) am I missing out by not using dockers?

No, I think in your use case you're good. A lot of the key features of containers, such as immutability, reproduceability, scaling, portability, etc. don't really apply to your use case.

If you reach a point where you find you want a stand-alone linux server, or an auto-reconfiguring reverse proxy to map domains to your services, or something like that - then it starts to have some additional benefit and I'd recommend it.

In fact, using native builds of this software on Windows is probably much more performant.

[–] [email protected] 7 points 2 years ago (7 children)

Containers can be based on operating systems that are different to your computer.

Containers utilise the host's kernel - which is why there needs to be some hoops to run Linux container on Windows (VM/WSL).

That's one of the most key differences between VMs and containers. VMs virtualise all the hardware, so you can have a totally different guest and host operating systems; whereas because a container is using the host kernel, it must use the same kind of operating system and accesses the host's hardware through the kernel.

The big advantage of that approach, over VMs, is that containers are much more lightweight and performant because they don't have a virtual kernel/hardware/etc. I find its best to think of them as a process wrapper, kind of like chroot for a specific application - you're just giving the application you're running a box to run in - but the host OS is still doing the heavy lifting.

[–] [email protected] 2 points 2 years ago

I was using file merging, but one issue I found was that arrays don't get merged - and since switching to use Traefik (which is great) there are a lot of arrays in the config! And I've since started using labels for my own tooling too.

[–] [email protected] 2 points 2 years ago (1 children)

I was recently helping someone working on a mini-project to do a bit of parsing of docker compose files, when I discovered that the docker compose spec is published as JSON Schema here.

I converted that into TypeScript types using JSON Schema to TypeScript. So I can create docker compose config in code and then just export it as yaml - I have a build/deploy script that does this at the end.

But now the great thing is that I can export/import that config, share it between projects, extend configs, mix-in, and so on. I've just started doing it and it's been really nice so far, when I get a chance and it's stabilised a bit I'm going to tidy it up and share it. But there's not much I've added beyond the above at the moment (just some bits to mix-in arrays, which was what set me off on this whole thing!)

[–] [email protected] 10 points 2 years ago* (last edited 2 years ago)

I just have a static page that I randomly change - you can see mine here. In this case I was testing the idea of having text within an SVG for better scaling from mobile to desktop, and also I'm loving orange and purple at the moment for some reason! Oh, and I was testing automated deployments from CI/CD, so I always use my own base domain with those first tests!

 

I run my own small software development company and I'd like somewhere my clients can login and get access to things like:

  • Access to documents from their repo(s) (GitHub, all contracts/etc. are kept here)
  • Links to invoices and to pay.
  • Milestone progress from their repo(s) (GitHub)
  • Links to their test, staging, and production services.
  • Ability to get in touch (potentially raising an issue in GitHub).

We're just doing things manually for now, but before we reinvent the wheel I thought it would be useful to see what's out there to either use directly or extend.

 

Is there some formal way(s) of quantifying potential flaws, or risk, and ensuring there's sufficient spread of tests to cover them? Perhaps using some kind of complexity measure? Or a risk assessment of some kind?

Experience tells me I need to be extra careful around certain things - user input, code generation, anything with a publicly exposed surface, third-party libraries/services, financial data, personal information (especially of minors), batch data manipulation/migration, and so on.

But is there any accepted means of formally measuring a system and ensuring that some level of test quality exists?

view more: next ›