Kissaki

joined 2 years ago
MODERATOR OF
[–] Kissaki 5 points 2 days ago

The main things a social media presence does is give exposure, which can then lead to interest and [mental] investment.

Without those things, the game disappears in a mass of titles on a huge platform. How are people supposed to find it? Luck?

Even if it's a good game, it's a harsh environment.

Traditionally, publishers did promotional work to increase exposure. If you publish as an indie onto Steam, without an exposure strategy, you have to have an exceptional product or exceptional luck to be successful.

I can empathize with their frustrations though. Working long and hard on something, especially if it's a good product, being satisfied with just the fact that it's a good product, with no/little audience, success, or financial success, is hard to swallow.

[–] Kissaki 5 points 2 days ago (1 children)

The strength of open-source

"Killing" seems like an inappropriate term for end-of-support though. That's clickbait to me.

[–] Kissaki 2 points 4 days ago* (last edited 4 days ago)

How do you intend to present your GitHub portfolio to your potential employers? Nobody's going to do a full, in-depth, or even basic analysis of your repos unless maybe with automated tools or what GitHub itself provides.

Your CV and interview are much more important. Solutions [and projects] matter much more than details. Experience and that you can talk about your work or experience is much more important than technical details.

A hash table library doesn't sound like particularly noteworthy expertise. Adding a dependency and calling simple documented methods on it in a simple, standard behavior manner isn't noteworthy.

If you're implementing your own, I wonder if "simply" implies a non-noteworthy implementation, or in-depth exploration of hashing and storage indexing. The latter would be a different project though, putting your other on hold.

I don't see it making a difference for employers what you pick here specifically.

If you're interested in implementing one or learning about the technicalities of it go ahead. Otherwise use a library and continue with your project or other interests.


Disclaimer: I'm not in the recruiting space nor do I have that much or recent experience being interviewed/the broader companies hiring processes.

[–] Kissaki 5 points 4 days ago (1 children)

Yes, every unfamiliar language requires some learning. But I don't think the bash syntax is particularly approachable.

I searched and picked the first result, but this seems to present what I mean pretty well https://unix.stackexchange.com/questions/248164/bash-if-syntax-confusion which doesn't even include the alternative if parens https://stackoverflow.com/questions/12765340/difference-between-parentheses-and-brackets-in-bash-conditionals

I find other languages syntaxes much more approachable.

I also mentioned the magic variable expansion operators. https://www.gnu.org/software/bash/manual/html_node/Shell-Parameter-Expansion.html

Most other languages are more expressive.

[–] Kissaki 1 points 4 days ago

OP could pick any language and have the same problem. Except maybe Python, but even that strays into symbolic line noise once a project gets big enough.

Personally, I don't see python far off from bash. Decent for small scripts, bad for anything bigger. While not necessarily natively available, it's readily available and more portable (Windows), and has a rich library ecosystem.

Personally, I dislike the indent syntax. And the various tooling and complexities don't feel approachable or stable, and structuring not good.

But maybe that's me. Many people seem to enjoy or reach for python even for complex systems.

More structured and stable programming languages do not have these issues.

[–] Kissaki 2 points 4 days ago (1 children)

that assumes you don't write any SQL

[–] Kissaki 19 points 4 days ago

I love the description as well. "One." "Zero."

[–] Kissaki 8 points 4 days ago

I found the comments/answers about backwards compatibility of not defined booleans and negative true interesting and plausible.

What I first thought of was that TRUE and FALSE can be redefined, so it serves as ensurance that within the library consistent values are being used no matter what other libs and callers do with their typing and definitions.

[–] Kissaki 2 points 4 days ago* (last edited 4 days ago) (3 children)

Microsoft SQL Server has a bit type and you always use 0 and 1 and cast/convert them. No native bool type. It's a hassle.

[–] Kissaki 7 points 4 days ago (5 children)

In your own description you added a bunch of considerations, requirements of following specific practices, having specific knowledge, and a ton of environmental requirements.

For simple scripts or duck tape schedules all of that is fine. For anything else, I would be at least mindful if not skeptical of bash being a good tool for the job.

Bash is installed on all linux systems. I would not be very concerned about some dependencies like sqlite, if that is what you're using. But very concerned about others, like jq, which is an additional tool and requirement where you or others will eventually struggle with diffuse dependencies or managing a managed environment.

Even if you query sqlite or whatever tool with the command line query tool, you have to be aware that getting a value like that into bash means you lose a lot of typing and structure information. That's fine if you get only one or very few values. But I would have strong aversions when it goes beyond that.

You seem to be familiar with Bash syntax. But others may not be. It's not a simple syntax to get into and intuitively understand without mistakes. There's too many alternatives of if-ing and comparing values. It ends up as magic. In your example, if you read code, you may guess that :- means fallback, but it's not necessarily obvious. And certainly not other magic flags and operators.


As an anecdote, I guess the most complex thing I have done with Bash was scripting a deployment and starting test-runs onto a distributed system (and I think collecting results? I don't remember). Bash was available and copying and starting processes via ssh was simple and robust enough. Notably, the scope and env requirements were very limited.

[–] Kissaki 1 points 4 days ago* (last edited 4 days ago)

😎


Share yours here too?

 

GitHub repo

Examples

> (15 kg/m) * 7cm
# (((15 * kg) / m)) * 7 * cm
out = 1050 * g
> 1 |> cos |> log
# 1 |> cos |> log
out = -0.6156264703860141
> display dev
# Display mode: dev (Developer)
>>> 1.5
# 1.5
out = 1.5
    # IEEE 754 - double - 64-bit
    #
    = 0x_3FF80000_00000000
    = 0x____3____F____F____8____0____0____0____0____0____0____0____0____0____0____0____0
    #    seee eeee eeee ffff ffff ffff ffff ffff ffff ffff ffff ffff ffff ffff ffff ffff
    = 0b_0011_1111_1111_1000_0000_0000_0000_0000_0000_0000_0000_0000_0000_0000_0000_0000
    #   63                48                  32                  16                   0
    #
    # sign    exponent              |-------------------- fraction --------------------|
    =   1 * 2 ^ (1023 - 1023) * 0b1.1000000000000000000000000000000000000000000000000000
[–] Kissaki 2 points 6 days ago

Let's put a story point estimation on that. Then we can extrapolate time range and risk.

3
submitted 2 weeks ago* (last edited 2 weeks ago) by Kissaki to c/nushell
 

I track and version my Nushell environment and configuration in a public repository.

I added a GitHub Actions workflow that tests these files. That will ensure a more defined environment and prerequisites/assumptions, given that they have to be set up in the workflow configuration. Given that I mostly work on Windows, but set up the CI to run on Linux/Ubuntu, it will also ensure platform neutrality.


Since Nushell version 0.101.0, there's no need for a default, base env.nu or config.nu, and the nu binary can be called with only the custom, minimal env and config files.

The nu binary offers --env-config and --config parameters.

I noticed that when using them, errors do not lead to error exist codes; nu will continue execution and report success despite env or config not loading [correctly]. (Bug Ticket #14745)


Do you version your environment configuration? Only locally, or with a hosted repository as a backup or to share? Do you run automated tests on it?

 

One of two Azure CDN providers was Edgio, which filed for bankruptcy.

azureedge.net dotnet CDN URLs will cease to work sometime next year after January 15th.


We expect that most users will not be directly affected, however, it is critical that you validate if you are affected and to watch for downtime or other kinds of breakage.

We maintain multiple Content Delivery Network (CDN) instances for delivering .NET builds. Some end in azureedge.net. These domains are hosted by edg.io, which will soon cease operations due to bankruptcy. We are required to migrate to a new CDN and will be using new domains going forward.

Affected domains:

  • dotnetcli.azureedge.net
  • dotnetbuilds.azureedge.net

Unaffected domains:

  • dotnet.microsoft.com
  • download.visualstudio.microsoft.com
6
submitted 3 weeks ago* (last edited 3 weeks ago) by Kissaki to c/dotnet
 

azureedge.net dotnet CDN URLs will cease to work sometime next year after January 15th.

One of two Azure CDN providers was Edgio, which filed for bankruptcy. CDN migration is in progress.


We expect that most users will not be directly affected, however, it is critical that you validate if you are affected and to watch for downtime or other kinds of breakage.

We maintain multiple Content Delivery Network (CDN) instances for delivering .NET builds. Some end in azureedge.net. These domains are hosted by edg.io, which will soon cease operations due to bankruptcy. We are required to migrate to a new CDN and will be using new domains going forward.

Affected domains:

  • dotnetcli.azureedge.net
  • dotnetbuilds.azureedge.net

Unaffected domains:

  • dotnet.microsoft.com
  • download.visualstudio.microsoft.com

  • Update dotnetcli.azureedge.net to builds.dotnet.microsoft.com
  • Update dotnetcli.blob.core.windows.net to builds.dotnet.microsoft.com

We also noticed that there is a lot of use of our storage account: dotnetcli.blob.core.windows.net. Please also search for it. The storage account is unaffected, however, it would be much better for everyone if you used our new CDN. It will deliver better performance.

9
submitted 3 weeks ago by Kissaki to c/nushell
 
  • Simplified Startup Configuration
  • path self
  • chunk-by
  • term query
  • merge deep
  • WASM support (again)
  • sys net inclueds new columns mac and ip
  • raw string pattern matching
  • dates can now be added to durations
  • additional explore keybinds
  • version in startup banner
  • input --default
  • PowerShell script invocation on Windows
  • new introspection tools

Breaking Changes:

  • ++ operator, stricter command signature parsing (resolves silent parse errors)
  • group-by now supports "groupers" (multiple criteria)
  • timeit
  • sys cpu
  • from csv and from tsv
  • std/iter scan
  • completion sorting in custom completers, import module naming with normalization
  • display_output hook
  • du flag changes
  • Code specific environment variables updated during source
3
nmake cancel (programming.dev)
submitted 1 month ago* (last edited 1 month ago) by Kissaki to c/programming_horror
 

ffaattaall eerrrroorr UU11005588: : tteerrmmiinnaatteedd bbyy uusseerr

 

a new NuGet dependency graph resolver built to dramatically improve performance

The new algorithm they developed uses a more streamlined approach, representing the graph as a flattened set where each node is created only once. This makes the in-memory dependency graph much smaller and easier to work with. Conflicts are resolved as the graph is being built, which avoids the need for the repetitive passes that the old dependency graph resolution algorithm required.

This new approach had dramatic results. The original dependency graph, which in our testing would create 1.6 million nodes for a complex project, was reduced to just 1,200 nodes. With fewer nodes to process, restore times dropped significantly; from 16 minutes down to just 2 minutes.

 

On November 22, 2024, Deno formally filed a petition with the USPTO to cancel Oracle’s trademark for “JavaScript.” This marks a pivotal step toward freeing “JavaScript” from legal entanglements and recognizing it as a shared public good.

Oracle has until January 4, 2025, to respond. If they fail to act, the case will go into default, and the trademark will likely be canceled.

 

On November 22, 2024, Deno formally filed a petition with the USPTO to cancel Oracle’s trademark for “JavaScript.” This marks a pivotal step toward freeing “JavaScript” from legal entanglements and recognizing it as a shared public good.

Oracle has until January 4, 2025, to respond. If they fail to act, the case will go into default, and the trademark will likely be canceled.

 

On November 22, 2024, Deno formally filed a petition with the USPTO to cancel Oracle’s trademark for “JavaScript.” This marks a pivotal step toward freeing “JavaScript” from legal entanglements and recognizing it as a shared public good.

Oracle has until January 4, 2025, to respond. If they fail to act, the case will go into default, and the trademark will likely be canceled.

view more: next ›