PaintedSnail

joined 1 year ago
[–] [email protected] 4 points 3 weeks ago (1 children)

We used to have them until several years ago they were deliberately removed. Their roots were destroying the roads and sidewalks, as well as infiltrating the underground infrastructure.

[–] [email protected] 1 points 3 weeks ago

How can I make sure that the citations are real and actually useful? Citations-cartels are already a thing.

I'm thinking that citations in papers can be actual links (akin to hyperlinks) to the location in the cited paper itself. This way it can be automatically verified that there are no citation loops, that citations reference current revisions, that the papers cited have not been retracted or otherwise discredited, and following citation trails becomes much easier. Would that help the citation-carcel issue, you think?

How can the review process be ported to that approach without losing the independence of the reviews? They are supposed to be anonymous and not affiliated with the authors in any way?

How important is anonymity in reviews? My thought process is going the opposite way: by linking reviews and comments on papers to the person/institution making it, it encourages them to be more responsible with their words and may indicate potential biases with regards to institution affiliations.

How can the amount of articles be reduced? Currently, you’re forced to publish as much as possible, published articles in “good” journals are your currency as a reseacher.

Here I'm also thinking the exact opposite: the issue isn't the numbers of papers, it's how the papers are organized that's the problem. We actually want MORE papers for the reasons hinted at here: important papers are going unpublished because they are (for lack of a better word) uninteresting. A null result is not an invalid result, and its important to get that data out there. By having journals gate-keep the data that gets released, we are doing the scientific community a disservice.

Of course, more papers increases the number of junk papers published, but that's where having the papers available openly and having citations linked electronically comes in. The data can be fed in to large data mining algorithms for meta analysis, indexing and searching, and categorization. Plus, if it later turns out that a paper is junk, any papers that cite it (and any papers that cite those, and so on) can all be flagged for review or just automatically retracted.

Thoughts?

[–] [email protected] 6 points 4 weeks ago (2 children)

I know little of the ins-and-outs of scientific publishing, but that didn't stop me from having a dumb thought: could the fediverse be a potential solution? Each university or research group could host their own instance of some software specifically for publishing papers, papers can cross-link citations to papers on other instances, people can make comments across instances that are tied to their own identities from their home instance, paper revisions can be tracked easily and bad citations spotted when a paper is updated or retracted, that kind of thing. The currency then becomes the reputation of the organizations and individuals, and this opens up a ton of data for automated analysis. I just don't know enough to know what problems would arise.

[–] [email protected] 2 points 1 month ago* (last edited 1 month ago) (1 children)

Yeah, the side quests are rather unimaginative in their tasks. For the most part they're not worth doing unless you want a bit more of a dive into the world lore. (A few give unique rewards, though.) Even some story quests are "player does menial chores for good karma with the locals because the devs need to pad things out a bit."

Shadowbringers is worth it, though. It's my personal favorite story arc. Endwalker, the one after, is my second favorite.

[–] [email protected] 2 points 1 month ago (3 children)

Come to FFXIV! We have decent stories, a voluntary PvP arena with multiple modes, and cat girls!

[–] [email protected] 2 points 1 month ago

The metrics are the only important part! How else are we supposed to know how good the line is unless we constantly stress test the line by collecting data? Your ability to use the line is not a useful metric, so we don't worry about that.

[–] [email protected] 9 points 1 month ago (2 children)

We know when you lie. We can see uptime stats.

[–] [email protected] 2 points 1 month ago* (last edited 1 month ago)

Basically, yes, though I think they have special hydraulic pullers, too. I forget the exact name. They have to take special measures if the day is too cold.

https://youtu.be/zqmOSMAtadc?si=FCG7HxiPWXNQY6Uj

[–] [email protected] 28 points 1 month ago (2 children)

According to Practical Engineering, tracks are no longer given a gap. The gap causes premature wear and excess noise. Instead, they lay the track under tension, and weld the joins between sections.

There is still a limit on how much heat they can handle before buckling, of course. I just thought that was a neat innovation.

[–] [email protected] 4 points 2 months ago

And it took a lot of hard work by a lot of people to adopt new date standards to avoid that problem. Now it's time to adopt new IP standards, and it's going to take a lot of hard work by a lot of people.

[–] [email protected] 7 points 2 months ago

I'm not saying planned obsolescence isn't a thing (because it is), but that's not the only reason. Making phones smaller, lighter, faster, and more feature-dense all mean that the phone has to be built with tighter manufacturing and operating tolerances. Faster chips are more prone to heat and vibration damage. Higher power requirements means the battery has a larger charge/discharge cycle. And unfortunately, tighter operating tolerances mean that they can fall out of those tolerances much more easily.

They get dropped, shaken, exposed to large environmental temperature swings, charged in wonky ways, exposed to hand oils and other kinds of dirt, and a slew of other evils. Older phones that didn't have such tight tolerances could handle all that better. Old Nokia phones weren't built to be indestructible, they are just such simple phones that there isn't much to break; but there's a reason people don't use them much anymore. You can still get simple feature phones, but the fact remains that they don't sell well, so not many are made, and the ones that are made don't have a lot of time and money invested in them.

Now Voyager is an extremely simple computer, made with technology that has huge tolerances, in an environment that is mostly consistent and known ahead of time so the design can deliberately account for it, had lots of testing, didn't have to take mass production into its design consideration, didn't have to make cost trade-offs, and has a dedicated engineering team to keep it going. It is still impressive that it has lasted this long, but that is more a testament to the incredible work that was and is being put into it than to the technology behind it.

[–] [email protected] 4 points 2 months ago* (last edited 2 months ago) (1 children)

If I interpret your question correctly, you are basically asking what the practical difference is between interpreting a model as a reflection of reality and interpreting a model as merely a mathematical tool.

A mathematical model, at its core, is used to allow us to make testable predictions about our observations. Interpretations of that model into some kind of explanation about the fundamental nature of reality is more the realm of philosophy. That philosophy can loop back into producing more mathematical models, but the models themselves only describe behavior, not nature.

A model by nature is an analogy, and analogies are always reductionist. Like any analogy, if you poke it hard enough, it starts to fall apart. They make assumptions, they do their best to plug holes, they try to come as close as they can to mirroring the behavior of our observations, but they always fall short somewhere. Relativity and Quantum Chromodynamics are both good examples. Both are very, very good at describing behavior within certain boundaries, but fall completely apart when you step outside of them. (Both, to expand on the example, use constants that are impericaly determined, but we have no idea where they come from.)

The danger is in when you start to assume that a model of reality is reality itself, and you forget that it's just a best guess of behaviors. Then you get statements like you first made. "Relativity assumes time is a dimension. The model for that works. Therefore time must be a dimension in reality. That must mean that not treating time as a dimension anywhere must be wrong." That line of thinking, though, forgets that a model is only correct within the scope of the model itself. As soon as you introduce a new model, any assumptions made by other models are no longer relevant. That will pigeonhole your thinking and lead you to incorrect conclusions due to mixed analogies.

That is how you get statements like your first one. "Model A treats time like an illusion, but model B treats time like a dimension. Ergo, all dimensions are illusions ." That is mixing analogies.

view more: next ›