blakestacey

joined 2 years ago
MODERATOR OF
[–] blakestacey@awful.systems 8 points 2 weeks ago

Good catch; thanks. I think I had too many awful.system tabs open at once.

[–] blakestacey@awful.systems 13 points 2 weeks ago

Highlight the space just after the abstract of my own most recent arXiv preprint for a surprise. :-)

[–] blakestacey@awful.systems 17 points 2 weeks ago* (last edited 2 weeks ago) (2 children)

Last week, we learned that area transphobe Sabine Hossenfelder is using her arXiv-posting privileges to shill Eric Weinstein's bullshit. I have poked around the places where I'd expect to find technical discussion of a physics preprint, and I've come up with nothing. The Stubsack thread, as superficial as it was, has been the most substantive conversation about her post's actual content.

[–] blakestacey@awful.systems 10 points 2 weeks ago (2 children)

NeurIPS is one of the big conferences for machine learning. Having your work accepted there is purportedly equivalent to getting a paper published in a top-notch journal in physics (a field that holds big conferences but treats journals as more the venues of record). Today I learned that NeurIPS endorses peer reviewers asking questions to chatbots during the review process. On their FAQ page for reviewers, they include the question

I often use LLMs to help me understand concepts and draft my writing. Can I use LLMs during the review process?

And their response is not shut the fuck up, the worms have reached your brain and we will have to operate. You know, the bare minimum that any decent person would ask for.

You can use resources (e.g. publications on Google Scholar, Wikipedia articles, interactions with LLMs and/or human experts without sharing the paper submissions) to enhance your understanding of certain concepts and to check the grammaticality and phrasing of your written review. Please exercise caution in these cases so you do not accidentally leak confidential information in the process.

"Yeah, go ahead, ask 'Grok is this true', but pretty please don't use the exact words from the paper you are reviewing. We are confident that the same people who turn to a machine to paraphrase their own writing will do so by hand first this time."

Please remember that you are responsible for the quality and accuracy of your submitted review regardless of any tools, resources, or other help you used to construct the final review.

"Having positioned yourself at the outlet pipe of the bullshit fountain and opened your mouth, please imbibe responsibly."

Far be it for me to suggest that NeurIPS taking an actually ethical stance about bullshit-fountain technology would call into question the presentations being made there and thus imperil their funding stream. But, I mean, if the shoe fits....

[–] blakestacey@awful.systems 10 points 2 weeks ago (3 children)

A question from ejwillingham:

Google seems to have turned off the -ai in search on iPhone (Safari browser)and overrides it to return an AI-generated result now. Anyone got a fucking workaround on this bc I do not want to see that shit

[–] blakestacey@awful.systems 7 points 2 weeks ago (2 children)

something something Ed Zitron really needs an editldr

[–] blakestacey@awful.systems 15 points 2 weeks ago

That's a ghastly amount of money to burn, particularly when it comes to children's education.

[–] blakestacey@awful.systems 12 points 2 weeks ago (6 children)

Sometimes they file reports against regulars, accusing them of "ableism" for being anti-slop-machine. That's also entertaining.

[–] blakestacey@awful.systems 8 points 2 weeks ago

Disclaimer: abstract above, content and main ideas are human-written; the full text below is written with significant help of AI but is human-verified as well as by other AIs.

"Oh, that pizza sauce recipe that calls for glue? It's totally OK, I checked it out with MechaHitler."

[–] blakestacey@awful.systems 3 points 2 weeks ago (1 children)

The trouble is that the overlap between the circles for "stuff I am excited enough to talk about" and "stuff that anybody would subscribe to a podcast for" is about the size of a rice grain.

[–] blakestacey@awful.systems 3 points 2 weeks ago (3 children)

Nope, not a clue.

[–] blakestacey@awful.systems 4 points 2 weeks ago (5 children)

Dang, maybe I should get in on that. I already own a microphone and everything.

 

Was anyone else getting a 503 error for a little while today?

 

Need to make a primal scream without gathering footnotes first? Have a sneer percolating in your system but not enough time/energy to make a whole post about it? Go forth and be mid!

Any awful.systems sub may be subsneered in this subthread, techtakes or no.

If your sneer seems higher quality than you thought, feel free to cut’n’paste it into its own post — there’s no quota for posting and the bar really isn’t that high.

The post Xitter web has spawned soo many “esoteric” right wing freaks, but there’s no appropriate sneer-space for them. I’m talking redscare-ish, reality challenged “culture critics” who write about everything but understand nothing. I’m talking about reply-guys who make the same 6 tweets about the same 3 subjects. They’re inescapable at this point, yet I don’t see them mocked (as much as they should be)
Like, there was one dude a while back who insisted that women couldn’t be surgeons because they didn’t believe in the moon or in stars? I think each and every one of these guys is uniquely fucked up and if I can’t escape them, I would love to sneer at them.

 

Have a sneer percolating in your system but not enough time/energy to make a whole post about it? Go forth and be mid!

Any awful.systems sub may be subsneered in this subthread, techtakes or no.

If your sneer seems higher quality than you thought, feel free to cut’n’paste it into its own post — there’s no quota for posting and the bar really isn’t that high.

The post Xitter web has spawned soo many “esoteric” right wing freaks, but there’s no appropriate sneer-space for them. I’m talking redscare-ish, reality challenged “culture critics” who write about everything but understand nothing. I’m talking about reply-guys who make the same 6 tweets about the same 3 subjects. They’re inescapable at this point, yet I don’t see them mocked (as much as they should be)
Like, there was one dude a while back who insisted that women couldn’t be surgeons because they didn’t believe in the moon or in stars? I think each and every one of these guys is uniquely fucked up and if I can’t escape them, I would love to sneer at them.

 

Many magazines have closed their submission portals because people thought they could send in AI-written stories.

For years I would tell people who wanted to be writers that the only way to be a writer was to write your own stories because elves would not come in the night and do it for you.

With AI, drunk plagiaristic elves who cannot actually write and would not know an idea or a sentence if it bit their little elvish arses will actually turn up and write something unpublishable for you. This is not a good thing.

 

Tesla's troubled Cybertruck appears to have hit yet another speed bump. Over the weekend, dozens of waiting customers reported that their impending deliveries had been canceled due to "an unexpected delay regarding the preparation of your vehicle."

Tesla has not announced an official stop sale or recall, and as of now, the reason for the suspended deliveries is unknown. But it's possible the electric pickup truck has a problem with its accelerator. [...] Yesterday, a Cybertruck owner on TikTok posted a video showing how the metal cover of his accelerator pedal allegedly worked itself partially loose and became jammed underneath part of the dash. The driver was able to stop the car with the brakes and put it in park. At the beginning of the month, another Cybertruck owner claimed to have crashed into a light pole due to an unintended acceleration problem.

Meanwhile, layoffs!

 

Google Books is indexing low quality, AI-generated books that will turn up in search results, and could possibly impact Google Ngram viewer, an important tool used by researchers to track language use throughout history.

 

[Eupalinos of Megara appears out of a time portal from ancient Ionia] Wow, you guys must be really good at digging tunnels by now, right?

 

In October, New York City announced a plan to harness the power of artificial intelligence to improve the business of government. The announcement included a surprising centerpiece: an AI-powered chatbot that would provide New Yorkers with information on starting and operating a business in the city.

The problem, however, is that the city’s chatbot is telling businesses to break the law.

 

a lesswrong: 47-minute read extolling the ambition and insights of Christopher Langan's "CTMU"

a science blogger back in the day: not so impressed

[I]t’s sort of like saying “I’m going to fix the sink in my bathroom by replacing the leaky washer with the color blue”, or “I’m going to fly to the moon by correctly spelling my left leg.”

Langan, incidentally, is a 9/11 truther, a believer in the "white genocide" conspiracy theory and much more besides.

 

Have a sneer percolating in your system but not enough time/energy to make a whole post about it? Go forth and be mid!

Any awful.systems sub may be subsneered in this subthread, techtakes or no.

If your sneer seems higher quality than you thought, feel free to cut'n'paste it into its own post, there’s no quota here and the bar really isn't that high

The post Xitter web has spawned soo many “esoteric” right wing freaks, but there’s no appropriate sneer-space for them. I’m talking redscare-ish, reality challenged “culture critics” who write about everything but understand nothing. I’m talking about reply-guys who make the same 6 tweets about the same 3 subjects. They’re inescapable at this point, yet I don’t see them mocked (as much as they should be)
Like, there was one dude a while back who insisted that women couldn’t be surgeons because they didn’t believe in the moon or in stars? I think each and every one of these guys is uniquely fucked up and if I can’t escape them, I would love to sneer at them.

 

If you've been around, you may know Elsevier for surveillance publishing. Old hands will recall their running arms fairs. To this storied history we can add "automated bullshit pipeline".

In Surfaces and Interfaces, online 17 February 2024:

Certainly, here is a possible introduction for your topic:Lithium-metal batteries are promising candidates for high-energy-density rechargeable batteries due to their low electrode potentials and high theoretical capacities [1], [2].

In Radiology Case Reports, online 8 March 2024:

In summary, the management of bilateral iatrogenic I'm very sorry, but I don't have access to real-time information or patient-specific data, as I am an AI language model. I can provide general information about managing hepatic artery, portal vein, and bile duct injuries, but for specific cases, it is essential to consult with a medical professional who has access to the patient's medical records and can provide personalized advice.

Edit to add this erratum:

The authors apologize for including the AI language model statement on page 4 of the above-named article, below Table 3, and for failing to include the Declaration of Generative AI and AI-assisted Technologies in Scientific Writing, as required by the journal’s policies and recommended by reviewers during revision.

Edit again to add this article in Urban Climate:

The World Health Organization (WHO) defines HW as “Sustained periods of uncharacteristically high temperatures that increase morbidity and mortality”. Certainly, here are a few examples of evidence supporting the WHO definition of heatwaves as periods of uncharacteristically high temperatures that increase morbidity and mortality

And this one in Energy:

Certainly, here are some potential areas for future research that could be explored.

Can't forget this one in TrAC Trends in Analytical Chemistry:

Certainly, here are some key research gaps in the current field of MNPs research

Or this one in Trends in Food Science & Technology:

Certainly, here are some areas for future research regarding eggplant peel anthocyanins,

And we mustn't ignore this item in Waste Management Bulletin:

When all the information is combined, this report will assist us in making more informed decisions for a more sustainable and brighter future. Certainly, here are some matters of potential concern to consider.

The authors of this article in Journal of Energy Storage seems to have used GlurgeBot as a replacement for basic formatting:

Certainly, here's the text without bullet points:

 

In which a man disappearing up his own asshole somehow fails to be interesting.

view more: ‹ prev next ›