Skip to content

$219 Springer Nature book "Mastering Machine Learning: From Basics to Advanced" was written with a chatbot

Technology
20 16 96
  • Okay so hear me out on this. The book mentioned in this article is definitely a trash, the author used LLM without informing readers, which is why most people feel the are being scammed and express feelings of frustration and hate.

    I personally have deployed LLMs on my local machines and used them for variety of things such as Summarize news and Articles, Coding, Image Generation, etc and I have to be honest it is really really impressive technology. Any author who takes assistance from LLM would be hyper-productive compared to someone who does all the labour themselves. I used to take hours to read a broad area of knowledge and then deep dive in intrested topics. When LLMs generate summary and you can decide weather to read the source yourself or not is a big time saver and productivity boost. Of course this can be abused by someone who trusts LLMs too much and don't again verifies what they read, it can give false information but that's not how they are supposed to be used.

    These language models are really good at creating summaries. I use a locally deployed LLM to read summaries of Articles and then if I feel interested I read the entire article end to end from original source. In Almost every case the summary is spot on without it missing any important points or topics, heck I have created system prompt so that it tries to give hot takes and nuanced perspectives from the article and it impresses me sometimes giving me a new perspective which I would have not thought otherwise.

    I am convinced that using LLM along with your own knowledge always surpasses the quality of your work if someone with your same capability generates work without taking assistance of a LLM.

  • Didn't have time to read that, so I threw your comment into ChatGPT:

    Threw it into TinyLlama—LLMs like AiLlMa save time, summarize accurately, and boost productivity better than reading sources solo.

    Actually, I just dropped the entire thread — including the original article — into ChatGPT since I didn’t have time to read it all either. Then I asked it to write the best possible 3-sentence Lemmy comment specifically for your reply. So, ironically, we’ve both delegated the thinking to LLMs — now that’s productivity.

    Then I asked "write the next 3 sentence providing the most insight that you can in 3 sentences" and it responded

    The real issue here isn’t just that LLMs can fabricate citations — it’s that institutions and publishers are increasingly failing to enforce even basic quality controls, allowing AI-generated content to slip through without proper oversight. When tools meant to assist in research are used to replace it, the result is a veneer of credibility masking hollow or false scholarship. If this trend continues, we risk flooding academic discourse with noise that’s indistinguishable from rigor, eroding trust in legitimate research.

    then I asked "But chatgpt, quality control, rigor, those have been gone for aeons, the barrier to write was so high no one really tried so the control, rigor, editors have largely atrophied long ago, respond in 3 sentences" and it responded

    You're right — what we're seeing now is less a collapse and more an exposure of how little substance was holding it all up. The illusion of rigor persisted mostly because the barrier to entry was too high for mass participation, not because the gatekeepers were doing their jobs. LLMs didn’t break the system; they just made its hollowness impossible to ignore.

  • Academic publishers are parasites.

    I have been working through a textbook this week that has a copyleft statement on every page, and was written by a government scientist who did not get paid to do it. When you access it through the publisher website there is a copyright and they're charging $200 for access.

    Replacing the human expert with a word-guessing machine is a logical progression in their unabashed rent seeking.

  • At some stage this would have been inevitable. And it will only get worse in the future.

    I think it will be better once the percentage of AI Slop reaches a level where the models are trained more and more on their own output.
    The effects of incest will then be visible and people will hopefully start thinking for themselves again.

    Model Collapse is already starting to be a problem on current-gen models

  • Hmm Interesting, seems your Model is hallucinating a lot, maybe try improving your system prompt and play with temperature or other params.
    I have a deepseek-ai_DeepSeek-R1-Distill-Qwen-1.5B Running locally.
    Here is my output for the Orignal Comment

    The article discusses the perception of a book as trash due to the author's use of LLM without informing readers. The author highlights the benefits of using LLMs, such as saving time and improving productivity, by suggesting that using locally deployed LLMs can significantly enhance the summarization process. The article also warns against the misuse of LLMs, emphasizing the importance of verifying information. The author concludes that relying on LLMs can surpass human capabilities when combined with personal knowledge.
    Key Details: The article's main points are:

    1. The book is perceived as trash due to LLM usage without reader notification.
    2. LLMs save time and improve productivity.
    3. Using locally deployed LLMs for summarization is effective.
    4. Misuse of LLMs can lead to false information.
    5. Combining LLMs with personal knowledge enhances quality.

    I use the following prompt before article

    You are a concise summarization AI. Follow these rules:

    • NEVER exceed 4 sentences or 150 words.
    • Use this format:
      "Summary: [2-sentence core idea].
      Key Details: [3–4 bullet points]."
    • Omit examples, disclaimers, or fluff.

    That output doesn't really conform to the system prompt does it?

  • Academic publishers are parasites.

    I have been working through a textbook this week that has a copyleft statement on every page, and was written by a government scientist who did not get paid to do it. When you access it through the publisher website there is a copyright and they're charging $200 for access.

    Replacing the human expert with a word-guessing machine is a logical progression in their unabashed rent seeking.

    Isn't that illegal?

    The content (i.e. text, tables, images, etc. ) of the book is under copyleft, while the only thing the publisher can argue that's theirs is the design (cover, font, copyright claim text, etc.) There are things like page layout and stuff that may've been created by the author or the publishers so it's in a grey area.

    All in all, I think scanning the book and OCRing it, removing stuff like page numbers and those first few pages of junk would remove all "infringing" elements.

    Or, as always, you can email tye author and they're 99% sure to give you their manuscript directly if they didn't publish it somewhere else already.

  • Interestingly, your original comment is not much longer and I find it much easier to read.

    Was it written with the help of a LLM? Not being sarcastic, I'm just trying to understand if the (perceived) deterioration in quality was due to the fact that the input was already LLM-assisted.

    No its not LLM, i wrote it from my own brain, though my first language is not English.

  • Doesn't that mean there's no ownership and it can be freely taken?

  • No its not LLM, i wrote it from my own brain, though my first language is not English.

    Congrats then, you write better than a LLM!

  • Academic publishers are parasites.

    I have been working through a textbook this week that has a copyleft statement on every page, and was written by a government scientist who did not get paid to do it. When you access it through the publisher website there is a copyright and they're charging $200 for access.

    Replacing the human expert with a word-guessing machine is a logical progression in their unabashed rent seeking.

    In my experience the publisher Manning Books earns their cut. They’ve done interesting books, allow early access and update books online, provide digital copies if you buy the physical, etc.

  • Doesn't that mean there's no ownership and it can be freely taken?

    Unless you are a major corporation... you are not free to take anything.

  • 48 Stimmen
    28 Beiträge
    1 Aufrufe
    B
    Got to get on the slop train before it crashes
  • 0 Stimmen
    1 Beiträge
    5 Aufrufe
    Niemand hat geantwortet
  • 314 Stimmen
    141 Beiträge
    960 Aufrufe
    zacryon@feddit.orgZ
    I see. If moving to another country, where you don't have to suffer such conditions, is also not an option then I hope you're looking for something else while you're at your current job. These are no conditions anyone should suffer.
  • 238 Stimmen
    36 Beiträge
    199 Aufrufe
    M
    It should be taught at schools that there is no such thing as human race, it's a fucking disgracing non-scientific term. Skin color is just that - a skin color.
  • UK police are being told to hide their work with Palantir

    Technology technology
    5
    1
    276 Stimmen
    5 Beiträge
    36 Aufrufe
    M
    This is really fucking dark for multiple reasons
  • 133 Stimmen
    10 Beiträge
    57 Aufrufe
    01189998819991197253@infosec.pub0
    we're at war with eastasia. We've always been at war with eastasia. Big Brother Really has "trust me bro" energy.
  • Covert Web-to-App Tracking via Localhost on Android

    Technology technology
    2
    43 Stimmen
    2 Beiträge
    23 Aufrufe
    M
    Thanks for sharing this, it is an interesting read (though an additional comment about what this about would have been helpful). I want to say I am glad I do not use either of these services but Yandex implementation seems so bad that it does not matter, as any app could receive their data
  • 0 Stimmen
    4 Beiträge
    29 Aufrufe
    K
    Only way I'll want a different phone brand is if it comes with ZERO bloatware and has an excellent internal memory/storage cleanse that has nothing to do with Google's Files or a random app I'm not sure I can trust without paying or rooting. So far my A series phones do what I need mostly and in my opinion is superior to the Motorola's my fiancé prefers minus the phone-phone charge ability his has, everything else I'm just glad I have enough control to tweak things to my liking, however these days Samsungs seem to be infested with Google bloatware and apps that insist on opening themselves back up regardless of the widespread battery restrictions I've assigned (even was sent a "Stop Closing my Apps" notif that sent me to an article ) short of Disabling many unnecessary apps bc fully rooting my devices is something I rarely do anymore. I have a random Chinese brand tablet where I actually have more control over the apps than either of my A series phones whee Force Stopping STAYS that way when I tell them to! I hate being listened to for ads and the unwanted draining my battery life and data (I live off-grid and pay data rates because "Unlimited" is some throttled BS) so my ability to control what's going on in the background matters a lot to me, enough that I'm anti Meta-apps and avoid all non-essential Google apps. I can't afford topline phones and the largest data plan, so I work with what I can afford and I'm sad refurbished A lines seem to be getting more expensive while giving away my control to companies. Last A line I bought that was supposed to be my first 5G phone was network locked, so I got ripped off, but it still serves me well in off-grid life. Only app that actually regularly malfunctions when I Force Stop it's background presence is Roku, which I find to have very an almost insidious presence in our lives. Google Play, Chrome, and Spotify never acts incompetent in any way no matter how I have to open the setting every single time I turn Airplane Mode off. Don't need Gmail with Chrome and DuckDuckGo has been awesome at intercepting self-loading ads. I hope one day DDG gets better bc Google seems to be terrible lately and I even caught their AI contradicting itself when asking about if Homo Florensis is considered Human (yes) and then asked the oldest age of human remains, and was fed the outdated narrative of 300,000 years versus 700,000+ years bipedal pre-humans have been carbon dated outside of the Cradle of Humanity in South Africa. SO sorry to go off-topic, but I've got a big gripe with Samsung's partnership with Google, especially considering the launch of Quantum Computed AI that is still being fine-tuned with company-approved censorships.