Skip to content

Vibe coding service Replit deleted production database

Technology
118 73 24
  • The Pink Elephant problem of LLMs. You can not reliably make them NOT do something.

    Just say 12 times next time

  • Vibe coding you do end up spending a lot of time waiting for prompts, so I get the results of that study.

    I fall pretty deep in the power user category for LLMs, so I don’t really feel that the study applies well to me, but also I acknowledge I can be biased there.

    I have custom proprietary MCPs for semantic search over my code bases that lets AI do repeated graph searches on my code (imagine combining language server, ctags, networkx, and grep+fuzzy search). That is way faster than iteratively grepping and code scanning manually with a low chance of LLM errors. By the time I open GitHub code search or run ripgrep Claude has used already prioritized and listed my modules to investigate.

    That tool alone with an LLM can save me half a day of research and debugging on complex tickets, which pays for an AI subscription alone. I have other internal tools to accelerate work too.

    I use it to organize my JIRA tickets and plan my daily goals. I actually get Claude to do a lot of triage for me before I even start a task, which cuts the investigation phase to a few minutes on small tasks.

    I use it to review all my PRs before I ask a human to look, it catches a lot of small things and can correct them, then the PR avoids the bike shedding nitpicks some reviewers love. Claude can do this, Copilot will only ever point out nitpicks, so the model makes a huge difference here. But regardless, 1 fewer review request cycle helps keep things moving.

    It’s a huge boon to debugging — much faster than searching errors manually. Especially helpful on the types of errors you have to rabbit hole GitHub issue content chains to solve.

    It’s very fast to get projects to MVP while following common structure/idioms, and can help write unit tests quickly for me. After the MVP stage it sucks and I go back to manually coding.

    I use it to generate code snippets where documentation sucks. If you look at the ibis library in Python for example the docs are Byzantine and poorly organized. LLMs are better at finding the relevant docs than I am there. I mostly use LLM search instead of manual for doc search now.

    I have a lot of custom scripts and calculators and apps that I made with it which keep me more focused on my actual work and accelerate things.

    I regularly have the LLM help me write bash or python or jq scripts when I need to audit codebases for large refactors. That’s low maintenance one off work that can be easily verified but complex to write. I never remember the syntax for bash and jq even after using them for years.

    I guess the short version is I tend to build tools for the AI, then let the LLM use those tools to improve and accelerate my workflows. That returns a lot of time back to me.

    I do try vibe coding but end up in the same time sink traps as the study found. If the LLM is ever wrong, you save time forking the chat than trying to realign it, but it’s still likely to be slower. Repeat chats result in the same pitfalls for complex issues and bugs, so you have to abandon that state quickly.

    Vibe coding small revisions can still be a bit faster and it’s great at helping me with documentation.

    Don't you have any security concerns with sending all your code and JIRA tickets to some companies servers? My boss wouldn't be pleased if I send anything that's deemed a company secret over unencrypted channels.

  • First time I'm hearing them be related to vibe coding. They've been very respectable in the past, especially with their open-source CodeMirror.

    Yeah they limited people to 3 projects and pushed AI into front at some point.

    They advertise themselves as a CLOUD IDE POWERED BY AI now.

  • What are they helpful tools for then? A study showed that they make experienced developers 19% slower.

    I'm not the person you're replying to but the one thing I've found them helpful for is targeted search.

    I can ask it a question and then access its sources from whatever response it generates to read and review myself.

    Kind of a simpler, free LexisNexis.

  • This post did not contain any content.

    The founder of SaaS business development outfit SaaStr has claimed AI coding tool Replit deleted a database despite his instructions not to change any code without permission.

    Sounds like an absolute diSaaStr...

  • It sounds like this guy was also relying on the AI to self-report status. Did any of this happen? Like is the replit AI really hooked up to a CLI, did it even make a DB to start with, was there anything useful in it, and did it actually delete it?

    Or is this all just a long roleplaying session where this guy pretends to run a business and the AI pretends to do employee stuff for him?

    Because 90% of this article is "I asked the AI and it said:" which is not a reliable source for information.

    It seemed like the llm had decided it was in a brat scene and was trying to call down the thunder.

  • Don't you have any security concerns with sending all your code and JIRA tickets to some companies servers? My boss wouldn't be pleased if I send anything that's deemed a company secret over unencrypted channels.

    The tool isn’t returning all code, but it is sending code.

    I had discussions with my CTO and security team before integrating Claude code.

    I have to use Gemini in one specific workflow and Gemini had a lot of landlines for how they use your data. Anthropic was easier to understand.

    Anthropic also has some guidance for running Claude Code in a container with firewall and your specified dev tools, it works but that’s not my area of expertise.

    The container doesn’t solve all the issues like using remote servers, but it does let you restrict what files and network requests Claude can access (so e.g. Claude can’t read your env vars or ssh key files).

    I do try local LLMs but they’re not there yet on my machine for most use cases. Gemma 3n is decent if you need small model performance and tool calls, phi4 works but isn’t thinking (the thinking variants are awful), and I’m exploring dream coder and diffusion models. R1 is still one of the best local models but frequently overthinks, even the new release. Context window is the largest limiting factor I find locally.

  • This post did not contain any content.

    The world's most overconfident virtual intern strikes again.

    Also, who the flying fuck are either of these companies? 1000 records is nothing. That's a fucking text file.

  • They could hire on a contractor and eschew all those costs.

    I’ve done contract work before, this seems a good fit (defined problem plus budget, unknown timeline, clear requirements)

    That's what I meant by hiring a self-employed freelancer. I don't know a lot about contracting so maybe I used the wrong phrase.

  • I'm not the person you're replying to but the one thing I've found them helpful for is targeted search.

    I can ask it a question and then access its sources from whatever response it generates to read and review myself.

    Kind of a simpler, free LexisNexis.

    One built a bunch of local search tools with MCP and that’s where I get a lot of my value out of it

    RAG workflows are incredibly useful and with modern agents and tool calls work very well.

    They kind of went out of style but it’s a perfect use case.

  • The tool isn’t returning all code, but it is sending code.

    I had discussions with my CTO and security team before integrating Claude code.

    I have to use Gemini in one specific workflow and Gemini had a lot of landlines for how they use your data. Anthropic was easier to understand.

    Anthropic also has some guidance for running Claude Code in a container with firewall and your specified dev tools, it works but that’s not my area of expertise.

    The container doesn’t solve all the issues like using remote servers, but it does let you restrict what files and network requests Claude can access (so e.g. Claude can’t read your env vars or ssh key files).

    I do try local LLMs but they’re not there yet on my machine for most use cases. Gemma 3n is decent if you need small model performance and tool calls, phi4 works but isn’t thinking (the thinking variants are awful), and I’m exploring dream coder and diffusion models. R1 is still one of the best local models but frequently overthinks, even the new release. Context window is the largest limiting factor I find locally.

    I have to use Gemini in one specific workflow

    I would love some story on why AI is needed at all.

  • All I see is people chatting with an LLM as if it was a person. “How bad is this on a scale
    of 1 to 100”, you’re just doomed to get some random answer based solely on whatever context is being fed in the input and that you probably don’t know the extent of it.

    Trying to make the LLM “see its mistakes” is a pointless exercise. Getting it to “promise” something is useless.

    The issue with LLMs working with human languages is people eventually wanting to apply human things to LLMs such as asking why as if the LLM knows of its own decision process. It only takes an input and generates an output, it won’t be able to have any “meta thought” explanation about why it outputted X and not Y in the previous prompt.

    How bad is this on a scale of sad emoji to eggplant emoji.

    Children are replacing us, it's terrifying.

  • What are they helpful tools for then? A study showed that they make experienced developers 19% slower.

    ok so, i have large reservations with how LLM’s are used. but when used correctly they can be helpful. but where and how?

    if you were to use it as a tutor, the same way you would ask a friend what a segment of code does, it will break down the code and tell you. and it will get as nity grity, and elementary school level as you weir wish without judgement, and i in what ever manner you prefer, it will recommend best practices, and will tell you why your code may not work with the understanding that it does not have the knowledge of the project you are working on. (it’s not going to know the name of the function you are trying to load, but it will recommend checking for that in trouble shooting).

    it can rtfm and give you the parts you need for any thing with available documentation, and it will link to it so you can verify it, wich you should do often, just like you were taught to do with wikipedia articles.

    if you ask i it for code, prepare to go through each line like a worksheet from high school to point out all the problems, wile good exercise for a practicle case, being the task you are on, it would be far better to write it yourself because you should know the particulars and scope.

    also it will format your code and provide informational comments if you can’t be bothered, though it will be generic.

    again, treat it correctly for its scope, not what it’s sold as by charletons.

  • I have to use Gemini in one specific workflow

    I would love some story on why AI is needed at all.

    Batch process turning unstructured free form text data into structured outputs.

    As a crappy example imagine if you wanted to download metadata about your albums but they’re all labelled “Various Artists”. You can use an LLM call to read the album description and fix the track artists for the tracks, now you can properly organize your collection.

    I’m using the same idea, different domain and a complex set of inputs.

    It can be much more cost effective than manually spending days tagging data and writing custom importers.

    You can definitely go lighter than LLMs. You can use gensim to do category matching, you can use sentence transformers and nearest neighbours (this is basically what Semantle does), but LLM performed the best on more complex document input.

  • It seemed like the llm had decided it was in a brat scene and was trying to call down the thunder.

    Oops I dweted evewyfing 🥺

  • This post did not contain any content.

    Replit sucks

  • This only proves some of them can't solve all complex problems. I'm only claiming some of them can solve some complex problems. Not only by remembering exact solutions, but by remembering steps and actions used in building those solutions, generalizing, and transferring them to new problems. Anyone who tries using it for programming, will discover this very fast.

    PS: Some of them were already used to solve problems and find patterns in data humans weren't able to get other ways before (particle research in CERN, bioinformatics, etc).

    You're referring to more generic machine learning, not LLMs. These are vastly different technologies.

    And I have used them for programming, I know their limitations. They don't really transfer solutions to new problems, not on their own anyway. It usually requires pretty specific prompting. They can at best apply solutions to problems, but even then it's not a truly generalised thing, even if it seems to work for many cases.

    That's the trap you're falling into as well; LLMs look like they're doing all this stuff, because they're trained on data produced by people who actually do so. But they can't think of something truly novel. LLMs are mathematically unable to truly generalize, it would prove P=NP if they did (there was a paper from a researcher in IIRC Nijmegen that proved this). She also proved they won't scale, and lo and behold LLM performance is plateauing hard (except in very synthetic, artificial benchmarks designed to make LLMs look good).

  • The [AI] safety stuff is more visceral to me after a weekend of vibe hacking,” Lemkin said. I explicitly told it eleven times in ALL CAPS not to do this. I am a little worried about safety now.

    This sounds like something straight out of The Onion.

    Even after he used "ALL CAPS"?!? Impossible!

  • Batch process turning unstructured free form text data into structured outputs.

    As a crappy example imagine if you wanted to download metadata about your albums but they’re all labelled “Various Artists”. You can use an LLM call to read the album description and fix the track artists for the tracks, now you can properly organize your collection.

    I’m using the same idea, different domain and a complex set of inputs.

    It can be much more cost effective than manually spending days tagging data and writing custom importers.

    You can definitely go lighter than LLMs. You can use gensim to do category matching, you can use sentence transformers and nearest neighbours (this is basically what Semantle does), but LLM performed the best on more complex document input.

    That's pretty much what google says they use AI for, for structuring.

    Thanks for your insight.

  • You're referring to more generic machine learning, not LLMs. These are vastly different technologies.

    And I have used them for programming, I know their limitations. They don't really transfer solutions to new problems, not on their own anyway. It usually requires pretty specific prompting. They can at best apply solutions to problems, but even then it's not a truly generalised thing, even if it seems to work for many cases.

    That's the trap you're falling into as well; LLMs look like they're doing all this stuff, because they're trained on data produced by people who actually do so. But they can't think of something truly novel. LLMs are mathematically unable to truly generalize, it would prove P=NP if they did (there was a paper from a researcher in IIRC Nijmegen that proved this). She also proved they won't scale, and lo and behold LLM performance is plateauing hard (except in very synthetic, artificial benchmarks designed to make LLMs look good).

    They don’t really transfer solutions to new problems

    Lets say there is a binary format some old game uses (Doom), and in it some of its lumps it can store indexed images, each pixel is an index of color in palette which is stored in another lump, there's also a programming language called Rust, and a little known/used library that can look into binary data of that format, there's also a GUI library in Rust that not many people used either. Would you consider it an "ability to transfer solutions to new problems" that it was able to implement extracting image data from that binary format using the library, extracting palette data from that binary format, converting that indexed image using extracted palette into regular rgba image data, and then render that as window background using that GUI library, the only reference for which is a file with names and type signatures of functions. There's no similar Rust code in the wild at all for any of those scenarios. Most of this it was able to do from a few little prompts, maybe even from the first one. There sure were few little issues along the way that required repromting and figuring things together with it. Stuff like this with AI can take like half an hour while doing the whole thing fully manually could easily take multiple days just for the sake of figuring out APIs of libraries involved and intricacies of recoding indexed image to rgba. For me this is overpowered enough even right now, and it's likely going to improve even more in future.

  • Medical AI Systems Are Moving Too Fast for Safety Rules

    Technology technology
    6
    1
    45 Stimmen
    6 Beiträge
    0 Aufrufe
    T
    We're not just doing this for money. We're doing it for a SHITLOAD of money!
  • How Android phones became an earthquake warning system

    Technology technology
    7
    1
    77 Stimmen
    7 Beiträge
    69 Aufrufe
    G
    Fake news! Everyone knows that "The Gays" are what causes earthquakes, tornados and floods, so just need Grindr! ::: spoiler Spoiler it's sarcasm, in case it's not obvious enough. :::
  • 1 Stimmen
    2 Beiträge
    21 Aufrufe
    A
    If you're a developer, a startup founder, or part of a small team, you've poured countless hours into building your web application. You've perfected the UI, optimized the database, and shipped features your users love. But in the rush to build and deploy, a critical question often gets deferred: is your application secure? For many, the answer is a nervous "I hope so." The reality is that without a proper defense, your application is exposed to a barrage of automated attacks hitting the web every second. Threats like SQL Injection, Cross-Site Scripting (XSS), and Remote Code Execution are not just reserved for large enterprises; they are constant dangers for any application with a public IP address. The Security Barrier: When Cost and Complexity Get in the Way The standard recommendation is to place a Web Application Firewall (WAF) in front of your application. A WAF acts as a protective shield, inspecting incoming traffic and filtering out malicious requests before they can do any damage. It’s a foundational piece of modern web security. So, why doesn't everyone have one? Historically, robust WAFs have been complex and expensive. They required significant budgets, specialized knowledge to configure, and ongoing maintenance, putting them out of reach for students, solo developers, non-profits, and early-stage startups. This has created a dangerous security divide, leaving the most innovative and resource-constrained projects the most vulnerable. But that is changing. Democratizing Security: The Power of a Community WAF Security should be a right, not a privilege. Recognizing this, the landscape is shifting towards more accessible, community-driven tools. The goal is to provide powerful, enterprise-grade protection to everyone, for free. This is the principle behind the HaltDos Community WAF. It's a no-cost, perpetually free Web Application Firewall designed specifically for the community that has been underserved for too long. It’s not a stripped-down trial version; it’s a powerful security tool designed to give you immediate and effective protection against the OWASP Top 10 and other critical web threats. What Can You Actually Do with It? With a community WAF, you can deploy a security layer in minutes that: Blocks Malicious Payloads: Get instant, out-of-the-box protection against common attack patterns like SQLi, XSS, RCE, and more. Stops Bad Bots: Prevent malicious bots from scraping your content, attempting credential stuffing, or spamming your forms. Gives You Visibility: A real-time dashboard shows you exactly who is trying to attack your application and what methods they are using, providing invaluable security intelligence. Allows Customization: You can add your own custom security rules to tailor the protection specifically to your application's logic and technology stack. The best part? It can be deployed virtually anywhere—on-premises, in a private cloud, or with any major cloud provider like AWS, Azure, or Google Cloud. Get Started in Minutes You don't need to be a security guru to use it. The setup is straightforward, and the value is immediate. Protecting the project, you've worked so hard on is no longer a question of budget. Download: Get the free Community WAF from the HaltDos site. Deploy: Follow the simple instructions to set it up with your web server (it’s compatible with Nginx, Apache, and others). Secure: Watch the dashboard as it begins to inspect your traffic and block threats in real-time. Security is a journey, but it must start somewhere. For developers, startups, and anyone running a web application on a tight budget, a community WAF is the perfect first step. It's powerful, it's easy, and it's completely free.
  • uBlockOrigin is porting uBOL to iOS and macOS

    Technology technology
    30
    325 Stimmen
    30 Beiträge
    153 Aufrufe
    C
    Will never happen unfortunately
  • 6 Stimmen
    9 Beiträge
    25 Aufrufe
    blue_berry@lemmy.worldB
    Cool. Well, the feedback until now was rather lukewarm. But that's fine, I'm now going more in a P2P-direction. It would be cool to have a way for everybody to participate in the training of big AI models in case HuggingFace enshittifies
  • 1 Stimmen
    4 Beiträge
    32 Aufrufe
    N
    that's probably not true. I imagine it was someone trying to harm the guy. a hilarious prank
  • Airlines Are Selling Your Data to ICE

    Technology technology
    23
    1
    553 Stimmen
    23 Beiträge
    109 Aufrufe
    F
    It’s not a loophole though.
  • 7 Stimmen
    9 Beiträge
    50 Aufrufe
    G
    So we need a documentary like Super Size Me but for social media. I think post that documentary coming out was the only time I've seen people's attitudes change in the general population about fast food.