We invested in automation… so why does it still feel like manual work?
Posted by Such_Rhubarb8095@reddit | sysadmin | View on Reddit | 89 comments
We're running IT for around 800 users and over the last 12–14 months we made a big push toward automation, we built onboarding workflows (account creation, permissions, device setup), set up patching schedules across departments, and added alerting rules for most critical systems.
On paper, everything is “automated.” In reality, it still feels like we're doing everything manually, just with extra steps.
Examples: Onboarding workflows fail halfway if one field is off so someone has to step in and finish manually, patch jobs complete but leave a percentage of devices in a weird state, manual cleanup again and alerts trigger but don't connect to any action tech has to interpret, investigate, then create a ticket.
So now instead of just doing tasks, we're constantly checking if automation worked… and fixing it when it didn't. My team literally has a morning routine where they go through “what broke overnight.”
It's frustrating because we invested time to reduce workload, but it feels like we just shifted the work into monitoring and maintenance.
OneSeaworthiness7768@reddit
This is a research post from a vibecoder. Don’t engage.
Note: it’s funny I’ve noticed a pattern they all seem to be flailing themselves as “jr sysadmin” lately.
ImportantMud9749@reddit
I thought so too when I read your post. However, after going through the user's history (just search their username, can't see on their account page), I really don't know. It could be related to whatever system they were forced into rather than them trying to vibe code their own.
They seem to be wary enough about AI not to try vibe coding a whole SaaS system, but then again who knows.
At least it doesn't appear to be a bot account?
OneSeaworthiness7768@reddit
Agree to disagree. I don’t think they’re a bot, I think they’re likely a human fishing for problems to market their apps to. If you look through the comments on their posts, Monday and Atera are mentioned multiple times in different subreddits by different users which makes me lean towards fake organic marketing/seo. When the same user makes multiple posts complaining of problems in a similar space with a slightly different story each time, they hardly ever reply to anyone and don’t offer any actual details, and there’s a pattern of the same one or two products recommended in the comments that are not anything well-known that you usually see come up in real conversation, that’s usually what’s happening. It’s either someone pushing their own app or a marketing company doing it as a service. This is a very common tactic being used everywhere across Reddit right now.
ImportantMud9749@reddit
Great insight, seeing as they've changed only their post history to private now is another sign of non-organic content.
No agree to disagree, you've convinced me.
Automatic_Beat_1446@reddit
how are they hiding those comments from their account comments page?
OneSeaworthiness7768@reddit
Reddit lets users hide their post history on their profile, but the posts are still visible elsewhere. You can search their history by going to old.reddit and search author:username. I hesitate to even draw attention to that, hopefully Reddit doesn’t “fix” that too since they removed the ability to do it in the new interface.
Automatic_Beat_1446@reddit
Ah, I knew that users could hide your history, but your example seems to indicate you can pick and choose which comments to hide?
the profile page i linked only has 2 comments, so i was a little confused.
OneSeaworthiness7768@reddit
Yes Reddit lets you choose which subreddits are visible on your profile.
Automatic_Beat_1446@reddit
that is so ripe for abuse, i almost dont know if i want to continue using this website.
thanks for the explanations.
immortalsteve@reddit
I was like "who the fuck automates a system without error handling this cannot possibly be real" and yet here we are. Good to know it's just guerilla marketing.
godspeedfx@reddit
I knew it immediately after reading the 2nd sentence. I'm so close to leaving this sub - all these bullshit research / product placement / AI posts disguised as professionals in duress are ruining this place.
OneSeaworthiness7768@reddit
Yeah it’s really bad lately. The last two months or so especially. I’m hoping that by calling them out it helps mods clean them up quicker and helps other users learn how to identify them quicker. It’s always super obvious to me, but I guess most people just don’t pay attention to the signs or see the common patterns in these posts.
I_cut_the_brakes@reddit
I see less of these on this sub now, or maybe I spend less time here, but I swear for a while every other post was just an ad.
OneSeaworthiness7768@reddit
I feel like it’s the opposite. I’m seeing more and more. Though credit to the mods, they remove most of them quickly. Some of the more subtle ones still fall through the cracks, partly due to their post histories being hidden so it’s harder to identify their repeat behavior. That’s why I use workarounds to view their hidden post histories because I know most people won’t bother to check. Calling them out might help mods remove/ban them.
Ma7h1@reddit
This is something I’ve seen quite often — automation doesn’t remove work, it shifts it.
Instead of doing tasks manually, you end up maintaining and validating the automation. And if that layer isn’t reliable or well-observed, it quickly turns into exactly what you described: “what broke overnight?”
From my experience, the missing piece is often visibility into the automation itself.
It’s not enough to automate onboarding, patching, etc. — you also need to monitor:
• whether workflows completed successfully end-to-end
• where they failed (and why)
• and whether the outcome is actually in the expected state
Otherwise, you’re stuck manually validating automated processes.
What helped us a lot here was using Checkmk to actually monitor the automation layer itself. Instead of just knowing that a job ran, we track if it really succeeded, if systems reached the desired state, and trigger alerts only when something is actionable.
Especially with things like mk-job (from checkmk), you can directly monitor automation jobs and workflows, which makes it much easier to see if something failed and where. Combined with proper checks and thresholds, this gives you a much clearer picture than just “job executed”.
For example:
That way, you move from “checking everything in the morning” to only reacting when something actually needs attention.
I’m also doing similar things in my homelab, and it’s a good reality check — if automation isn’t observable and reliable, it just creates hidden manual work.
So I’d say:
👉 automation without observability = hidden manual effort
Tools like Checkmk (and extensions like mk-job) really help close that gap by making automation outcomes visible and actionable — not just “executed.” 👍
OneSeaworthiness7768@reddit
Be gone, SEO marketer.
TheDevauto@reddit
If it feels that way the automation was not done well. You will always have maintenance, but breaking flows more than 2-5% of the time at most likely mean the flow that was automated was the on paper perfect flow. The exception paths for workflows should be captured and if they occur often, automated as well.
It is an iterative process and isnt magic.
Process mining tools can help you capture what the actual processes are and you can then fix or adjust things before automating that works as well.
Pls_submit_a_ticket@reddit
You have done automation, it just isn’t perfect yet. A field being off in your onboarding/off boarding workflow happens, especially when you first start. But you can make it so it doesn’t happen again.
You ideally create roles with HR, have templates for roles. Or the really lazy way that I don’t recommend because it’s bad security practice, is to just copy the permissions of another person from that role.
There should also be an approval flow, ours goes to HR, they can either approve it, or reject with comments. Then IT gets an approval. If adjustments need to be made, there’s a link they click and edit fields. Then IT approves it and the accounts get created.
A field breaking the workflow should happen once or twice before making it idiot proof. Eventually, you’ll have onboarding automation that you never have to touch. Then you apply what you learned to the next task you want to automate your way out of.
MumfordAdvisory@reddit
This happens more than people admit, automation reduces visible work but increases hidden maintenance. Teams end up spending just as much time, it’s just harder to see where it’s going.
jwalker55@reddit
I built my onboarding automation in n8n and focused heavily on input validation so that the chances of success are much greater. Also, there are times when it still might not work for other reasons, so building proper error handling into the workflow and adding a human-in-the-loop where needed makes it easier to remediate, rather than just blasting a "no worky" alert and forcing the user into manual work. The workflow can just retry the action with the corrected input, or maybe there was just an API outage and we need to retry the existing input.
doblephaeton@reddit
Look into error handling of your automations, if a field is left off, do you build in a default setting,
look into logging of automation steps to see where things are breaking,
Look at notifications of the errors to teams or other tools to be alerted to when things break.
Centimane@reddit
Highly recommend don't treat a blank field as default, instead have a default option that is labeled
<default>. Not only is it more clear, but will save you from silly things like a single space instead of blank, or a tab character, or a.that the user didn't notice or some other cursed input.sambodia85@reddit
Programming is classic 80/20 problems.
You can get 80% of the outcome for 20% of the effort, but the final 20% will take 80% effort.
Input Validation, Tests, error handling are what separates the pretenders from the pros.
I’m a pretender.
ScriptMonkey78@reddit
So much this!
"Please enter a number between 1 and 10" User spells the number five and script errors out "Why is it broken?"
ProblyAThrowawayAcct@reddit
Threeve.
braytag@reddit
Yep this. You automated wrong. It's like programming, you should have safety, fallbacks and logs.
Infinite-Stress2508@reddit
Yep, seems to me the automations aren't implemented correctly.
jrsherrod@reddit
I hate to say it, but while there is routine upkeep in a patch management workflow and maintaining agent stack compatibility across security controls and balancing that against liberty to get work done, if your imaging is failing halfway through your desktop engineers who maintain the backend are failing to put the right drivers in place and write clean scripts. Imaging should only be failing on things like manufacturers surprise switching SSDs so you have to slip in a new RST driver to the server, but even then that happens on a delay from deployment and shouldn't put hitches into anything.
I hate to say it, but while it is true that environments require constant maintenance, it sounds like the dysfunction where you are is more in "git gud" territory.
stufforstuff@reddit
Sound like your "automation" was done half assed at best. Fix it.
Junior_Investment514@reddit
i actually invested heavy in automation just to look over brittle scripts lol. but i still think for some automation agencies and businesses with very boring tasks like getting invoices or quotes from diffrent websites things like skyvern could work and can actually deliver that saving time promise
TerrificVixen5693@reddit
Need to fix the automation with better error handling so that one field doesn’t trigger a failure.
And welcome to IT. They can never fire us, because despite all the MDMs agents and automation, 6% of the systems are failing to automatically change to the new DNS servers via the automated package, please walk server to server manually changing the ones that won’t take the package.
discosoc@reddit
What you describe is failed attempts at automation, not actual automation. Do it correctly this time.
ExceptionEX@reddit
Well it sounds like you failed to build out the right automation, you need to build in validation, error handling, and reporting notification.
Right now it sounds like your system is so fragile, it fails often, requires manual intervention, and because confidence is so low you are hovering over the process.
Until you resolve those issues you aren't automated.
phoenix823@reddit
Well further investment in automation would figure out the root cause of what's breaking patching and fixing that so it stops breaking. Not allowing one field to be off to break onboarding. But responding to broken automation rather than doing all the work by hand and THEN responding to breakage is a big step in the right direction. You just need to keep on pushing.
doubleUsee@reddit
automating a process is easy. Automating a process so that it runs independently and without intervention 98% of the time is a skill and takes a lot more time and experience. Automating it so that it runs without intervention 100% of the time is a dream that's almost never achievable.
Onboarding can work quite smoothly automatically, but if it's just take fields a, b and c and plonk them into AD, done, it'll fail a lot. The automation needs to account for as many possible scenarios as possible. Every time you're called in to manually fix something, you need to evaluate if it's ever gonna happen again, if it will - then automate your response accordingly.
In addition it pays boons to work on fixing the input. If people put garbage into your automation, it's gonna be hard to get it to work. validation on the input side, training, and awareness - and sometimes some malicious motivation ('if you don't put in the data right, it'll take a week before it's corrected' isn't a nice or good thing to say, but it does motivate people to be more careful).
In some of the scripts I maintain that face user inputs, three quarters of it is various sorts of error handling so that they almost never need my input. That's part of automation.
At the same time, the reality of advanced automation is that your job will become babysitting the automation. Even perfect automation will need updating, adjusting and editing over time. I personally much prefer spending an hour fixing automation on onboarding than spending an hour creating 24 accounts in exactly the same way, during which the risk of human error is much bigger, and then it'll be my fault.
Fallingdamage@reddit
I like to have most of my important automations email me a report containing job results when complete. Sometimes they contain a table of values, sometimes its a single sentance. I just want to know it was done or not done. Still takes a little of my time in the morning, but glancing at a quick email or subject line containing a positive outcome is still faster than doing the job myself by a lot.
WizardsOfXanthus@reddit
This all the way. I work in a hospital and manage our onboarding jobs, along with 600 other SSIS jobs. We onboard anywhere between 150-200 new employees every two weeks. Over trial and error, we have so many IF ELSE statements set up to handle things we’ve both expected and then never expected, so they became enhancements. It’s dialed in pretty well after two years now, but man it’s scary waking up and seeing it fail and having to take care of it. We also have email alerts set up upon failure, a web GUI dashboard to show all jobs running and which ones failed, bringing the error message to the top of each log for a quick scan. Opening it up will show the entirety of the log. It’s extremely helpful and would be such a pain if we don’t have something like this.
IOUAPIZZA@reddit
This is an underrated and maybe not brought up enough in discussion for a solution(s). What is your scale/volume? Op said they have about 800 users, and I have that much in my environment too. We are not "automated", but our scale is lower. Even if the guys on and off board 5 staff in a week, would automating the whole thing and tying it into the HR system work for us? Maybe. In my situation:
For my two techs, I do value the experience of working through the processes and procedures of things like account creation, I want them to be familiar with this, but I still wanted to cut the tedium a bit. I need them to be competent backups. So I did several things.
A Visio diagram with the sync/access flow, of user creation to having AD attributes, to sync to Entra and dynamic group membership. The visual shows that licenses are attached to groups too.
A PS script they can run as needed to make a user account, but they have to type the attributes and details in from the request. They can copy/paste from the ticket, and they have a reference sheet for Departments, Job Titles, Extension attributes, etc., and what they give access to.
If our volume demanded it, I'd absolutely work on scaling it. With a 5 person team and two techs underneath me, my value comes from being able to update our documentation, even if I'm the majority maintainer, and teach and help troubleshoot, so they are not as reliant on me later. With you, it sounds like your scales demand the automation so your team can focus and pick up on the issues that keep the place running, which is exactly how automation should be used. To scale and meet your demands to be able to take care of important issues and details.
WizardsOfXanthus@reddit
We have just over 11,000 employees. Our source data comes from our parent hospital. Combined, it's around 55,000 employees, but the hospital I work in, just over 11,000. We pull in the source data, then my jobs run as middleware for any new hires, job changes, terms, retires, etc. We have multiple staging tables that we dump the data into, and then many of our downstream systems (Orchestrator, K2 dashboard, etc.) use these tables for their processes.
Source data refreshes by 4:00 am. My jobs run and complete by 5:10 am, so when I come in and see a failure, my job also reverts back to the previous day's data until it's fixed. Sometimes when this happens on a Sunday or Monday (every other Monday is new employee orientation), HR is NOT happy with coming in and not finding their updated list. Can't please everyone, but it is what it is.
Yupsec@reddit
Like Google told their first group of SRE's, "We want you to try and automate yourself out of a job, good luck." You can't. Like you said, it's just a dream.
Fallingdamage@reddit
You keep building automations and refining what doesn't work. 800 employees isnt too hard to manage.
If you just got rid of all the automations, would you have less work to do?
We use some on boarding automation and part of the system is to prepare the data then check the data against what would be expected first. If there are conflicts, do the thing that would be done to resolve conflicts. "Our automation breaks when something unexpected happens." - Then build your automation to tolerate that and when the next unexpected thing happens, update your automation to handle that as well, etc.
I have automated about 60% of my regular routine work. Each one of my tools will tell me what failed while I use it. Any scheduled tasks/scripts will send me an email if they fail and some email me with results regardless of success or failure. No need to poke around and see what broke. Just check a log that your automatons should be writing to and see how they did.
If you're teams spends all their time figuring out if things worked or not, I think you might have the wrong approach.
National-Cricket7469@reddit
Honestly that sounds pretty normal. A lot of automation works great until something slightly different happens, then it breaks and someone has to step in. Sometimes the issue is just that the workflows are trying to handle too many edge cases.
Keeping some automations smaller and simpler tends to make them more reliable. We actually ran into something similar before. For some repetitive tasks we stopped over engineering it and just used Workbeaver, here we can create our own template by recording the task's process once, saving it and that's it we can reuse it whenever we need that task done. Not the normal automation we use to know, but it definitely works pretty good.
just_a_user@reddit
This paper, The Ironies of Automation, is from 1983 and still holds true today.
cantuse@reddit
Damn, the first 2-3 paragraphs are just spitting the absolute truth:
thortgot@reddit
Improve your on-boarding to handle the edge cases or better yet solve your input so an invalid field would stop the user first.
MajStealth@reddit
if the onboarding can fail by "being off" prevent it being off my making the decision a dropdown with set options.
btw what you did is sort of what you wanted, you now dont have to do 100% of the work, but before you just did not check if it worked, or if it did not work it slipped through.
like our update to our itsm, which also triggered an agentupdate, which at some clients sort of failed, no service installed, but that only get noticed when you search for it.
Sunsparc@reddit
I have a "manual block" in my onboarding script that presents all of the information gathered from the onboarding ticket to the admin running the script. They can choose a number to change the information of that particular field and in order to proceed, they have to type out PROCEED (all caps) exactly. The script will stay at the manual block until PROCEED exactly is entered. If there is missing information, a big red warning appears in the value of that particular field to draw attention to it.
MajStealth@reddit
Entered "PROCEED exactly", script did not work, advice please.
Sunsparc@reddit
Thankfully the desktop techs I support with automation are intelligent.
Khue@reddit
Input validation is super crucial for doing automation. If input relies on an end user, it's super tough because users are INSANE with what they come up with sometimes.
MajStealth@reddit
Starting with the new user name.... the amount of double work i had with marcus/markus and carsten/karsten..... if even HR can not be trusted, who can?
Khue@reddit
One of the more frustrating things I've had to deal with in a past life is punctuation. There seems to be more free form use of various types of punctuations in names these days and typically you could just regex out a lot of harmful injection based things. User names are getting crazy inundated with unique punctuation.
Going back to the HR thing, we've resorted to placing pop up texts with massive warnings about how once entered, the name is set. Far too often did HR put in the wrong last name and then everything from the user's UPN to email address are messed up and it's almost easier to delete the entire account and redo it from scratch.
Frothyleet@reddit
For one, you just made my son Bobby Drop Tables; cry, so I hope you're happy.
Second, I don't know anything about development, or SQL, but an AI told me that you fix injection by parameterizing, so just do that!
fresh-dork@reddit
my favorite version of that was at a place where we did an email migration. my coworker is mexican and she's got two last names - they got the order wrong and then tried to tell her to just deal with it. lovely stuff
pdp10@reddit
Link the user database to the HRIS, with the HRIS authoritative, and an explicit use-case of changing names or the spelling of names. This means that the account name isn't the primary key, but instead a "UID" or "SID".
It's substantial work, but when it's done, HR gets to fix their own inaccuracies.
MajStealth@reddit
Oh my little padawan. It soinds nice, until you learn of SAP and their prices or 20year old software that also does the timesheets, sidedoors and the truckgates.....
sobrique@reddit
Ugh yeah. We had a 'super legacy' payroll/time clock system running on an entirely different platform (OpenVMS).
And it worked really well, and NEVER failed for literally years.
It just had precisely zero interoperability with anything else, because plenty of people haven't even heard of it.
That was a good few years ago now, but I'm honestly not sure it's not still in service, because like I say it never broke.
TheOhNoNotAgain@reddit
https://xkcd.com/1319/
ConstanceJill@reddit
OK, I've only just noticed that there are various options under the comic itself, such as "Dark mode". So I tried it and thought "Heh, not good enough". So let's try "Darkest Mode", then… welp, should have seen it coming, can't see anything. But fair enough, it is the darkest… "Space Opera Mode", what's this? LMAO.
FlickKnocker@reddit
Since the dawn of time, humans have tried to automate away mundane tasks, but it always requires maintenance/refactoring. Think of the first dudes that hooked a plow to an oxen. It probably worked 80% of the time, but then the plow would break, the ox would kick the guy and kill him, a wolf would eat the baby oxen...
chickibumbum_byomde@reddit
quite normal there, you didn’t remove the work, you moved it from execution to validation. Automation only helps if it’s reliable and observable. If workflows fail halfway, patches leave edge cases, and alerts don’t lead to action, you end up babysitting automation instead of benefiting from it.
The missing piece here is proper clarity/monitoring and feedback loops. You need to clearly see what worked, what failed, and why, without manually checking everything every morning.i personally too lazy to track manually, automation therefore is a lifesaver, sometimes literally heheh, because you can track automation outcomes, surface failures as real incidents, and avoid digging through everything manually.
Until automation is predictable and transparent, it will always feel like extra work instead of less.
viral-architect@reddit
Nothing creates more manual work than new automation.
Geminii27@reddit
So it was automated, just... poorly, with no error checking.
knightofargh@reddit
You got the first 50% of automation in place. Now you need to make it self-healing and reliable. Agile sucks, but it teaches one thing correctly “start small and iterate”.
Take a programming course and learn about exception handling and input validation. Knowing how to program makes automation scripting easier because you know how to gracefully handle errors and more importantly how to ensure a bad input validates or refuses to even start the automation.
420GB@reddit
So clearly you're not done automating?
Why does your automation even accept a "field that is off"? Do you know validate your data? That's terrifying. Also, retries and error handling. These are solved problems, you just haven't done the work it sounds like.
Why do you do the cleanup manually? Why isn't the recovery and cleanup process after a failure automated?
It sounds like you are just getting started with your automation.
pdp10@reddit
Automation is supposed to have error handling. For example, command-line programs, and perhaps even Windows GUI programs have exit codes that indicate error conditions.
This isn't a good substitute for programs and systems that handle errors with aplomb. For example, package management systems that are transactional, and back out an update if things go wrong. Downloading programs that resume partial downloads with HTTP Byte-range requests.
APIs should always allow query of current-state so that automation can check its own work. For example, querying the BMC user list with IPMI protocol.
When it comes to updates, systems tend to right themselves with a bit of time. Perhaps in some cases it would be sufficient to check update status after 48 hours, instead of 24 hours.
bossman1337@reddit
In the programming world it is called CI, Continuouse Integration. You should just keep improving your automation until it handles all possibilities.
OtterwiseOccupied@reddit
You need an end to end testing suite - you know what your outcomes should be and what your required inputs should look like.
If your automation is in a git repo (and no keys are present) have Claude introspect your code, give it a framework of inputs and desired outcomes and use the brainstorm skill to step through it. I bet you could have this up with an mvp in a week.
The flows sound great that you’ve built - with a testing suite and some validation, I bet it starts to feel better.
bu3nno@reddit
There are likely solutions but it depends on your workflow. For user onboarding, what is the source of data? I'd look I to making the required fields mandatory. If you can't do that, or the creation fails, you can set default values and/or notify the relevant party to fix and retry.
If you are having to review the automation at every failure, then you need to add logging, send meaningful notifications with the failure reason. There's no reason to spend all morning figuring these things out.
OneSeaworthiness7768@reddit
Yeah OP is trying to make that solution, that’s why they make these silly posts. Very unlikely they even work in the field. They’re trying to build a SaaS they can sell to IT departments. Their post history is all vague AI written posts asking about how they solve their problems. Classic vocoder research.
Remarkable_Papaya_79@reddit
Change your strategy before deploying automation. This is a big problem if your team has to check every automation process. Consult someone to help in forming new plans to implement automation.
jekksy@reddit
"Automation applied to an efficient operation will magnify the efficiency. The second is that automation applied to an inefficient operation will magnify the inefficiency". - Bill Gates
iama_bad_person@reddit
You're thinking about this the wrong way. The way I see you, you HAVE reduced your workload. Do you really think chasing up the occasional failed automated onboarding (which, with proper error prevention on the inboarding side, should be few and far between) is more work than doing all of them manually? The percentage of devices needing manual cleanup vs all of them needing manual patching?
GoldTap9957@reddit
If you have to check it every morning, it's not automation. It's just delayed manual work. Just saying!
Likely_a_bot@reddit
A bad process is the enemy of automation.
Rushing to automate a process without first defining and refining that process can actually negate any efforts to automate it.
hankhalfhead@reddit
We treated our onboarding automation like a provisioning queue. We racked up the onboarding requests from hr in our ticketing system, and made a tool to consume them. the tool populates an account creation script. We therefore can review and adjust each account creation before it fires. The tool validates the request and fails on any prechecks such as username collision etc.
Not perfect but it is definitely a net benefit
Max-P@reddit
Bad automation is worse than no automation in a lot of cases, especially if it's not trivial to fix and you have to do the rest manually. The fact you can't just correct the field that's off and continue deployment is a major red flag for the automation.
At most you should have like, oops Slack didn't install automatically, but everything else did so it's a 5 minutes fix to just remote in and install it manually.
Good automation is, user unboxes brand new laptop shipped straight from the factory, log in with their corporate email, give it an hour and all the apps, settings, VPN and updates are installed and the user is ready to work with zero further interaction. Automation failure is supposed to be the rare exception not the rule.
CaseClosedEmail@reddit
Make those fields mandatory and keep responsibility in that team. We had something similar where the HR Admin didn’t care about the fields.
With a new HR Admin team, zero issues
statix85@reddit
If you need to check automation daily it’s not automation. Focus on the solution first and make sure it works 99% of the time
mb194dc@reddit
Hasn't onboarding been automated for 20+ years in most big orgs.
Welcome to the 21st century, moving where bottlenecks are = efficiency...
Vegetable-Ad-1817@reddit
I remeber some article a few years back that supposed there was microsoft phlosophy on something like this, they would offer all these products cheap as chips, but the money they made was on integrations, if you want your kanban in project to integrate with your calendar in outlook ka-ching. If you want the version of something that works with datalake ka-ching. And other followed fast, servicenow is utterly terrible until you integrate it with, oh wait no I cant I need the premium plugin for that app. And that flows down and down. AI is only making it worse, if the industry is based on that premium collaboration model then its only going to helpfully tell you exactly what premium product you need to integrate it, or risk it having to write an entire middleware app in C# and you spend 3 days figuring out why some regex it wrote doesnt work, only to find they change some verb in the API from 'run' to 'execute' in version 47.3.7845. TLDR everything about IT is based on deception
Paulina8097@reddit
I wouldn't expect to have my whole onboarding workflow 100% automated. Nor I would want this
I'd automate just small chunks of the workflow and make sure to set up a test and QA processes to make sure the automations work.
Something to have in mind while you do this is that there are steps where it's good to have manual work. For example, just before submitting a payroll, having someone manually check each value is fundamental.
Do you have any metrics to measure the impact? Is the onboarding faster in any way? If yes, then the automation might be actually helping, even if it takes a lot of manual work.
EgoistHedonist@reddit
That's just badly built automation. Shit code does shit things.
affordable_firepower@reddit
This describes the 'automation' where I work.
We've managed to automate tasks, but not processes. each task requires the input to be in a specific format.
What lessons can I pass on to you from this?
Validate the input. if you start with data in a format that some script isnt expecting, then you'll hit errors. Trim leading and trailing spaces. Force the input to use only the characters you need.
have the scripts/workflows constructed in a way that can handle exceptions. we have a server build script that can't handle a server with an additional, unused network card in it. what the script should do is recognise that, log it, carry on and then end with a distinct return code.
Don't assume anything - always double check what each step of the process is expecting in terms of input and output formats
log everything. And return that log to the process initiator to sign off
valryuu@reddit
Limit possible inputs.
USarpe@reddit
Die Frage ist, was war vorher mit den fehlgeschlagenen Task? Hattet Ihr darüber einen Überblick und könntet Ihr alles lösen? Der Vorteil einer Automation ist ja auch, die Dokumentation und das Monitoring. Wenn jetzt statt 80% der Systeme 99% überwacht und gewartet werden, wäre das schon ein riesen Fortschritt. Und es sind immer die letzten 10-20%, welche 80-90% Aufwand verursachen.
N7Valor@reddit
Because automation itself needs to be maintained and improved upon. You've never heard of a self-maintaining / self-healing / self-updating Powershell script right?
gumbrilla@reddit
Yeah, sounds brittle.. It's got to have a % success, over x%.. I'd say 90%. If it's not, you've not got automation, you've got shit automation.
It should also alert. you shouldn't have to check, even when it does go wrong. It's an incident, and then a problem ticket.
If I was guessing, that you didn't add operational requirements to the project/work. School boy error. It's been implemented with functional requirements only, which is very very ironic.
j9wxmwsujrmtxk8vcyte@reddit
So you built bad automations and you are surprised they don't magically make life easier?
If a field can be off, there is not sufficient validation.
If an automation does not fail gracefully, it is trash.
If an automation does not include a notification when it fails in a problematic state, it was built by an incompetent person.
There can be growing pains but if things are not getting better, get a professional to help you.