2020-01-23T14:08:00Z

What is your primary use case for Tidal Automation?


How do you or your organization use this solution?

Please share with us so that your peers can learn from your experiences.

Thank you!

Guest
1414 Answers

author avatar
Top 10Real User

We use it for a host of standard/general stuff, like batch workflow automation, in the front and back offices. We have also centralized all of our SQL Server maintenance that is running on it. Instead of having SQL Server maintenance plans or jobs running on 300 or 400 disparate servers, we run them through Tidal so we have consolidated administration and reporting that feeds straight into ServiceNow. Last year, we made a step change with our DR recovery process. We had a bunch of people running manual scripts and different things where you have networks: Wintel, DBAs, or application support teams. They were running their own separate scripts to do application failover. This is different when it's active-active or active-passive replication. What we did was integrate it with different command line driven jobs, like PowerShell commands, to effectively failover applications and infrastructure into a sequenced set of dependant jobs. Therefore, if we need DR, we were not relying on a mix of SMEs saying, "Where was that script or how do we fail this over?" Instead we can just push a button and the thing fails over, which is beautiful. Additionally we do compliance reporting from within Tidal and like many people we are regulated from PWC. Everyone has the technology control frameworks that they have to evidence. Instead of people taking screenshots, we will effectively find out what information PWC need and build the job using CLI which runs on either month or quarter end. The job will go off, collect that evidence, come back, and be formatted. Then, we just drop it in SharePoint or use Tidal to save it to a file share, sending an email off to say, "Your evidence is collected. You need to review it, then sent it onto audit." We use it for a vast array of housekeeping jobs. It is not that Tidal is a monitoring tool, but automation is basically as far as your imagination can take you with anything that runs by a command line, which is virtually anything you can do. We previously had a use case for it to give us a quick alert for when some of our infrastructure became unavailable. We just had it running every minute. Typically, it's not an enterprise monitoring tool, but if you have some deficiencies or things that you need to enhance, or give a different sort of dimension to, we've used it for that in the past. We also run it against our infrastructure using PowerShell to pull a whole host of reporting from our infrastructure daily, which is useful. We use Tidal to run SQL Server and Windows. There is not really any Unix. Since we start using it, they do more stuff in AWS. They now have a whole bunch of different cloud capabilities. We are moving towards private cloud. We're in the sandbox at the moment.

2020-04-05T09:13:00Z
author avatar
Top 5LeaderboardReal User

It's a company-wide batch scheduler. It runs tons for us. It runs Windows, Unix/Linux. We connect with a lot of databases: Oracle, SQL, Sybase. We have BusinessObjects BI adapters, we scan emails, and we incorporate it with TriZetto Facets healthcare solutions. There's so much. It's our core enterprise scheduler.

2020-03-03T08:47:00Z
author avatar
Top 5LeaderboardReal User

We use Tidal to run jobs across multiple application platforms, such as SAP, ECC, PDN, and Informatica, as well as jobs that run in Azure cloud. We also use it for several warehouse management jobs with OS/400 and AS/400 connectors. We have a lot of different types of connectors, then we are bringing all these jobs into Tidal so we can set up dependencies between jobs that run, e.g., an SAP job and a OS400 job may be dependent on each other in some way, allowing a cross-platform job flow. We are currently on the most recent version.

2020-02-12T17:16:00Z
author avatar
Top 10Real User

I have three installs of Tidal: production, qual and dev. I have a portfolio of 12,000 unique job definitions in production, 13,500 definitions in qual, and about 8,000 in dev. The Tidal adapters I use are for Windows and Linux agents, as well as Informatica, Cognos, and mSQL.

2020-02-09T12:23:00Z
author avatar
Top 10Real User

We use Tidal extensively to run our health and welfare claims processing throughout the day. That's the reason we got Tidal back in 2011. We receive 15,000 to 20,000 claims a day and we use Tidal to process the whole thing, all the way through to creating checks at the end of the day. Since 2011, we've expanded it to other applications and other processes: mostly reports, and files that come in electronically from other companies that feed other applications. And in a roundabout way, what we use Tidal for is to execute the applications to load whatever needs to be done on those applications. The transfer function we used to do with Tidal has been switched over to another software product called Cleo. And that is run by our network team. That way they can control all the information that comes in and out of our building. They can put secure FTP on it, encrypt and decrypt the information, and set password protections. Cleo has its own scheduler, like Tidal, but they don't use it. They let Tidal execute the Cleo commands to bring the data in and Tidal will execute any application programs after that. Overall we run 1,100 to 1,200 steps every day, depending on day of the week. I call them "steps," but they're actually multiple steps. Before you get to the actual processing of a program there might be a move, a copy, or a delete when we're clearing out folders, using DOS commands. We then move data around to certain directories so that either the TriZetto software that we use can find that data or any internal programs that we use in VBS, .NET, Oracle, or MS SQL stored procedures can find that data. We're also starting to use this new MDM application which captures addresses from various databases, verifies they are correct, and pulls them together into one database. After all of our nightly processing, we have Tidal kick off the main MDM master so that all those addresses are in sync. Tidal sits on its own database and then it talks, through agents, to the other applications.

2020-02-05T10:15:00Z
author avatar
Top 5LeaderboardReal User

We have a product called J.D. Edwards which is our ERP system. Our biggest use case for Tidal is to automate jobs that we submit through J.D. Edwards. Our second use case would be automating maintenance — stopping services, deleting logs — your "keeping the lights on" type of stuff. And our third use case is using it for any automation tasks that we come across. Tidal is our product of choice at the moment. If we're going to automate something, we're going to use Tidal to automate it. We integrate Tidal with Linux, Windows, iSeries, SQL Server, and Oracle, in addition to J.D. Edwards.

2020-02-05T10:15:00Z
author avatar
Top 10Real User

We use it to call multiple source systems, such as Informatica workflows, Unix scripts, Windows scripts, PowerShell, batch files, and a few SAP web programs. We use it for certain file events and monitors. Tidal, by itself, can't monitor, but we create a script and job for that, then schedule it in Tidal. We use Tidal for multi-purposes. We use Tidal for our SQL Server, where we call from Tidal any procedures, statements, SQLs, or jobs. We also call a few HANA Stored Procedures from it. As of today, Tidal doesn't have an adapter, but we have some internal scripts which call Stored Procedures from Tidal. We run around 2000 to 3000 jobs per day. The infrastructure is in Azure.

2020-01-30T11:44:00Z
author avatar
Top 10Real User

We use it to manage our batch processing. For us, it came in as a replacement for a lot of different systems running crontab. In our case it's primarily for Unix/Linux systems that don't have their own mechanism for kicking off all these batch processes. It's the coordinator of all of our background processes and batch jobs that are running overnight and during the day. We use it to kick off custom Unix/Linux scripts that will launch our application processes. It's almost entirely Windows and Linux shell scripts that it's kicking off.

2020-01-29T11:22:00Z
author avatar
Top 10Real User

Our use of Tidal is mostly file-event driven. We use it to manage our ingestion, processing, and loading of data. Tidal has a hook and it runs ETL for us. It runs jobs and SQL and some of our database appliances like IIAS, the new version of Netezza Teradata. We have a file gateway that receives a file and drops it in a location. That file event picks it up and drops it over to the ETL tool. The ETL tool will run and aggregate a number of source files and turn it into a properly formatted input file. That file then goes through data hygiene and data analysis. Then it goes through a matching process. It is then put back out and runs an ETL process to stick it into a SQL database. And then there are a number of jobs that are run in the SQL database to manipulate that file. We don't have a lot of calendared events or scheduled windows. We have a central location for Tidal in our data center, and then we have client-hosted solutions where we run smaller instances of Tidal, and those are in the cloud. We use AWS, Azure, and GCP.

2020-01-29T11:22:00Z
author avatar
Top 5LeaderboardConsultant

For most of the companies where I have put Tidal in, it runs everything. It does back office, handles trading, reporting of time, doing a lot of file transfers between vendors and regulatory bodies, etc. We use it to do a whole variety of things. File transfers are our most valuable use cases because those are the ones where we tend to have service level agreements and potential fines. Right now, we are just in a traditional installation with local servers. We use the solution from Hadoop and Workday and are not using adapters from them.

2020-01-29T08:35:00Z
author avatar
Top 10Real User

We use it primarily to run SAP jobs. While there is other minor stuff it runs in, 98 percent is SAP. We have a number of different types of SAP systems. There are different teams who are responsible for configuring, managing, and setting up jobs. They are the ones who define the jobs and schedule them. There is an administrative team who is responsible for maintaining the system landscape and providing training for Tidal. They also provide standards, guidance, guidelines, and jobs. We use the solution for cross-platform, cross-application workloads within SAP. Therefore, within SAP, we might run a job on one system, but wait for the job on other systems to finish first. That is our interdependency between SAP systems. However, we don't do things like run something on SAP, then go do something on a non-SAP system. We may have a bit of that, but that's not a big part of what we do. It's mostly within SAP systems or within an SAP system.

2020-01-29T08:35:00Z
author avatar
Top 10Real User

We're running jobs on a global scale. Being a global company, we're running scheduled jobs and ad hoc jobs across different regions. Jobs cover backend processing, financials, and the like. We're running on an SAP ERP system and we're also running Informatica for data warehouse. We're running BusinessObjects web reports as well as a lot of straight Windows and Unix command-line things. We run FTP processing, PGP encryption processing, and data services jobs. We're running about seven or eight of the different adapter types that Tidal has available. We have it on-prem. Both our test and production environments are on fault-tolerant setups.

2020-01-27T06:39:00Z
author avatar
Top 5LeaderboardReal User

We primarily use it for scheduling our JD Edwards ERP software batch jobs. The solution runs on Windows. It also integrates with our Unix & AIX systems. We use it for automating EDI transactions, so it reaches out to FTP sites as well.

2020-01-27T06:39:00Z
author avatar
Real User

We are mainly using it for triggering data jobs. It does a lot of ITIL stuff and data movement from systems into Hadoop. We use it because it has the capability of dependency triggering or dependency running. That's the main idea behind it. Also, it helps us to centralize and organize jobs across the organization. We use Tidal to run Hadoop backup system, SAP HANA, and SAP BusinessObjects. We also trigger a lot of jobs into SnapLogic, Salesforce, ServiceNow, Workday, and Tableau, along with a couple of dashboards. We run a couple of batches from our Unix and Windows machines: the stuff that the developers are working on and want to run in ITIL. But, SAP is the main thing. The main goal is to use Tidal for managing and monitoring cross-platform, cross-application workloads. The ability to manage those loads is what they do well. I can put a job to run in SAP, and once the job ends successfully, I can run that job in Hadoop. Or, I can run that job in Salesforce.

2020-01-23T14:08:00Z
Learn what your peers think about Tidal Automation. Get advice and tips from experienced pros sharing their opinions. Updated: April 2020.
442,041 professionals have used our research since 2012.