Subscribe to Blog via Email
I’ve been discussing for years about the importance of network to database performance, especially once I started working on VLDBs, (Very Large Databases) but its a topic that often is disregarded. Now that I’m working more and more in the cloud, it’s become more evident the importance of the network to our survival.
For each and every cloud project I’ve been involved in, there is evidently going to be multiple challenges that turn to the network administrator for a solution. I don’t blame the administrator in any way when he becomes exasperated by our requests. As it is my solemn duty to protect the database, the network administrator is the sole protector of the network. You’ll hear a frustrated DBA say, “just open the &^$# network up! Let’s just get this connected to our cloud provider!” I have to admit that this request must be akin to someone asking a DBA to provide SYSDBA to a developer in production.
So yes, there are a lot of moving parts in a cloud environment. No, not all of them are at the database level, but many of them could be at the network level. This means that your new cloud environment must connect past firewalls, proxies, blocked ports and authentication steps that may not have been required back in the sole on-premise days.
Yeah, there’s a bit more to the network than demonstrated in the picture above.
The database connection needs a secure connection past the firewall and may require proxy configurations to access via a web browser. The application interface to manage them may require proxy settings in browsers that may have automated processes to manage outside a manual proxy setting. You may have network configurations that are different from one local office to another. We’ve only discussed configuration and haven’t even considered speed, packet size and bandwidth.
So here is my recommendation- make friends with your network administrator. In fact, take the ol’ chap out for a beer or two. Learn about what it takes to master, protect and ensure the company’s network from the threats outside. Learning about the network will provide you with incredible value as a cloud administrator and you may get a great friend out of the venture, too. For those of you that don’t make friends with your network admin, I don’t want to be hearing about any mishaps with phenobarbital to get the information, OK? 🙂
The sales, support and technical teams were brought into the Denver Tech Center office to do some advanced training in Hybrid Cloud. There were many take-aways from the days we spent in the office, (which is saying a lot- most of you likely know how much I hate working anywhere but from home… :)) and I thought I would share a bit of this information with those that are aching for more details on this new and impressive offering from release 5.
If you’re new to Hybrid Cloud and want to know the high level info, please see my blog post on the topic.
Cloud Control now includes a number of new options for database targets in EM12c. These new drop down options include cloning to ease access to the new hybrid cloning. Once you’ve logged into Cloud Control, go to Targets à Databases and then choose a database you wish to implement cloning features for. Right click on the target and the drop downs will take you to the cloning options under Oracle Database –> Cloning.
There will be the following choices from this drop down:
Using a test master is essential for snap clones, which are a great way to offer great space savings and eliminates the time that is required for standard cloning processes. The test master is in a read only mode, so it will need to be refreshed or recreated with an up to date copy, (which will then be another option in the drop down, “Disable Test Master”) for new cloning procedures.
For the example today, we’ll use the following production database:
We’ll use an existing test master database to perform our clone from:
We can right click on the database and choose to create a clone. This is going to be an artifact via a snapclone, so keep this in mind as we inspect the times and results of this process.
Upon choosing to create a snapshot clone of a pluggable database. This will then create snapshot clone, each clone is just a copy of the file header with block changes involved on the read only or read-write clone.
Once you fill out the pertinent data for the clone, using the correct preferred credentials with SYSDBA privileges, name the new pluggable database, the name you’d like it displayed in Cloud Control as and enter the PDB administration credentials, password and confirm the password. Once that’s done, choose if you’d like to clone it to a different container, (CDB) than what the source resides on and then add the database host and ASM credentials.
Once you click next, you have more advanced options to view and/or setup:
The advanced options allow you to change the sparse disk group to create the clone on, storage limits and database options.
You can then choose to have data masking to protect any sensitive data, but keep in mind, once you do so, you will no longer be using a snapclone due to the masking, but the option to implement it at this step is an option. You can also set up any pre, post or SQL scripts that need to be run as part of the clone. This could include resetting sequences, user passwords, etc.
The next step allows you to schedule the clone in the future or run immediately.
You can also choose the type of notifications, as this is simply an EM Job that is submitted to perform the cloning process. Once you’ve reviewed the cloning steps chosen via the wizard, you can then submit.
Once the jobs been submitted, the submitter can monitor the job steps:
Once the clone has completed, you can view each of the steps, including the time each took.
The source database was over 500 GB and was cloned in less than one minute! You also will see the new cloned database in the targets list:
If curious, note that this is a fully cloned database that is on ASM, which you can view, just as you would for any other database.
Again, note the size and that this can be managed like any other database that you would have created via a DBCA template or through a standard creation process.
More to come soon and thanks to Oracle for letting us get our hands on the new 184.108.40.206 hybrid cloning!