Monday Jul 06, 2015

Understanding Plans, Profiles and Policies in Enterprise Manager Ops Center

Enterprise Manager Ops Center uses a combination of Plans and Profiles to maximize repeatability and reuse, while giving you the degree of flexibility to provision and update what you need to in your environment. The structured building block approach will allow reuse of many of the components instead of re-entering data each time, but does make the whole thing look very confusing until you understand the relationship between plans and profiles.

The sort activities covered by plans and profiles are:

  • OS provisioning and configuration
  • BIOS and Firmware updates
  • LDOM creation
  • Zone creation
  • Boot environment creation
  • Patching and adding packages (S10 and S11)
  • Configuration files and pre/post action scripts (S10)
  • Automation using operational plans (scripts)

The Building Blocks

So, firstly, let's look at the building blocks and see what is the difference between a Plan and a Profile.


Profiles contain the configuration information required to perform an action eg:

  • OS Provisioning plan - contains type/architecture, resourcing, OS version, software group, language, timezone, password, filesystems, name services
  • OS configuration - contains type/architecture, agent/agentless, multipathing, network configuration
  • Discovery - contains asset type, tags, IP address, credentials, grouping
  • Logical Domains - contains type, name, CPU/Cores, memory, automatic recovery, storage, networks

These are just some examples of the 16 types of profiles available in the BUI. 


Plans are objects that can be run (executed) to make something happen. Plans contain profiles, other plans, policies, or a combination of these.


In addition to these profiles, there are update and monitoring policies and credentials that can also be created and edited here.


Let's look at a couple of examples from a process prospective and how they actually look in the BUI. A side note here is the screenshots have been taken from an Ops Center 12.2.2 environment. Ops Center 12.3.0 introduces new styles of icons, but the principles are still all the same.

Process overview

For example, if we were going to provision a bare metal physical sever, you would have 2 choices:

  • A simple provision that would just lay down an operating system
  • A complex provision that
    • could update the BIOS on an X86
    • lay down the OS
    • update (patch) the OS
    • Add applications
    • Run scripts
    • Apply a non default monitoring profile
    • etc.

You choose the one that best suits what you are trying to do.

Simple OS Provisioning

If you just wanted to get an OS laid down, basic network/filesystems configured and possibly an agent installed, you would choose a "Provision OS" plan

This plan contains 2 profiles, an "OS Provisioning" profile and an "OS Configuration" profile. These profiles contain the answers to the same questions that would have been asked if you provisioned the server manually. A point to remember: it is required that you create your profiles (answering the questions that the wizard presents) before they are available to be added to a plan.

In the BUI it looks like this:

Complex OS provisioning

This plan must have the same 2 plans that the simple approach did, but has the option to add many other plans to be able to patch the deployed OS (Update OS, Update software), add software (Install, Operational plans), and further configure the OS/Application using scripts (Pre-install, Post-install, Operational plans)

In the BUI it looks like this:

The steps (profiles) you choose will be determined by what you want to achieve and if you are provisioning Solaris 10/Linux or Solaris11. 

Duplicating steps

In addition, most of these optional steps can be duplicated to allow you to execute more than one profile. To do this, add your first profile for that step, then select (highlight) that step and if it is available, the copy icon (with 1, 2 shown on it) will become active.

 Click that icon and the step will be duplicated allowing you to run more than one profile.

This makes the whole operation much more flexible, as you could have an update profile for your OS , one for web servers and one for databases. So if this plan was to build a web server, your plan would contain both the OS update profile and the web server update profile, avoiding the need to have the OS patches in 2 profiles.

 Other examples

 Another example of this would be if you wanted to build and LDOM or if you wanted to build an LDOM and deploy an OS into it (complex or simple), you would choose the appropriate plan.

Building an LDOM

Building an LDOM requires a "Create Logical Domain" plan

which only has a single step, which is a "Logical Domain" profile.

Building an LDOM and provisioning an OS

You can build the LDOM and provision the OS into it in a single action by creating a "Configure and Install Logical Domains" plan

which contains two steps, which is a "Logical Domain" profile and an "Install Server Profile"


By now, hopefully, the pattern has become clear. Plans and profiles are just the building blocks that allow you to deploy your system in the way you want. Each of the components are modular, so they can be re-used as much as possible and make it easier to maintain as you have fewer places you need to change when you want to change your configurations. There are many other types of plans offered by Ops Center that will create zones , build M-series physical domains and deploy OVM for X86 virtual machines, individually or combined with OS deployment, but they all follow the same basic structure. While how to do this is all laid out in the online documentaion, my best advice is to get yourself some test systems and try it out. There is often no substitute for having actually done it.


Rodney Lindner

    Wednesday Jun 10, 2015

    New Enterprise Manager Release Allows Customers to Manage Their Hybrid Cloud from a Single Interface

    Oracle is today announcing Oracle Enterprise Manager 12c Release 5, which simplifies the journey to the cloud.  For the first time, you will be able to manage your hybrid cloud as one, from a single pane of glass.

    According to a recent Computerworld survey, cloud consumers are struggling with security issues, integration headaches and performance concerns in public and hybrid cloud environments.

    The latest release of Oracle Enterprise Manager addresses these concerns through new enhancements that manage, migrate, test and deploy applications and databases across hybrid clouds, with 100% reuse of existing management tools and practices.  These enhancements will provide centralized visibility and control to Oracle customers on their cloud journey, while helping to ensure that their existing company-wide standards are applied across the estate.  The following set of hybrid cloud features is now available for those using Oracle Database and Oracle Java Cloud Services:

    • Workload portability & secure, bi-directional cloning: Replication of application artifacts and data across the hybrid estate with automated workload migration
    • Quality of service management: deep, expert performance management, diagnostics, tuning, leveraging the full breadth of Oracle’s extensive best practices for optimizing our stack.
    • Lifecycle and Cloud Management: automated patching and rapid provisioning of database and middleware with self-service access for agility.
    • Data governance & configuration compliance controls:  Configuration and compliance management, data masking, compliance frameworks and rich notifications for control.

    Clone database to Oracle Cloud with the click of a single button
    Migrate on-premises IT assets to Oracle Cloud, or from Oracle Cloud to on-premises, with a single click

    A simple, secure deployment architecture

    This rich functionality is exposed through an extremely simple deployment architecture.  Administrators will  install the Oracle Enterprise Manager Hybrid Cloud Gateway to manage and encrypt all required communication between their existing on-premises Enterprise Manager and the Oracle Cloud Platform...and that’s it!  With no additional network reconfiguration, Oracle customers can immediately begin to manage their hybrid estate as if it were one.

    Customers, partners and analysts welcome Oracle’s hybrid cloud management approach

    "e-DBA, an Oracle Platinum Partner and customer, has seen the benefits Oracle Enterprise Manager brings to our clients regardless of the size and complexity of their estate, which is why we also use it internally,” said Phil Brown, Global Lead for Enterprise Management Technologies, e-DBA.  “Reusing our existing skills will accelerate adoption of the Oracle Cloud Platform since our hybrid cloud will be a natural extension of our existing estate, all seamlessly managed as one." 

    “As Bridge Consulting has expanded our Oracle footprint to include Oracle Cloud Platform, we have recognized the importance of centralized management control over our entire hybrid estate,” said Marco Bettini, Co-founder and CTO of Bridge Consulting.  “The new Oracle Enterprise Manager capabilities are exactly what we need to ensure our hybrid cloud can be managed and consumed as one single environment, and we are especially glad we can leverage our existing best practices and knowledge across the entirety of our hybrid estate.” 

    “Our customers have consistently told us that they want to treat their hybrid cloud as a single unit, with workload portability as well as consistent governance across the entire estate,” said Dan Koloski, Senior Director of Product Management, Systems and Cloud Management, Oracle.  “The latest release of Oracle Enterprise Manager provides customers the ability to lift-and-shift applications from on-premises to cloud and from cloud to on-premises, while leveraging all of the rich quality of service and lifecycle management capabilities they use today.  We are excited to enable a hybrid cloud that truly functions as an extension of an on-premises data center.”   

    “IDC predicts that the majority of IT organizations will adopt a hybrid cloud strategy in the next five years.  For these companies, a key success factor is the ability to consistently support workload monitoring, management and portability across on premises IT and public cloud services,” said Mary Johnston Turner, Research Vice President, IDC.  “Oracle customers who are evaluating a hybrid cloud strategy spanning Oracle on premises and public cloud databases, web services, backup services and development resources will be well served to consider the new Enterprise Manager hybrid cloud management capabilities.”


    One of the most appealing aspects of this solution is that it is so simple to deploy and operate

    Clear benefits

    The value of Oracle Enterprise Manager has been empirically demonstrated by numerous studies from Crimson Consulting, Pique Solutions, Forrester and IDC.  Some of the recorded benefits are:

    Clear benefits Oracle Enterprise Manager

    Oracle Enterprise Manager 12c release 5 will help you simplify your journey to cloud by managing your hybrid estate as one.  A LIVE webcast with more information and Q&A will be aired on June 25, 2015.  Please register here to learn how to overcome the challenges involved with managing IT environments where public- and private clouds and on-premises infrastructure can thrive as one 


    Stay Connected:
    Twitter |
    Facebook | YouTube | Linkedin | Newsletter
    Download the Oracle Enterprise Manager 12c Mobile app


    Tuesday May 26, 2015

    Data to collect when logging a Service Request on Enterprise Manager Ops Center

    If you ever have to log a Service Request (SR) on your Enterprise Manager Ops Center instance, the question often arises as to what data you should include with the SR. Including the right information when you log the SR will greatly reduce your Time To Resolution (TTR).

    So, what data should you add to the SR? That will depend on what type of issue you are experiencing.

    Basic Information for All SR's

    While the My Oracle Support (MOS) portal will help by asking you relevent questions, here is a list of things to think about while answering them:

    1. A clear problem description - Tell us about what you were attempting to do when you encountered the issue, what host(s) were involved and what error message you saw. You would be surprised how many SR's get logged with "It didn't work" or just the error message as the problem description without telling us what you were actually trying to do. 
    2. Time and date - Tell us the time and date you saw the issue. The log files are time stamped, so knowing the time the issue occurred (at least approximately) will reduce the extent of searching of the log files that will need to be done.
    3. Software Version - Always tell us what version of Enterprise Manager Ops Center you are using. To find out your version, look on your Ops Center BUI under [Administration] ==> [Enterprise Controller] and at the top of the middle panel (Summary Tab) will be listed the Enterprise Controller Version. Don't forget to include if there are any IDR's (Patches) that have been applied to your Enterprise Manager Ops Center instance.

    Additional Data to include with your SR

    The most common thing to include with your SR is an OCDoctor --collectlogs output (# /var/opt/sun/xvm/OCDoctor/ --collectlogs), but it will often depend on which part of the process Ops Center was at when the issue occurred. If your issue falls under multiple sections below, please supply the data from all relevant sections.

    It should be noted that all of the Ops Center logs will rollover over time and on a busy environment that may not be a long time. So, it is important to collect any log files as soon as possible after you have seen your issue or you should reproduce the issue just before collecting the logs.

    Browser User Interface (BUI) issues

    • Collect screen shots of what you are seeing in the browser
    • Run # /var/opt/sun/xvm/OCDoctor/ --collectlogs on your Enterprise Controller and collect the output file

    Job related issues (Job fails)

    • Provide the job number
    • Provide the full job logs
      • In the BUI, click through to the failed task on the job and then export the logs
      • Select the radio button for "Full Job Log" and save the job log
    • Run # /var/opt/sun/xvm/OCDoctor/ --collectlogs on your Enterprise Controller and collect the output file

    OS provisioning issues

    • Capture the full job log (as described above)
    • Run # /var/opt/sun/xvm/OCDoctor/ --collectlogs on the relevant Proxy Controller and collect the output file
    • Capture any console logs of either the physical server or, if deploying to LDOMs, capture /var/log/vtnds/[LDOM name]/console.log
    • If deploying to an LDOM, run # /var/opt/sun/xvm/OCDoctor/ --collectlogs on the relevant CDOM and collect the output file

    Patching issues

    • Capture the full job log (as described above)
    • Run # /var/opt/sun/xvm/OCDoctor/ --collectlogs on the system being patched and collect the output file
    • Run # /var/opt/sun/xvm/OCDoctor/ --collectlogs on your Enterprise Controller and collect the output file

    Discovery issues

    • Capture the full job log (as described above)
    • Run # /var/opt/sun/xvm/OCDoctor/ --collectlogs on your Enterprise Controller and collect the output file
    • Run # /var/opt/sun/xvm/OCDoctor/ --collectlogs on the relevant Proxy Controller and collect the output file

    Agent deployment issues

    • Capture the agent install output and logfile (/var/tmp/agent_install, /var/scn/install/log)
    • Capture the agent configuration output and logfile (/var/tmp/agentadm.log)
    • If the agent has actually been installed, run # /var/opt/sun/xvm/OCDoctor/ --collectlogs on the system you are installing the agent on and collect the output file
    • Run # /var/opt/sun/xvm/OCDoctor/ --collectlogs on the relevant Proxy Controller and collect the output file

    Domain Model issues

    • Run # /var/opt/sun/xvm/OCDoctor/ --collectlogs -u [EC Master Admin User] on your Enterprise Controller and collect the output file
      • This will collect the domain model information as well as the normal output from OCDoctor log collect. Note:  this can take a long time to run
      • The [EC Master Admin User] is the user that you specified as part of the initial install. On some customer systems, this will be root but it does not have to be.

    Collecting the right information and providing it as part of the initial logging of the SR will make it easier for the engineer to analyze your issue and avoid a lot of unnecessary back and forth. Overall, it will get your issue resolved quicker. I hope this information will help you if you ever have to log a SR on Enterprise Manager Ops Center.


    Rodney Lindner

    Wednesday Apr 22, 2015

    Test Drive Oracle’s Application Platform as a Service Solution

    We've just launched a new set of workshops in several U.S. cities: Oracle Platform as a Service for Application Development.

    This follows the success of the Database as a Service workshop series (see previous blog entry). It’s another opportunity to test drive new Oracle Enterprise Manager capabilities, but it goes far beyond Enterprise Manager. This time we focus on Java development and testing in the private and public cloud, and the cloud operations needed to support them.

    So bring your laptop, connect to our live environment and try it for yourself!

    The day begins with an overview of APaaS benefits and the architectural choices for building your enterprise private or public cloud (or both). You then use step-by-step workbooks that guide you through creating an application platform / middleware cloud environment. The event is perfect for application developers, IT managers and anyone developing, testing and deploying Java applications.

    The time is evenly split between private and public cloud labs. These are the workbooks we’ll go through:

    • Middleware as a Service
    • SOA as a Service
    • Fusion Middleware Provisioning
    • Creating and Exploring Java Cloud Service
    • Building and Deploying an Application with Java Cloud Service
    • Managing Java Cloud Service Operations

    Looks interesting? Register for an event near you.

    Wednesday Feb 25, 2015

    Editing EM12c Jobs in Bulk

    I recently received requests for suggestions on how to edit EM12c jobs in bulk. The two examples that were presented to me were:

    1. Change the start time of 70+ jobs
    2. Change the oracle home path in 80+ RMAN script jobs

    So how do we do this?

    Read More

    Monday Feb 09, 2015

    Take the IOUG Database Manageability Survey

    The Independent Oracle Users Group (IOUG), is conducting its annual database manageability survey. If you are a database professional responsible for managing and operating your organization's database applications and systems, then the IOUG wants to hear from you.

    Take the next few minutes to answer simply questions around your common manageability practices and challenges. Share your experience and insight with the IOUG and Oracle communities and you'll be entered into a draw for a chance to win an Apple iPad from the IOUG. You will also receive a free copy of the final report.

    Take the 2015 IOUG Database Manageability Survey

    Stay Connected:
    Twitter |
    Facebook | YouTube | Linkedin | Newsletter
    Download the Oracle Enterprise Manager 12c Mobile app

    Monday Jan 26, 2015

    Register Now for the Cloud Platform Online Forum

    January 28th, 2015 10:00 a.m. PST/1:00 p.m. EST
    Register Today

    Modern Business. Modern Cloud.
    Is Your PaaS Delivering the Agility Your Users Demand?

    Don't miss the opportunity to join Oracle and IDC in a series of deep discussions on Cloud Platform as a Service. This online event kicks off on Wednesday January 28th at 10:00 a.m. PST. Take advantage of 20+ sessions, 10+ demos and 100+ tips and techniques to plan your PaaS adoption effectively. Topics encompass private, public and hybrid cloud. You will also learn how to plan your transition from traditional IT delivery to cloud delivery using Oracle technologies you use every day. Register at

    Oracle Enterprise Manager discussions during this event include tips and techniques on workload consolidation using database as a service and application platform as a service; transforming traditional IT delivery using self-service; effectively collaborating with business users with showback/chargeback, and more. Our speaker Sudip Datta, Vice President of Product Management will also provide a glimpse into the future of PaaS management.

    Stay Connected:
    Twitter |
    Facebook | YouTube | Linkedin | Newsletter
    Download the Oracle Enterprise Manager 12c Mobile app

    Sunday Jan 18, 2015

    Enterprise Manager Ops Center - Ops Center EC/PC sizing

    Counting Access Points Vs Assets

    When designing a customer's Ops Center environment, one of the first questions I get asked is how many servers can an Enterprise Controller (EC) or a Proxy Controller (PC) support.  Customers think of sizing, as being the number of servers,  LDOMs, etc. and these are lumped together into a total number of assets. In the real world, the answer to the "How many assets can my Enterprise Controller /Proxy Controller support?" question is, "It Depends".  How you place your Proxy Controllers in your network and what you manage from each Proxy Controller can impact the total number of assets that can be supported from an Enterprise Controller.

    The amount of load placed on an Enterprise Controller or Proxy Controller is not directly determined by the number of assets it manages, but by the number of Access Points (AP) it is supporting. A simple view of Access Points is that they are a connection. But it is a little more complex in that they are also a reflection of the way Ops Center internally models its data structures. The actual load on any given Enterprise Controller or Proxy Controller closely links to the number of Access Points it is managing. The number of Access Points seen by the Enterprise Controller can differ depending on whether an asset is connected via a single or multiple Proxy Controllers. Therefore, the Access Point count on the Enterprise Controller will be higher than the actual number of managed assets, if assets are managed by multiple Proxy Controllers. When the documentation is referring to Enterprise Controller /Proxy Controller capacity and uses the term "Asset", it is actually counting "Access Points".

    Let's look at a couple examples of this to make it clearer:

    Physical Server ILOM

    In this example, a single physical server has multiple data collection methods. A Proxy Controller can gather LOM data by accessing the service processor and by using the agent running inside the operating system.

    If the same Proxy Controller is used to access a single asset, the two data feeds are consolidated into a single Access Point. The Enterprise Controllers total load will be a single Access Point.

    If different Proxy Controllers are used to access a single asset,  each Proxy Controller will record it as a separate Access Point. The Enterprise Controllers total load will be the sum of both Proxy Controllers and will be a total of 2 Access Points.


    Another example would be an LDOM guest where we obtain data from the Control Domain agent and the agent running inside the LDOM guest OS.

    Once again, if  both data feeds are via a single Proxy Controller they only count as 1 Access Point on both the Proxy Controller and the Enterprise Controller.

    Where as, if each data feed is via a separate Proxy Controller they each count as 1 Access Point on each Proxy Controller and the total Access Point count on the Enterprise Controller will be 2.

    With the release of Enterprise Manager Ops Center - 12.2.2, we have updated and renamed the Enterprise Manager Ops Center Sizing and Performance Guide ( This revised document is easier to understand and has been updated to reflect current releases of Ops Center. Note: Since the term "Access Point" was not considered to be commonly understood, the term "Asset" has been used in the documentation. 

    To make the counting of Access Points easier, we have added a program to the OCDoctor toolbox. This program "" is new in OCDoctor version 4.45 and can be found on any EC/PC or system that has an agent deployed. The path to the OCDoctor toolbox is /var/opt/sun/xvm/OCDoctor/toolbox.

    # /var/opt/sun/xvm/OCDoctor/toolbox/
   [-v] { standard | machine | agent | all }
            standard - show Asset count
            machine - show machine output with Assets names
            agent - show information about agents, agentless Assets, Service Processors
            all - show all the information together
            -v : verbose output, present additional notes around the counting algorithm
            (Version 1.0 2015/01/08) 

    Let's just look at the standard output (we are running this on the Enterprise Controller). The output shows:

    • The total Access Point count for the EC (72)
    • The number of Access Points for each type of asset, for each Proxy Controller ( Note: the total Access Point count for each Proxy Controller is labeled as Assets )
    root@ec1:~# /var/opt/sun/xvm/OCDoctor/toolbox/ standard
    EC 72
    Proxy Assets Zones Ldoms OVMGuests Servers Storages Switches ExadataCells MSeriesChassis MSeriesDomain
    pc4   32     5     25    0         2       0        0        0            0              0
    pc1   28     0     26    0         2       0        0        0            0              0
    pc0   12     2     4     0         6       0        0        0            0              0
    Use option '-v' to see additional notes on the counting algorithm.


    You can also use the "machine" option to list out which asset is attached to which proxy.

    # /var/opt/sun/xvm/OCDoctor/toolbox/ machine
    EC 72
    Proxy 32 pc4
    Zones 5 S11zone101 S11zone102 S11zone100 S11zone103 S11zone104
    Ldoms 25 stdldom21 stdldom34 stdldom36 stdldom22 stdldom45 stdldom47 stdldom25 stdldom43 stdldom49 stdldom30 stdldom23 stdldom29 stdldom42 stdldom20 stdldom32 stdldom46 stdldom44 stdldom26 stdldom31 stdldom27 stdldom48 stdldom40 stdldom35 stdldom41 stdldom28
    OVMGuests 0
    Servers 2 pc4
    Proxy 28 pc1
    Zones 0
    Ldoms 26 stdldom21 stdldom34 stdldom36 stdldom22 stdldom45 stdldom47 stdldom25 stdldom43 stdldom49 stdldom30 stdldom23 stdldom29 stdldom42 stdldom33 stdldom20 stdldom32 stdldom46 stdldom44 stdldom26 stdldom31 stdldom27 stdldom48 stdldom40 stdldom35 stdldom41 stdldom28
    OVMGuests 0
    Servers 2 pc1

    You can see this would be incredibly verbose on a large environment ( I have truncated it here). You can clearly see that the LDOMs (stdldomXX) are being reported by both "pc1" (via Control Domain) and "pc4" (via OS agent). Note: the differing LDOM count on "pc1" and "pc4"  as "stdldom33" has no OS or agent on it, so it only reports against "pc1".

    You can also use the "agent" option to display agent/agentless/LOM totals for each Proxy Controller.

    # /var/opt/sun/xvm/OCDoctor/toolbox/ agent
    EC 72
    Proxy Agents Agentless SPs
    pc4   25     2         0
    pc1   1      1         0
    pc0   5      5         5
    Use option '-v' to see additional notes on the counting algorithm.

    In addition, --troubleshoot will also be checking for the number of configured Access Points and comparing this number against the recommend Access Point count. It will print a warning at approximately 90% of the recommended capacity and a warning if you exceed the recommended capacity.

    Armed with the information above, you should be able to better design your Ops Center deployments and to monitor the growing load on the Enterprise Controller/Proxy Controllers as you add more assets.



    Monday Jan 12, 2015

    Enterprise Manager Ops Center - The Power of Groups

    To Group or not to Group

    Any customer with more than a handful of servers will struggle to take advantage of Ops Center's full features, unless they embrace the use of groups. The ability to group a number of like servers into a logical target, that can then be operated on, means you can do a single action instead of running the same action hundreds of times, once for each server in the group. Grouping also allows you to view and manage your environment in line with your current organizational units and business practices.

    What can be done with Groups

    Groups avoid the need to select individual servers and can be used as a target for:

    • Update profiles (Patching) - Update patches/packages and files and apply pre/post actions
    • Operational plans - Run a script, update config files etc
    • Perform actions - Reboot/halt/refresh/power-off/power-on actions to all assets in the group
    • Monitoring profiles - Apply customized monitoring profiles
    • Reporting -  Run a single report that includes multiple servers

    Groups can also:

    • Display membership - The Membership Graph or the asset tree view both show the servers that make up a group
    • Display incidents/alerts - Groups work as a roll-up point for incidents on any server in the group
    • Display Top CPU/Memory/Network consumers - The "Asset summary" of a group shows top consumers of CPU/Memory/Network resources
    • Restrict the assets that a given user has access to in Ops Center

    Types of Groups

    Groups are totally logical constructs. An asset (Server, OS, Zone, LDOM) can be a member of as many groups as you like. Deleting a group, does not delete the assets it contains. While most often groups will contain assets of all the same type (eg: System LOMs), as this will give you a group where an action like "power off" makes sense to all the members of the group, it is also possible to create a group that is made up of differing asset types eg: all the assets (OS/LOM/Zones) that are part of a physical server. This type of group would normally be used to restrict the permissions of users so that they could only view/manage the servers for which they are responsible.  A key fact to remember when thinking about groups is that an asset that is a member of one group is not precluded from being a member of other groups.

    Ops Center Predefined Groups

    As a starting point, Ops Center provides a number of predefined groups found under the [ Navigation ==> Assets ] menu.

    While most of the groupings are what you would expect, there are a few that require a little more explanation.

    Standard Views 

    [ All Assets ] - Not really a group as everything turns up here

    [ Engineered Systems ] - A grouping of all discovered Engineered Systems (SPARC SuperCluster). Note that each Engineered System is also its own sub-group

    [ Operating Systems ] -  Grouping based on OS type, release and version

    [ Servers ] - Grouping based on physical servers

    • SPARC
      • M-Series -  M[3/4/5/8/0]000, M6 and M10 servers
      • Other SPARC - servers that are not sun4u or sun4v architecture or non Oracle/Sun servers
      • U-Class - servers that have sun4u architecture CPU's (V240/V490/SF6800 etc.)
      • V-Class - servers that have sun4v architecture CPU's (T1000/T2000/T5XX0/T3/T4/T5 etc.) - not V-series servers as you might first think
    • x86
      • Other x86 - Non Oracle/Sun servers 
      • x64 - 64 bit servers
      • x86 32-bit - 32 bit servers

    [ Chassis ] - 6000/8000 blade based chassis and their server blades 

    [ Network Switches ] - Managed InfiniBand and network switches. Ops Center only manages a limited number of switch models and these will normally be found as part of an Engineered System (Exadata/Exalogic/SPARC Super Cluster).

    [ Racks ]  - Both Engineered System racks and manually declared racks. It is not commonly known that you can declare all the racks in your data center in Ops Center and place all your servers in their respective racks, giving you a useful data center view.

    All the predefined groups are useful but as you can see, they are based on broad brush physical characteristics of a server and its OS. There is no allowance for how you actually use your servers. For that you will need to build your own "User Defined Groups".

    User Defined Groups

    User Defined Groups are an extremely powerful addition to Ops Center and allow you to model your application, organizational units and business constraints into Ops Center's management interface. Basically, it makes Ops Center capable of working much more in alignment with the way your business operates. Before we go onto how to create "User Defined Groups", let's go over, in a little more detail, what you could use them for:

    • Applications - create a group of all your web servers to report on patch levels, apply patches, update application configurations, restart applications, list top resource consumers.
    • Prod/Dev/Test - create a group based on production status, so you can  apply differing monitoring/alerting profiles, produce management reports and restrict admin access.
    • Admin By - create a group of all the servers that an admin(s) is responsible for, so they can quickly respond to alerts or you can limit the functions they are authorized to perform.
    • Patching - create a group based on the servers that are in the 1st round of patching, so you can easily and consistently patch, do before and after patching reports and maintain consistent patch levels across key applications.

    These are just a few of the many things for which groups can be used. Setting up groups will greatly decrease your day to day workload and increase the manageability of your environment. Without the use of grouping, it is unlikely that you will be able to scale your Ops Center environment efficiently beyond about 30 servers.

    Creating a User Defined Group 

    First select the "Create Group" action [ Navigation ==> All Assets ==> Create Group ]

    Static Groups 

    Static groups are just as the name suggests, you define a group and place static members in it.

    The default action of the "Configure Group" wizard is to create a Static Group. As long  as the "Configure group rules" checkbox is unchecked this will be a static group.

    Give the group a name (mandatory), a description (Optional), and one or more group tags (Optional) and click "Next" and "Finish" to complete the wizard and launch the job that creates the group.

    Tagging is another powerful feature that will be the topic of another blog, but in summary, it is a way of storing an arbitrary tag (value pair) with an asset or group, which means you can store any important information with the asset, such as Asset Number, Production status, etc.

    Now, one by one, navigate to your servers and manually add the server to the group you have created.

    Select your individual servers page and select the "Add Asset to Group" action.

    Select the Group you want to add to (our example group is "Static Group") and the click then [Add Assets to Group] button.

    Dynamic (Smart) Groups 

    Dynamic (smart) groups are once again much as the label says. An asset(server/OS etc) will become part of the group based on it matching one or many criteria. The criteria is evaluated every time the group is accessed. So if you deploy a new server, create a zone or update any other matched attribute, it will change the group membership. The next time you access the group its membership will be automatically updated to include the current view of the environment. There is a large number of attributes that can be used to make criteria and the criteria can be combined to make complex grouping rules. There is more than enough to discuss on building these rules for another blog, but today, let's just go with a single example to give you the feel for the capabilities of dynamic grouping.

    We will launch the "Create Group" wizard, as we did for the static group, but this time we will give it a more descriptive name and description. Last but not least, we will check the "Configure group rules" check-box, which will make the group we create a dynamic group.

    Rules can be as simple as "hostname" starts with "prod" or as complex as having multiple rules each with multiple criteria matches. This is why I will be going into more details on building these rule sets in another blog in the next few weeks.

    For this example, I have chosen a moderate level of complexity. We have a single rule, but we will only match on any asset that has all 4 of the attributes set.

    • OS Version contains 11 ( I also could have used Version equals 5.11)
    • Has an IP address is on subnet
    • Is a Non Global Zone
    • Is managed by Ops Center (It would be unusual to not be in a managed state, but a Proxy Controller in a NGZ is an example of a non managed asset. )

    Click [Next] to see the preview screen and to check that we matched the assets we want.

    You can see that we have matched on 4 Solaris 11 zones. Now let's see how that looks in the BUI [Navigation==>Assets ==>All User Defined Groups (pull-down)].

    You see we have our static group and our dynamic group we have just created.

    OK, let's create a second group, but this time for managed Non Global zones of the network.

    Create a new name and description.

    Configure our rule, but this time look for Non Global Zones on the network.

    Preview shows no matching hosts, which in this case is correct, as I have not deployed a zone on that network yet. Finish the wizard and now let's look in the BUI to see what we have.

    Checking the [All User Defined Groups] pull-down, we now see our static group and 2 different S11 NGZ groups, one with 4 members and one with no members.  (I was not quite consistent with the group naming, but I could fix that using the [Modify Group action].)

    Now if I go off and deploy a few new zones, we can then see what our smart groups look like. I have deployed 2 zones on the subnet and one more zone on the subnet.

    As you can see, the new zones automatically appear in the correct groups.

    Dynamic (Smart) Groups - Criteria

    There are far too many attributes to go through here ( a few are listed below) and I will be writing a further blog to show you how to use some of the more useful ones.

    Semantic Tags

    But I will call out one special criteria (attribute) that is probably the most useful one of all - the Semantic tag. A Semantic tag is an arbitrary tag or a tag/value pair that can be added to any asset to basically store descriptive information about that asset. You can add a tag to an asset by simply clicking the [Edit Tag] action.

    Examples of Semantic Tags (Key):

     Tag Name
     PRODUCTION  Describes the production status
     SALES  Describes the business unit that owns the server

    Examples of Semantic Tags/Value pairs (Key and Value):

    Tag Name  Value Description
     PRODUCTION_STATUS  DEV/UAT/PROD The value describes the production status DEV/UAT/PROD
     Admin_By The value describes the name/group of the administrator of the system (could even be their email)
     Business_Unit  SALES/ENGINEERING/ACCOUNTS The value describes the business unit that owns the server
     Application  DataBase/Web/ApplicationServer The value describes the application running on the asset

    As you can see, you can create a Semantic Tag to store any information about your server that you require. These tags and tag/value pair scan be used as attributes to create Dynamic groups.

    Configure a group using a "Semantic Tag Key & Value".

    And the group based on the tag/value pair is ready to use.

    Nesting Groups

    One final feature of groups is that you can nest them (have a group that contains 1 or more other groups).

    Create a group as before. This time click the check-box for "Configure subgroups".

    Then you must drag the subgroups you want to include to the "Selected Group" icon.

    Repeat this procedure until you have all your required groups selected.

    Now click [Next], [Next] and [Finish], then check what our new group looks like.

    You can see the S11-Zones group contains both S11 NGZ groups.

    And by highlighting the S11-Zones group, we can see its membership and incident information for all included assets.


    I hope this has given you a good understanding of groups and how they can make your use of Ops Center more productive.



    Tuesday Nov 25, 2014

    Monitoring NFS mounted file systems using EM12c

    A customer recently asked me how they could monitor and alert against all the NFS mounted file systems across their datacenter. Here is a quick guide to do the same.

    Read More

    Friday Oct 24, 2014

    Get Compliant with Oracle Java VM Database PSU OCT 2014 using EM12c

    Check for compliance and automate patching of Oracle Database fleet using EM12c

    Oracle along with its regular Quarterly Database PSU/SPU/CPU update this October 2014 released Oracle JAVA VM PSU patch, the patch is recommended to be applied to all databases in your fleet (esp. the ones that uses JAVA).  (For more information, support Note- 1929745.1 explains it in detail).
    The mandate primarily is to apply the patches against the databases that use JAVAVM option. Ideally, you would need to apply it against all databases, so in case a new database is created in the ORACLE_HOME it is covered.

    Oracle Enterprise Manager 12c provides support features of Compliance Management and automated Patching of databases. Using both these features you could identify the databases in your fleet that needs the patch and automate applying them.  

    To get started, download the starter kit here. It contains utilities to jump start on your compliance, the kit contains a readily importable Compliance Standard and a step-by-step guide.

    [Read More]

    Tuesday Sep 30, 2014

    Using JVM Diagnostics (JVMD) to help tune production JVMs

    Contributing Author: Shiraz Kanga, Consulting Member of Technical Staff, Oracle

    Tuning a production JVM involves more than merely adding more RAM to it via the -Xmx parameter. It depends upon an understanding of how your application truly behaves in a production environment. Most JVM tuning is done by developers with a simulated load in their Development or QA environment. This is unlikely to be truly representative of the production load running on production hardware with regards to proper JVM tuning.

    One of the tools that actually contains real-world production data is JVM Diagnostics (JVMD). Hence it is a good idea to use data collected by JVMD for tuning your production JVMs. Note that JVMD is a component of Oracle Enterprise Manager, licensed as part of both Weblogic Management Pack and the Non-oracle Middleware Management Pack.

    Figure 1. Heap Utilization and Garbage Collections for a specific JVM

    In this document we are primarily addressing the Hotspot JVM. There are several aspects of tuning this JVM that we will look into:

    Tuning Heap Size

    The main parameters needed to tune heap size are:
    • -Xms<n>[g|m|k] is the initial and minimum size of the Java heap
    • -Xmx<n>[g|m|k] is the maximum possible size that the Java heap can grow upto

    Figure 2. Heap after GC and Garbage Collection Overhead for a specific JVM

    The Java Heap size here refers to the total size of the young and the old generation spaces. To start, take a look at the Heap usage chart (Figure 1) of your production JVM under maximum load in the JVMD Performance Diagnostics page. You should see some patterns in the minimum and the maximum heap sizes over time. You can use this data as a rough guide for your choice of -Xms and -Xmx with a reasonable amount of padding. After setting these you should start monitoring the garbage collection charts of your production JVMs (Figure 2) in the JVMD Live Heap Analysis page. It is useful to look into the JVMD metric called "Heap use after GC" which provides a good reflection of the actual amount of heap memory being used by your application. Ideally this metric should remain relatively steady over time with only few full garbage collections occuring. If there are too many full garbage collections then performance of your production application is impacted since GC is done by blocking threads that take a while to scan the entire heap. You can monitor this metric with the JVM GC Overhead% chart on the same page of JVMD. Garbage collection overhead is the percentage of total time spent in garbage collection. Increasing -Xmx can help to make these happen less frequently but actually it is time to dig deeper into your tuning options.

    The key questions that you need to answer are - How frequently does garbage collection take place, How long does each collection take and what is the actual memory used (i.e. heap after GC). Also be sure that you NEVER make the heap size larger than the available free RAM on your system as disk will decrease performance as RAM will start getting swapped to disk.

    The Sun HotSpot JVM relies on generational garbage collection to achieve optimum performance. The -XX:SurvivorRatio command line parameter could further help in tuning garbage collection.

    The Java heap has a young generation for newly created objects and an old generation for long lived objects. The young generation is further subdivided into the Eden space where new objects are allocated and the Survivor space where new objects that are still in use can survive their first few garbage collections before being promoted to old generations. The Survivor Ratio is the ratio of Eden to Survivor space in the young object area of the heap. Increasing this setting optimizes the JVM for applications with high object creation and low object preservation. In applications that generate more medium and long lived objects, this setting should be lowered from the default and vice versa.

    For example, -XX:SurvivorRatio=10 sets the ratio between each survivor space and eden space to be 1:10. If survivor spaces are too small, they will overflow directly into the old generation. If survivor spaces are too large, they will be empty. At each GC, the JVM determines the number of times an object can be copied before it is tenured, called the tenure threshold. This threshold should be set to keep the survivor space half full.

    Most tuning operations represent a trade-off of some type or another. In the case of garbage collection the trade-off usually involves the memory used v/s throughput and latency.
    • The throughput of a JVM is measured in terms of the time spent doing garbage collection vs. the time spent outside of garbage collection (referred to as application time). It is the inverse of GC overhead mentioned above and represents the amount of work done by an application as a ratio of time spent in GC. Throughput can be tuned with -XX:GCTimeRatio=99 where 99 is the default which represents a 1% GC overhead.
    • Latency is the amount of time delay that is caused by garbage collection. Latency for GC pauses can be tuned by specifying rhe maximum pause time goal with the command line option -XX:MaxGCPauseMillis=<N>. This is interpreted as a hint that pause times of <N> milliseconds or less are desired. By default, there is no maximum pause time goal. If a pause time goal is specified, the heap size and other garbage collection related parameters are adjusted in an attempt to keep garbage collection pauses shorter than the specified value. Note that these adjustments may cause the garbage collector to reduce the overall throughput of the application and in some cases the desired pause time goal cannot be met.
    Some lesser-known options are about permanent generation space which is used by the JVM itself to hold metadata, classes structures and so on:
    • -XX:PermSize=<n>[g|m|k] is the initial and minimum size of the permanent generation space.
    • -XX:MaxPermSize=<n>[g|m|k] is the maximum size of the permanent generation space. If you ever get the message java.lang.OutOfMemoryError: PermGen space then it means that your application is loading a very large number of classes and this should be raised.
    • -Xss=<n>[g|m|k]is the size of the thread stack. Each thread in a Java application has its own stack. The stack is used to hold return addresses, arguments to functions and method calls, and so on. The default stack size setting for a thread in Java is 1MB. In a highly multi-threaded system, like an application server at any given point in time there are multiple thread pools and threads that are in use so this may need to be reduced. Since stack size has to be allocated in contiguous blocks and if the machine is being used actively and there are many threads running in the system you may encounter an OutOfMemory error even when you have sufficient heap space. Recursive code can quickly exhaust the stack and if you use such code then you may need to increase the -Xss setting. However, if you see java.lang.OutOfMemoryError: unable to create new native thread then you may have too many threads, or each thread has a large stack; so you may need to decrease it.

    Tuning Garbage Collection Algorithm

    Garbage collection is expensive. Generational garbage collectors have the JVM  memory divided into several spaces.
    • Eden space: All objects are placed here when first created
    • Survivor spaces: One or more regions where objects mature
    • Tenured space: Where long lived objects are stored
    • Permanent generation: This area is only used by the JVM itself to hold metadata, such as data structures for classes, methods, interned strings
    One thing that people often forget to try, is to lower the amount of garbage being created in the first place. There are a lot of ways to do this which are specific to the application/code that is being written. This often involves techniques such as using StringBuilder/StringBuffer instead of Strings, lowering the amount of logging, etc.

    There are several GC algorithms which are available to be used in a Java VM. The following command line options allow to use a specific GC algorithm:
    • -XX:+UseSerialGC uses a single threaded, young generation, and old generation garbage collector (Normally this is a poor choice and should be used only for small Java heap sizes such as -Xmx256m or smaller)
    • -XX:+UseParallelGC utilizes a multithreaded (parallel) garbage collector for the young generation and a single-threaded garbage collector for the old generation space in parallel.
    • -XX:+UseParallelOldGC uses a multithread garbage collector for both the young and old generations.
    • -XX:+UseParNewGC -> enables a multithreaded, young generation garbage collector
    • -XX:+UseConcMarkSweepGC -> enables the VM’s mostly concurrent garbage collector. It also auto-enables -XX:+UseParNewGC (use if If you are not able to meet your application’s worst case latency requirements due to full garbage collection duration being too long)
    • -XX:+UseG1GC -> garbage first collector (default in java 7, can be also used in latest releases of Java 6)
    In practice, the default in Java 6 is ParallelGC and in Java 7 it is the G1GC. Changing the algorithm requires detailed analysis of the application behavior. If you see a nice regular sawtooth chart in the heap usage you may not need any changes at all. If not, we recommend trying out each GC algorithm under a realistic load and then comparing it to the default algorithm's behavior under the same load. Usually you will find that the default algorithm outperforms the new setting and that there is no reason to change it.

    As you can see, tuning the JVM and it's garbage collectors is largely a trade-off between space and time. If you had infinite heap space then you would never need to collect garbage. Inversely, if you could tolerate infinite time delays, then you could run a cleanup as frequently as you like and keep the heap compact. Clearly, both those situations are impossible. Finding the right middle ground that is right for you requires careful balancing act based on understanding how GC works and what the application requires.


    Java SE 6 HotSpot Virtual Machine Garbage Collection Tuning

    Summary of Enterprise Manager general session at Oracle OpenWorld 2014

    Oracle OpenWorld is at full speed now and this morning we just delivered our Enterprise Manager general session. We realize that probably the majority of the thousands of readers of our blog are not attending OpenWorld so that's why we have created a summary of the Tweets that were sent during the session
    (in case you are not familiar with Twitter timelines, please read from the bottom to the top for a chronological order of the presentation) 

    For those that want to see the full presentation, we've uploaded them to our SlideShare account for your convenience .

    For all Enterprise Manager related sessions at Oracle OpenWorld 2014, please take look here and let us know which one you'd like to see summarized or even see the full deck?

    Wednesday Sep 17, 2014

    Stay Connected at Oracle OpenWorld

    Follow Oracle Enterprise Manager 12c @ Oracle OpenWorld

    If you are planning to attend OpenWorld this year, make sure you maximize the full conference experience—stay connected to all the Oracle Enterprise Manager 12c activities.

    For the complete list of OpenWorld events and activities, read the Oracle Enterprise Manager 12c Focus on Doc.

    Stay Connected:
    Twitter |
    Facebook | YouTube | Linkedin | Newsletter
    Download the Oracle Enterprise Manager 12c Mobile app

    Thursday Sep 11, 2014

    Simplify deployment of JVMD Agents to command line Java applications

    Contributing Author: Shiraz Kanga, Consulting Member of Technical Staff, Oracle

    Most customers of Oracle Enterprise Manager using JVM Diagnostics use the tool to monitor their Java Applications servers like Weblogic, Websphere, Tomcat, etc. In this environment it is fairly easy to deploy the JVMD Agent. Since it is distributed as a war file, you merely deploy the agent into a running application server using the management GUI or command line tools. Then you can start monitoring with no need for a restart of the app server or for the modification of any startup commands or scripts. However, with other types of Java applications that do not allow for any code deployment at runtime such as AWT/Swing or command line java applications these steps are necessary. Modifying startup scripts is complex because each application comes with its own custom and unique launch script. Additionally, the command that actually launches the runtime needs to have the java command with its related parameters (like -Xmx) the JVMD Agent with its own parameters (like console host/port) and the application itself which may have some more custom parameters. People often get confused due to the complexity that is seen here.

    I've recently had customers that needed to monitor Hadoop, HDFS, Zookeeper, Kafka, Cassandra and Solr with JVMD. In order to simplify some of the complexity discussed above, I created a simple script based framework that makes things a bit easier. Feel free to use my approach to quickly setup JVMD with these or any other command line java programs. You can also use it as the basis for your own modifications. The framework modifies the startup scripts supplied with these tools in order to add the JVMD agent. All the code/scripts are attached in a zip file. Both original and modified versions of all changed scripts are included so you can easily see the modifications I made with a simple diff.

    Here's how these scripts are setup. Everything is configured using 4 environment variables as shown below:

        export JVMD_AGENT_HOME=/home/skanga/servers
        export JVMD_MANAGER_PORT=3800
        export JVMD_UNIQUE_ID=<unique name for each server process>

    where the JVMD_AGENT_HOME must contain the (from the attached zip file) and jamagent.war (which can be downloaded from your JVMD console). The first three of these are likely to remain unchanged for all the JVMs being monitored so you can easily add them directly into if needed.

    The JVMD_UNIQUE_ID will always be unique so it must not be placed there. However it has two other modes where you can use a pointer to the unique ID instead of specifying it directly. You can point to either an environment variable or to a JVM system property that holds the actual unique ID. If you are using these cases then you could add this one to the script too.

    If JVMD_UNIQUE_ID starts with the string "sysprop-" then the actual unique ID will be read from the JVM system property named by the string following "sysprop-". For example if JVMD_UNIQUE_ID is "sysprop-server_name" and we have a system property -Dserver_name=MyTestingServer then JVMD will use MyTestingServer as the JVM unique identifier.

    If JVMD_UNIQUE_ID starts with the string "envvar-" then the actual unique ID will be read from the environment variable named by the string following "envvar-". For example if JVMD_UNIQUE_ID is "envvar-server_name" and we have an environment variable called server_name=MyTestingServer then JVMD will use MyTestingServer as the JVM unique identifier.

    Caution: Do not use dash (minus) character in the environment variable setup of unique id. Use underscore instead.

    Generic Launch Script Modifications

    After these four environment variables are set we need to modify our launch scripts. Make sure you have a backup of all files before you proceed. In the main script that you use to launch your java application look for a line that has a format that is similar to the one below: 
    and replace it with

    So we simply added a $JVMD_AGENT_INSERT just before the name of the Main class. If there are multiple such lines then you should modify them all in the same way. And in order to configure $JVMD_AGENT_INSERT we also need to source (with some error checking). So we insert a snippet like this in the line just before the JAVA invocation. 

    # add JVMD Agent Env settings
    [[ -e "${JVMD_AGENT_HOME}/" ]] 
    && source "${JVMD_AGENT_HOME}/" ||
    { echo "ERROR: JVMD_AGENT_HOME undefined or does not contain" 1>&2 ; exit 1; } 

    NOTE: Everything after the comment above should in a single line of code in your launch script. This line gets mangled by the blogging software so it is best to cut & paste it from it from one of the scripts in the attached zip file.

    We will now look at how I used these techniques to add JVMD monitoring to Kafka, Hadoop, Zookeeper, Cassandra and Solr. 

    1) Kafka 2.8.0-

    I used Kafka 2.8.0- and downloaded it directly from the Kafka site. In Kafka, ALL processes are initiated through a common launcher called in the bin folder. All the other shell scripts (including the built-in Zookeeper) call this one. So this single insertion point is the only place that we will need to modify in order to add JVMD monitoring to Kafka. Pretty simple. Using the modified script (inside the attached zip file) you can run the servers as shown below:

    TEST - with mods to use JVMD
    cd /home/skanga/servers/kafka_2.8.0-
    export JVMD_AGENT_HOME=/home/skanga/servers
    export JVMD_MANAGER_PORT=3800
    # start a zookeeper server
    export JVMD_UNIQUE_ID=zookeeper-server
    ./ ../config/
    # start a kafka server
    export JVMD_UNIQUE_ID=kafka-server
    ./ ../config/

    2) Hadoop 2.4.1

    The scripts called hadoop, hfds, mapred and yarn in the hadoop bin directory will ALL need to be modified for JVMD monitoring. Using the modified scripts (inside the attached zip file) you can run all the servers as shown below:

    TEST - with mods for hadoop command to use JVMD

    cd /home/skanga/servers/hadoop-2.4.1
    export JVMD_AGENT_HOME=/home/skanga/servers
    export JVMD_MANAGER_PORT=3802
    # Launch the hdfs nfs gateway
    export JVMD_UNIQUE_ID=hdfs-nfs3-gateway
    ./bin/hdfs nfs3
    # Run a mapreduce history server
    export JVMD_UNIQUE_ID=mapred-historyserver
    ./bin/mapred historyserver
    # Run a yarn resource manager
    export JVMD_UNIQUE_ID=yarn-resourcemanager
    ./bin/yarn resourcemanager
    # Run a hadoop map-reduce job to find the value of PI (QuasiMonteCarlo method)
    export JVMD_UNIQUE_ID=hadoop-test-pi-montecarlo
    ./bin/hadoop jar ./share/hadoop/mapreduce/hadoop-mapreduce-examples-2.4.1.jar pi 1024 100

    3) Zookeeper 3.4.6

    The standalone version of zookeeper has a common environment setup script called where most JVMD setup can be done. After that a minor modification is needed in the java launch command in after which all JVMD monitoring works fine. The scripts called and probably do not need monitoring but can be easily added if really needed.

    TEST - with mods for command to use JVMD

    cd /home/skanga/servers/zookeeper-3.4.6/bin
    export JVMD_AGENT_HOME=/home/skanga/servers
    export JVMD_MANAGER_PORT=3800
    export JVMD_UNIQUE_ID=zk-server
    # start the zookeeper server
    ./ start
    ./ status
    ./ stop

    4) Cassandra 2.0.9

    The Apache Cassandra data store has a common environment setup script called conf/ where we can add the command to source our include script. Then a minor modification is needed to the java launch command in bin/cassandra after which all JVMD monitoring works fine. The other scripts probably do not need monitoring but can be easily added if really needed. 

    TEST - with mods for cassandra command to use JVMD

    cd /home/skanga/servers/apache-cassandra-2.0.9/bin
    export JVMD_AGENT_HOME=/home/skanga/servers
    export JVMD_MANAGER_PORT=3800
    export JVMD_UNIQUE_ID=cassandra-server
    # start cassandra
    ./cassandra -f

    5) Solr 4.9.0

    The Solr search server is an interesting case. In production scenarios, users will probably use the Solr war file in their own application server. In this scenario the standard JVMD warfile can be deployed to the same application server and monitored easily. However, the Solr distribution also include an embedded mode which may be used by simply running java -jar start.jar and for this scenario we have converted this java command into a simple script called and added it to the same folder as start.jar in order to run it. Using this script (inside the attached zip file) you can run a test as shown below:

    TEST - with addition of command to use JVMD with Solr

    cd /home/skanga/servers/solr-4.9.0/example
    export JVMD_AGENT_HOME=/home/skanga/servers
    export JVMD_MANAGER_PORT=3800
    export JVMD_UNIQUE_ID=solr-server
    # start solr

    After everything is setup properly for your servers you should see all the relevant JVMs in the default pool with the proper ID as shown in the image below.

    JVMs in Default Pool (with hostnames & ip addresses blanked out)
    Click image to expand it in a new tab

    Remember to be a bit patient and wait a few seconds until the connections are established and the servers appear in the console.


    Latest information and perspectives on Oracle Enterprise Manager.

    Related Blogs


    « October 2015