Hi all,
So, this week we’ll have an Autonomous Health Framework series o posts! One post per day, stay tuned!
Today I’ll review some points of the Autonomous Health Framework and how it can help us on the day to day tasks as well as on specific problem resolutions. This framework is a bundle of tools that were already available like orachk/exachk, tfa plus other features that I will try to cover focusing on practical cases.
So, first of all, you need to install it, if you have a RAC cluster most likely you already have it but it is always good to update it, as Oracle security recommendations as also part of the orachk analyze. For reference: In an ideal world you have root access or sudo, as its daemons under root provides a more wide log collection capabilities.
- Download the latest version from Autonomous Health Framework (AHF) – Including TFA and ORAchk/EXAChk (Doc ID 2550798.1)
- Transfer and unzip the file to your server and as root follow the installation as below
root@servertst01 oracrs]# ./ahf_setup -ahf_loc /oraadm/oracrs/product/19.0.0/ahf -data_dir /oraadm/oracrs/product/19.0.0/ahf/data AHF Installer for Platform Linux Architecture x86_64 AHF Installation Log : /tmp/ahf_install_120962_2020_05_07-22_41_11.log Starting Autonomous Health Framework (AHF) Installation AHF Version: 20.1.3 Build Date: 202004290950 TFA is already installed at : /oraadm/oracrs/product/19.0.0/tfa/servertst01/tfa_home Installed TFA Version : 192000 Build ID : 20190426041420 AHF Location : /oraadm/oracrs/product/19.0.0/ahf/oracle.ahf AHF Data Directory : /oraadm/oracrs/product/19.0.0/ahf/oracle.ahf/data Shutting down TFA : /oraadm/oracrs/product/19.0.0/tfa/servertst01/tfa_home Copying TFA Data Files from /oraadm/oracrs/product/19.0.0/tfa/servertst01/tfa_home Uninstalling TFA : /oraadm/oracrs/product/19.0.0/tfa/servertst01/tfa_home Do you want to add AHF Notification Email IDs ? [Y]|N : n Login using root is disabled in sshd config. Installing AHF only on Local Node Extracting AHF to /oraadm/oracrs/product/19.0.0/ahf/oracle.ahf Configuring TFA Services Copying TFA Data Files to AHF Discovering Nodes and Oracle Resources Starting TFA Services Created symlink from /etc/systemd/system/multi-user.target.wants/oracle-tfa.service to /etc/systemd/system/oracle-tfa.service. Created symlink from /etc/systemd/system/graphical.target.wants/oracle-tfa.service to /etc/systemd/system/oracle-tfa.service. .---------------------------------------------------------------------------------. | Host | Status of TFA | PID | Port | Version | Build ID | +-------------+---------------+--------+------+------------+----------------------+ | servertst01 | RUNNING | 135195 | 5000 | 20.1.3.0.0 | 20130020200429095054 | '-------------+---------------+--------+------+------------+----------------------' Running TFA Inventory... Adding default users to TFA Access list... .--------------------------------------------------------------------------------------. | Summary of AHF Configuration | +-----------------+--------------------------------------------------------------------+ | Parameter | Value | +-----------------+--------------------------------------------------------------------+ | AHF Location | /oraadm/oracrs/product/19.0.0/ahf/oracle.ahf | | TFA Location | /oraadm/oracrs/product/19.0.0/ahf/oracle.ahf/tfa | | Orachk Location | /oraadm/oracrs/product/19.0.0/ahf/oracle.ahf/orachk | | Data Directory | /oraadm/oracrs/product/19.0.0/ahf/oracle.ahf/data | | Repository | /oraadm/oracrs/product/19.0.0/ahf/oracle.ahf/data/repository | | Diag Directory | /oraadm/oracrs/product/19.0.0/ahf/oracle.ahf/data/servertst01/diag | '-----------------+--------------------------------------------------------------------' Starting orachk daemon from AHF ... AHF binaries are available in /oraadm/oracrs/product/19.0.0/ahf/oracle.ahf/bin AHF is successfully installed Moving /tmp/ahf_install_120962_2020_05_07-22_41_11.log to /oraadm/oracrs/product/19.0.0/ahf/oracle.ahf/data/servertst01/diag/ahf/ As you can see as the root user is not allowed to ssh, the installation was done only on the local node. Repeat the process on the remaning ones Once you have it completed you will be able to to a tfactl status and a toolstatus and see the processes and tools available. [root@servertst01 bin]# ./tfactl status .----------------------------------------------------------------------------------------------------. | Host | Status of TFA | PID | Port | Version | Build ID | Inventory Status | +-------------+---------------+--------+------+------------+----------------------+------------------+ | servertst01 | RUNNING | 135195 | 5000 | 20.1.3.0.0 | 20130020200429095054 | COMPLETE | | servertst02 | RUNNING | 241723 | 5000 | 20.1.3.0.0 | 20130020200429095054 | COMPLETE | '-------------+---------------+--------+------+------------+----------------------+------------------' [root@servertst01 bin]# [root@servertst01 bin]# ./tfactl toolstatus .------------------------------------------------------------------. | TOOLS STATUS - HOST : servertst01 | +----------------------+--------------+--------------+-------------+ | Tool Type | Tool | Version | Status | +----------------------+--------------+--------------+-------------+ | Development Tools | orachk | 19.3.0.0.0 | DEPLOYED | | | oratop | 14.1.2 | DEPLOYED | +----------------------+--------------+--------------+-------------+ | Support Tools Bundle | darda | 2.10.0.R6036 | DEPLOYED | | | oswbb | 8.3.2 | RUNNING | | | prw | 12.1.13.11.4 | NOT RUNNING | +----------------------+--------------+--------------+-------------+ | TFA Utilities | alertsummary | 19.3.0.0.0 | DEPLOYED | | | calog | 19.3.0.0.0 | DEPLOYED | | | dbcheck | 18.3.0.0.0 | DEPLOYED | | | dbglevel | 19.3.0.0.0 | DEPLOYED | | | grep | 19.3.0.0.0 | DEPLOYED | | | history | 19.3.0.0.0 | DEPLOYED | | | ls | 19.3.0.0.0 | DEPLOYED | | | managelogs | 19.3.0.0.0 | DEPLOYED | | | menu | 19.3.0.0.0 | DEPLOYED | | | param | 19.3.0.0.0 | DEPLOYED | | | ps | 19.3.0.0.0 | DEPLOYED | | | pstack | 19.3.0.0.0 | DEPLOYED | | | summary | 19.3.0.0.0 | DEPLOYED | | | tail | 19.3.0.0.0 | DEPLOYED | | | triage | 19.3.0.0.0 | DEPLOYED | | | vi | 19.3.0.0.0 | DEPLOYED | '----------------------+--------------+--------------+-------------' Note :- DEPLOYED : Installed and Available - To be configured or run interactively. NOT RUNNING : Configured and Available - Currently turned off interactively. RUNNING : Configured and Available. [root@servertst01 bin]#
3. Using it you can ‘query’ all cluster nodes for information from a single point which can really speed up your information collection. Here is an example of the summary command:
[root@servertst01 bin]# ./tfactl summary -h --------------------------------------------------------------------------------- Usage : TFACTL [run] summary -help --------------------------------------------------------------------------------- Command : /oraptp/oracrs/product/19.0.0/ahf/oracle.ahf/tfa/bin/tfactl [run] summary [OPTIONS] Following Options are supported: [no_components] : [Default] Complete Summary Collection -overview : [Optional/Default] Complete Summary Collection - Overview -crs : [Optional/Default] CRS Status Summary -asm : [Optional/Default] ASM Status Summary -acfs : [Optional/Default] ACFS Status Summary -database : [Optional/Default] DATABASE Status Summary -exadata : [Optional/Default] EXADATA Status Summary Not enabled/ignored in Windows and Non-Exadata machine -patch : [Optional/Default] Patch Details -listener : [Optional/Default] LISTENER Status Summary -network : [Optional/Default] NETWORK Status Summary -os : [Optional/Default] OS Status Summary -tfa : [Optional/Default] TFA Status Summary -summary : [Optional/Default] Summary Tool Metadata -json : [Optional] - Prepare json report -html : [Optional] - Prepare html report -print : [Optional] - Display [html or json] Report at Console -silent : [Optional] - Interactive console by defauly -history <num> : [Optional] - View Previous <numberof> Summary Collection History in Interpreter -node <node(s)> : [Optional] - local or Comma Separated Node Name(s) -help : Usage/Help. --------------------------------------------------------------------------------- - Data Collection From Node - servertst02 .. Done. [root@servertst01 bin]# [root@servertst01 bin]# ./tfactl summary Executing Summary in Parallel on Following Nodes: Node : servertst01 Node : servertst02 LOGFILE LOCATION : /oraadm/oracrs/product/19.0.0/ahf/oracle.ahf/data/repository/suptools/servertst01/summary/root/20200508124309/log/summary_command_20200508124309_servertst01_35756.log Component Specific Summary collection : - Collecting CRS details ... Done. - Collecting ASM details ... Done. - Collecting ACFS details ... Done. - Collecting DATABASE details ... Done. - Collecting PATCH details ... Done. - Collecting LISTENER details ... Done. - Collecting NETWORK details ... Done. - Collecting OS details ... Done. - Collecting TFA details ... Done. - Collecting SUMMARY details ... Done. Remote Summary Data Collection : In-Progress - Please wait ... - Data Collection From Node - servertst02 .. Done. Prepare Clusterwide Summary Overview ... Done cluster_status_summary COMPONENT STATUS DETAILS +-----------+---------+---------------------------------------------------------------------------------------------------+ CRS PROBLEM .-----------------------------------------------. | CRS_SERVER_STATUS : ONLINE | | CRS_STATE : ONLINE | | CRS_INTEGRITY_CHECK : FAIL | | CRS_RESOURCE_STATUS : OFFLINE Resources Found | '-----------------------------------------------' ASM PROBLEM .-------------------------------------------------------. | ASM_DISK_SIZE_STATUS : WARNING - Available Size < 20% | | ASM_BLOCK_STATUS : PASS | | ASM_CHAIN_STATUS : PASS | | ASM_INCIDENTS : PASS | | ASM_PROBLEMS : PASS | '-------------------------------------------------------' ACFS OFFLINE .-----------------------. | ACFS_STATUS : OFFLINE | '-----------------------' DATABASE PROBLEM .-----------------------------------------------------------------------------------------------. | ORACLE_HOME_DETAILS | ORACLE_HOME_NAME | +----------------------------------------------------------------------------+------------------+ | .------------------------------------------------------------------------. | OraDb11g_home1 | | | PROBLEMS | INCIDENTS | DB_BLOCKS | DATABASE_NAME | STATUS | DB_CHAINS | | | | +----------+-----------+-----------+---------------+---------+-----------+ | | | | PROBLEM | PROBLEM | PASS | FSIGNTST | PROBLEM | PROBLEM | | | | '----------+-----------+-----------+---------------+---------+-----------' | | '----------------------------------------------------------------------------+------------------' PATCH OK .----------------------------------------------. | CRS_PATCH_CONSISTENCY_ACROSS_NODES : OK | | DATABASE_PATCH_CONSISTENCY_ACROSS_NODES : OK | '----------------------------------------------' LISTENER OK .-----------------------. | LISTNER_STATUS : OK | '-----------------------' NETWORK OK .---------------------------. | CLUSTER_NETWORK_STATUS : | '---------------------------' OS OK .-----------------------. | MEM_USAGE_STATUS : OK | '-----------------------' TFA OK .----------------------. | TFA_STATUS : RUNNING | '----------------------' SUMMARY OK .-----------------------------------. | SUMMARY_EXECUTION_TIME : 0H:2M:9S | '-----------------------------------' +-----------+---------+---------------------------------------------------------------------------------------------------+ ### Entering in to SUMMARY Command-Line Interface ### tfactl_summary>list Components : Select Component - select [component_number|component_name] 1 => overview 2 => crs_overview 3 => asm_overview 4 => acfs_overview 5 => database_overview 6 => patch_overview 7 => listener_overview 8 => network_overview 9 => os_overview 10 => tfa_overview 11 => summary_overview
See you tomorrow at the next post o this series!
Elisson Almeida