Solución de problemas de lentitud con el tráfico, gestión
Environment
- Palo Alto Firewall .
- Cualquier PAN-OS archivo .
- Solución de problemas general.
Resolution
A continuación se presentan algunos comandos (con una breve descripción) que pueden ser útiles para solucionar problemas de administración o problemas relacionados con el tráfico. Los problemas pueden variar de naturaleza persistente a intermitente o esporádica.
NOTE: Este documento es una directriz general y no debe tomarse como el diagnóstico final de la cuestión.
mostrar la información del sistema - Este comando nos proporcionará una instantánea del PAN-OS modelo, , actualizaciones dinámicas (aplicación, amenazas, AV , , ) WF URL versiones, entre otras cosas. El'tiempo de actividad'mencionado aquí se refiere al tiempo de actividad del plan de datos. Esto se restablecerá si se ha reiniciado el plano de datos o todo el dispositivo.
admin@anuragFW> show system info
hostname: anuragFW
ip-address: 10.21.56.125
netmask: 255.255.254.0
default-gateway: 10.21.56.1
ip-assignment: static
ipv6-address: unknown
ipv6-link-local-address: fe80::20c:29ff:fecb:50ab/64
ipv6-default-gateway:
mac-address: 00:0c:29:cb:50:ab
time: Mon May 29 23:18:55 2017
uptime: 18 days, 22:00:54
family: vm
model: PA-VM
serial: 007000006243
vm-mac-base: 00:1B:17:F4:9B:00
vm-mac-count: 256
vm-uuid: 564DF4C3-7C1D-36AC-B58E-37AE79CB50AB
vm-cpuid: E4060300FFFBAB1F
vm-license: VM-100
vm-mode: VMWare ESXi
sw-version: 8.0.2
global-protect-client-package-version: 4.0.0
app-version: 703-4048
app-release-date: 2017/05/25 21:17:50
av-version: 2256-2743
av-release-date: 2017/05/27 04:01:19
threat-version: 703-4048
threat-release-date: 2017/05/25 21:17:50
wf-private-version: 0
wf-private-release-date: unknown
url-db: paloaltonetworks
wildfire-version: 144067-145602
wildfire-release-date: 2017/05/29 10:22:50
url-filtering-version: 20170529.20212
global-protect-datafile-version: unknown
global-protect-datafile-release-date: unknown
global-protect-clientless-vpn-version: 63-78
global-protect-clientless-vpn-release-date: 2017/05/15 13:47:50
logdb-version: 8.0.15
platform-family: vm
vpn-disable-mode: off
multi-vsys: off
operational-mode: normal
mostrar los <follow></follow> recursos del sistema - Este comando proporciona el uso en tiempo real del uso de la CPU administración. El'up'mencionado aquí se refiere al tiempo de actividad del plano de administración. Este comando también se puede utilizar para buscar el uso de memoria e intercambiar el uso si existe. Idealmente, el uso de la memoria del intercambio no debe ser demasiado o degradar, que indicaría la pérdida de memoria o simplemente demasiada carga. Este comando sigue el mismo formato que ejecutar el comando ' Top ' en máquinas Linux. Si las confirmaciones tardan demasiado (más tiempo que una "línea base" establecida), la alta administración CPU puede ser una de las causas.
admin@anuragFW> show system resources follow
top - 23:20:35 up 18 days, 22:02, 1 user, load average: 0.24, 0.64, 0.56
Tasks: 130 total, 2 running, 128 sleeping, 0 stopped, 0 zombie
Cpu(s): 2.9%us, 5.4%sy, 0.7%ni, 90.7%id, 0.0%wa, 0.0%hi, 0.2%si, 0.0%st
Mem: 6577072k total, 6235716k used, 341356k free, 112720k buffers
Swap: 0k total, 0k used, 0k free, 3751268k cached
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
30823 root 30 10 23464 1388 1008 S 3.9 0.0 0:00.02 top
30822 admin 20 0 23464 1416 1016 R 2.0 0.0 0:00.02 top
1 root 20 0 16548 148 0 S 0.0 0.0 0:22.39 init
2 root 20 0 0 0 0 S 0.0 0.0 0:00.02 kthreadd
3 root 20 0 0 0 0 S 0.0 0.0 5:39.32 ksoftirqd/0
5 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 kworker/0:0H
7 root RT 0 0 0 0 S 0.0 0.0 0:02.44 migration/0
8 root 20 0 0 0 0 S 0.0 0.0 0:00.00 rcu_bh
9 root 20 0 0 0 0 S 0.0 0.0 2:54.19 rcu_sched
10 root RT 0 0 0 0 S 0.0 0.0 0:00.02 migration/1
11 root 20 0 0 0 0 S 0.0 0.0 0:57.69 ksoftirqd/1
12 root 20 0 0 0 0 S 0.0 0.0 1:26.35 kworker/1:0
13 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 kworker/1:0H
14 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 khelper
230 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 writeback
233 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 bioset
234 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 kblockd
451 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 xenbus_frontend
457 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 ata_sff
464 root 20 0 0 0 0 S 0.0 0.0 0:00.00 khubd
473 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 md
476 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 edac-poller
576 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 rpciod
578 root 20 0 0 0 0 S 0.0 0.0 3:25.40 kworker/0:1
589 root 20 0 0 0 0 S 0.0 0.0 0:16.02 kswapd0
590 root 39 19 0 0 0 S 0.0 0.0 0:30.91 khugepaged
591 root 20 0 0 0 0 S 0.0 0.0 0:00.00 fsnotify_mark
592 root 1 -19 0 0 0 S 0.0 0.0 0:00.00 nfsiod
738 root 20 0 0 0 0 S 0.0 0.0 0:00.01 scsi_eh_0
741 root 20 0 0 0 0 S 0.0 0.0 0:00.00 scsi_eh_1
759 root 20 0 0 0 0 S 0.0 0.0 0:00.00 kworker/0:2
766 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 mpt_poll_0
767 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 mpt/0
773 root 20 0 0 0 0 S 0.0 0.0 0:00.00 kworker/1:2
776 root 20 0 0 0 0 S 0.0 0.0 0:00.00 scsi_eh_2
802 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 kpsmoused
830 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 deferwq
845 root 0 -20 0 0 0 S 0.0 0.0 0:00.47 kworker/1:1H
846 root 0 -20 0 0 0 S 0.0 0.0 0:20.87 kworker/0:1H
847 root 20 0 0 0 0 S 0.0 0.0 1:15.67 kjournald
902 root 16 -4 18992 424 8 S 0.0 0.0 0:00.53 udevd
mostrar información de la sesión - Este comando proporciona información sobre los parámetros de sesión establecidos junto con los contadores para la velocidad de paquetes, nuevas conexiones, etc. Los parámetros de sesión incluyen, pero no limitado a, el total y el número actual de sesiones, tiempos de espera, configuración. La primera sección de la salida es dinámica, lo que significa que produciría salidas diferentes en cada ejecución de este comando.
admin@anuragFW> show session info
target-dp: *.dp0
--------------------------------------------------------------------------------
Number of sessions supported: 256000
Number of allocated sessions: 176
Number of active TCP sessions: 166
Number of active UDP sessions: 10
Number of active ICMP sessions: 0
Number of active GTPc sessions: 0
Number of active GTPu sessions: 0
Number of pending GTPu sessions: 0
Number of active BCAST sessions: 0
Number of active MCAST sessions: 0
Number of active predict sessions: 0
Session table utilization: 0%
Number of sessions created since bootup: 374948
Packet rate: 12/s
Throughput: 8 kbps
New connection establish rate: 0 cps
--------------------------------------------------------------------------------
Session timeout
TCP default timeout: 3600 secs
TCP session timeout before SYN-ACK received: 5 secs
TCP session timeout before 3-way handshaking: 10 secs
TCP half-closed session timeout: 120 secs
TCP session timeout in TIME_WAIT: 15 secs
TCP session timeout for unverified RST: 30 secs
UDP default timeout: 30 secs
ICMP default timeout: 6 secs
other IP default timeout: 30 secs
Captive Portal session timeout: 30 secs
Session timeout in discard state:
TCP: 90 secs, UDP: 60 secs, other IP protocols: 60 secs
--------------------------------------------------------------------------------
Session accelerated aging: True
Accelerated aging threshold: 80% of utilization
Scaling factor: 2 X
--------------------------------------------------------------------------------
Session setup
TCP - reject non-SYN first packet: True
Hardware session offloading: True
IPv6 firewalling: True
Strict TCP/IP checksum: True
ICMP Unreachable Packet Rate: 200 pps
--------------------------------------------------------------------------------
Application trickling scan parameters:
Timeout to determine application trickling: 10 secs
Resource utilization threshold to start scan: 80%
Scan scaling factor over regular aging: 8
--------------------------------------------------------------------------------
Session behavior when resource limit is reached: drop
--------------------------------------------------------------------------------
Pcap token bucket rate : 10485760
--------------------------------------------------------------------------------
Max pending queued mcast packets per session : 0
--------------------------------------------------------------------------------
mostrar sesión de estadísticas del sistema - Este comando muestra valores en tiempo real para el recuento de sesiones activas, rendimiento, velocidad de paquetes y tiempo de actividad (plano de datos) (tiempo de actividad del plan de datos). Esta ventana de salida se actualizará cada pocos segundos para actualizar los valores mostrados.
admin@anuragFW> show system statistics session
System Statistics: ('q' to quit, 'h' for help)
Device is up : 18 days 22 hours 3 mins 54 sec
Packet rate : 32/s
Throughput : 136 Kbps
Total active sessions : 166
Active TCP sessions : 157
Active UDP sessions : 9
Active ICMP sessions : 0
You can type the following key to switch what to display
--------------------------------------------------------
'a' - Display application statistics
'h' - Display this help page
'q' - Quit this program
's' - Display system statistics
estadísticas del grupo de planes de datos de depuración: la salida de este comando se ha cambiado significativamente con respecto a las versiones anteriores. Ahora muestra los búferes de paquetes, los grupos de recursos y los usos de caché de memoria por diferentes procesos. Si los grupos se agotan, el rendimiento del tráfico se verá afectado correspondiente a ese grupo de recursos en particular. En cuanto a los grupos, el número de la izquierda muestra el resto, mientras que el número de la derecha muestra la capacidad total. La capacidad total puede variar en función de plataformas, modelos y OS versiones. Del mismo modo, si un determinado proceso utiliza demasiada memoria, también puede causar problemas relacionados con ese proceso.
admin@anuragFW> debug dataplane pool statistics
Pow Atomic Memory Pools
[ 0] Work Queue Entries : 98300/98304 0xe028378340
[ 1] Packet Buffers : 38474/38912 0xc000a61780
Software Pools
[ 0] Shared Pool 24 ( 24): 659564/660000 0xe000467500
[ 1] Shared Pool 32 ( 32): 659521/660000 0xe001607200
[ 2] Shared Pool 40 ( 40): 169989/170000 0xe002cb0000
[ 3] Shared Pool 192 ( 192): 1253777/1255000 0xe0033d2480
[ 4] Shared Pool 256 ( 256): 139968/140000 0xe011e68180
[ 5] software packet buffer 0 ( 512): 16384/16384 0xe02a038900
[ 6] software packet buffer 1 ( 1024): 16384/16384 0xe02a848a80
[ 7] software packet buffer 2 ( 2048): 32768/32768 0xe02b858c00
[ 8] software packet buffer 3 (33280): 8192/8192 0xe02f878d80
[ 9] software packet buffer 4 (66048): 304/304 0xe03fc80f00
[10] CTD AV Block ( 1024): 32/32 0xe0d46fa400
[11] Regex Results (11544): 8000/8000 0xe0d4e80200 2
[12] SSH Handshake State ( 6512): 16/16 0xe0eee36b80
[13] SSH State ( 3200): 128/128 0xe0eee50480
[14] TCP host connections ( 176): 15/16 0xe0eeeb4d00
[15] DFA Result ( 1024): 1024/1024 0xe0f13a6180
[16] GTP Context Entries ( 256): 85336/85336 0xe0f1a3e400
Shared Pools Statistics
Current local reuse cache counts for each pool:
core 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
24 0 384 0 0 0 0 0 0 0 0 0 0 0 0 0 0
32 0 384 0 0 0 0 0 0 0 0 0 0 0 0 0 0
40 0 11 0 0 0 0 0 0 0 0 0 0 0 0 0 0
192 0 765 0 0 0 0 0 0 0 0 0 0 0 0 0 0
256 0 32 0 0 0 0 0 0 0 0 0 0 0 0 0 0
Local Reuse Pools Shared Pool 24 Shared Pool 32 Shared Pool 40 Shared Pool 192 Shared Pool 256
Cached / Max 384/6144 384/6144 11/512 765/12288 32/512
Cached + Free / Total 659948/660000 659905/660000 170000/170000 1254542/1255000 140000/140000
User Quota Threshold Min.Alloc Cur.Alloc Max.Alloc Total-Alloc Fail-Thresh Fail-Nomem Local-Reuse Data(Pool)-SZ
fptcp_seg 65536 0 0 0 0 537140 0 0 536649 16 (24)
inner_decode 4000 0 0 0 0 0 0 0 0 16 (24)
detector_threat 233016 0 0 22 0 114079 0 0 114023 24 (24)
spyware_state 51200 0 0 0 0 0 0 0 0 24 (24)
vm_vcheck 81920 0 0 0 0 1322271 0 0 1322148 24 (24)
ctd_patmatch 192001 0 0 30 0 14000 0 0 14000 24 (24)
proxy_l2info 6400 0 0 0 0 0 0 0 0 24 (24)
proxy_pktmr 6400 0 0 0 0 0 0 0 0 16 (24)
vm_field 500000 0 0 95 0 5135529 0 0 5134623 32 (32)
prl_cookie 6400 0 0 0 0 0 0 0 0 32 (32)
decode_filter 81920 0 0 0 0 2194 0 0 2183 40 (40)
hash_decode 1024 0 0 0 0 0 0 0 0 104 (192)
appid_session 256002 0 0 0 0 314584 0 0 311524 104 (192)
appid_dfa_state 256002 0 0 0 0 4 0 0 4 184 (192)
cpat_state 64000 1004000 32000 0 0 0 0 0 0 184 (192)
sml_regfile 512004 0 0 122 0 560187 0 0 554372 192 (192)
ctd_flow 256002 0 0 61 0 280118 0 0 279134 192 (192)
ctd_flow_state 256002 0 0 53 0 17008 0 0 15066 176 (192)
ctd_dlp_flow 64000 878500 32000 0 0 84 0 0 83 192 (192)
proxy_flow 12800 0 0 74 0 13000 0 0 11479 192 (192)
prl_st 6400 0 0 0 0 0 0 0 0 192 (192)
ssl_hs_st 6400 0 0 0 0 13000 0 0 9727 192 (192)
ssl_key_block 12800 0 0 74 0 12908 0 0 12510 192 (192)
ssl_st 12800 0 0 74 0 13000 0 0 9815 192 (192)
ssl_hs_mac 19200 0 0 0 0 49924 0 0 35277 variable
timer_chunk 131072 0 0 0 0 99555 0 0 89397 256 (256)
Memory Pool Size 126400KB, start address 0xe000000000
alloc size 853757, max 1315990
fixed buf allocator, size 129430024
sz allocator, page size 32768, max alloc 4096 quant 64
pool 0 element size 64 avail list 5 full list 2
pool 1 element size 128 avail list 7 full list 1
pool 2 element size 192 avail list 1 full list 0
pool 3 element size 256 avail list 1 full list 7
pool 4 element size 320 avail list 2 full list 0
pool 5 element size 384 avail list 1 full list 0
pool 8 element size 576 avail list 1 full list 0
pool 10 element size 704 avail list 15 full list 0
pool 16 element size 1088 avail list 1 full list 0
pool 23 element size 1536 avail list 1 full list 0
pool 24 element size 1600 avail list 1 full list 0
pool 25 element size 1664 avail list 1 full list 2
pool 26 element size 1728 avail list 1 full list 0
pool 27 element size 1792 avail list 1 full list 0
pool 28 element size 1856 avail list 1 full list 0
pool 29 element size 1920 avail list 1 full list 0
pool 31 element size 2048 avail list 1 full list 0
pool 33 element size 2176 avail list 1 full list 0
pool 34 element size 2240 avail list 1 full list 0
pool 35 element size 2304 avail list 1 full list 0
pool 37 element size 2432 avail list 1 full list 0
pool 44 element size 2880 avail list 1 full list 0
pool 53 element size 3456 avail list 1 full list 0
pool 57 element size 3712 avail list 1 full list 0
pool 58 element size 3776 avail list 1 full list 0
parent allocator
alloc size 2067189, max 2165493
malloc allocator
current usage 2097152 max. usage 2195456, free chunks 3885, total chunks 3949
Mem-Pool-Type MaxSz(KB) Threshold MinSz(KB) CurSz(B) Cur.Alloc Total-Alloc Fail-Thresh Fail-Nomem Local-Reuse(cache)
ctd_dlp_buf 3968 63200 1984 0 0 0 0 0 0 (0)
proxy 51862 0 0 678056 5333 187433 0 0 84 (1)
clientless_sslvpn 51862 0 0 0 0 0 0 0 0 (0)
l7_data 2712 0 0 0 0 0 0 0 0 (0)
l7_misc 30491 88480 15245 111464 1417 2799 0 0 62 (1)
cfg_name_cache 553 88480 276 48 1 1 0 0 0 (0)
scantracker 480 75840 240 0 0 0 0 0 0 (0)
appinfo 105 88480 52 0 0 0 0 0 0 (0)
user 16750 0 0 1544 22 22 0 0 0 (0)
userpolicy 8000 0 0 0 0 1602 0 0 0 (0)
dns 2048 88480 1024 0 0 0 0 0 0 (0)
credential 8192 75840 4096 0 0 0 0 0 0 (0)
Cache-Type MAX-Entries Cur-Entries Cur.SZ(B) Insert-Failure Mem-Pool-Type
ssl_server_cert 16384 305 24400 0 l7_misc
ssl_cert_cn 1250 1105 86488 0 l7_misc
ssl_cert_cache 256 94 6016 0 proxy
ssl_sess_cache 1000 1000 208000 0 proxy
proxy_exclude 1024 5 1200 0 proxy
proxy_notify 8192 0 0 0 proxy
ctd_block_answer 16384 0 0 0 l7_misc
username_cache 4096 1 48 0 cfg_name_cache
threatname_cache 4096 0 0 0 cfg_name_cache
hipname_cache 256 0 0 0 cfg_name_cache
ctd_cp 16384 0 0 0 l7_misc
ctd_driveby 4096 0 0 0 l7_misc
ctd_pcap 1024 0 0 0 l7_misc
ctd_sml 4096 2 96 0 l7_misc
ctd_url 50000 5 480 0 l7_misc
app_tracker 8192 0 0 0 l7_misc
threat_tracker 4096 0 0 0 l7_misc
scan_tracker 4096 0 0 0 scantracker
app_info 500 0 0 0 appinfo
dns 6400 0 0 0 proxy
dns_v4 10000 0 0 0 dns
dns_v6 10000 0 0 0 dns
dns_id 1024 0 0 0 dns
tcp_mcb 8256 0 0 0 l7_misc
sslvpn_ck_cache 25 0 0 0 clientless_sslvpn
user_cache 64000 22 1544 0 user
userpolicy_cache 8000 0 0 0 userpolicy
mostrar el monitor de recursos en ejecución: este es el comando más importante para obtener usos de planes de datos CPU en diferentes intervalos de tiempo. Por lo general, si las CPU estancias altas (>90), el tráfico se sentiría lento, la latencia también aumentaría. La mejor estrategia es determinar un uso regular de 24 horas ("línea base") y luego compararlo con los momentos en que se experimentan picos.
admin@anuragFW> show running resource-monitor
> day Per-day monitoring statistics
> hour Per-hour monitoring statistics
> minute Per-minute monitoring statistics
> second Per-second monitoring statistics
> week Per-week monitoring statistics
| Pipe through a command
<Enter> Finish input
admin@anuragFW> show running resource-monitor
Resource monitoring sampling data (per second):
CPU load sampling by group:
flow_lookup : 3%
flow_fastpath : 3%
flow_slowpath : 3%
flow_forwarding : 3%
flow_mgmt : 3%
flow_ctrl : 3%
nac_result : 0%
flow_np : 3%
dfa_result : 0%
module_internal : 3%
aho_result : 0%
zip_result : 0%
pktlog_forwarding : 3%
send_out : 3%
flow_host : 3%
send_host : 3%
CPU load (%) during last 60 seconds:
core 0 1
* 3
* 3
* 3
* 3
* 3
* 3
* 3
* 3
* 3
* 3
* 3
* 3
* 3
* 3
* 3
* 3
* 3
* 3
* 3
* 3
* 2
* 3
* 3
* 3
* 3
* 3
* 3
* 3
* 2
* 2
* 3
* 3
* 3
* 3
* 3
* 3
* 3
* 3
* 3
* 4
* 3
* 3
* 3
* 3
* 3
* 3
* 2
* 2
* 2
* 1
* 3
* 3
* 3
* 3
* 3
* 3
* 3
* 3
* 3
* 3
Resource utilization (%) during last 60 seconds:
session:
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
packet buffer:
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
packet descriptor:
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
sw tags descriptor:
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
Resource monitoring sampling data (per minute):
CPU load (%) during last 60 minutes:
core 0 1
avg max avg max
* * 3 3
* * 3 8
* * 11 50
* * 8 36
* * 3 4
* * 4 22
* * 3 5
* * 3 4
* * 3 3
* * 3 4
* * 3 3
* * 3 4
* * 3 3
* * 2 3
* * 3 4
* * 3 3
* * 3 4
* * 3 4
* * 3 4
* * 3 4
* * 3 4
* * 3 4
* * 3 4
* * 3 4
* * 3 4
* * 3 4
* * 3 4
* * 3 4
* * 3 4
* * 3 4
* * 3 4
* * 3 4
* * 3 4
* * 3 4
* * 3 4
* * 3 4
* * 3 4
* * 3 4
* * 3 4
* * 3 4
* * 3 4
* * 3 4
* * 3 3
* * 3 3
* * 3 4
* * 3 4
* * 3 4
* * 3 4
* * 3 4
* * 3 4
* * 3 4
* * 3 4
* * 3 4
* * 3 4
* * 3 4
* * 3 4
* * 3 3
* * 3 4
* * 3 4
* * 3 4
Resource utilization (%) during last 60 minutes:
session (average):
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
session (maximum):
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
packet buffer (average):
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
packet buffer (maximum):
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
packet descriptor (average):
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
packet descriptor (maximum):
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
sw tags descriptor (average):
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
sw tags descriptor (maximum):
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
Resource monitoring sampling data (per hour):
CPU load (%) during last 24 hours:
core 0 1
avg max avg max
* * 3 36
* * 3 4
* * 3 4
* * 3 6
* * 3 8
* * 3 7
* * 3 5
* * 3 5
* * 3 6
* * 3 5
* * 3 5
* * 3 6
* * 3 8
* * 3 7
* * 3 5
* * 3 9
* * 3 5
* * 3 7
* * 3 6
* * 3 6
* * 3 6
* * 3 6
* * 3 7
* * 3 7
Resource utilization (%) during last 24 hours:
session (average):
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0
session (maximum):
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0
packet buffer (average):
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0
packet buffer (maximum):
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0
packet descriptor (average):
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0
packet descriptor (maximum):
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0
sw tags descriptor (average):
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0
sw tags descriptor (maximum):
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
0 0 0 0 0 0 0 0 0
Resource monitoring sampling data (per day):
CPU load (%) during last 7 days:
core 0 1
avg max avg max
* * 3 10
* * 3 12
* * 3 30
* * 3 20
* * 3 49
* * 3 77
* * 3 13
Resource utilization (%) during last 7 days:
session (average):
0 0 0 0 0 0 0
session (maximum):
0 0 0 0 0 0 0
packet buffer (average):
0 0 0 0 0 0 0
packet buffer (maximum):
0 0 0 0 0 0 0
packet descriptor (average):
0 0 0 0 0 0 0
packet descriptor (maximum):
0 0 0 0 0 0 0
sw tags descriptor (average):
0 0 0 0 0 0 0
sw tags descriptor (maximum):
0 0 0 0 0 0 0
Resource monitoring sampling data (per week):
CPU load (%) during last 13 weeks:
core 0 1
avg max avg max
* * 3 77
* * 3 74
* * * *
* * * *
* * * *
* * * *
* * * *
* * * *
* * * *
* * * *
* * * *
* * * *
* * * *
Resource utilization (%) during last 13 weeks:
session (average):
0 0 0 0 0 0 0 0 0 0 0 0 0
session (maximum):
0 0 0 0 0 0 0 0 0 0 0 0 0
packet buffer (average):
0 0 0 0 0 0 0 0 0 0 0 0 0
packet buffer (maximum):
0 0 0 0 0 0 0 0 0 0 0 0 0
packet descriptor (average):
0 0 0 0 0 0 0 0 0 0 0 0 0
packet descriptor (maximum):
0 0 0 0 0 0 0 0 0 0 0 0 0
sw tags descriptor (average):
0 0 0 0 0 0 0 0 0 0 0 0 0
sw tags descriptor (maximum):
0 0 0 0 0 0 0 0 0 0 0 0 0
show counter global - Este comando enumera todos los contadores disponibles en la firewall versión OS dada. Por cada paquete que llega, atraviesa o incluso se cae, deberíamos ver uno o más contadores subir. Éstos son extremadamente poderosos en resolver problemas los problemas relacionados con el tráfico cuando se combinan con el filtro de paquetes.
admin@anuragFW> show counter global
Global counters:
Elapsed time since last sampling: 491.920 seconds
name value rate severity category aspect description
--------------------------------------------------------------------------------
pkt_recv 57490275 67 info packet pktproc Packets received
pkt_sent 1402176 44 info packet pktproc Packets transmitted
pkt_sent_host 359410 0 info packet pktproc Packets successfully transmitted to host interface
pkt_stp_rcv 1085307 0 info packet pktproc STP BPDU packets received
session_allocated 375497 1 info session resource Sessions allocated
session_freed 375449 2 info session resource Sessions freed
session_installed 375019 1 info session resource Sessions installed
session_discard 85940 0 info session resource Session set to discard by security policy check
session_unverified_rst 1034 0 info session pktproc Session aging timer modified by unverified RST
...cut for brevity...
Se pueden establecer diferentes filtros para restringir el enfoque en los contadores relevantes.
admin@anuragFW> show counter global filter severity drop
Global counters:
Elapsed time since last sampling: 13.407 seconds
name value rate severity category aspect description
-----------------------------------------------------------------------------------------------------------------
flow_rcv_dot1q_tag_err 46215426 13 drop flow parse Packets dropped: 802.1q tag not configured
flow_no_interface 46215426 13 drop flow parse Packets dropped: invalid interface
-----------------------------------------------------------------------------------------------------------------
También podemos usar el subcomité 'match' para buscar resultados basados en la coincidencia de cadenas con el argumento de 'match'.
admin@anuragFW> show counter global | match deny
flow_policy_deny 23905 0 drop flow session Session setup: denied by policy
flow_rematch_parent 10 0 info flow pktproc number of rematch deny for parent sessions
flow_host_service_deny 1871180 2 drop flow mgmt Device management session denied