mirror of
				https://github.com/Telecominfraproject/wlan-lanforge-scripts.git
				synced 2025-11-04 12:48:00 +00:00 
			
		
		
		
	
		
			
				
	
	
		
			1037 lines
		
	
	
		
			45 KiB
		
	
	
	
		
			Python
		
	
	
		
			Executable File
		
	
	
	
	
			
		
		
	
	
			1037 lines
		
	
	
		
			45 KiB
		
	
	
	
		
			Python
		
	
	
		
			Executable File
		
	
	
	
	
#!/usr/bin/env python3
 | 
						|
'''
 | 
						|
This will create a station, create TCP and UDP traffic, run it a short amount of time, and verify whether traffic
 | 
						|
 was sent and received.  It also verifies the station connected to the requested BSSID if bssid is specified
 | 
						|
 as an argument. The script will clean up the station and connections at the end of the test.
 | 
						|
 | 
						|
CLI Example:
 | 
						|
./sta_connect2.py --mgr localhost --dut_ssid <ssid> --dut_passwd <passwd> --dut_security wpa2
 | 
						|
--upstream_port eth1 --radio wiphy1
 | 
						|
 | 
						|
CLI Example for kpi.csv report output:
 | 
						|
./sta_connect2.py --mgr localhost --dut_ssid <ssid> --dut_passwd <passwd> --dut_security wpa2
 | 
						|
--upstream_port eth2  --csv_outfile sta_connect2.csv --test_rig LF-Lab --test_tag L3 --dut_hw_version Linux
 | 
						|
--dut_model_num 1 --dut_sw_version 5.4.5 --dut_serial_num 1234
 | 
						|
 | 
						|
CLI Example for kpi.csv, variable tx/rx rates, and pdu size:
 | 
						|
./sta_connect2.py --mgr localhost --dut_ssid <ssid> --dut_passwd <passwd> --dut_security wpa2
 | 
						|
--upstream_port eth2  --download_bps 768000 --upload_bps 256000 --side_a_pdu 300 --side_b_pdu 750
 | 
						|
--csv_outfile sta_connect2.csv --test_rig LF-Lab --test_tag L3 --dut_hw_version Linux --dut_model_num 1
 | 
						|
--dut_sw_version 5.4.5 --dut_serial_num 1234
 | 
						|
 | 
						|
Copyright 2021 Candela Technologies Inc
 | 
						|
License: Free to distribute and modify. LANforge systems must be licensed.
 | 
						|
'''
 | 
						|
 | 
						|
import sys
 | 
						|
import os
 | 
						|
import csv
 | 
						|
import importlib
 | 
						|
import argparse
 | 
						|
from pprint import pformat
 | 
						|
import time
 | 
						|
import logging
 | 
						|
import datetime
 | 
						|
 | 
						|
if sys.version_info[0] != 3:
 | 
						|
    print("This script requires Python 3")
 | 
						|
    exit(1)
 | 
						|
 | 
						|
sys.path.append(os.path.join(os.path.abspath(__file__ + "../../../")))
 | 
						|
 | 
						|
LFUtils = importlib.import_module("py-json.LANforge.LFUtils")
 | 
						|
removeCX = LFUtils.removeCX
 | 
						|
removeEndps = LFUtils.removeEndps
 | 
						|
realm = importlib.import_module("py-json.realm")
 | 
						|
Realm = realm.Realm
 | 
						|
influx = importlib.import_module("py-scripts.influx_utils")
 | 
						|
RecordInflux = influx.RecordInflux
 | 
						|
lf_report = importlib.import_module("py-scripts.lf_report")
 | 
						|
lf_graph = importlib.import_module("py-scripts.lf_graph")
 | 
						|
lf_kpi_csv = importlib.import_module("py-scripts.lf_kpi_csv")
 | 
						|
logger = logging.getLogger(__name__)
 | 
						|
lf_logger_config = importlib.import_module("py-scripts.lf_logger_config")
 | 
						|
 | 
						|
OPEN = "open"
 | 
						|
WEP = "wep"
 | 
						|
WPA = "wpa"
 | 
						|
WPA2 = "wpa2"
 | 
						|
WPA3 = "wpa3"
 | 
						|
MODE_AUTO = 0
 | 
						|
 | 
						|
 | 
						|
class StaConnect2(Realm):
 | 
						|
    def __init__(self, host, port, _dut_ssid="jedway-open-1", _dut_passwd="NA", _dut_bssid="",
 | 
						|
                 _user="", _passwd="", _sta_mode="0", _radio="wiphy0",
 | 
						|
                 _influx_host=None, _influx_db=None, _influx_user=None,
 | 
						|
                 _influx_passwd=None,
 | 
						|
                 _resource=1, _upstream_resource=1, _upstream_port="eth1",
 | 
						|
                 _sta_name=None, _sta_prefix='sta', _bringup_time_sec=300,
 | 
						|
                 debug_=False, _dut_security=OPEN, _exit_on_error=False,
 | 
						|
                 _cleanup_on_exit=True, _clean_all_sta=False, _runtime_sec=None, kpi_csv=None, outfile=None,
 | 
						|
                 download_bps=None, upload_bps=None, side_a_pdu=None, side_b_pdu=None, _exit_on_fail=False):
 | 
						|
        # do not use `super(LFCLiBase,self).__init__(self, host, port, _debugOn)
 | 
						|
        # that is py2 era syntax and will force self into the host variable, making you
 | 
						|
        # very confused.
 | 
						|
        super().__init__(host, port, debug_=debug_, _exit_on_fail=_exit_on_fail)
 | 
						|
        self.host = host
 | 
						|
        self.port = port
 | 
						|
        self.debug = debug_
 | 
						|
        self.dut_security = _dut_security
 | 
						|
        self.dut_ssid = _dut_ssid
 | 
						|
        self.dut_passwd = _dut_passwd
 | 
						|
        self.dut_bssid = _dut_bssid
 | 
						|
        self.user = _user
 | 
						|
        self.passwd = _passwd
 | 
						|
        self.sta_mode = _sta_mode  # See add_sta LANforge CLI users guide entry
 | 
						|
        self.radio = _radio
 | 
						|
        self.resource = _resource
 | 
						|
        self.outfile = outfile
 | 
						|
        self.kpi_csv = kpi_csv
 | 
						|
        self.download_bps = download_bps
 | 
						|
        self.upload_bps = upload_bps
 | 
						|
        self.side_a_pdu = side_a_pdu
 | 
						|
        self.side_b_pdu = side_b_pdu
 | 
						|
        self.epoch_time = int(time.time())
 | 
						|
        self.upstream_resource = _upstream_resource
 | 
						|
        self.upstream_port = _upstream_port
 | 
						|
        self.runtime_secs = _runtime_sec
 | 
						|
        self.cleanup_on_exit = _cleanup_on_exit
 | 
						|
        self.clean_all_sta = _clean_all_sta
 | 
						|
        self.sta_url_map = None  # defer construction
 | 
						|
        self.upstream_url = None  # defer construction
 | 
						|
        self.station_names = []
 | 
						|
        if _sta_name:
 | 
						|
            self.station_names = [_sta_name]
 | 
						|
        self.sta_prefix = _sta_prefix
 | 
						|
        self.bringup_time_sec = _bringup_time_sec
 | 
						|
        self.resulting_stations = {}
 | 
						|
        self.resulting_endpoints = {}
 | 
						|
        self.station_profile = None
 | 
						|
        self.l3_udp_profile = None
 | 
						|
        self.l3_tcp_profile = None
 | 
						|
        self.influx_host = _influx_host
 | 
						|
        self.influx_db = _influx_db
 | 
						|
        self.influx_user = _influx_user
 | 
						|
        self.influx_passwd = _influx_passwd
 | 
						|
        self.name_prefix = "tcp"
 | 
						|
        self.use_existing_sta = False
 | 
						|
 | 
						|
        self.cx_profile = self.new_l3_cx_profile()
 | 
						|
        self.cx_profile.host = self.host
 | 
						|
        self.cx_profile.port = self.port
 | 
						|
        self.cx_profile.name_prefix = self.name_prefix
 | 
						|
 | 
						|
        if self.outfile is not None:
 | 
						|
            results = self.outfile[:-4]
 | 
						|
            results = results + "-results.csv"
 | 
						|
            self.csv_results_file = open(results, "w")
 | 
						|
            self.csv_results_writer = csv.writer(self.csv_results_file, delimiter=",")
 | 
						|
 | 
						|
    def get_kpi_results(self):
 | 
						|
        # make json call to get kpi results
 | 
						|
        endp_list = self.json_get(
 | 
						|
            "endp?fields=name,eid,delay,jitter,rx+rate,rx+rate+ll,rx+bytes,rx+drop+%25,rx+pkts+ll",
 | 
						|
            debug_=False)
 | 
						|
        logger.info("endp_list: {endp_list}".format(endp_list=endp_list))
 | 
						|
 | 
						|
    def get_csv_name(self):
 | 
						|
        logger.info("self.csv_results_file {}".format(self.csv_results_file.name))
 | 
						|
        return self.csv_results_file.name
 | 
						|
 | 
						|
    # Query all endpoints to generate rx and other stats, returned
 | 
						|
    # as an array of objects.
 | 
						|
    def get_rx_values(self):
 | 
						|
        endp_list = self.json_get(
 | 
						|
            "endp?fields=name,eid,delay,jitter,rx+rate,rx+rate+ll,rx+bytes,rx+drop+%25,rx+pkts+ll",
 | 
						|
            debug_=True)
 | 
						|
        # logger.info("endp_list: {endp_list}".format(endp_list=endp_list))
 | 
						|
        endp_rx_drop_map = {}
 | 
						|
        endp_rx_map = {}
 | 
						|
        our_endps = {}
 | 
						|
        endps = []
 | 
						|
 | 
						|
        total_ul = 0
 | 
						|
        total_ul_ll = 0
 | 
						|
        total_dl = 0
 | 
						|
        total_dl_ll = 0
 | 
						|
        udp_dl = 0
 | 
						|
        tcp_dl = 0
 | 
						|
        udp_ul = 0
 | 
						|
        tcp_ul = 0
 | 
						|
 | 
						|
        '''
 | 
						|
        for e in self.cx_profile.created_endp.keys():
 | 
						|
            our_endps[e] = e
 | 
						|
        print("our_endps {our_endps}".format(our_endps=our_endps))
 | 
						|
        '''
 | 
						|
        for endp_name in endp_list['endpoint']:
 | 
						|
            if endp_name != 'uri' and endp_name != 'handler':
 | 
						|
                for item, endp_value in endp_name.items():
 | 
						|
                    # if item in our_endps:
 | 
						|
                    if True:
 | 
						|
                        endps.append(endp_value)
 | 
						|
                        logger.debug("endpoint: {item} value:\n".format(item=item))
 | 
						|
                        logger.debug(endp_value)
 | 
						|
                        # print("item {item}".format(item=item))
 | 
						|
 | 
						|
                        for value_name, value in endp_value.items():
 | 
						|
                            if value_name == 'rx bytes':
 | 
						|
                                endp_rx_map[item] = value
 | 
						|
                            if value_name == 'rx rate':
 | 
						|
                                endp_rx_map[item] = value
 | 
						|
                                # print("value {value}".format(value=value))
 | 
						|
                            if value_name == 'rx rate ll':
 | 
						|
                                endp_rx_map[item] = value
 | 
						|
                                # print("value {value}".format(value=value))
 | 
						|
                            if value_name == 'rx pkts ll':
 | 
						|
                                endp_rx_map[item] = value
 | 
						|
                            if value_name == 'rx drop %':
 | 
						|
                                endp_rx_drop_map[item] = value
 | 
						|
                            if value_name == 'rx rate':
 | 
						|
                                # This hack breaks for mcast or if someone names endpoints weirdly.
 | 
						|
                                # logger.info("item: ", item, " rx-bps: ", value_rx_bps)
 | 
						|
 | 
						|
                                # info for download test data
 | 
						|
                                if item.startswith("udp") and item.endswith("-A"):
 | 
						|
                                    udp_ul += int(value)
 | 
						|
                                    # logger.info(udp_udl)
 | 
						|
                                elif item.startswith("tcp") and item.endswith("-A"):
 | 
						|
                                    tcp_ul += int(value)
 | 
						|
                                    # logger.info(tcp_dl)
 | 
						|
 | 
						|
                                # info for upload test data
 | 
						|
                                if item.startswith("udp") and item.endswith("-B"):
 | 
						|
                                    udp_dl += int(value)
 | 
						|
                                    # logger.info(udp_ul)
 | 
						|
                                elif item.startswith("tcp") and item.endswith("-B"):
 | 
						|
                                    tcp_dl += int(value)
 | 
						|
                                    # logger.info(tcp_ul)
 | 
						|
                                # combine total download (tcp&udp) and upload (tcp&udp) test data
 | 
						|
                                if item.endswith("-A"):
 | 
						|
                                    total_dl += int(value)
 | 
						|
                                    # logger.info(total_dl)
 | 
						|
                                else:
 | 
						|
                                    total_ul += int(value)
 | 
						|
                            if value_name == 'rx rate ll':
 | 
						|
                                # This hack breaks for mcast or if someone
 | 
						|
                                # names endpoints weirdly.
 | 
						|
                                if item.endswith("-A"):
 | 
						|
                                    total_dl_ll += int(value)
 | 
						|
                                else:
 | 
						|
                                    total_ul_ll += int(value)
 | 
						|
 | 
						|
        # logger.debug("total-dl: ", total_dl, " total-ul: ", total_ul, "\n")
 | 
						|
        return endp_rx_map, endp_rx_drop_map, endps, udp_dl, tcp_dl, udp_ul, tcp_ul, total_dl, total_ul, total_dl_ll, total_ul_ll
 | 
						|
 | 
						|
    # Common code to generate timestamp for CSV files.
 | 
						|
    def time_stamp(self):
 | 
						|
        return time.strftime('%m_%d_%Y_%H_%M_%S',
 | 
						|
                             time.localtime(self.epoch_time))
 | 
						|
 | 
						|
    def get_station_url(self, sta_name_=None):
 | 
						|
        if sta_name_ is None:
 | 
						|
            raise ValueError("get_station_url wants a station name")
 | 
						|
        if self.sta_url_map is None:
 | 
						|
            self.sta_url_map = {}
 | 
						|
            for sta_name in self.station_names:
 | 
						|
                self.sta_url_map[sta_name] = "port/1/%s/%s" % (self.resource, sta_name)
 | 
						|
        return self.sta_url_map[sta_name_]
 | 
						|
 | 
						|
    def get_upstream_url(self):
 | 
						|
        if self.upstream_url is None:
 | 
						|
            self.upstream_url = "port/1/%s/%s" % (self.upstream_resource, self.upstream_port)
 | 
						|
        return self.upstream_url
 | 
						|
 | 
						|
    # Compare pre-test values to post-test values
 | 
						|
    def compare_vals(self, name, post_val, print_pass=False, print_fail=True):
 | 
						|
        # print(f"Comparing {name}")
 | 
						|
        # logger.info(name)
 | 
						|
        # logger.info(postVal)
 | 
						|
        if post_val > 0:
 | 
						|
            pass_msg = str(post_val) + " bps"
 | 
						|
            self._pass("%s %s" % (name, pass_msg), print_pass)
 | 
						|
        else:
 | 
						|
            fail_msg = str(post_val) + " bps"
 | 
						|
            self._fail("%s did not report traffic: %s" % (name, fail_msg), print_fail)
 | 
						|
 | 
						|
    def remove_stations(self):
 | 
						|
        for name in self.station_names:
 | 
						|
            LFUtils.removePort(self.resource, name, self.lfclient_url)
 | 
						|
 | 
						|
    def num_associated(self, bssid):
 | 
						|
        counter = 0
 | 
						|
        # print("there are %d results" % len(self.station_results))
 | 
						|
        fields = "_links,port,alias,ip,ap,port+type"
 | 
						|
        self.station_results = self.find_ports_like("%s*" % self.sta_prefix, fields, debug_=False)
 | 
						|
        logger.info("num_associated() - self.station_results: {}".format(self.station_results))
 | 
						|
        if (self.station_results is None) or (len(self.station_results) < 1):
 | 
						|
            self.get_failed_result_list()
 | 
						|
        for eid, record in self.station_results.items():
 | 
						|
            # logger.info("-- -- -- -- -- -- -- -- -- -- -- -- -- -- -- -- -- -- ")
 | 
						|
            logger.debug(pformat(eid))
 | 
						|
            # logger.debug(pformat(record))
 | 
						|
            if record["ap"] == bssid:
 | 
						|
                counter += 1
 | 
						|
            # logger.info("-- -- -- -- -- -- -- -- -- -- -- -- -- -- -- -- -- -- ")
 | 
						|
        return counter
 | 
						|
 | 
						|
    def clear_test_results(self):
 | 
						|
        self.resulting_stations = {}
 | 
						|
        self.resulting_endpoints = {}
 | 
						|
        super().clear_test_results()
 | 
						|
        # super(StaConnect, self).clear_test_results().test_results.clear()
 | 
						|
 | 
						|
    def setup(self, extra_securities=None):
 | 
						|
        if extra_securities is None:
 | 
						|
            extra_securities = []
 | 
						|
        self.clear_test_results()
 | 
						|
        self.check_connect()
 | 
						|
        upstream_json = self.json_get("%s?fields=alias,phantom,down,port,ip" % self.get_upstream_url(), debug_=False)
 | 
						|
 | 
						|
        if upstream_json is None:
 | 
						|
            self._fail(message="Unable to query %s, bye" % self.upstream_port, print_=True)
 | 
						|
            return False
 | 
						|
 | 
						|
        if upstream_json['interface']['ip'] == "0.0.0.0":
 | 
						|
            if self.debug:
 | 
						|
                logger.debug(pformat(upstream_json))
 | 
						|
            self._fail("Warning: %s lacks ip address" % self.get_upstream_url(), print_=True)
 | 
						|
            return False
 | 
						|
 | 
						|
        self.pre_cleanup()
 | 
						|
 | 
						|
        # Create stations and turn dhcp on
 | 
						|
        self.station_profile = self.new_station_profile()
 | 
						|
        self.station_profile.sta_prefix = self.sta_prefix
 | 
						|
 | 
						|
        if self.dut_security == WPA2:
 | 
						|
            self.station_profile.use_security(security_type="wpa2", ssid=self.dut_ssid, passwd=self.dut_passwd)
 | 
						|
        elif self.dut_security == OPEN:
 | 
						|
            self.station_profile.use_security(security_type="open", ssid=self.dut_ssid, passwd="[BLANK]")
 | 
						|
        elif self.dut_security == WPA:
 | 
						|
            self.station_profile.use_security(security_type="wpa", ssid=self.dut_ssid, passwd=self.dut_passwd)
 | 
						|
        elif self.dut_security == WEP:
 | 
						|
            self.station_profile.use_security(security_type="wep", ssid=self.dut_ssid, passwd=self.dut_passwd)
 | 
						|
        elif self.dut_security == WPA3:
 | 
						|
            self.station_profile.use_security(security_type="wpa3", ssid=self.dut_ssid, passwd=self.dut_passwd)
 | 
						|
        self.station_profile.set_command_flag("add_sta", "create_admin_down", 1)
 | 
						|
        for security in extra_securities:
 | 
						|
            self.station_profile.add_security_extra(security=security)
 | 
						|
        # print("Adding new stations ", end="")
 | 
						|
        logger.info("Adding new stations ")
 | 
						|
        self.station_profile.create(radio=self.radio, sta_names_=self.station_names, up_=False, debug=self.debug,
 | 
						|
                                    suppress_related_commands_=True)
 | 
						|
        LFUtils.wait_until_ports_appear(self.lfclient_url, self.station_names, debug=self.debug)
 | 
						|
 | 
						|
        # Create UDP endpoints
 | 
						|
        self.l3_udp_profile = self.new_l3_cx_profile()
 | 
						|
        '''
 | 
						|
        self.l3_udp_profile.side_a_min_bps = 128000
 | 
						|
        self.l3_udp_profile.side_b_min_bps = 128000
 | 
						|
        self.l3_udp_profile.side_a_min_pdu = 1200
 | 
						|
        self.l3_udp_profile.side_b_min_pdu = 1500
 | 
						|
        '''
 | 
						|
        self.l3_udp_profile.side_a_min_bps = self.download_bps
 | 
						|
        self.l3_udp_profile.side_b_min_bps = self.upload_bps
 | 
						|
        self.l3_udp_profile.side_a_min_pdu = self.side_a_pdu
 | 
						|
        self.l3_udp_profile.side_b_min_pdu = self.side_b_pdu
 | 
						|
        self.l3_udp_profile.report_timer = 1000
 | 
						|
        self.l3_udp_profile.name_prefix = "udp"
 | 
						|
        port_list = list(self.find_ports_like("%s+" % self.sta_prefix))
 | 
						|
        if (port_list is None) or (len(port_list) < 1):
 | 
						|
            raise ValueError("Unable to find ports named '%s'+" % self.sta_prefix)
 | 
						|
        self.l3_udp_profile.create(endp_type="lf_udp",
 | 
						|
                                   side_a=port_list,
 | 
						|
                                   side_b="%d.%s" % (self.resource, self.upstream_port),
 | 
						|
                                   suppress_related_commands=True)
 | 
						|
 | 
						|
        # Create TCP endpoints
 | 
						|
        self.l3_tcp_profile = self.new_l3_cx_profile()
 | 
						|
        '''
 | 
						|
        self.l3_tcp_profile.side_a_min_bps = 128000
 | 
						|
        self.l3_tcp_profile.side_b_min_bps = 56000
 | 
						|
        '''
 | 
						|
        self.l3_tcp_profile.side_a_min_bps = self.download_bps
 | 
						|
        self.l3_tcp_profile.side_b_min_bps = self.upload_bps
 | 
						|
        self.l3_tcp_profile.side_a_min_pdu = self.side_a_pdu
 | 
						|
        self.l3_tcp_profile.side_b_min_pdu = self.side_b_pdu
 | 
						|
        self.l3_tcp_profile.name_prefix = self.name_prefix
 | 
						|
        self.l3_tcp_profile.report_timer = 1000
 | 
						|
        self.l3_tcp_profile.create(endp_type="lf_tcp",
 | 
						|
                                   side_a=list(self.find_ports_like("%s+" % self.sta_prefix)),
 | 
						|
                                   side_b="%d.%s" % (self.resource, self.upstream_port),
 | 
						|
                                   suppress_related_commands=True)
 | 
						|
 | 
						|
    def start(self):
 | 
						|
        if self.station_profile is None:
 | 
						|
            self._fail("Incorrect setup")
 | 
						|
        logger.debug(pformat(self.station_profile))
 | 
						|
        if self.station_profile.up is None:
 | 
						|
            self._fail("Incorrect station profile, missing profile.up")
 | 
						|
        if not self.station_profile.up:
 | 
						|
            logger.info("Bringing ports up...")
 | 
						|
            # logger.info("start() - self.lf_client_url: %s", self.lfclient_url)
 | 
						|
            data = {"shelf": 1,
 | 
						|
                    "resource": self.resource,
 | 
						|
                    "port": "ALL",
 | 
						|
                    "probe_flags": 1}
 | 
						|
            # logger.info("start() - data1: %s", data)
 | 
						|
            self.json_post("/cli-json/nc_show_ports", data)
 | 
						|
            #print("start() - ln 252 - self.j")
 | 
						|
            self.station_profile.admin_up()
 | 
						|
            LFUtils.waitUntilPortsAdminUp(self.resource, self.lfclient_url, self.station_names)
 | 
						|
 | 
						|
        self.csv_add_column_headers()
 | 
						|
 | 
						|
        if self.influx_db:
 | 
						|
            grapher = RecordInflux(_influx_host=self.influx_host,
 | 
						|
                                   _influx_db=self.influx_db,
 | 
						|
                                   _influx_user=self.influx_user,
 | 
						|
                                   _influx_passwd=self.influx_passwd,
 | 
						|
                                   _longevity=1,
 | 
						|
                                   _devices=self.station_names,
 | 
						|
                                   _monitor_interval=1,
 | 
						|
                                   _target_kpi=['bps rx'])
 | 
						|
 | 
						|
        # station_info = self.jsonGet(self.mgr_url, "%s?fields=port,ip,ap" % (self.getStaUrl()))
 | 
						|
        duration = 0
 | 
						|
        maxTime = self.bringup_time_sec
 | 
						|
        ip = "0.0.0.0"
 | 
						|
        ap = ""
 | 
						|
        # print("Waiting for %s stations to associate to AP: " % len(self.station_names), end="")
 | 
						|
        logger.info("Waiting for %s stations to associate to AP: ", len(self.station_names))
 | 
						|
        connected_stations = {}
 | 
						|
        while (len(connected_stations.keys()) < len(self.station_names)) and (duration < maxTime):
 | 
						|
            duration += 3
 | 
						|
            time.sleep(3)
 | 
						|
            print(".", end="")
 | 
						|
            for sta_name in self.station_names:
 | 
						|
                sta_url = self.get_station_url(sta_name)
 | 
						|
                station_info = self.json_get(sta_url + "?fields=port,ip,ap")
 | 
						|
 | 
						|
                # logger.info("start() - sta_url: %s", sta_url)
 | 
						|
                # logger.info("start() - station_info: %s", station_info)
 | 
						|
 | 
						|
                # LFUtils.debug_printer.pprint(station_info)
 | 
						|
                if (station_info) and ("interface" in station_info):
 | 
						|
                    if "ip" in station_info["interface"]:
 | 
						|
                        ip = station_info["interface"]["ip"]
 | 
						|
                    if "ap" in station_info["interface"]:
 | 
						|
                        ap = station_info["interface"]["ap"]
 | 
						|
 | 
						|
                if (ap == "Not-Associated") or (ap == ""):
 | 
						|
                    if self.debug:
 | 
						|
                        # print(" -%s," % sta_name, end="")
 | 
						|
                        logger.info(" -%s,", sta_name)
 | 
						|
                else:
 | 
						|
                    if ip == "0.0.0.0":
 | 
						|
                        if self.debug:
 | 
						|
                            # print(" %s (0.0.0.0)" % sta_name, end="")
 | 
						|
                            logger.info(" %s (0.0.0.0)", sta_name)
 | 
						|
                    else:
 | 
						|
                        connected_stations[sta_name] = sta_url
 | 
						|
            data = {
 | 
						|
                "shelf": 1,
 | 
						|
                "resource": self.resource,
 | 
						|
                "port": "ALL",
 | 
						|
                "probe_flags": 1
 | 
						|
            }
 | 
						|
            # logger.info("start() - data2: %s", data)
 | 
						|
            self.json_post("/cli-json/nc_show_ports", data)
 | 
						|
            if self.influx_db:
 | 
						|
                grapher.getdata()
 | 
						|
        # LFUtils.wait_until_ports_appear(port_list=self.station_names, debug=self.debug)
 | 
						|
        LFUtils.wait_until_ports_appear(self.lfclient_url, port_list=self.station_names, debug=self.debug)
 | 
						|
 | 
						|
        for sta_name in self.station_names:
 | 
						|
            sta_url = self.get_station_url(sta_name)
 | 
						|
            station_info = self.json_get(sta_url)  # + "?fields=port,ip,ap")
 | 
						|
            if station_info is None:
 | 
						|
                logger.info("unable to query %s", sta_url)
 | 
						|
            self.resulting_stations[sta_url] = station_info
 | 
						|
            try:
 | 
						|
                ap = station_info["interface"]["ap"]
 | 
						|
            except Exception as e:
 | 
						|
                logger.info(e)
 | 
						|
                ap = "NULL"
 | 
						|
            ip = station_info["interface"]["ip"]
 | 
						|
            if (ap != "") and (ap != "Not-Associated"):
 | 
						|
                print(" %s +AP %s, " % (sta_name, ap), end="")
 | 
						|
                if self.dut_bssid != "":
 | 
						|
                    if self.dut_bssid.lower() == ap.lower():
 | 
						|
                        self._pass(sta_name + " connected to BSSID: " + ap)
 | 
						|
                        # self.test_results.append("PASSED: )
 | 
						|
                        logger.info("PASSED: Connected to BSSID: %s", ap)
 | 
						|
                    else:
 | 
						|
                        self._fail(
 | 
						|
                            "%s connected to wrong BSSID, requested: %s  Actual: %s" % (sta_name, self.dut_bssid, ap))
 | 
						|
            else:
 | 
						|
                self._fail(sta_name + " did not connect to AP")
 | 
						|
                return False
 | 
						|
 | 
						|
            if ip == "0.0.0.0":
 | 
						|
                self._fail("%s did not get an ip. Ending test" % sta_name)
 | 
						|
            else:
 | 
						|
                self._pass("%s connected to AP: %s  With IP: %s" % (sta_name, ap, ip))
 | 
						|
 | 
						|
        if not self.passes():
 | 
						|
            if self.cleanup_on_exit:
 | 
						|
                logger.info("Cleaning up...")
 | 
						|
                self.remove_stations()
 | 
						|
            return False
 | 
						|
 | 
						|
        # start cx traffic
 | 
						|
        logger.info("Starting CX Traffic")
 | 
						|
        self.l3_udp_profile.start_cx()
 | 
						|
        self.l3_tcp_profile.start_cx()
 | 
						|
        time.sleep(1)
 | 
						|
        # Refresh stats
 | 
						|
        self.l3_udp_profile.refresh_cx()
 | 
						|
        self.l3_tcp_profile.refresh_cx()
 | 
						|
 | 
						|
    def collect_endp_stats(self, endp_map):
 | 
						|
        logger.info("Collecting Data")
 | 
						|
        fields = "/all"
 | 
						|
        for (cx_name, endps) in endp_map.items():
 | 
						|
            try:
 | 
						|
                endp_url = "/endp/%s%s" % (endps[0], fields)
 | 
						|
                endp_json = self.json_get(endp_url)
 | 
						|
                self.resulting_endpoints[endp_url] = endp_json
 | 
						|
                ptest_a_tx = endp_json['endpoint']['tx bytes']
 | 
						|
                ptest_a_rx = endp_json['endpoint']['rx bytes']
 | 
						|
 | 
						|
                # ptest = self.json_get("/endp/%s?fields=tx+bytes,rx+bytes" % cx_names[cx_name]["b"])
 | 
						|
                endp_url = "/endp/%s%s" % (endps[1], fields)
 | 
						|
                endp_json = self.json_get(endp_url)
 | 
						|
                self.resulting_endpoints[endp_url] = endp_json
 | 
						|
 | 
						|
                ptest_b_tx = endp_json['endpoint']['tx bytes']
 | 
						|
                ptest_b_rx = endp_json['endpoint']['rx bytes']
 | 
						|
 | 
						|
                self.compare_vals(endps[0], ptest_a_tx)
 | 
						|
                self.compare_vals(endps[0], ptest_a_rx)
 | 
						|
 | 
						|
                self.compare_vals(endps[1], ptest_b_tx)
 | 
						|
                self.compare_vals(endps[1], ptest_b_rx)
 | 
						|
 | 
						|
            except Exception as e:
 | 
						|
                self.error(e)
 | 
						|
 | 
						|
    def stop(self):
 | 
						|
        # stop cx traffic
 | 
						|
        logger.info("Stopping CX Traffic")
 | 
						|
        self.l3_udp_profile.stop_cx()
 | 
						|
        self.l3_tcp_profile.stop_cx()
 | 
						|
 | 
						|
        # Refresh stats
 | 
						|
        logger.info("Refresh CX stats")
 | 
						|
        self.l3_udp_profile.refresh_cx()
 | 
						|
        self.l3_tcp_profile.refresh_cx()
 | 
						|
 | 
						|
        logger.info("Sleeping for 5 seconds")
 | 
						|
        time.sleep(5)
 | 
						|
 | 
						|
        # get data for endpoints JSON
 | 
						|
        self.collect_endp_stats(self.l3_udp_profile.created_cx)
 | 
						|
        self.collect_endp_stats(self.l3_tcp_profile.created_cx)
 | 
						|
 | 
						|
        # print("self.l3_udp_profile.created_cx -ln 412-: ", self.l3_udp_profile.created_cx)
 | 
						|
        # print("self.l3_tcp_profile.created_cx: ", self.l3_tcp_profile.created_cx)
 | 
						|
 | 
						|
 | 
						|
    def cleanup(self):
 | 
						|
        # remove all endpoints and cxs
 | 
						|
        if self.cleanup_on_exit:
 | 
						|
            for sta_name in self.station_names:
 | 
						|
                LFUtils.removePort(self.resource, sta_name, self.lfclient_url)
 | 
						|
            curr_endp_names = []
 | 
						|
            removeCX(self.lfclient_url, self.l3_udp_profile.get_cx_names())
 | 
						|
            removeCX(self.lfclient_url, self.l3_tcp_profile.get_cx_names())
 | 
						|
            for (cx_name, endp_names) in self.l3_udp_profile.created_cx.items():
 | 
						|
                curr_endp_names.append(endp_names[0])
 | 
						|
                curr_endp_names.append(endp_names[1])
 | 
						|
            for (cx_name, endp_names) in self.l3_tcp_profile.created_cx.items():
 | 
						|
                curr_endp_names.append(endp_names[0])
 | 
						|
                curr_endp_names.append(endp_names[1])
 | 
						|
            removeEndps(self.lfclient_url, curr_endp_names, debug=self.debug)
 | 
						|
 | 
						|
    def pre_cleanup(self):
 | 
						|
        self.cx_profile.cleanup_prefix()
 | 
						|
        # do not clean up station if existed prior to test
 | 
						|
        if not self.use_existing_sta:
 | 
						|
            for sta in self.station_names:
 | 
						|
                self.rm_port(sta, check_exists=True, debug_=self.debug)
 | 
						|
 | 
						|
    # builds test data into kpi.csv report
 | 
						|
    def record_kpi_csv(
 | 
						|
            self,
 | 
						|
            sta_list,
 | 
						|
            total_test,
 | 
						|
            total_pass,
 | 
						|
            udp_dl,
 | 
						|
            tcp_dl,
 | 
						|
            udp_ul,
 | 
						|
            tcp_ul,
 | 
						|
            total_dl,
 | 
						|
            total_ul):
 | 
						|
        '''
 | 
						|
 | 
						|
        logger.debug(
 | 
						|
            "NOTE:  Adding kpi to kpi.csv, sta_count {sta_list}  total-download-bps:{total_dl}  upload: {total_ul}  bi-directional: {total}\n".format(
 | 
						|
                sta_list=sta_list, total_dl=total_dl, total_ul=total_ul,
 | 
						|
                total=(total_ul + total_dl)))
 | 
						|
 | 
						|
        logger.debug(
 | 
						|
            "NOTE:  Adding kpi to kpi.csv, sta_count {sta_list}  total-download-bps:{total_dl_ll_bps}  upload: {total_ul_ll_bps}  bi-directional: {total_ll}\n".format(
 | 
						|
                sta_list=sta_list, total_dl_ll_bps=total_dl_ll_bps, total_ul_ll_bps=total_ul_ll_bps,
 | 
						|
                total_ll=(total_ul_ll_bps + total_dl_ll_bps)))
 | 
						|
        '''
 | 
						|
        # the short description will allow for more data to show up in one test-tag graph
 | 
						|
 | 
						|
        # logic for Subtest-Pass & Subtest-Fail columns
 | 
						|
        subpass_udp_dl = 0
 | 
						|
        subpass_udp_ul = 0
 | 
						|
        subpass_tcp_dl = 0
 | 
						|
        subpass_tcp_ul = 0
 | 
						|
        subfail_udp_dl = 1
 | 
						|
        subfail_udp_ul = 1
 | 
						|
        subfail_tcp_dl = 1
 | 
						|
        subfail_tcp_ul = 1
 | 
						|
 | 
						|
        if udp_dl > 0:
 | 
						|
            subpass_udp_dl = 1
 | 
						|
            subfail_udp_dl = 0
 | 
						|
        if udp_ul > 0:
 | 
						|
            subpass_udp_ul = 1
 | 
						|
            subfail_udp_ul = 0
 | 
						|
        if tcp_dl > 0:
 | 
						|
            subpass_tcp_dl = 1
 | 
						|
            subfail_tcp_dl = 0
 | 
						|
        if tcp_ul > 0:
 | 
						|
            subpass_tcp_ul = 1
 | 
						|
            subfail_tcp_ul = 0
 | 
						|
 | 
						|
        # logic for pass/fail column
 | 
						|
        # total_test & total_pass values from lfcli_base.py
 | 
						|
        if total_test == total_pass:
 | 
						|
            pass_fail = "PASS"
 | 
						|
        else:
 | 
						|
            pass_fail = "FAIL"
 | 
						|
 | 
						|
        # kpi data for UDP download traffic
 | 
						|
        results_dict = self.kpi_csv.kpi_csv_get_dict_update_time()
 | 
						|
        results_dict['Graph-Group'] = "UDP Download Rate"
 | 
						|
        results_dict['pass/fail'] = pass_fail
 | 
						|
        results_dict['Subtest-Pass'] = subpass_udp_dl
 | 
						|
        results_dict['Subtest-Fail'] = subfail_udp_dl
 | 
						|
        results_dict['short-description'] = "UDP-DL {download_bps} bps  pdu {side_a_pdu}  {sta_list} STA".format(
 | 
						|
            download_bps=self.download_bps, side_a_pdu=self.side_a_pdu, sta_list=sta_list)
 | 
						|
        results_dict['numeric-score'] = "{}".format(udp_dl)
 | 
						|
        results_dict['Units'] = "bps"
 | 
						|
        self.kpi_csv.kpi_csv_write_dict(results_dict)
 | 
						|
 | 
						|
        # kpi data for UDP upload traffic
 | 
						|
        results_dict['Graph-Group'] = "UDP Upload Rate"
 | 
						|
        results_dict['pass/fail'] = pass_fail
 | 
						|
        results_dict['Subtest-Pass'] = subpass_udp_ul
 | 
						|
        results_dict['Subtest-Fail'] = subfail_udp_ul
 | 
						|
        results_dict['short-description'] = "UDP-UL {upload_bps} bps  pdu {side_b_pdu}  {sta_list} STA".format(
 | 
						|
            upload_bps=self.upload_bps, side_b_pdu=self.side_b_pdu, sta_list=sta_list)
 | 
						|
        results_dict['numeric-score'] = "{}".format(udp_ul)
 | 
						|
        results_dict['Units'] = "bps"
 | 
						|
        self.kpi_csv.kpi_csv_write_dict(results_dict)
 | 
						|
 | 
						|
        # kpi data for TCP download traffic
 | 
						|
        results_dict = self.kpi_csv.kpi_csv_get_dict_update_time()
 | 
						|
        results_dict['Graph-Group'] = "TCP Download Rate"
 | 
						|
        results_dict['pass/fail'] = pass_fail
 | 
						|
        results_dict['Subtest-Pass'] = subpass_tcp_dl
 | 
						|
        results_dict['Subtest-Fail'] = subfail_tcp_dl
 | 
						|
        results_dict['short-description'] = "TCP-DL {download_bps} bps  pdu {side_a_pdu}  {sta_list} STA".format(
 | 
						|
            download_bps=self.download_bps, side_a_pdu=self.side_a_pdu, sta_list=sta_list)
 | 
						|
        results_dict['numeric-score'] = "{}".format(tcp_dl)
 | 
						|
        results_dict['Units'] = "bps"
 | 
						|
        self.kpi_csv.kpi_csv_write_dict(results_dict)
 | 
						|
 | 
						|
        # kpi data for TCP upload traffic
 | 
						|
        results_dict['Graph-Group'] = "TCP Upload Rate"
 | 
						|
        results_dict['pass/fail'] = pass_fail
 | 
						|
        results_dict['Subtest-Pass'] = subpass_tcp_ul
 | 
						|
        results_dict['Subtest-Fail'] = subfail_tcp_ul
 | 
						|
        results_dict['short-description'] = "TCP-UL {upload_bps} bps  pdu {side_b_pdu}  {sta_list} STA".format(
 | 
						|
            upload_bps=self.upload_bps, side_b_pdu=self.side_b_pdu, sta_list=sta_list)
 | 
						|
        results_dict['numeric-score'] = "{}".format(tcp_ul)
 | 
						|
        results_dict['Units'] = "bps"
 | 
						|
        self.kpi_csv.kpi_csv_write_dict(results_dict)
 | 
						|
 | 
						|
    # record results for .html & .pdf reports
 | 
						|
    def record_results(
 | 
						|
            self,
 | 
						|
            sta_count,
 | 
						|
            udp_dl,
 | 
						|
            tcp_dl,
 | 
						|
            udp_ul,
 | 
						|
            tcp_ul,
 | 
						|
            total_dl_bps,
 | 
						|
            total_ul_bps):
 | 
						|
 | 
						|
        dl = self.download_bps
 | 
						|
        ul = self.upload_bps
 | 
						|
        dl_pdu = self.side_a_pdu
 | 
						|
        ul_pdu = self.side_b_pdu
 | 
						|
 | 
						|
        tags = dict()
 | 
						|
        tags['requested-ul-bps'] = ul
 | 
						|
        tags['requested-dl-bps'] = dl
 | 
						|
        tags['ul-pdu-size'] = ul_pdu
 | 
						|
        tags['dl-pdu-size'] = dl_pdu
 | 
						|
        tags['station-count'] = sta_count
 | 
						|
        # tags['attenuation'] = atten
 | 
						|
        tags["script"] = 'sta_connect2'
 | 
						|
 | 
						|
        '''
 | 
						|
        # Add user specified tags
 | 
						|
        for k in self.user_tags:
 | 
						|
            tags[k[0]] = k[1]
 | 
						|
        '''
 | 
						|
 | 
						|
        # now = str(datetime.datetime.utcnow().isoformat())
 | 
						|
 | 
						|
        print(
 | 
						|
            "NOTE:  Adding results to influx, total-download-bps: %s  upload: %s  bi-directional: %s\n" %
 | 
						|
            (total_dl_bps, total_ul_bps, (total_ul_bps + total_dl_bps)))
 | 
						|
 | 
						|
        '''
 | 
						|
        if self.influxdb is not None:
 | 
						|
            self.influxdb.post_to_influx(
 | 
						|
                "total-download-bps", total_dl_bps, tags, now)
 | 
						|
            self.influxdb.post_to_influx(
 | 
						|
                "total-upload-bps", total_ul_bps, tags, now)
 | 
						|
            self.influxdb.post_to_influx(
 | 
						|
                "total-bi-directional-bps",
 | 
						|
                total_ul_bps +
 | 
						|
                total_dl_bps,
 | 
						|
                tags,
 | 
						|
                now)
 | 
						|
        '''
 | 
						|
 | 
						|
        if self.csv_results_file:
 | 
						|
            row = [self.epoch_time, self.time_stamp(), sta_count,
 | 
						|
                   ul, ul, dl, dl, dl_pdu, dl_pdu, ul_pdu, ul_pdu,
 | 
						|
                   udp_ul, tcp_ul, udp_dl, tcp_dl,
 | 
						|
                   total_dl_bps, total_ul_bps, (total_ul_bps + total_dl_bps)
 | 
						|
                   ]
 | 
						|
            '''
 | 
						|
            # Add values for any user specified tags
 | 
						|
            for k in self.user_tags:
 | 
						|
                row.append(k[1])
 | 
						|
            '''
 | 
						|
 | 
						|
            self.csv_results_writer.writerow(row)
 | 
						|
            self.csv_results_file.flush()
 | 
						|
 | 
						|
    def csv_generate_results_column_headers(self):
 | 
						|
        csv_rx_headers = [
 | 
						|
            'Time epoch',
 | 
						|
            'Time',
 | 
						|
            'Station-Count',
 | 
						|
            'UL-Min-Requested',
 | 
						|
            'UL-Max-Requested',
 | 
						|
            'DL-Min-Requested',
 | 
						|
            'DL-Max-Requested',
 | 
						|
            'UL-Min-PDU',
 | 
						|
            'UL-Max-PDU',
 | 
						|
            'DL-Min-PDU',
 | 
						|
            'DL-Max-PDU',
 | 
						|
            # 'Attenuation',
 | 
						|
            'UDP-Upload-bps',
 | 
						|
            'TCP-Upload-bps',
 | 
						|
            'UDP-Download-bps',
 | 
						|
            'TCP-Download-bps',
 | 
						|
            'Total-UDP/TCP-Upload-bps',
 | 
						|
            'Total-UDP/TCP-Download-bps',
 | 
						|
            'Total-UDP/TCP-UL/DL-bps']
 | 
						|
 | 
						|
        return csv_rx_headers
 | 
						|
 | 
						|
    # Write initial headers to csv file.
 | 
						|
    def csv_add_column_headers(self):
 | 
						|
        if self.csv_results_file is not None:
 | 
						|
            self.csv_results_writer.writerow(
 | 
						|
                self.csv_generate_results_column_headers())
 | 
						|
            self.csv_results_file.flush()
 | 
						|
 | 
						|
 | 
						|
# ~class
 | 
						|
 | 
						|
 | 
						|
# - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
 | 
						|
 | 
						|
 | 
						|
def main():
 | 
						|
    lfjson_host = "localhost"
 | 
						|
    lfjson_port = 8080
 | 
						|
    parser = argparse.ArgumentParser(
 | 
						|
        prog="sta_connect2.py",
 | 
						|
        formatter_class=argparse.RawTextHelpFormatter,
 | 
						|
        description="""
 | 
						|
---------------------------
 | 
						|
LANforge Unit Test:  Connect Station to AP - sta_connect2.py
 | 
						|
---------------------------
 | 
						|
Summary:
 | 
						|
This will create a station, create TCP and UDP traffic, run it a short amount of time, and verify whether traffic 
 | 
						|
was sent and received.  It also verifies the station connected to the requested BSSID if bssid is specified 
 | 
						|
as an argument. The script will clean up the station and connections at the end of the test.
 | 
						|
---------------------------
 | 
						|
CLI Example: 
 | 
						|
./sta_connect2.py --mgr localhost --dut_ssid <ssid> --dut_passwd <passwd> --dut_security wpa2 
 | 
						|
--upstream_port eth1 --radio wiphy1
 | 
						|
 | 
						|
CLI Example for kpi.csv report output:
 | 
						|
./sta_connect2.py --mgr localhost --dut_ssid <ssid> --dut_passwd <passwd> --dut_security wpa2 
 | 
						|
--upstream_port eth2  --csv_outfile sta_connect2.csv --test_rig LF-Lab --test_tag L3 --dut_hw_version Linux 
 | 
						|
--dut_model_num 1 --dut_sw_version 5.4.5 --dut_serial_num 1234
 | 
						|
 | 
						|
CLI Example for kpi.csv, variable tx/rx rates, and pdu size:
 | 
						|
./sta_connect2.py --mgr localhost --dut_ssid <ssid> --dut_passwd <passwd> --dut_security wpa2 
 | 
						|
--upstream_port eth2  --download_bps 768000 --upload_bps 256000 --side_a_pdu 300 --side_b_pdu 750 
 | 
						|
--csv_outfile sta_connect2.csv --test_rig LF-Lab --test_tag L3 --dut_hw_version Linux --dut_model_num 1 
 | 
						|
--dut_sw_version 5.4.5 --dut_serial_num 1234
 | 
						|
 | 
						|
--------------------------- 
 | 
						|
""")
 | 
						|
    parser.add_argument("-m", "--mgr", type=str, help="address of the LANforge GUI machine (localhost is default)",
 | 
						|
                        default='localhost')
 | 
						|
    parser.add_argument("-o", "--port", type=int, help="IP Port the LANforge GUI is listening on (8080 is default)",
 | 
						|
                        default=8080)
 | 
						|
    parser.add_argument("-u", "--user", type=str, help="TBD: credential login/username")
 | 
						|
    parser.add_argument("-p", "--passwd", type=str, help="TBD: credential password")
 | 
						|
    parser.add_argument("--resource", type=str, help="LANforge Station resource ID to use, default is 1", default=1)
 | 
						|
    parser.add_argument("--upstream_resource", type=str, help="LANforge Ethernet port resource ID to use, default is 1",
 | 
						|
                        default=None)
 | 
						|
    parser.add_argument("--upstream_port", type=str, help="LANforge Ethernet port name, default is eth2",
 | 
						|
                        default='1.1.eth2')
 | 
						|
    parser.add_argument("--radio", type=str, help="LANforge radio to use, default is wiphy0", default='wiphy0')
 | 
						|
    parser.add_argument("--sta_mode", type=str,
 | 
						|
                        help="LANforge station-mode setting (see add_sta LANforge CLI documentation, default is 0 (auto))",
 | 
						|
                        default=0)
 | 
						|
    parser.add_argument("--dut_ssid", type=str, help="DUT SSID")
 | 
						|
    parser.add_argument("--dut_security", type=str, help="DUT security: openLF, wpa, wpa2, wpa3")
 | 
						|
    parser.add_argument("--dut_passwd", type=str, help="DUT PSK password.  Do not set for OPEN auth")
 | 
						|
    parser.add_argument("--dut_bssid", type=str, help="DUT BSSID to which we expect to connect.")
 | 
						|
    parser.add_argument("--download_bps", type=int, help="Set the minimum bps value on test endpoint A. Default: 128000", default=128000)
 | 
						|
    parser.add_argument("--upload_bps", type=int, help="Set the minimum bps value on test endpoint B. Default: 128000", default=128000)
 | 
						|
    parser.add_argument("--side_a_pdu", type=int, help="Set the minimum pdu value on test endpoint A. Default: 1200", default=1200)
 | 
						|
    parser.add_argument("--side_b_pdu", type=int, help="Set the minimum pdu value on test endpoint B. Default: 1500", default=1500)
 | 
						|
    parser.add_argument("--runtime_sec", type=int, help="Set test duration time. Default: 60 seconds", default=60)
 | 
						|
    parser.add_argument("--debug", help="enable debugging", action="store_true")
 | 
						|
    parser.add_argument("--prefix", type=str, help="Station prefix. Default: 'sta'", default='sta')
 | 
						|
    parser.add_argument("--bringup_time", type=int,
 | 
						|
                        help="Seconds to wait for stations to associate and aquire IP. Default: 300", default=300)
 | 
						|
    parser.add_argument('--influx_user', help='Username for your Influx database', default=None)
 | 
						|
    parser.add_argument('--influx_passwd', help='Password for your Influx database', default=None)
 | 
						|
    parser.add_argument('--influx_db', help='Name of your Influx database', default=None)
 | 
						|
    parser.add_argument('--influx_host',
 | 
						|
                        help='Host of your influx database if different from the system you are running on',
 | 
						|
                        default='localhost')
 | 
						|
    parser.add_argument('--monitor_interval', help='How frequently you want to append to your database', default='5s')
 | 
						|
    parser.add_argument('--debug_log', default=None, help="Specify a file to send debug output to")
 | 
						|
    parser.add_argument('--no_cleanup', help='Do not cleanup before exit', action='store_true')
 | 
						|
    parser.add_argument('--local_lf_report_dir', help='--local_lf_report_dir override the report path, primary use when running test in test suite', default="")
 | 
						|
 | 
						|
    # kpi_csv arguments:
 | 
						|
    parser.add_argument(
 | 
						|
        "--test_rig",
 | 
						|
        default="",
 | 
						|
        help="test rig for kpi.csv, testbed that the tests are run on")
 | 
						|
    parser.add_argument(
 | 
						|
        "--test_tag",
 | 
						|
        default="",
 | 
						|
        help="test tag for kpi.csv,  test specific information to differenciate the test")
 | 
						|
    parser.add_argument(
 | 
						|
        "--dut_hw_version",
 | 
						|
        default="",
 | 
						|
        help="dut hw version for kpi.csv, hardware version of the device under test")
 | 
						|
    parser.add_argument(
 | 
						|
        "--dut_sw_version",
 | 
						|
        default="",
 | 
						|
        help="dut sw version for kpi.csv, software version of the device under test")
 | 
						|
    parser.add_argument(
 | 
						|
        "--dut_model_num",
 | 
						|
        default="",
 | 
						|
        help="dut model for kpi.csv,  model number / name of the device under test")
 | 
						|
    parser.add_argument(
 | 
						|
        "--dut_serial_num",
 | 
						|
        default="",
 | 
						|
        help="dut serial for kpi.csv, serial number / serial number of the device under test")
 | 
						|
    parser.add_argument(
 | 
						|
        "--test_priority",
 | 
						|
        default="",
 | 
						|
        help="dut model for kpi.csv,  test-priority is arbitrary number")
 | 
						|
    parser.add_argument(
 | 
						|
        "--test_id",
 | 
						|
        default="L3 Data",
 | 
						|
        help="test-id for kpi.csv,  script or test name")
 | 
						|
    parser.add_argument(
 | 
						|
        '--csv_outfile',
 | 
						|
        help="--csv_outfile <Output file for csv data>",
 | 
						|
        default="")
 | 
						|
    # logging configuration:
 | 
						|
    parser.add_argument("--lf_logger_config_json",
 | 
						|
                        help="--lf_logger_config_json <json file> , json configuration of logger")
 | 
						|
 | 
						|
    args = parser.parse_args()
 | 
						|
 | 
						|
    # set up logger
 | 
						|
    logger_config = lf_logger_config.lf_logger_config()
 | 
						|
    if args.lf_logger_config_json:
 | 
						|
        # logger_config.lf_logger_config_json = "lf_logger_config.json"
 | 
						|
        logger_config.lf_logger_config_json = args.lf_logger_config_json
 | 
						|
        logger_config.load_lf_logger_config()
 | 
						|
 | 
						|
    # for kpi.csv generation
 | 
						|
    local_lf_report_dir = args.local_lf_report_dir
 | 
						|
    test_rig = args.test_rig
 | 
						|
    test_tag = args.test_tag
 | 
						|
    dut_hw_version = args.dut_hw_version
 | 
						|
    dut_sw_version = args.dut_sw_version
 | 
						|
    dut_model_num = args.dut_model_num
 | 
						|
    dut_serial_num = args.dut_serial_num
 | 
						|
    # test_priority = args.test_priority  # this may need to be set per test
 | 
						|
    test_id = args.test_id
 | 
						|
 | 
						|
    if local_lf_report_dir != "":
 | 
						|
        report = lf_report.lf_report(
 | 
						|
            _path=local_lf_report_dir,
 | 
						|
            _results_dir_name="sta_connect2",
 | 
						|
            _output_html="sta_connect2.html",
 | 
						|
            _output_pdf="sta_connect2.pdf")
 | 
						|
    else:
 | 
						|
        report = lf_report.lf_report(
 | 
						|
            _results_dir_name="sta_connect2",
 | 
						|
            _output_html="sta_connect2.html",
 | 
						|
            _output_pdf="sta_connect2.pdf")
 | 
						|
 | 
						|
    kpi_path = report.get_report_path()
 | 
						|
    # kpi_filename = "kpi.csv"
 | 
						|
    logger.info("kpi_path :{kpi_path}".format(kpi_path=kpi_path))
 | 
						|
 | 
						|
    kpi_csv = lf_kpi_csv.lf_kpi_csv(
 | 
						|
        _kpi_path=kpi_path,
 | 
						|
        _kpi_test_rig=test_rig,
 | 
						|
        _kpi_test_tag=test_tag,
 | 
						|
        _kpi_dut_hw_version=dut_hw_version,
 | 
						|
        _kpi_dut_sw_version=dut_sw_version,
 | 
						|
        _kpi_dut_model_num=dut_model_num,
 | 
						|
        _kpi_dut_serial_num=dut_serial_num,
 | 
						|
        _kpi_test_id=test_id)
 | 
						|
 | 
						|
    if args.csv_outfile is not None:
 | 
						|
        current_time = time.strftime("%Y-%m-%d-%H-%M-%S", time.localtime())
 | 
						|
        csv_outfile = "{}_{}-sta_connect2.csv".format(
 | 
						|
            args.csv_outfile, current_time)
 | 
						|
        csv_outfile = report.file_add_path(csv_outfile)
 | 
						|
        logger.info("csv output file : {}".format(csv_outfile))
 | 
						|
 | 
						|
    upstream_port = LFUtils.name_to_eid(args.upstream_port)
 | 
						|
    if args.upstream_resource:
 | 
						|
        upstream_resource = args.upstream_resource
 | 
						|
    else:
 | 
						|
        upstream_resource = upstream_port[1]
 | 
						|
 | 
						|
    staConnect = StaConnect2(args.mgr, args.port,
 | 
						|
                             _resource=args.resource,
 | 
						|
                             _upstream_resource=upstream_resource,
 | 
						|
                             _upstream_port=upstream_port[2],
 | 
						|
                             _radio=args.radio,
 | 
						|
                             _sta_mode=args.sta_mode,
 | 
						|
                             debug_=args.debug,
 | 
						|
                             _influx_db=args.influx_db,
 | 
						|
                             _influx_passwd=args.influx_passwd,
 | 
						|
                             _influx_user=args.influx_user,
 | 
						|
                             _influx_host=args.influx_host,
 | 
						|
                             kpi_csv=kpi_csv,
 | 
						|
                             outfile=args.csv_outfile,
 | 
						|
                             download_bps=args.download_bps,
 | 
						|
                             upload_bps=args.upload_bps,
 | 
						|
                             side_a_pdu=args.side_a_pdu,
 | 
						|
                             side_b_pdu=args.side_b_pdu,
 | 
						|
                             _runtime_sec=args.runtime_sec,
 | 
						|
                             _exit_on_fail=True,
 | 
						|
                             _exit_on_error=False)
 | 
						|
 | 
						|
    if args.user:
 | 
						|
        staConnect.user = args.user
 | 
						|
    if args.passwd:
 | 
						|
        staConnect.passwd = args.passwd
 | 
						|
    if args.dut_ssid:
 | 
						|
        staConnect.dut_ssid = args.dut_ssid
 | 
						|
    if args.dut_passwd:
 | 
						|
        staConnect.dut_passwd = args.dut_passwd
 | 
						|
    if args.dut_bssid:
 | 
						|
        staConnect.dut_bssid = args.dut_bssid
 | 
						|
    if args.dut_security:
 | 
						|
        staConnect.dut_security = args.dut_security
 | 
						|
    if args.prefix or (args.prefix != "sta"):
 | 
						|
        staConnect.sta_prefix = args.prefix
 | 
						|
 | 
						|
    staConnect.station_names = ["%s0000" % args.prefix]
 | 
						|
    staConnect.bringup_time_sec = args.bringup_time
 | 
						|
    staConnect.setup()
 | 
						|
    staConnect.start()
 | 
						|
    logger.info("napping %f sec", staConnect.runtime_secs)
 | 
						|
 | 
						|
    time.sleep(staConnect.runtime_secs)
 | 
						|
 | 
						|
    # fill out data kpi.csv and results reports
 | 
						|
    temp_stations_list = []
 | 
						|
    temp_stations_list.extend(staConnect.station_profile.station_names.copy())
 | 
						|
    total_test = len(staConnect.get_result_list())
 | 
						|
    total_pass = len(staConnect.get_passed_result_list())
 | 
						|
    endp_rx_map, endp_rx_drop_map, endps, udp_dl, tcp_dl, udp_ul, tcp_ul, total_dl, total_ul, total_dl_ll, total_ul_ll = staConnect.get_rx_values()
 | 
						|
    staConnect.record_kpi_csv(len(temp_stations_list), total_test, total_pass, udp_dl, tcp_dl, udp_ul, tcp_ul, total_dl, total_ul)
 | 
						|
    staConnect.record_results(len(temp_stations_list), udp_dl, tcp_dl, udp_ul, tcp_ul, total_dl, total_ul)
 | 
						|
 | 
						|
    # Reporting Results (.pdf & .html)
 | 
						|
    csv_results_file = staConnect.get_csv_name()
 | 
						|
    logger.info("csv_results_file: {}".format(csv_results_file))
 | 
						|
    # csv_results_file = kpi_path + "/" + kpi_filename
 | 
						|
    report.set_title("L3 Station Connect 2")
 | 
						|
    report.build_banner()
 | 
						|
    report.set_table_title("L3 Stations Connect 2 Key Performance Indexes")
 | 
						|
    report.build_table_title()
 | 
						|
    report.set_table_dataframe_from_csv(csv_results_file)
 | 
						|
    report.build_table()
 | 
						|
    report.write_html_with_timestamp()
 | 
						|
    report.write_index_html()
 | 
						|
    # report.write_pdf(_page_size = 'A3', _orientation='Landscape')
 | 
						|
    # report.write_pdf_with_timestamp(_page_size='A4', _orientation='Portrait')
 | 
						|
    report.write_pdf_with_timestamp(_page_size='A4', _orientation='Landscape')
 | 
						|
 | 
						|
    staConnect.stop()
 | 
						|
    # exit(1)
 | 
						|
 | 
						|
    # py-json/lanforge/lfcli_base.py - get_result_list():
 | 
						|
    staConnect.get_result_list()
 | 
						|
    is_passing = staConnect.passes()
 | 
						|
 | 
						|
    # py-json/lanforge/lfcli_base.py - get_all_message():
 | 
						|
    logger.info(staConnect.get_all_message())
 | 
						|
 | 
						|
    # TODO clean up pass fail  to use realm
 | 
						|
    if not is_passing:
 | 
						|
        logger.info("FAIL:  Some tests failed")
 | 
						|
    else:
 | 
						|
        logger.info("PASS:  All tests pass")
 | 
						|
 | 
						|
    # cleanup stations
 | 
						|
    if not args.no_cleanup:
 | 
						|
        staConnect.pre_cleanup()
 | 
						|
 | 
						|
# - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
 | 
						|
 | 
						|
 | 
						|
if __name__ == "__main__":
 | 
						|
    main()
 |